commit stringlengths 40 40 | old_file stringlengths 4 118 | new_file stringlengths 4 118 | old_contents stringlengths 0 2.94k | new_contents stringlengths 1 4.43k | subject stringlengths 15 444 | message stringlengths 16 3.45k | lang stringclasses 1 value | license stringclasses 13 values | repos stringlengths 5 43.2k | prompt stringlengths 17 4.58k | response stringlengths 1 4.43k | prompt_tagged stringlengths 58 4.62k | response_tagged stringlengths 1 4.43k | text stringlengths 132 7.29k | text_tagged stringlengths 173 7.33k |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
a2ee6106a6c98dae102cf14902c6b82f480e6cbe | python/main.py | python/main.py | import sys
from enum import Enum
class Furniture(Enum):
bed = 1
couce = 2
desk = 3
chair = 4
tv = 5
table = 6
rug = 7
shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
| import sys
from enum import Enum
from furniture import *
#class Furniture(Enum):
# bed = 1
# couce = 2
# desk = 3
# chair = 4
# tv = 5
# table = 6
# rug = 7
# shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
placeDesksAndChairs()
placeCouchesTablesAndTv()
placeBeds()
placeShelves()
placeRugs()
| Add calls to furniture placement functions | Add calls to furniture placement functions
| Python | apache-2.0 | TheZoq2/VRHack,TheZoq2/VRHack,TheZoq2/VRHack,TheZoq2/VRHack | import sys
from enum import Enum
class Furniture(Enum):
bed = 1
couce = 2
desk = 3
chair = 4
tv = 5
table = 6
rug = 7
shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
Add calls to furniture placement functions | import sys
from enum import Enum
from furniture import *
#class Furniture(Enum):
# bed = 1
# couce = 2
# desk = 3
# chair = 4
# tv = 5
# table = 6
# rug = 7
# shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
placeDesksAndChairs()
placeCouchesTablesAndTv()
placeBeds()
placeShelves()
placeRugs()
| <commit_before>import sys
from enum import Enum
class Furniture(Enum):
bed = 1
couce = 2
desk = 3
chair = 4
tv = 5
table = 6
rug = 7
shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
<commit_msg>Add calls to furniture placement functions<commit_after> | import sys
from enum import Enum
from furniture import *
#class Furniture(Enum):
# bed = 1
# couce = 2
# desk = 3
# chair = 4
# tv = 5
# table = 6
# rug = 7
# shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
placeDesksAndChairs()
placeCouchesTablesAndTv()
placeBeds()
placeShelves()
placeRugs()
| import sys
from enum import Enum
class Furniture(Enum):
bed = 1
couce = 2
desk = 3
chair = 4
tv = 5
table = 6
rug = 7
shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
Add calls to furniture placement functionsimport sys
from enum import Enum
from furniture import *
#class Furniture(Enum):
# bed = 1
# couce = 2
# desk = 3
# chair = 4
# tv = 5
# table = 6
# rug = 7
# shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
placeDesksAndChairs()
placeCouchesTablesAndTv()
placeBeds()
placeShelves()
placeRugs()
| <commit_before>import sys
from enum import Enum
class Furniture(Enum):
bed = 1
couce = 2
desk = 3
chair = 4
tv = 5
table = 6
rug = 7
shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
<commit_msg>Add calls to furniture placement functions<commit_after>import sys
from enum import Enum
from furniture import *
#class Furniture(Enum):
# bed = 1
# couce = 2
# desk = 3
# chair = 4
# tv = 5
# table = 6
# rug = 7
# shelf = 8
f = open(sys.argv[1], 'r')
print(f.read())
placeDesksAndChairs()
placeCouchesTablesAndTv()
placeBeds()
placeShelves()
placeRugs()
|
c7ed3e2a39c7de1120a33cd0253d9ac3bd9e7984 | redcliff/cli.py | redcliff/cli.py | from sys import exit
import argparse
from .commands import dispatch
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url')
parser.add_argument('-k', '--api-key')
parser.add_argument('-C', '--config-file')
parser.add_argument('cmd')
parser.add_argument('args', nargs=argparse.REMAINDER)
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
| from sys import exit
import argparse
from .commands import dispatch, choices
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url',
metavar='https://redmine.example.com',
help='Base URL of your Redmine installation.')
parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify',
action='store_const', const=False)
parser.add_argument('-k', '--api-key',
help='Your Redmine API key.')
parser.add_argument('-C', '--config-file',
help='Override default config path.')
parser.add_argument('cmd',
choices=choices,
help='Command to execute.')
parser.add_argument('args',
nargs=argparse.REMAINDER,
help='Arguments to command. Use --help to get '
'command-specific help.')
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
| Update main arguments parser config | Update main arguments parser config
| Python | mit | dmedvinsky/redcliff | from sys import exit
import argparse
from .commands import dispatch
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url')
parser.add_argument('-k', '--api-key')
parser.add_argument('-C', '--config-file')
parser.add_argument('cmd')
parser.add_argument('args', nargs=argparse.REMAINDER)
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
Update main arguments parser config | from sys import exit
import argparse
from .commands import dispatch, choices
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url',
metavar='https://redmine.example.com',
help='Base URL of your Redmine installation.')
parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify',
action='store_const', const=False)
parser.add_argument('-k', '--api-key',
help='Your Redmine API key.')
parser.add_argument('-C', '--config-file',
help='Override default config path.')
parser.add_argument('cmd',
choices=choices,
help='Command to execute.')
parser.add_argument('args',
nargs=argparse.REMAINDER,
help='Arguments to command. Use --help to get '
'command-specific help.')
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
| <commit_before>from sys import exit
import argparse
from .commands import dispatch
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url')
parser.add_argument('-k', '--api-key')
parser.add_argument('-C', '--config-file')
parser.add_argument('cmd')
parser.add_argument('args', nargs=argparse.REMAINDER)
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
<commit_msg>Update main arguments parser config<commit_after> | from sys import exit
import argparse
from .commands import dispatch, choices
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url',
metavar='https://redmine.example.com',
help='Base URL of your Redmine installation.')
parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify',
action='store_const', const=False)
parser.add_argument('-k', '--api-key',
help='Your Redmine API key.')
parser.add_argument('-C', '--config-file',
help='Override default config path.')
parser.add_argument('cmd',
choices=choices,
help='Command to execute.')
parser.add_argument('args',
nargs=argparse.REMAINDER,
help='Arguments to command. Use --help to get '
'command-specific help.')
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
| from sys import exit
import argparse
from .commands import dispatch
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url')
parser.add_argument('-k', '--api-key')
parser.add_argument('-C', '--config-file')
parser.add_argument('cmd')
parser.add_argument('args', nargs=argparse.REMAINDER)
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
Update main arguments parser configfrom sys import exit
import argparse
from .commands import dispatch, choices
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url',
metavar='https://redmine.example.com',
help='Base URL of your Redmine installation.')
parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify',
action='store_const', const=False)
parser.add_argument('-k', '--api-key',
help='Your Redmine API key.')
parser.add_argument('-C', '--config-file',
help='Override default config path.')
parser.add_argument('cmd',
choices=choices,
help='Command to execute.')
parser.add_argument('args',
nargs=argparse.REMAINDER,
help='Arguments to command. Use --help to get '
'command-specific help.')
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
| <commit_before>from sys import exit
import argparse
from .commands import dispatch
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url')
parser.add_argument('-k', '--api-key')
parser.add_argument('-C', '--config-file')
parser.add_argument('cmd')
parser.add_argument('args', nargs=argparse.REMAINDER)
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
<commit_msg>Update main arguments parser config<commit_after>from sys import exit
import argparse
from .commands import dispatch, choices
from .config import get_config
from .utils import merge
def main():
parser = argparse.ArgumentParser()
parser.add_argument('-u', '--base-url',
metavar='https://redmine.example.com',
help='Base URL of your Redmine installation.')
parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify',
action='store_const', const=False)
parser.add_argument('-k', '--api-key',
help='Your Redmine API key.')
parser.add_argument('-C', '--config-file',
help='Override default config path.')
parser.add_argument('cmd',
choices=choices,
help='Command to execute.')
parser.add_argument('args',
nargs=argparse.REMAINDER,
help='Arguments to command. Use --help to get '
'command-specific help.')
args = vars(parser.parse_args())
conf = get_config(args.pop('config_file'))
cmd = args.pop('cmd')
cmd_args = args.pop('args')
merged_conf = merge(conf, args)
return dispatch(cmd, cmd_args, merged_conf)
if __name__ == '__main__':
exit(main())
|
02ca88129430044f4202991358939d87f8c6da0b | simple-cipher/simple_cipher.py | simple-cipher/simple_cipher.py | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| Refactor to reuse the encode method for decoding | Refactor to reuse the encode method for decoding
| Python | agpl-3.0 | CubicComet/exercism-python-solutions | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
Refactor to reuse the encode method for decoding | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| <commit_before>import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
<commit_msg>Refactor to reuse the encode method for decoding<commit_after> | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
Refactor to reuse the encode method for decodingimport math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| <commit_before>import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
<commit_msg>Refactor to reuse the encode method for decoding<commit_after>import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
|
3a86ea268b7cb9f00968e7dcb228d6821dafda99 | simple-cipher/simple_cipher.py | simple-cipher/simple_cipher.py | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| Refactor to reuse the encode method for decoding | Refactor to reuse the encode method for decoding
| Python | agpl-3.0 | CubicComet/exercism-python-solutions | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
Refactor to reuse the encode method for decoding | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| <commit_before>import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
<commit_msg>Refactor to reuse the encode method for decoding<commit_after> | import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
Refactor to reuse the encode method for decodingimport math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
| <commit_before>import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
<commit_msg>Refactor to reuse the encode method for decoding<commit_after>import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s, dirn=1):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key))
def decode(self, s):
return self.encode(s, dirn=-1)
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
|
16c9563a75792aba7ccc0d979f579d64dc0140c1 | common_rg_bar.py | common_rg_bar.py | #!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
esc = chr(27)
print (''.join((
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
''' % dict(
prog_name=sys.argv[0],
))
| #!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
3. (optional) Text to display (without color)
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
if len(sys.argv) >= 4:
start_text = sys.argv[3] + ' '
else:
start_text = ''
esc = chr(27)
print (''.join((
start_text,
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code) - len(start_text)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
3. (optional) Text to display
''' % dict(
prog_name=sys.argv[0],
))
| Add optional text to display | Add optional text to display
| Python | mit | kwadrat/rgb_tdd | #!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
esc = chr(27)
print (''.join((
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
''' % dict(
prog_name=sys.argv[0],
))
Add optional text to display | #!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
3. (optional) Text to display (without color)
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
if len(sys.argv) >= 4:
start_text = sys.argv[3] + ' '
else:
start_text = ''
esc = chr(27)
print (''.join((
start_text,
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code) - len(start_text)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
3. (optional) Text to display
''' % dict(
prog_name=sys.argv[0],
))
| <commit_before>#!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
esc = chr(27)
print (''.join((
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
''' % dict(
prog_name=sys.argv[0],
))
<commit_msg>Add optional text to display<commit_after> | #!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
3. (optional) Text to display (without color)
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
if len(sys.argv) >= 4:
start_text = sys.argv[3] + ' '
else:
start_text = ''
esc = chr(27)
print (''.join((
start_text,
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code) - len(start_text)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
3. (optional) Text to display
''' % dict(
prog_name=sys.argv[0],
))
| #!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
esc = chr(27)
print (''.join((
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
''' % dict(
prog_name=sys.argv[0],
))
Add optional text to display#!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
3. (optional) Text to display (without color)
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
if len(sys.argv) >= 4:
start_text = sys.argv[3] + ' '
else:
start_text = ''
esc = chr(27)
print (''.join((
start_text,
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code) - len(start_text)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
3. (optional) Text to display
''' % dict(
prog_name=sys.argv[0],
))
| <commit_before>#!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
esc = chr(27)
print (''.join((
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
''' % dict(
prog_name=sys.argv[0],
))
<commit_msg>Add optional text to display<commit_after>#!/usr/bin/env python3
'''
Given:
1. status code: (0 - OK, other value - BAD)
2. terminal window width
3. (optional) Text to display (without color)
shows red/green bar to visualize return code of previous command
'''
import sys
def main():
if len(sys.argv) >= 2:
code = sys.argv[1]
if code == 'x':
col_char = '3'
cols_limit = 78
code = '' # No code provided - only yellow bar
else:
if code == 'y':
col_char = '3'
else:
value = int(code)
if value:
col_char = '1'
else:
col_char = '2'
cols_limit = int(sys.argv[2])
if len(sys.argv) >= 4:
start_text = sys.argv[3] + ' '
else:
start_text = ''
esc = chr(27)
print (''.join((
start_text,
esc,
'[4',
col_char,
'm',
' ' * (cols_limit - 2 - len(code) - len(start_text)),
code,
esc,
'[0m',
)))
else:
print ('''
Usage: %(prog_name)s status_code number_of_columns
1. status code: 0 - OK (green color), other values - BAD (red color)
2. number of columns: the width of text console
3. (optional) Text to display
''' % dict(
prog_name=sys.argv[0],
))
|
ff391fc302b6d4e9fab0653522fa2fe47d8e8faa | beavy_modules/url_extractor/lib.py | beavy_modules/url_extractor/lib.py | import lassie
from pyembed.core import PyEmbed
from beavy.app import cache
pyembed = PyEmbed()
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs)
|
from pyembed.core import PyEmbed
from beavy.app import cache
from lassie import Lassie
import re
# lassie by default isn't extensive enough for us
# configure it so that it is.
from lassie.filters import FILTER_MAPS
FILTER_MAPS['meta']['open_graph']['map'].update({
# general
"og:type": "type",
"og:site_name": "site_name",
})
FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I)
FILTER_MAPS['meta']['generic']['map'].update({
# articles
"article:published_time": "published_time",
"article:modified_time": "modified_time",
"article:expiration_time": "expiration_time",
"article:section": "section",
"article:section_url": "section_url",
# music
"music:duration": "duration",
"music:release_date": "release_date",
# video
"video:duration": "duration",
"video:release_date": "release_date",
# author
"author": "author",
# book
"book:author": "author",
"book:isbn": "isbn",
"book:release_date": "release_date",
})
# general configuration
pyembed = PyEmbed()
lassie = Lassie()
lassie.request_opts = {
'headers':{
# tell Lassie to tell others it is facebook
'User-Agent': 'facebookexternalhit/1.1'
}
}
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed(url, **kwargs)
| Configure Lassie for more information | Configure Lassie for more information
| Python | mpl-2.0 | beavyHQ/beavy,beavyHQ/beavy,beavyHQ/beavy,beavyHQ/beavy | import lassie
from pyembed.core import PyEmbed
from beavy.app import cache
pyembed = PyEmbed()
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs)
Configure Lassie for more information |
from pyembed.core import PyEmbed
from beavy.app import cache
from lassie import Lassie
import re
# lassie by default isn't extensive enough for us
# configure it so that it is.
from lassie.filters import FILTER_MAPS
FILTER_MAPS['meta']['open_graph']['map'].update({
# general
"og:type": "type",
"og:site_name": "site_name",
})
FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I)
FILTER_MAPS['meta']['generic']['map'].update({
# articles
"article:published_time": "published_time",
"article:modified_time": "modified_time",
"article:expiration_time": "expiration_time",
"article:section": "section",
"article:section_url": "section_url",
# music
"music:duration": "duration",
"music:release_date": "release_date",
# video
"video:duration": "duration",
"video:release_date": "release_date",
# author
"author": "author",
# book
"book:author": "author",
"book:isbn": "isbn",
"book:release_date": "release_date",
})
# general configuration
pyembed = PyEmbed()
lassie = Lassie()
lassie.request_opts = {
'headers':{
# tell Lassie to tell others it is facebook
'User-Agent': 'facebookexternalhit/1.1'
}
}
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed(url, **kwargs)
| <commit_before>import lassie
from pyembed.core import PyEmbed
from beavy.app import cache
pyembed = PyEmbed()
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs)
<commit_msg>Configure Lassie for more information<commit_after> |
from pyembed.core import PyEmbed
from beavy.app import cache
from lassie import Lassie
import re
# lassie by default isn't extensive enough for us
# configure it so that it is.
from lassie.filters import FILTER_MAPS
FILTER_MAPS['meta']['open_graph']['map'].update({
# general
"og:type": "type",
"og:site_name": "site_name",
})
FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I)
FILTER_MAPS['meta']['generic']['map'].update({
# articles
"article:published_time": "published_time",
"article:modified_time": "modified_time",
"article:expiration_time": "expiration_time",
"article:section": "section",
"article:section_url": "section_url",
# music
"music:duration": "duration",
"music:release_date": "release_date",
# video
"video:duration": "duration",
"video:release_date": "release_date",
# author
"author": "author",
# book
"book:author": "author",
"book:isbn": "isbn",
"book:release_date": "release_date",
})
# general configuration
pyembed = PyEmbed()
lassie = Lassie()
lassie.request_opts = {
'headers':{
# tell Lassie to tell others it is facebook
'User-Agent': 'facebookexternalhit/1.1'
}
}
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed(url, **kwargs)
| import lassie
from pyembed.core import PyEmbed
from beavy.app import cache
pyembed = PyEmbed()
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs)
Configure Lassie for more information
from pyembed.core import PyEmbed
from beavy.app import cache
from lassie import Lassie
import re
# lassie by default isn't extensive enough for us
# configure it so that it is.
from lassie.filters import FILTER_MAPS
FILTER_MAPS['meta']['open_graph']['map'].update({
# general
"og:type": "type",
"og:site_name": "site_name",
})
FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I)
FILTER_MAPS['meta']['generic']['map'].update({
# articles
"article:published_time": "published_time",
"article:modified_time": "modified_time",
"article:expiration_time": "expiration_time",
"article:section": "section",
"article:section_url": "section_url",
# music
"music:duration": "duration",
"music:release_date": "release_date",
# video
"video:duration": "duration",
"video:release_date": "release_date",
# author
"author": "author",
# book
"book:author": "author",
"book:isbn": "isbn",
"book:release_date": "release_date",
})
# general configuration
pyembed = PyEmbed()
lassie = Lassie()
lassie.request_opts = {
'headers':{
# tell Lassie to tell others it is facebook
'User-Agent': 'facebookexternalhit/1.1'
}
}
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed(url, **kwargs)
| <commit_before>import lassie
from pyembed.core import PyEmbed
from beavy.app import cache
pyembed = PyEmbed()
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs)
<commit_msg>Configure Lassie for more information<commit_after>
from pyembed.core import PyEmbed
from beavy.app import cache
from lassie import Lassie
import re
# lassie by default isn't extensive enough for us
# configure it so that it is.
from lassie.filters import FILTER_MAPS
FILTER_MAPS['meta']['open_graph']['map'].update({
# general
"og:type": "type",
"og:site_name": "site_name",
})
FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I)
FILTER_MAPS['meta']['generic']['map'].update({
# articles
"article:published_time": "published_time",
"article:modified_time": "modified_time",
"article:expiration_time": "expiration_time",
"article:section": "section",
"article:section_url": "section_url",
# music
"music:duration": "duration",
"music:release_date": "release_date",
# video
"video:duration": "duration",
"video:release_date": "release_date",
# author
"author": "author",
# book
"book:author": "author",
"book:isbn": "isbn",
"book:release_date": "release_date",
})
# general configuration
pyembed = PyEmbed()
lassie = Lassie()
lassie.request_opts = {
'headers':{
# tell Lassie to tell others it is facebook
'User-Agent': 'facebookexternalhit/1.1'
}
}
@cache.memoize()
def extract_info(url):
return lassie.fetch(url)
@cache.memoize()
def extract_oembed(url, **kwargs):
return pyembed.embed(url, **kwargs)
|
c58b2dd49ad5c73d49b496025d13116da30b3b9a | examples/qm7_long.py | examples/qm7_long.py | import numpy
from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
train_feats = []
test_feats = []
for tf in feats:
X_train = tf.fit_transform(Xin_train)
X_test = tf.transform(Xin_test)
train_feats.append(X_train)
test_feats.append(X_test)
X_train = numpy.hstack(train_feats)
X_test = numpy.hstack(test_feats)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
| from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity, MultiFeature
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
full_feat = MultiFeature(feats)
X_train = full_feat.fit_transform(Xin_train)
X_test = full_feat.transform(Xin_test)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
| Change qm7 example to use MultiFeature | Change qm7 example to use MultiFeature
| Python | mit | crcollins/molml | import numpy
from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
train_feats = []
test_feats = []
for tf in feats:
X_train = tf.fit_transform(Xin_train)
X_test = tf.transform(Xin_test)
train_feats.append(X_train)
test_feats.append(X_test)
X_train = numpy.hstack(train_feats)
X_test = numpy.hstack(test_feats)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
Change qm7 example to use MultiFeature | from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity, MultiFeature
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
full_feat = MultiFeature(feats)
X_train = full_feat.fit_transform(Xin_train)
X_test = full_feat.transform(Xin_test)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
| <commit_before>import numpy
from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
train_feats = []
test_feats = []
for tf in feats:
X_train = tf.fit_transform(Xin_train)
X_test = tf.transform(Xin_test)
train_feats.append(X_train)
test_feats.append(X_test)
X_train = numpy.hstack(train_feats)
X_test = numpy.hstack(test_feats)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
<commit_msg>Change qm7 example to use MultiFeature<commit_after> | from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity, MultiFeature
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
full_feat = MultiFeature(feats)
X_train = full_feat.fit_transform(Xin_train)
X_test = full_feat.transform(Xin_test)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
| import numpy
from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
train_feats = []
test_feats = []
for tf in feats:
X_train = tf.fit_transform(Xin_train)
X_test = tf.transform(Xin_test)
train_feats.append(X_train)
test_feats.append(X_test)
X_train = numpy.hstack(train_feats)
X_test = numpy.hstack(test_feats)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
Change qm7 example to use MultiFeaturefrom sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity, MultiFeature
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
full_feat = MultiFeature(feats)
X_train = full_feat.fit_transform(Xin_train)
X_test = full_feat.transform(Xin_test)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
| <commit_before>import numpy
from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
train_feats = []
test_feats = []
for tf in feats:
X_train = tf.fit_transform(Xin_train)
X_test = tf.transform(Xin_test)
train_feats.append(X_train)
test_feats.append(X_test)
X_train = numpy.hstack(train_feats)
X_test = numpy.hstack(test_feats)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
<commit_msg>Change qm7 example to use MultiFeature<commit_after>from sklearn.linear_model import Ridge
from sklearn.kernel_ridge import KernelRidge
from sklearn.metrics import mean_absolute_error as MAE
from molml.features import EncodedBond, Connectivity, MultiFeature
from utils import load_qm7
if __name__ == "__main__":
# This is just boiler plate code to load the data
Xin_train, Xin_test, y_train, y_test = load_qm7()
feats = [
EncodedBond(n_jobs=-1, max_depth=3),
Connectivity(depth=1, n_jobs=-1),
Connectivity(depth=3, use_coordination=True, n_jobs=-1),
]
full_feat = MultiFeature(feats)
X_train = full_feat.fit_transform(Xin_train)
X_test = full_feat.transform(Xin_test)
clfs = [
Ridge(alpha=0.01),
KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"),
]
for clf in clfs:
print clf
clf.fit(X_train, y_train)
train_error = MAE(clf.predict(X_train), y_train)
test_error = MAE(clf.predict(X_test), y_test)
print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error)
print
|
67c444fb3603c234916b790d3dded3625f0512e5 | pivot/test/test_utils.py | pivot/test/test_utils.py | """
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_google_analytics_processor(self):
self.assertEquals(get_latest_term(), 'au12')
| """
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_get_latest_term(self):
self.assertEquals(get_latest_term(), 'au12')
| Rename test to something more descriptive. | Rename test to something more descriptive.
| Python | apache-2.0 | uw-it-aca/pivot,uw-it-aca/pivot,uw-it-aca/pivot,uw-it-aca/pivot,uw-it-aca/pivot | """
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_google_analytics_processor(self):
self.assertEquals(get_latest_term(), 'au12')
Rename test to something more descriptive. | """
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_get_latest_term(self):
self.assertEquals(get_latest_term(), 'au12')
| <commit_before>"""
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_google_analytics_processor(self):
self.assertEquals(get_latest_term(), 'au12')
<commit_msg>Rename test to something more descriptive.<commit_after> | """
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_get_latest_term(self):
self.assertEquals(get_latest_term(), 'au12')
| """
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_google_analytics_processor(self):
self.assertEquals(get_latest_term(), 'au12')
Rename test to something more descriptive."""
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_get_latest_term(self):
self.assertEquals(get_latest_term(), 'au12')
| <commit_before>"""
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_google_analytics_processor(self):
self.assertEquals(get_latest_term(), 'au12')
<commit_msg>Rename test to something more descriptive.<commit_after>"""
Tests utility scripts
"""
import os
from django.test import TestCase, RequestFactory
from django.test.utils import override_settings
import pivot
from pivot.utils import get_latest_term
TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__),
'test',
'test_resources',
'csvfiles/',)
class UtilsTest(TestCase):
@override_settings(CSV_ROOT=TEST_CSV_PATH)
def test_get_latest_term(self):
self.assertEquals(get_latest_term(), 'au12')
|
f7c03daa9ce803ec10e1c7cd9319840045f47663 | ddsc_core/management/commands/export_pi_xml.py | ddsc_core/management/commands/export_pi_xml.py | import sys
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = "help"
def handle(self, *args, **options):
try:
source = args[0]
except IndexError:
self.stdout.write(self.help)
return
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(sys.stdout)
| from optparse import make_option
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = (
"Create pi.xml from a template. A template is a valid pi.xml file " +
"without events (they are ignored if present) and per series a " +
"`comment` element containing a ddsc timeseries uuid."
)
option_list = BaseCommand.option_list + (
make_option(
'-f',
'--file',
dest='file',
help='write to file instead of stdout'
),
)
def handle(self, *args, **options):
# source is a pi.xml file that serves as a template: its headers are
# used, but any events are ignored. The `comment` element of each
# series is expected to contain a ddsc uuid.
try:
source = args[0]
except IndexError:
self.stderr.write(self.help)
return
# destination is the resulting pi.xml file. If no destination is
# given, output is written to stdout.
try:
destination = open(options.get('file'), 'w')
except TypeError:
destination = self.stdout
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(destination)
| Improve management command for exporting pi-xml | Improve management command for exporting pi-xml
| Python | mit | ddsc/ddsc-core,ddsc/ddsc-core | import sys
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = "help"
def handle(self, *args, **options):
try:
source = args[0]
except IndexError:
self.stdout.write(self.help)
return
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(sys.stdout)
Improve management command for exporting pi-xml | from optparse import make_option
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = (
"Create pi.xml from a template. A template is a valid pi.xml file " +
"without events (they are ignored if present) and per series a " +
"`comment` element containing a ddsc timeseries uuid."
)
option_list = BaseCommand.option_list + (
make_option(
'-f',
'--file',
dest='file',
help='write to file instead of stdout'
),
)
def handle(self, *args, **options):
# source is a pi.xml file that serves as a template: its headers are
# used, but any events are ignored. The `comment` element of each
# series is expected to contain a ddsc uuid.
try:
source = args[0]
except IndexError:
self.stderr.write(self.help)
return
# destination is the resulting pi.xml file. If no destination is
# given, output is written to stdout.
try:
destination = open(options.get('file'), 'w')
except TypeError:
destination = self.stdout
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(destination)
| <commit_before>import sys
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = "help"
def handle(self, *args, **options):
try:
source = args[0]
except IndexError:
self.stdout.write(self.help)
return
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(sys.stdout)
<commit_msg>Improve management command for exporting pi-xml<commit_after> | from optparse import make_option
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = (
"Create pi.xml from a template. A template is a valid pi.xml file " +
"without events (they are ignored if present) and per series a " +
"`comment` element containing a ddsc timeseries uuid."
)
option_list = BaseCommand.option_list + (
make_option(
'-f',
'--file',
dest='file',
help='write to file instead of stdout'
),
)
def handle(self, *args, **options):
# source is a pi.xml file that serves as a template: its headers are
# used, but any events are ignored. The `comment` element of each
# series is expected to contain a ddsc uuid.
try:
source = args[0]
except IndexError:
self.stderr.write(self.help)
return
# destination is the resulting pi.xml file. If no destination is
# given, output is written to stdout.
try:
destination = open(options.get('file'), 'w')
except TypeError:
destination = self.stdout
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(destination)
| import sys
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = "help"
def handle(self, *args, **options):
try:
source = args[0]
except IndexError:
self.stdout.write(self.help)
return
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(sys.stdout)
Improve management command for exporting pi-xmlfrom optparse import make_option
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = (
"Create pi.xml from a template. A template is a valid pi.xml file " +
"without events (they are ignored if present) and per series a " +
"`comment` element containing a ddsc timeseries uuid."
)
option_list = BaseCommand.option_list + (
make_option(
'-f',
'--file',
dest='file',
help='write to file instead of stdout'
),
)
def handle(self, *args, **options):
# source is a pi.xml file that serves as a template: its headers are
# used, but any events are ignored. The `comment` element of each
# series is expected to contain a ddsc uuid.
try:
source = args[0]
except IndexError:
self.stderr.write(self.help)
return
# destination is the resulting pi.xml file. If no destination is
# given, output is written to stdout.
try:
destination = open(options.get('file'), 'w')
except TypeError:
destination = self.stdout
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(destination)
| <commit_before>import sys
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = "help"
def handle(self, *args, **options):
try:
source = args[0]
except IndexError:
self.stdout.write(self.help)
return
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(sys.stdout)
<commit_msg>Improve management command for exporting pi-xml<commit_after>from optparse import make_option
from django.core.management.base import BaseCommand
import pandas as pd
from tslib.readers import PiXmlReader
from tslib.writers import PiXmlWriter
from ddsc_core.models import Timeseries
class Command(BaseCommand):
args = "<pi.xml>"
help = (
"Create pi.xml from a template. A template is a valid pi.xml file " +
"without events (they are ignored if present) and per series a " +
"`comment` element containing a ddsc timeseries uuid."
)
option_list = BaseCommand.option_list + (
make_option(
'-f',
'--file',
dest='file',
help='write to file instead of stdout'
),
)
def handle(self, *args, **options):
# source is a pi.xml file that serves as a template: its headers are
# used, but any events are ignored. The `comment` element of each
# series is expected to contain a ddsc uuid.
try:
source = args[0]
except IndexError:
self.stderr.write(self.help)
return
# destination is the resulting pi.xml file. If no destination is
# given, output is written to stdout.
try:
destination = open(options.get('file'), 'w')
except TypeError:
destination = self.stdout
reader = PiXmlReader(source)
writer = PiXmlWriter(reader.get_tz())
for md, _ in reader.get_series():
try:
uuid = md.pop('comment')
timeseries = Timeseries.objects.get(uuid=uuid)
df = timeseries.get_events()
writer.set_series(md, df)
except Timeseries.DoesNotExist:
df = pd.DataFrame()
writer.set_series(md, df)
except:
pass
writer.write(destination)
|
afb398094e207fdd338a492dbbe9fca3f041e2c7 | tests/test_postgres_processor.py | tests/test_postgres_processor.py | import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
| import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from django.test import TestCase
from scrapi.processing.postgres import PostgresProcessor, Document
from . import utils
from scrapi.linter.document import RawDocument
test_db = PostgresProcessor()
# NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
class DocumentTestCase(TestCase):
@pytest.mark.django_db
def test_Documents_can_speak(self):
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
| Make this test a django test case | Make this test a django test case
| Python | apache-2.0 | erinspace/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,erinspace/scrapi,felliott/scrapi,mehanig/scrapi,fabianvf/scrapi,fabianvf/scrapi,felliott/scrapi,mehanig/scrapi | import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
Make this test a django test case | import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from django.test import TestCase
from scrapi.processing.postgres import PostgresProcessor, Document
from . import utils
from scrapi.linter.document import RawDocument
test_db = PostgresProcessor()
# NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
class DocumentTestCase(TestCase):
@pytest.mark.django_db
def test_Documents_can_speak(self):
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
| <commit_before>import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
<commit_msg>Make this test a django test case<commit_after> | import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from django.test import TestCase
from scrapi.processing.postgres import PostgresProcessor, Document
from . import utils
from scrapi.linter.document import RawDocument
test_db = PostgresProcessor()
# NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
class DocumentTestCase(TestCase):
@pytest.mark.django_db
def test_Documents_can_speak(self):
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
| import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
Make this test a django test caseimport pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from django.test import TestCase
from scrapi.processing.postgres import PostgresProcessor, Document
from . import utils
from scrapi.linter.document import RawDocument
test_db = PostgresProcessor()
# NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
class DocumentTestCase(TestCase):
@pytest.mark.django_db
def test_Documents_can_speak(self):
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
| <commit_before>import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from . import utils
from scrapi.linter.document import NormalizedDocument, RawDocument
from scrapi.processing.postgres import PostgresProcessor, Document
test_db = PostgresProcessor()
NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
@pytest.mark.django_db
def test_process_raw():
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
<commit_msg>Make this test a django test case<commit_after>import pytest
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings")
from django.test import TestCase
from scrapi.processing.postgres import PostgresProcessor, Document
from . import utils
from scrapi.linter.document import RawDocument
test_db = PostgresProcessor()
# NORMALIZED = NormalizedDocument(utils.RECORD)
RAW = RawDocument(utils.RAW_DOC)
class DocumentTestCase(TestCase):
@pytest.mark.django_db
def test_Documents_can_speak(self):
test_db.process_raw(RAW)
queryset = Document(docID='someID', source=RAW['source'])
assert queryset.docID == RAW.attributes['docID']
|
5d59f800da9fb737cd87d47301793f750ca1cbdd | pysnow/exceptions.py | pysnow/exceptions.py | # -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
| # -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class UnexpectedResponseFormat(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
| Add missing UnexpectedResponseFormat for backward compatability | Add missing UnexpectedResponseFormat for backward compatability
Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com>
| Python | mit | rbw0/pysnow | # -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
Add missing UnexpectedResponseFormat for backward compatability
Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com> | # -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class UnexpectedResponseFormat(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
| <commit_before># -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
<commit_msg>Add missing UnexpectedResponseFormat for backward compatability
Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com><commit_after> | # -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class UnexpectedResponseFormat(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
| # -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
Add missing UnexpectedResponseFormat for backward compatability
Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com># -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class UnexpectedResponseFormat(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
| <commit_before># -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
<commit_msg>Add missing UnexpectedResponseFormat for backward compatability
Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com><commit_after># -*- coding: utf-8 -*-
class PysnowException(Exception):
pass
class InvalidUsage(PysnowException):
pass
class UnexpectedResponseFormat(PysnowException):
pass
class ResponseError(PysnowException):
message = "<empty>"
detail = "<empty>"
def __init__(self, error):
if "message" in error:
self.message = error["message"] or self.message
if "detail" in error:
self.detail = error["detail"] or self.detail
def __str__(self):
return "Error in response. Message: %s, Details: %s" % (
self.message,
self.detail,
)
class MissingResult(PysnowException):
pass
class NoResults(PysnowException):
pass
class EmptyContent(PysnowException):
pass
class MultipleResults(PysnowException):
pass
class MissingToken(PysnowException):
pass
class TokenCreateError(PysnowException):
def __init__(self, error, description, status_code):
self.error = error
self.description = description
self.snow_status_code = status_code
class QueryTypeError(PysnowException):
pass
class QueryMissingField(PysnowException):
pass
class QueryEmpty(PysnowException):
pass
class QueryExpressionError(PysnowException):
pass
class QueryMultipleExpressions(PysnowException):
pass
|
2e812af6b937091d65a0b83ead936894a2789d02 | rdflib/serializer.py | rdflib/serializer.py | """
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "UTF-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
| """
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "utf-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
| Change to preferred encoding style. | Change to preferred encoding style.
UTF-8 -> utf-8
| Python | bsd-3-clause | RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib | """
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "UTF-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
Change to preferred encoding style.
UTF-8 -> utf-8 | """
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "utf-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
| <commit_before>"""
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "UTF-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
<commit_msg>Change to preferred encoding style.
UTF-8 -> utf-8<commit_after> | """
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "utf-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
| """
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "UTF-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
Change to preferred encoding style.
UTF-8 -> utf-8"""
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "utf-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
| <commit_before>"""
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "UTF-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
<commit_msg>Change to preferred encoding style.
UTF-8 -> utf-8<commit_after>"""
Serializer plugin interface.
This module is useful for those wanting to write a serializer that can
plugin to rdflib. If you are wanting to invoke a serializer you likely
want to do so through the Graph class serialize method.
TODO: info for how to write a serializer that can plugin to rdflib.
See also rdflib.plugin
"""
from typing import IO, TYPE_CHECKING, Optional
from rdflib.term import URIRef
if TYPE_CHECKING:
from rdflib.graph import Graph
__all__ = ["Serializer"]
class Serializer:
def __init__(self, store: "Graph"):
self.store: "Graph" = store
self.encoding: str = "utf-8"
self.base: Optional[str] = None
def serialize(
self,
stream: IO[bytes],
base: Optional[str] = None,
encoding: Optional[str] = None,
**args
) -> None:
"""Abstract method"""
def relativize(self, uri: str):
base = self.base
if base is not None and uri.startswith(base):
uri = URIRef(uri.replace(base, "", 1))
return uri
|
6e32cfd9b2640b4f119a3a8e4138c883fd4bcef0 | _tests/test_scikit_ci_addons.py | _tests/test_scikit_ci_addons.py |
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = ci_addons.home() + '/' + addon
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci/noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
|
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = os.path.join(ci_addons.home(), addon)
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci' + os.path.sep + 'noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
| Fix failing tests on appveyor | ci: Fix failing tests on appveyor
| Python | apache-2.0 | scikit-build/scikit-ci-addons,scikit-build/scikit-ci-addons |
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = ci_addons.home() + '/' + addon
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci/noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
ci: Fix failing tests on appveyor |
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = os.path.join(ci_addons.home(), addon)
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci' + os.path.sep + 'noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
| <commit_before>
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = ci_addons.home() + '/' + addon
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci/noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
<commit_msg>ci: Fix failing tests on appveyor<commit_after> |
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = os.path.join(ci_addons.home(), addon)
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci' + os.path.sep + 'noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
|
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = ci_addons.home() + '/' + addon
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci/noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
ci: Fix failing tests on appveyor
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = os.path.join(ci_addons.home(), addon)
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci' + os.path.sep + 'noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
| <commit_before>
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = ci_addons.home() + '/' + addon
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci/noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
<commit_msg>ci: Fix failing tests on appveyor<commit_after>
import ci_addons
import os
import pytest
import subprocess
from . import captured_lines
def test_home():
expected_home = os.path.abspath(os.path.dirname(__file__) + '/..')
assert ci_addons.home() == expected_home
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_path(addon):
expected_path = os.path.join(ci_addons.home(), addon)
if not addon.endswith('.py'):
expected_path += '.py'
assert ci_addons.path(addon) == expected_path
def test_list(capsys):
ci_addons.list_addons()
output_lines, _ = captured_lines(capsys)
assert 'anyci' + os.path.sep + 'noop.py' in output_lines
@pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py'])
def test_execute(addon, capfd):
ci_addons.execute(addon, ['foo', 'bar'])
output_lines, _ = captured_lines(capfd)
assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines
def test_cli():
root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
environment = dict(os.environ)
environment['PYTHONPATH'] = root
subprocess.check_call(
"python -m ci_addons",
shell=True,
env=environment,
stderr=subprocess.STDOUT,
cwd=str(root)
)
|
43ab1500719665b44e3b4eca4def9002711c2ee8 | githublist/parser.py | githublist/parser.py | import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
| import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos?per_page=100'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
| Update api url for recent 100 instead of default 30 | Update api url for recent 100 instead of default 30
| Python | mit | kshvmdn/github-list,kshvmdn/github-list,kshvmdn/github-list | import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
Update api url for recent 100 instead of default 30 | import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos?per_page=100'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
| <commit_before>import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
<commit_msg>Update api url for recent 100 instead of default 30<commit_after> | import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos?per_page=100'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
| import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
Update api url for recent 100 instead of default 30import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos?per_page=100'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
| <commit_before>import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
<commit_msg>Update api url for recent 100 instead of default 30<commit_after>import requests
import collections
API_URL = 'https://api.github.com/users/{}/repos?per_page=100'
def main(user):
return parse(request(user))
def request(user):
return requests.get(url=API_URL.format(user))
def parse(response):
repos = response.json()
data = []
if repos is None:
return None
for repo in repos:
if 'name' in repo and not repo['fork']:
data.append(
collections.OrderedDict([('name', repo['name']),
('desc', repo['description']),
('lang', repo['language']),
('stars', repo['stargazers_count'])]))
return data
if __name__ == '__main__':
import pprint
u = 'kshvmdn'
pprint.pprint(main(u))
|
fdf0daefac50de71a8c4f80a9ef877669ebea48b | byceps/services/tourney/transfer/models.py | byceps/services/tourney/transfer/models.py | """
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
| """
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
| Change tourney match transfer model from `attrs` to `dataclass` | Change tourney match transfer model from `attrs` to `dataclass`
| Python | bsd-3-clause | homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps | """
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
Change tourney match transfer model from `attrs` to `dataclass` | """
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
| <commit_before>"""
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
<commit_msg>Change tourney match transfer model from `attrs` to `dataclass`<commit_after> | """
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
| """
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
Change tourney match transfer model from `attrs` to `dataclass`"""
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
| <commit_before>"""
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from typing import NewType
from uuid import UUID
from attr import attrs
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@attrs(auto_attribs=True, frozen=True, slots=True)
class Match:
id: MatchID
<commit_msg>Change tourney match transfer model from `attrs` to `dataclass`<commit_after>"""
byceps.services.tourney.transfer.models
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2019 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from dataclasses import dataclass
from typing import NewType
from uuid import UUID
TourneyCategoryID = NewType('TourneyCategoryID', UUID)
TourneyID = NewType('TourneyID', UUID)
MatchID = NewType('MatchID', UUID)
MatchCommentID = NewType('MatchCommentID', UUID)
ParticipantID = NewType('ParticipantID', UUID)
@dataclass(frozen=True)
class Match:
id: MatchID
|
50519406ac64766874ce9edf5cea69233461ffb2 | tests/test_config.py | tests/test_config.py | # -*- coding: utf-8 -*-
import pytest
import uuid
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.fixture
def dummyvalue():
return str(uuid.uuid4())
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam, dummyvalue):
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
| # -*- coding: utf-8 -*-
import pytest
import uuid
import tempfile
import os
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.yield_fixture
def dummy_config_file():
filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4()))
yield filename
if os.path.isfile(filename):
os.remove(filename)
@pytest.fixture
def custom_config_file(dummy_config_file, scope='module'):
return S3Keyring(profile_name='test', config_file=dummy_config_file).config
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam):
dummyvalue = str(uuid.uuid4())
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
def test_read_custom_config_file(custom_config_file, dummy_config_file):
"""Reads a parameter from a custom config file"""
profile_name = custom_config_file.get('default', 'profile')
assert profile_name == 'default'
assert custom_config_file.config_file == dummy_config_file
assert os.path.isfile(dummy_config_file)
def test_write_config_in_custom_config_file(custom_config_file, dummyparam,
config):
dummyvalue = str(uuid.uuid4())
custom_config_file.set('default', dummyparam, dummyvalue)
custom_config_file.save()
custom_config_file.load()
assert custom_config_file.get('default', dummyparam) == dummyvalue
assert config.config_file != custom_config_file.config_file
config.load()
assert config.get('default', dummyparam) != dummyvalue
| Test custom configuration file feature | Test custom configuration file feature
| Python | mit | InnovativeTravel/s3-keyring | # -*- coding: utf-8 -*-
import pytest
import uuid
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.fixture
def dummyvalue():
return str(uuid.uuid4())
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam, dummyvalue):
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
Test custom configuration file feature | # -*- coding: utf-8 -*-
import pytest
import uuid
import tempfile
import os
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.yield_fixture
def dummy_config_file():
filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4()))
yield filename
if os.path.isfile(filename):
os.remove(filename)
@pytest.fixture
def custom_config_file(dummy_config_file, scope='module'):
return S3Keyring(profile_name='test', config_file=dummy_config_file).config
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam):
dummyvalue = str(uuid.uuid4())
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
def test_read_custom_config_file(custom_config_file, dummy_config_file):
"""Reads a parameter from a custom config file"""
profile_name = custom_config_file.get('default', 'profile')
assert profile_name == 'default'
assert custom_config_file.config_file == dummy_config_file
assert os.path.isfile(dummy_config_file)
def test_write_config_in_custom_config_file(custom_config_file, dummyparam,
config):
dummyvalue = str(uuid.uuid4())
custom_config_file.set('default', dummyparam, dummyvalue)
custom_config_file.save()
custom_config_file.load()
assert custom_config_file.get('default', dummyparam) == dummyvalue
assert config.config_file != custom_config_file.config_file
config.load()
assert config.get('default', dummyparam) != dummyvalue
| <commit_before># -*- coding: utf-8 -*-
import pytest
import uuid
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.fixture
def dummyvalue():
return str(uuid.uuid4())
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam, dummyvalue):
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
<commit_msg>Test custom configuration file feature<commit_after> | # -*- coding: utf-8 -*-
import pytest
import uuid
import tempfile
import os
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.yield_fixture
def dummy_config_file():
filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4()))
yield filename
if os.path.isfile(filename):
os.remove(filename)
@pytest.fixture
def custom_config_file(dummy_config_file, scope='module'):
return S3Keyring(profile_name='test', config_file=dummy_config_file).config
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam):
dummyvalue = str(uuid.uuid4())
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
def test_read_custom_config_file(custom_config_file, dummy_config_file):
"""Reads a parameter from a custom config file"""
profile_name = custom_config_file.get('default', 'profile')
assert profile_name == 'default'
assert custom_config_file.config_file == dummy_config_file
assert os.path.isfile(dummy_config_file)
def test_write_config_in_custom_config_file(custom_config_file, dummyparam,
config):
dummyvalue = str(uuid.uuid4())
custom_config_file.set('default', dummyparam, dummyvalue)
custom_config_file.save()
custom_config_file.load()
assert custom_config_file.get('default', dummyparam) == dummyvalue
assert config.config_file != custom_config_file.config_file
config.load()
assert config.get('default', dummyparam) != dummyvalue
| # -*- coding: utf-8 -*-
import pytest
import uuid
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.fixture
def dummyvalue():
return str(uuid.uuid4())
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam, dummyvalue):
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
Test custom configuration file feature# -*- coding: utf-8 -*-
import pytest
import uuid
import tempfile
import os
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.yield_fixture
def dummy_config_file():
filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4()))
yield filename
if os.path.isfile(filename):
os.remove(filename)
@pytest.fixture
def custom_config_file(dummy_config_file, scope='module'):
return S3Keyring(profile_name='test', config_file=dummy_config_file).config
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam):
dummyvalue = str(uuid.uuid4())
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
def test_read_custom_config_file(custom_config_file, dummy_config_file):
"""Reads a parameter from a custom config file"""
profile_name = custom_config_file.get('default', 'profile')
assert profile_name == 'default'
assert custom_config_file.config_file == dummy_config_file
assert os.path.isfile(dummy_config_file)
def test_write_config_in_custom_config_file(custom_config_file, dummyparam,
config):
dummyvalue = str(uuid.uuid4())
custom_config_file.set('default', dummyparam, dummyvalue)
custom_config_file.save()
custom_config_file.load()
assert custom_config_file.get('default', dummyparam) == dummyvalue
assert config.config_file != custom_config_file.config_file
config.load()
assert config.get('default', dummyparam) != dummyvalue
| <commit_before># -*- coding: utf-8 -*-
import pytest
import uuid
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.fixture
def dummyvalue():
return str(uuid.uuid4())
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam, dummyvalue):
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
<commit_msg>Test custom configuration file feature<commit_after># -*- coding: utf-8 -*-
import pytest
import uuid
import tempfile
import os
from s3keyring.s3 import S3Keyring
@pytest.fixture
def config(scope='module'):
return S3Keyring(profile_name='test').config
@pytest.yield_fixture
def dummyparam(config, scope='module'):
yield 'dummyparam'
config.config.remove_option('default', 'dummyparam')
@pytest.yield_fixture
def dummy_config_file():
filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4()))
yield filename
if os.path.isfile(filename):
os.remove(filename)
@pytest.fixture
def custom_config_file(dummy_config_file, scope='module'):
return S3Keyring(profile_name='test', config_file=dummy_config_file).config
def test_read_config(config):
"""Sets value for an existing configuration option"""
profile_name = config.get('default', 'profile')
assert profile_name == 'default'
def test_write_config(config, dummyparam):
dummyvalue = str(uuid.uuid4())
config.set('default', dummyparam, dummyvalue)
config.save()
config.load()
assert config.get('default', dummyparam) == dummyvalue
def test_read_custom_config_file(custom_config_file, dummy_config_file):
"""Reads a parameter from a custom config file"""
profile_name = custom_config_file.get('default', 'profile')
assert profile_name == 'default'
assert custom_config_file.config_file == dummy_config_file
assert os.path.isfile(dummy_config_file)
def test_write_config_in_custom_config_file(custom_config_file, dummyparam,
config):
dummyvalue = str(uuid.uuid4())
custom_config_file.set('default', dummyparam, dummyvalue)
custom_config_file.save()
custom_config_file.load()
assert custom_config_file.get('default', dummyparam) == dummyvalue
assert config.config_file != custom_config_file.config_file
config.load()
assert config.get('default', dummyparam) != dummyvalue
|
025927fa19bb96095a2ea1c53524945f1f9590ce | spur/results.py | spur/results.py | def result(return_code, output, stderr_output, allow_error=False):
if allow_error or return_code == 0:
return ExecutionResult(return_code, output, stderr_output)
else:
raise RunProcessError(return_code, output, stderr_output)
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
| def result(return_code, output, stderr_output, allow_error=False):
result = ExecutionResult(return_code, output, stderr_output)
if allow_error or return_code == 0:
return result
else:
raise result.to_error()
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
def to_error(self):
return RunProcessError(
self.return_code,
self.output,
self.stderr_output
)
| Move logic for creating RunProcessError to ExecutionResult.to_error | Move logic for creating RunProcessError to ExecutionResult.to_error
| Python | bsd-2-clause | mwilliamson/spur.py | def result(return_code, output, stderr_output, allow_error=False):
if allow_error or return_code == 0:
return ExecutionResult(return_code, output, stderr_output)
else:
raise RunProcessError(return_code, output, stderr_output)
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
Move logic for creating RunProcessError to ExecutionResult.to_error | def result(return_code, output, stderr_output, allow_error=False):
result = ExecutionResult(return_code, output, stderr_output)
if allow_error or return_code == 0:
return result
else:
raise result.to_error()
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
def to_error(self):
return RunProcessError(
self.return_code,
self.output,
self.stderr_output
)
| <commit_before>def result(return_code, output, stderr_output, allow_error=False):
if allow_error or return_code == 0:
return ExecutionResult(return_code, output, stderr_output)
else:
raise RunProcessError(return_code, output, stderr_output)
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
<commit_msg>Move logic for creating RunProcessError to ExecutionResult.to_error<commit_after> | def result(return_code, output, stderr_output, allow_error=False):
result = ExecutionResult(return_code, output, stderr_output)
if allow_error or return_code == 0:
return result
else:
raise result.to_error()
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
def to_error(self):
return RunProcessError(
self.return_code,
self.output,
self.stderr_output
)
| def result(return_code, output, stderr_output, allow_error=False):
if allow_error or return_code == 0:
return ExecutionResult(return_code, output, stderr_output)
else:
raise RunProcessError(return_code, output, stderr_output)
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
Move logic for creating RunProcessError to ExecutionResult.to_errordef result(return_code, output, stderr_output, allow_error=False):
result = ExecutionResult(return_code, output, stderr_output)
if allow_error or return_code == 0:
return result
else:
raise result.to_error()
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
def to_error(self):
return RunProcessError(
self.return_code,
self.output,
self.stderr_output
)
| <commit_before>def result(return_code, output, stderr_output, allow_error=False):
if allow_error or return_code == 0:
return ExecutionResult(return_code, output, stderr_output)
else:
raise RunProcessError(return_code, output, stderr_output)
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
<commit_msg>Move logic for creating RunProcessError to ExecutionResult.to_error<commit_after>def result(return_code, output, stderr_output, allow_error=False):
result = ExecutionResult(return_code, output, stderr_output)
if allow_error or return_code == 0:
return result
else:
raise result.to_error()
class RunProcessError(RuntimeError):
def __init__(self, return_code, output, stderr_output):
message = "return code: {0}\noutput: {1}\nstderr output: {2}".format(
return_code, output, stderr_output)
super(type(self), self).__init__(message)
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
class ExecutionResult(object):
def __init__(self, return_code, output, stderr_output):
self.return_code = return_code
self.output = output
self.stderr_output = stderr_output
def to_error(self):
return RunProcessError(
self.return_code,
self.output,
self.stderr_output
)
|
936db17eed36284917395a6a8272351dabbc8168 | numpy/_array_api/_dtypes.py | numpy/_array_api/_dtypes.py | from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64
# Note: This name is changed
from .. import bool_ as bool
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
| import numpy as np
# Note: we use dtype objects instead of dtype classes. The spec does not
# require any behavior on dtypes other than equality.
int8 = np.dtype('int8')
int16 = np.dtype('int16')
int32 = np.dtype('int32')
int64 = np.dtype('int64')
uint8 = np.dtype('uint8')
uint16 = np.dtype('uint16')
uint32 = np.dtype('uint32')
uint64 = np.dtype('uint64')
float32 = np.dtype('float32')
float64 = np.dtype('float64')
# Note: This name is changed
bool = np.dtype('bool')
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
| Use dtype objects instead of classes in the array API | Use dtype objects instead of classes in the array API
The array API does not require any methods or behaviors on dtype objects,
other than that they be literals that can be compared for equality and passed
to dtype keywords in functions. Since dtype objects are already used by the
dtype attribute of ndarray, this makes it consistent, so that
func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the
same thing as numpy._array_api.<dtype>. This also fixes an issue in the array
API test suite due to the fact that dtype classes and objects are not equal as
dictionary keys.
| Python | mit | cupy/cupy,cupy/cupy,cupy/cupy,cupy/cupy | from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64
# Note: This name is changed
from .. import bool_ as bool
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
Use dtype objects instead of classes in the array API
The array API does not require any methods or behaviors on dtype objects,
other than that they be literals that can be compared for equality and passed
to dtype keywords in functions. Since dtype objects are already used by the
dtype attribute of ndarray, this makes it consistent, so that
func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the
same thing as numpy._array_api.<dtype>. This also fixes an issue in the array
API test suite due to the fact that dtype classes and objects are not equal as
dictionary keys. | import numpy as np
# Note: we use dtype objects instead of dtype classes. The spec does not
# require any behavior on dtypes other than equality.
int8 = np.dtype('int8')
int16 = np.dtype('int16')
int32 = np.dtype('int32')
int64 = np.dtype('int64')
uint8 = np.dtype('uint8')
uint16 = np.dtype('uint16')
uint32 = np.dtype('uint32')
uint64 = np.dtype('uint64')
float32 = np.dtype('float32')
float64 = np.dtype('float64')
# Note: This name is changed
bool = np.dtype('bool')
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
| <commit_before>from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64
# Note: This name is changed
from .. import bool_ as bool
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
<commit_msg>Use dtype objects instead of classes in the array API
The array API does not require any methods or behaviors on dtype objects,
other than that they be literals that can be compared for equality and passed
to dtype keywords in functions. Since dtype objects are already used by the
dtype attribute of ndarray, this makes it consistent, so that
func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the
same thing as numpy._array_api.<dtype>. This also fixes an issue in the array
API test suite due to the fact that dtype classes and objects are not equal as
dictionary keys.<commit_after> | import numpy as np
# Note: we use dtype objects instead of dtype classes. The spec does not
# require any behavior on dtypes other than equality.
int8 = np.dtype('int8')
int16 = np.dtype('int16')
int32 = np.dtype('int32')
int64 = np.dtype('int64')
uint8 = np.dtype('uint8')
uint16 = np.dtype('uint16')
uint32 = np.dtype('uint32')
uint64 = np.dtype('uint64')
float32 = np.dtype('float32')
float64 = np.dtype('float64')
# Note: This name is changed
bool = np.dtype('bool')
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
| from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64
# Note: This name is changed
from .. import bool_ as bool
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
Use dtype objects instead of classes in the array API
The array API does not require any methods or behaviors on dtype objects,
other than that they be literals that can be compared for equality and passed
to dtype keywords in functions. Since dtype objects are already used by the
dtype attribute of ndarray, this makes it consistent, so that
func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the
same thing as numpy._array_api.<dtype>. This also fixes an issue in the array
API test suite due to the fact that dtype classes and objects are not equal as
dictionary keys.import numpy as np
# Note: we use dtype objects instead of dtype classes. The spec does not
# require any behavior on dtypes other than equality.
int8 = np.dtype('int8')
int16 = np.dtype('int16')
int32 = np.dtype('int32')
int64 = np.dtype('int64')
uint8 = np.dtype('uint8')
uint16 = np.dtype('uint16')
uint32 = np.dtype('uint32')
uint64 = np.dtype('uint64')
float32 = np.dtype('float32')
float64 = np.dtype('float64')
# Note: This name is changed
bool = np.dtype('bool')
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
| <commit_before>from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64
# Note: This name is changed
from .. import bool_ as bool
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
<commit_msg>Use dtype objects instead of classes in the array API
The array API does not require any methods or behaviors on dtype objects,
other than that they be literals that can be compared for equality and passed
to dtype keywords in functions. Since dtype objects are already used by the
dtype attribute of ndarray, this makes it consistent, so that
func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the
same thing as numpy._array_api.<dtype>. This also fixes an issue in the array
API test suite due to the fact that dtype classes and objects are not equal as
dictionary keys.<commit_after>import numpy as np
# Note: we use dtype objects instead of dtype classes. The spec does not
# require any behavior on dtypes other than equality.
int8 = np.dtype('int8')
int16 = np.dtype('int16')
int32 = np.dtype('int32')
int64 = np.dtype('int64')
uint8 = np.dtype('uint8')
uint16 = np.dtype('uint16')
uint32 = np.dtype('uint32')
uint64 = np.dtype('uint64')
float32 = np.dtype('float32')
float64 = np.dtype('float64')
# Note: This name is changed
bool = np.dtype('bool')
_all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64,
float32, float64, bool]
_boolean_dtypes = [bool]
_floating_dtypes = [float32, float64]
_integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
_numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
|
ffd4c52155acd7d04939e766ebe63171b580a2fa | src/__init__.py | src/__init__.py | import os
import logging
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect' ]
# connected client object
_client = None
def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
if epgdb.find(':') >= 0:
# epg is remote: host:port
# TODO: create socket, pass it to client
_client = GuideClient("epg")
else:
# epg is local
_client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg")
return _client
| import os
import logging
from socket import gethostbyname, gethostname
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ]
# connected client object
_client = None
def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
if _client:
return _client
if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \
address != gethostbyname(gethostname()):
# epg is remote: host:port
if address.find(':') >= 0:
host, port = address.split(':', 1)
else:
host = address
port = DEFAULT_EPG_PORT
# create socket, pass it to client
_client = GuideClient((host, port))
else:
# EPG is local, only use unix socket
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
_client = ipc.launch([server, logfile, str(loglevel), epgdb, address],
2, GuideClient, "epg")
return _client
| Add the ability to use inet socket as well. | Add the ability to use inet socket as well.
git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e
| Python | lgpl-2.1 | freevo/kaa-epg | import os
import logging
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect' ]
# connected client object
_client = None
def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
if epgdb.find(':') >= 0:
# epg is remote: host:port
# TODO: create socket, pass it to client
_client = GuideClient("epg")
else:
# epg is local
_client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg")
return _client
Add the ability to use inet socket as well.
git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e | import os
import logging
from socket import gethostbyname, gethostname
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ]
# connected client object
_client = None
def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
if _client:
return _client
if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \
address != gethostbyname(gethostname()):
# epg is remote: host:port
if address.find(':') >= 0:
host, port = address.split(':', 1)
else:
host = address
port = DEFAULT_EPG_PORT
# create socket, pass it to client
_client = GuideClient((host, port))
else:
# EPG is local, only use unix socket
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
_client = ipc.launch([server, logfile, str(loglevel), epgdb, address],
2, GuideClient, "epg")
return _client
| <commit_before>import os
import logging
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect' ]
# connected client object
_client = None
def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
if epgdb.find(':') >= 0:
# epg is remote: host:port
# TODO: create socket, pass it to client
_client = GuideClient("epg")
else:
# epg is local
_client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg")
return _client
<commit_msg>Add the ability to use inet socket as well.
git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e<commit_after> | import os
import logging
from socket import gethostbyname, gethostname
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ]
# connected client object
_client = None
def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
if _client:
return _client
if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \
address != gethostbyname(gethostname()):
# epg is remote: host:port
if address.find(':') >= 0:
host, port = address.split(':', 1)
else:
host = address
port = DEFAULT_EPG_PORT
# create socket, pass it to client
_client = GuideClient((host, port))
else:
# EPG is local, only use unix socket
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
_client = ipc.launch([server, logfile, str(loglevel), epgdb, address],
2, GuideClient, "epg")
return _client
| import os
import logging
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect' ]
# connected client object
_client = None
def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
if epgdb.find(':') >= 0:
# epg is remote: host:port
# TODO: create socket, pass it to client
_client = GuideClient("epg")
else:
# epg is local
_client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg")
return _client
Add the ability to use inet socket as well.
git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8eimport os
import logging
from socket import gethostbyname, gethostname
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ]
# connected client object
_client = None
def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
if _client:
return _client
if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \
address != gethostbyname(gethostname()):
# epg is remote: host:port
if address.find(':') >= 0:
host, port = address.split(':', 1)
else:
host = address
port = DEFAULT_EPG_PORT
# create socket, pass it to client
_client = GuideClient((host, port))
else:
# EPG is local, only use unix socket
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
_client = ipc.launch([server, logfile, str(loglevel), epgdb, address],
2, GuideClient, "epg")
return _client
| <commit_before>import os
import logging
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect' ]
# connected client object
_client = None
def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
if epgdb.find(':') >= 0:
# epg is remote: host:port
# TODO: create socket, pass it to client
_client = GuideClient("epg")
else:
# epg is local
_client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg")
return _client
<commit_msg>Add the ability to use inet socket as well.
git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e<commit_after>import os
import logging
from socket import gethostbyname, gethostname
from kaa.base import ipc
from client import *
from server import *
__all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ]
# connected client object
_client = None
def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO):
"""
"""
global _client
if _client:
return _client
if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \
address != gethostbyname(gethostname()):
# epg is remote: host:port
if address.find(':') >= 0:
host, port = address.split(':', 1)
else:
host = address
port = DEFAULT_EPG_PORT
# create socket, pass it to client
_client = GuideClient((host, port))
else:
# EPG is local, only use unix socket
# get server filename
server = os.path.join(os.path.dirname(__file__), 'server.py')
_client = ipc.launch([server, logfile, str(loglevel), epgdb, address],
2, GuideClient, "epg")
return _client
|
9d960bfa74a09382839f9b671004bebaffe46611 | reui/Screen.py | reui/Screen.py | """
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
| """
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
def draw_pixel(self, box, x, y, on=True):
bx, by = self._boxMap[box]
self._bitmap.draw_pixel(bx + x, by + y, on)
def draw_text(self, box, x, y, string, font=box.font):
bx, by = self._boxMap[box]
self._bitmap.draw_text(bx + x, by + y, string, font)
| Support for Box direct drawing to screen bitmap | Support for Box direct drawing to screen bitmap | Python | mit | mharriger/reui | """
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
Support for Box direct drawing to screen bitmap | """
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
def draw_pixel(self, box, x, y, on=True):
bx, by = self._boxMap[box]
self._bitmap.draw_pixel(bx + x, by + y, on)
def draw_text(self, box, x, y, string, font=box.font):
bx, by = self._boxMap[box]
self._bitmap.draw_text(bx + x, by + y, string, font)
| <commit_before>"""
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
<commit_msg>Support for Box direct drawing to screen bitmap<commit_after> | """
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
def draw_pixel(self, box, x, y, on=True):
bx, by = self._boxMap[box]
self._bitmap.draw_pixel(bx + x, by + y, on)
def draw_text(self, box, x, y, string, font=box.font):
bx, by = self._boxMap[box]
self._bitmap.draw_text(bx + x, by + y, string, font)
| """
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
Support for Box direct drawing to screen bitmap"""
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
def draw_pixel(self, box, x, y, on=True):
bx, by = self._boxMap[box]
self._bitmap.draw_pixel(bx + x, by + y, on)
def draw_text(self, box, x, y, string, font=box.font):
bx, by = self._boxMap[box]
self._bitmap.draw_text(bx + x, by + y, string, font)
| <commit_before>"""
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
<commit_msg>Support for Box direct drawing to screen bitmap<commit_after>"""
A screen object contains a collection of boxes to be displayed on a
physical display device.
"""
from pydispatch import dispatcher
from reui import SGL_BOX_UPDATE
from gaugette import bitmap
class Screen:
_boxes = []
_boxMap = {}
_bitmap = None
def __init__(self, width, height, display):
#self._bitmap = bitmap.Bitmap(width, height, 'y')
self._display = display
#self._display.bitmap = self._bitmap
self._bitmap = self._display.bitmap
def add_box(self, x, y, box):
self._boxes.append((x, y, box))
self._boxMap[box] = (x, y)
dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE,
sender=box)
def draw(self):
for x, y, box in self._boxes:
self._bitmap.replace_rect(x, y, box._bitmap)
def on_box_update(self, **args):
if 'sender' in args:
sender = args['sender']
if sender in self._boxMap:
(x, y) = self._boxMap[sender]
self._bitmap.replace_rect(x, y, sender._bitmap)
self._display.display()
def draw_pixel(self, box, x, y, on=True):
bx, by = self._boxMap[box]
self._bitmap.draw_pixel(bx + x, by + y, on)
def draw_text(self, box, x, y, string, font=box.font):
bx, by = self._boxMap[box]
self._bitmap.draw_text(bx + x, by + y, string, font)
|
f087ea792b1e093e6ed49e3dd3b647f2f8276f64 | acme/_metadata.py | acme/_metadata.py | # python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '0'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
| # python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '1'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
| Update Acme version to 0.2.1. | Update Acme version to 0.2.1.
PiperOrigin-RevId: 375471102
Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4
| Python | apache-2.0 | deepmind/acme,deepmind/acme | # python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '0'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
Update Acme version to 0.2.1.
PiperOrigin-RevId: 375471102
Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4 | # python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '1'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
| <commit_before># python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '0'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
<commit_msg>Update Acme version to 0.2.1.
PiperOrigin-RevId: 375471102
Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4<commit_after> | # python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '1'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
| # python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '0'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
Update Acme version to 0.2.1.
PiperOrigin-RevId: 375471102
Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4# python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '1'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
| <commit_before># python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '0'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
<commit_msg>Update Acme version to 0.2.1.
PiperOrigin-RevId: 375471102
Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4<commit_after># python3
# Copyright 2018 DeepMind Technologies Limited. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""Package metadata for acme.
This is kept in a separate module so that it can be imported from setup.py, at
a time when acme's dependencies may not have been installed yet.
"""
# We follow Semantic Versioning (https://semver.org/)
_MAJOR_VERSION = '0'
_MINOR_VERSION = '2'
_PATCH_VERSION = '1'
# Example: '0.4.2'
__version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
|
0cf0f3de5879795fcd01b8d88bf11efb3362f530 | script/echo.py | script/echo.py | #!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
| #!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(keepalive=True, post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
| Make example bot use keepalive | [Instabot] Make example bot use keepalive
| Python | mit | CylonicRaider/Instant,CylonicRaider/Instant,CylonicRaider/Instant,CylonicRaider/Instant,CylonicRaider/Instant | #!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
[Instabot] Make example bot use keepalive | #!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(keepalive=True, post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
| <commit_before>#!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
<commit_msg>[Instabot] Make example bot use keepalive<commit_after> | #!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(keepalive=True, post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
| #!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
[Instabot] Make example bot use keepalive#!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(keepalive=True, post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
| <commit_before>#!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
<commit_msg>[Instabot] Make example bot use keepalive<commit_after>#!/usr/bin/env python3
# -*- coding: ascii -*-
# A small example bot for Instant.
import sys
import instabot
NICKNAME = 'Echo'
def post_cb(self, msg, meta):
if msg['text'].startswith('!echo '):
return msg['text'][6:]
def main():
b = instabot.CmdlineBotBuilder(defnick=NICKNAME)
b.make_parser(sys.argv[0],
desc='An Instant bot bouncing back received messages.')
b.parse(sys.argv[1:])
bot = b(keepalive=True, post_cb=post_cb)
try:
bot.run()
except KeyboardInterrupt:
sys.stderr.write('\n')
finally:
bot.close()
if __name__ == '__main__': main()
|
d90d35063f1a79916c20d32d3634842dd59798f1 | api/tests/conftest.py | api/tests/conftest.py | import pytest
@pytest.fixture(scope='module')
def app():
from api import app
return app
| import pytest
@pytest.fixture(scope='module')
def app():
from api import app, db
app.config['TESTING'] = True
db.create_all()
return app
| Fix default fixture to initialize database | Fix default fixture to initialize database
| Python | mit | Demotivated/loadstone | import pytest
@pytest.fixture(scope='module')
def app():
from api import app
return app
Fix default fixture to initialize database | import pytest
@pytest.fixture(scope='module')
def app():
from api import app, db
app.config['TESTING'] = True
db.create_all()
return app
| <commit_before>import pytest
@pytest.fixture(scope='module')
def app():
from api import app
return app
<commit_msg>Fix default fixture to initialize database<commit_after> | import pytest
@pytest.fixture(scope='module')
def app():
from api import app, db
app.config['TESTING'] = True
db.create_all()
return app
| import pytest
@pytest.fixture(scope='module')
def app():
from api import app
return app
Fix default fixture to initialize databaseimport pytest
@pytest.fixture(scope='module')
def app():
from api import app, db
app.config['TESTING'] = True
db.create_all()
return app
| <commit_before>import pytest
@pytest.fixture(scope='module')
def app():
from api import app
return app
<commit_msg>Fix default fixture to initialize database<commit_after>import pytest
@pytest.fixture(scope='module')
def app():
from api import app, db
app.config['TESTING'] = True
db.create_all()
return app
|
6decf1f48e56832b1d15d3fc26d92f9813d13353 | coop_cms/moves.py | coop_cms/moves.py | # -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| # -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| Fix HTMLParser compatibility in Python 3 | Fix HTMLParser compatibility in Python 3
| Python | bsd-3-clause | ljean/coop_cms,ljean/coop_cms,ljean/coop_cms | # -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
Fix HTMLParser compatibility in Python 3 | # -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| <commit_before># -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
<commit_msg>Fix HTMLParser compatibility in Python 3<commit_after> | # -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| # -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
Fix HTMLParser compatibility in Python 3# -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
| <commit_before># -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from HTMLParser import HTMLParser
from StringIO import StringIO
else:
# Python 3
from html.parser import HTMLParser
from io import BytesIO as StringIO
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
<commit_msg>Fix HTMLParser compatibility in Python 3<commit_after># -*- coding: utf-8 -*-
"""
coop_cms manage compatibilty with django and python versions
"""
import sys
from django import VERSION
if sys.version_info[0] < 3:
# Python 2
from StringIO import StringIO
from HTMLParser import HTMLParser
else:
# Python 3
from io import BytesIO as StringIO
from html.parser import HTMLParser as BaseHTMLParser
class HTMLParser(BaseHTMLParser):
def __init__(self):
BaseHTMLParser.__init__(self, convert_charrefs=False)
try:
from django.utils.deprecation import MiddlewareMixin
except ImportError:
MiddlewareMixin = object
if VERSION >= (1, 9, 0):
from wsgiref.util import FileWrapper
else:
from django.core.servers.basehttp import FileWrapper
if VERSION >= (1, 8, 0):
from unittest import SkipTest
else:
# Deprecated in Django 1.9
from django.utils.unittest import SkipTest
def make_context(request, context_dict):
""""""
if VERSION >= (1, 9, 0):
context = dict(context_dict)
if request:
context['request'] = request
else:
from django.template import RequestContext, Context
if request:
context = RequestContext(request, context_dict)
else:
context = Context(context_dict)
return context
|
9931bd1d5459a983717fb502826f3cca87225b96 | src/qrl/services/grpcHelper.py | src/qrl/services/grpcHelper.py | # coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
| # coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
context.set_code(StatusCode.INVALID_ARGUMENT)
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
| Set code to Invalid argument for ValueErrors | Set code to Invalid argument for ValueErrors
| Python | mit | jleni/QRL,cyyber/QRL,jleni/QRL,cyyber/QRL,theQRL/QRL,randomshinichi/QRL,theQRL/QRL,randomshinichi/QRL | # coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
Set code to Invalid argument for ValueErrors | # coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
context.set_code(StatusCode.INVALID_ARGUMENT)
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
| <commit_before># coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
<commit_msg>Set code to Invalid argument for ValueErrors<commit_after> | # coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
context.set_code(StatusCode.INVALID_ARGUMENT)
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
| # coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
Set code to Invalid argument for ValueErrors# coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
context.set_code(StatusCode.INVALID_ARGUMENT)
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
| <commit_before># coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
<commit_msg>Set code to Invalid argument for ValueErrors<commit_after># coding=utf-8
# Distributed under the MIT software license, see the accompanying
# file LICENSE or http://www.opensource.org/licenses/mit-license.php.
from grpc import StatusCode
from qrl.core.misc import logger
class GrpcExceptionWrapper(object):
def __init__(self, response_type, state_code=StatusCode.UNKNOWN):
self.response_type = response_type
self.state_code = state_code
def _set_context(self, context, exception):
if context is not None:
context.set_code(self.state_code)
context.set_details(str(exception))
def __call__(self, f):
def wrap_f(caller_self, request, context):
try:
return f(caller_self, request, context)
except ValueError as e:
context.set_code(StatusCode.INVALID_ARGUMENT)
self._set_context(context, e)
logger.info(str(e))
return self.response_type()
except Exception as e:
self._set_context(context, e)
logger.exception(e)
return self.response_type()
return wrap_f
|
ce1fb05e825e9be7589fd12ab798cae760b605e6 | sheldon/bot.py | sheldon/bot.py | # -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self):
"""
Function for loading bot.
:return:
"""
# Creating empty lists for plugins and adapters
self.plugins = []
self.adapters = []
| # -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self, command_line_arguments):
"""
Function for loading bot.
:param command_line_arguments: dict, arguments for start script
:return:
"""
self._load_config(command_line_arguments)
def _load_config(self, command_line_arguments):
"""
Сreate and load bot config.
:param command_line_arguments: dict, arguments for creating config:
config-prefix - prefix of environment
variables.
Default - 'SHELDON_'
:return:
"""
# Config class is imported from sheldon.config
if 'config-prefix' in command_line_arguments:
self.config = Config(prefix=command_line_arguments['config-prefix'])
else:
self.config = Config()
| Add basic load config function | Add basic load config function
| Python | mit | lises/sheldon | # -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self):
"""
Function for loading bot.
:return:
"""
# Creating empty lists for plugins and adapters
self.plugins = []
self.adapters = []
Add basic load config function | # -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self, command_line_arguments):
"""
Function for loading bot.
:param command_line_arguments: dict, arguments for start script
:return:
"""
self._load_config(command_line_arguments)
def _load_config(self, command_line_arguments):
"""
Сreate and load bot config.
:param command_line_arguments: dict, arguments for creating config:
config-prefix - prefix of environment
variables.
Default - 'SHELDON_'
:return:
"""
# Config class is imported from sheldon.config
if 'config-prefix' in command_line_arguments:
self.config = Config(prefix=command_line_arguments['config-prefix'])
else:
self.config = Config()
| <commit_before># -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self):
"""
Function for loading bot.
:return:
"""
# Creating empty lists for plugins and adapters
self.plugins = []
self.adapters = []
<commit_msg>Add basic load config function<commit_after> | # -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self, command_line_arguments):
"""
Function for loading bot.
:param command_line_arguments: dict, arguments for start script
:return:
"""
self._load_config(command_line_arguments)
def _load_config(self, command_line_arguments):
"""
Сreate and load bot config.
:param command_line_arguments: dict, arguments for creating config:
config-prefix - prefix of environment
variables.
Default - 'SHELDON_'
:return:
"""
# Config class is imported from sheldon.config
if 'config-prefix' in command_line_arguments:
self.config = Config(prefix=command_line_arguments['config-prefix'])
else:
self.config = Config()
| # -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self):
"""
Function for loading bot.
:return:
"""
# Creating empty lists for plugins and adapters
self.plugins = []
self.adapters = []
Add basic load config function# -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self, command_line_arguments):
"""
Function for loading bot.
:param command_line_arguments: dict, arguments for start script
:return:
"""
self._load_config(command_line_arguments)
def _load_config(self, command_line_arguments):
"""
Сreate and load bot config.
:param command_line_arguments: dict, arguments for creating config:
config-prefix - prefix of environment
variables.
Default - 'SHELDON_'
:return:
"""
# Config class is imported from sheldon.config
if 'config-prefix' in command_line_arguments:
self.config = Config(prefix=command_line_arguments['config-prefix'])
else:
self.config = Config()
| <commit_before># -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self):
"""
Function for loading bot.
:return:
"""
# Creating empty lists for plugins and adapters
self.plugins = []
self.adapters = []
<commit_msg>Add basic load config function<commit_after># -*- coding: utf-8 -*-
"""
@author: Seva Zhidkov
@contact: zhidkovseva@gmail.com
@license: The MIT license
Copyright (C) 2015
"""
from sheldon.exceptions import *
from sheldon.manager import *
from sheldon.config import *
from sheldon.adapter import *
from sheldon.storage import *
class Sheldon:
"""
Main class of the bot.
Run script creating new instance of this class and run it.
"""
def __init__(self, command_line_arguments):
"""
Function for loading bot.
:param command_line_arguments: dict, arguments for start script
:return:
"""
self._load_config(command_line_arguments)
def _load_config(self, command_line_arguments):
"""
Сreate and load bot config.
:param command_line_arguments: dict, arguments for creating config:
config-prefix - prefix of environment
variables.
Default - 'SHELDON_'
:return:
"""
# Config class is imported from sheldon.config
if 'config-prefix' in command_line_arguments:
self.config = Config(prefix=command_line_arguments['config-prefix'])
else:
self.config = Config()
|
74f82029223cc541beab98d7026abb1ec992be40 | createTodoFile.py | createTodoFile.py | """createTodoFile.py: Creates an todo file with title name as current date"""
import time
import os.path
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
else:
print("ERROR: " + filename + " already exist! Exiting..")
# TODO: To move files into archive if more than a week
def archiveFiles():
pass
def main():
createfile()
if __name__ == '__main__':
main()
| """createTodoFile.py: Creates an todo file with title name as current date"""
import os.path
import time
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
addfileToSublime(filename)
else:
print("ERROR: " + filename + " already exist! Exiting..")
def addfileToSublime(file):
os.system("subl --add " + file)
def main():
createfile()
if __name__ == '__main__':
main()
| Add created file to sublime | feat: Add created file to sublime
| Python | mit | prajesh-ananthan/Tools | """createTodoFile.py: Creates an todo file with title name as current date"""
import time
import os.path
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
else:
print("ERROR: " + filename + " already exist! Exiting..")
# TODO: To move files into archive if more than a week
def archiveFiles():
pass
def main():
createfile()
if __name__ == '__main__':
main()
feat: Add created file to sublime | """createTodoFile.py: Creates an todo file with title name as current date"""
import os.path
import time
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
addfileToSublime(filename)
else:
print("ERROR: " + filename + " already exist! Exiting..")
def addfileToSublime(file):
os.system("subl --add " + file)
def main():
createfile()
if __name__ == '__main__':
main()
| <commit_before>"""createTodoFile.py: Creates an todo file with title name as current date"""
import time
import os.path
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
else:
print("ERROR: " + filename + " already exist! Exiting..")
# TODO: To move files into archive if more than a week
def archiveFiles():
pass
def main():
createfile()
if __name__ == '__main__':
main()
<commit_msg>feat: Add created file to sublime<commit_after> | """createTodoFile.py: Creates an todo file with title name as current date"""
import os.path
import time
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
addfileToSublime(filename)
else:
print("ERROR: " + filename + " already exist! Exiting..")
def addfileToSublime(file):
os.system("subl --add " + file)
def main():
createfile()
if __name__ == '__main__':
main()
| """createTodoFile.py: Creates an todo file with title name as current date"""
import time
import os.path
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
else:
print("ERROR: " + filename + " already exist! Exiting..")
# TODO: To move files into archive if more than a week
def archiveFiles():
pass
def main():
createfile()
if __name__ == '__main__':
main()
feat: Add created file to sublime"""createTodoFile.py: Creates an todo file with title name as current date"""
import os.path
import time
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
addfileToSublime(filename)
else:
print("ERROR: " + filename + " already exist! Exiting..")
def addfileToSublime(file):
os.system("subl --add " + file)
def main():
createfile()
if __name__ == '__main__':
main()
| <commit_before>"""createTodoFile.py: Creates an todo file with title name as current date"""
import time
import os.path
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
else:
print("ERROR: " + filename + " already exist! Exiting..")
# TODO: To move files into archive if more than a week
def archiveFiles():
pass
def main():
createfile()
if __name__ == '__main__':
main()
<commit_msg>feat: Add created file to sublime<commit_after>"""createTodoFile.py: Creates an todo file with title name as current date"""
import os.path
import time
def createfile():
# My-File--2009-12-31--23-59-59.txt
date = time.strftime("%d-%m-%Y")
filename = "GOALS--" + date + ".todo"
if not os.path.exists(filename):
with open(filename, "a") as myfile:
myfile.write("[RESULTS - {}]".format(date))
print("INFO: " + filename + " created!")
addfileToSublime(filename)
else:
print("ERROR: " + filename + " already exist! Exiting..")
def addfileToSublime(file):
os.system("subl --add " + file)
def main():
createfile()
if __name__ == '__main__':
main()
|
0f114a144268bb611ff00db9917756a8c02f84b9 | project/api/signals.py | project/api/signals.py | # Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff and not instance.person:
activate_user(instance)
return
| # Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import activate_user
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff and not instance.person:
activate_user(instance)
return
| Connect person to user account | Connect person to user account
| Python | bsd-2-clause | barberscore/barberscore-api,dbinetti/barberscore-django,barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,dbinetti/barberscore,dbinetti/barberscore,barberscore/barberscore-api | # Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff and not instance.person:
activate_user(instance)
return
Connect person to user account | # Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import activate_user
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff and not instance.person:
activate_user(instance)
return
| <commit_before># Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff and not instance.person:
activate_user(instance)
return
<commit_msg>Connect person to user account<commit_after> | # Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import activate_user
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff and not instance.person:
activate_user(instance)
return
| # Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff and not instance.person:
activate_user(instance)
return
Connect person to user account# Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import activate_user
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff and not instance.person:
activate_user(instance)
return
| <commit_before># Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
if settings.DJANGO_SETTINGS_MODULE == 'settings.prod':
if not instance.is_staff and not instance.person:
activate_user(instance)
return
<commit_msg>Connect person to user account<commit_after># Django
# Third-Party
from django.db.models.signals import pre_delete
from django.db.models.signals import pre_save
from django.dispatch import receiver
from django.conf import settings
# Local
from .models import User
from .tasks import activate_user
from .tasks import delete_account
@receiver(pre_delete, sender=User)
def user_pre_delete(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff:
delete_account(instance)
return
@receiver(pre_save, sender=User)
def user_pre_save(sender, instance, **kwargs):
allowed = any([
settings.DJANGO_SETTINGS_MODULE == 'settings.prod',
settings.DJANGO_SETTINGS_MODULE == 'settings.dev',
])
if allowed:
if not instance.is_staff and not instance.person:
activate_user(instance)
return
|
9568efceab48f87ed8302ec4f9bad4b15aac4c5a | tests/test_action.py | tests/test_action.py | import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
| import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
def test_connection_closed_after_sending_mail(self, mock_smtp_class):
mock_smtp = mock_smtp_class.return_value
self.action.execute("MSFT has crossed $10 price level")
mock_smtp.send_message.assert_called_with(mock.ANY)
self.assertTrue(mock_smtp.quit.called)
mock_smtp.assert_has_calls([
mock.call.send_message(mock.ANY),
mock.call.quit()
])
| Add test to check if connection is closed after email is sent. | Add test to check if connection is closed after email is sent.
| Python | mit | bsmukasa/stock_alerter | import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
Add test to check if connection is closed after email is sent. | import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
def test_connection_closed_after_sending_mail(self, mock_smtp_class):
mock_smtp = mock_smtp_class.return_value
self.action.execute("MSFT has crossed $10 price level")
mock_smtp.send_message.assert_called_with(mock.ANY)
self.assertTrue(mock_smtp.quit.called)
mock_smtp.assert_has_calls([
mock.call.send_message(mock.ANY),
mock.call.quit()
])
| <commit_before>import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
<commit_msg>Add test to check if connection is closed after email is sent.<commit_after> | import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
def test_connection_closed_after_sending_mail(self, mock_smtp_class):
mock_smtp = mock_smtp_class.return_value
self.action.execute("MSFT has crossed $10 price level")
mock_smtp.send_message.assert_called_with(mock.ANY)
self.assertTrue(mock_smtp.quit.called)
mock_smtp.assert_has_calls([
mock.call.send_message(mock.ANY),
mock.call.quit()
])
| import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
Add test to check if connection is closed after email is sent.import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
def test_connection_closed_after_sending_mail(self, mock_smtp_class):
mock_smtp = mock_smtp_class.return_value
self.action.execute("MSFT has crossed $10 price level")
mock_smtp.send_message.assert_called_with(mock.ANY)
self.assertTrue(mock_smtp.quit.called)
mock_smtp.assert_has_calls([
mock.call.send_message(mock.ANY),
mock.call.quit()
])
| <commit_before>import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
<commit_msg>Add test to check if connection is closed after email is sent.<commit_after>import smtplib
import unittest
from unittest import mock
from action import PrintAction, EmailAction
@mock.patch("builtins.print")
class PrintActionTest(unittest.TestCase):
def test_executing_action_prints_message(self, mock_print):
action = PrintAction()
action.execute("GOOG > $10")
mock_print.assert_called_with("GOOG > $10")
@mock.patch("smtplib.SMTP")
class EmailActionTest(unittest.TestCase):
def setUp(self):
self.action = EmailAction(to="bsmukasa@gmail.com")
def test_email_is_sent_to_the_right_server(self, mock_smtp_class):
self.action.execute("MSFT has crossed $10 price level")
mock_smtp_class.assert_called_with("email.stocks.com")
def test_connection_closed_after_sending_mail(self, mock_smtp_class):
mock_smtp = mock_smtp_class.return_value
self.action.execute("MSFT has crossed $10 price level")
mock_smtp.send_message.assert_called_with(mock.ANY)
self.assertTrue(mock_smtp.quit.called)
mock_smtp.assert_has_calls([
mock.call.send_message(mock.ANY),
mock.call.quit()
])
|
5bc4aa60be988abc98ba76ca4b790b259d75af37 | capstone/rl/policies/egreedy.py | capstone/rl/policies/egreedy.py | import random
from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(vf, state, actions)
| from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(state, actions, vf)
| Fix EGreedy policy call order | Fix EGreedy policy call order
| Python | mit | davidrobles/mlnd-capstone-code | import random
from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(vf, state, actions)
Fix EGreedy policy call order | from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(state, actions, vf)
| <commit_before>import random
from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(vf, state, actions)
<commit_msg>Fix EGreedy policy call order<commit_after> | from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(state, actions, vf)
| import random
from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(vf, state, actions)
Fix EGreedy policy call orderfrom .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(state, actions, vf)
| <commit_before>import random
from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(vf, state, actions)
<commit_msg>Fix EGreedy policy call order<commit_after>from .greedy import Greedy
from .random_policy import RandomPolicy
from ..policy import Policy
from ...utils import check_random_state
class EGreedy(Policy):
def __init__(self, e, random_state=None):
self.e = e
self.greedy = Greedy()
self.rand = RandomPolicy()
self.random_state = check_random_state(random_state)
def action(self, state, actions=None, vf=None):
policy = self.rand if self.random_state.rand() < self.e else self.greedy
return policy.action(state, actions, vf)
|
547787272a6945bfefd086504e4c3dcaf483bc37 | tests/test_logger.py | tests/test_logger.py | """Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
| """Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
| Remove one blank to be friends with the lama. | Remove one blank to be friends with the lama.
| Python | mit | openspending/gobble | """Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
Remove one blank to be friends with the lama. | """Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
| <commit_before>"""Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
<commit_msg>Remove one blank to be friends with the lama.<commit_after> | """Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
| """Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
Remove one blank to be friends with the lama."""Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
| <commit_before>"""Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
<commit_msg>Remove one blank to be friends with the lama.<commit_after>"""Test the logger module"""
from gobble.logger import log
def test_gobble_logger_exists():
assert log.name == 'Gobble'
|
7be79e544eecf158a6ff1bde2f9f76f5145e4ae1 | tests/tools_tests.py | tests/tools_tests.py | """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
| """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
| Add a test that calls exec_cmd | Add a test that calls exec_cmd
| Python | bsd-3-clause | ftao/python-ifcfg | """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
Add a test that calls exec_cmd | """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
| <commit_before>"""Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
<commit_msg>Add a test that calls exec_cmd<commit_after> | """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
| """Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
Add a test that calls exec_cmd"""Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
| <commit_before>"""Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
<commit_msg>Add a test that calls exec_cmd<commit_after>"""Tests for ifcfg.tools."""
import logging
import os
import unittest
import ifcfg
from ifcfg.tools import exec_cmd
from nose.tools import eq_
class IfcfgToolsTestCase(unittest.TestCase):
def test_minimal_logger(self):
os.environ['IFCFG_DEBUG'] = '1'
log = ifcfg.tools.minimal_logger(__name__)
eq_(log.level, logging.DEBUG)
os.environ['IFCFG_DEBUG'] = '0'
def test_command(self):
output, __, __ = exec_cmd("echo -n 'this is a test'")
self.assertEqual(output, "this is a test")
|
30dbda17bfa3b52dc2aace6eba6b8c1e4b3f7542 | robot-name/robot_name.py | robot-name/robot_name.py | # File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
| # File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
def factory_name(self):
char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2))
num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3))
self.name = char + num
return self.name
R1 = Robot()
print(R1.factory_name())
| Add methord to generate unique robot name | Add methord to generate unique robot name
| Python | mit | amalshehu/exercism-python | # File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
Add methord to generate unique robot name | # File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
def factory_name(self):
char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2))
num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3))
self.name = char + num
return self.name
R1 = Robot()
print(R1.factory_name())
| <commit_before># File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
<commit_msg>Add methord to generate unique robot name<commit_after> | # File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
def factory_name(self):
char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2))
num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3))
self.name = char + num
return self.name
R1 = Robot()
print(R1.factory_name())
| # File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
Add methord to generate unique robot name# File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
def factory_name(self):
char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2))
num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3))
self.name = char + num
return self.name
R1 = Robot()
print(R1.factory_name())
| <commit_before># File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
<commit_msg>Add methord to generate unique robot name<commit_after># File: robot_name.py
# Purpose: Write a program that manages robot factory settings.
# Programmer: Amal Shehu
# Course: Exercism
# Date: Friday 30 September 2016, 03:00 PM
import string
import random
class Robot():
"""Robot facory settings"""
def __init__(self):
self.name = ""
def factory_name(self):
char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2))
num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3))
self.name = char + num
return self.name
R1 = Robot()
print(R1.factory_name())
|
3134af98d2fcf88752170d628400a7e863d4c959 | was/artists/views.py | was/artists/views.py | from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'register.html'
form_class = UpdateArtistForm
success_url = '/'
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/') | from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists, User
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'update.html'
form_class = UpdateArtistForm
success_url = '/'
def get_initial(self):
initial = {}
user = User.objects.get(username=self.request.user.username)
initial['username'] = user.username
initial['email'] = user.email
return initial
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/') | Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model. | Create update view (extends UpdateView generic view).
Define get_initail in order to pre populate two custom fields username and email which are not in the original
Artists model.
| Python | mit | KeserOner/where-artists-share,KeserOner/where-artists-share | from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'register.html'
form_class = UpdateArtistForm
success_url = '/'
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/')Create update view (extends UpdateView generic view).
Define get_initail in order to pre populate two custom fields username and email which are not in the original
Artists model. | from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists, User
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'update.html'
form_class = UpdateArtistForm
success_url = '/'
def get_initial(self):
initial = {}
user = User.objects.get(username=self.request.user.username)
initial['username'] = user.username
initial['email'] = user.email
return initial
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/') | <commit_before>from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'register.html'
form_class = UpdateArtistForm
success_url = '/'
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/')<commit_msg>Create update view (extends UpdateView generic view).
Define get_initail in order to pre populate two custom fields username and email which are not in the original
Artists model.<commit_after> | from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists, User
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'update.html'
form_class = UpdateArtistForm
success_url = '/'
def get_initial(self):
initial = {}
user = User.objects.get(username=self.request.user.username)
initial['username'] = user.username
initial['email'] = user.email
return initial
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/') | from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'register.html'
form_class = UpdateArtistForm
success_url = '/'
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/')Create update view (extends UpdateView generic view).
Define get_initail in order to pre populate two custom fields username and email which are not in the original
Artists model.from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists, User
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'update.html'
form_class = UpdateArtistForm
success_url = '/'
def get_initial(self):
initial = {}
user = User.objects.get(username=self.request.user.username)
initial['username'] = user.username
initial['email'] = user.email
return initial
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/') | <commit_before>from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'register.html'
form_class = UpdateArtistForm
success_url = '/'
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/')<commit_msg>Create update view (extends UpdateView generic view).
Define get_initail in order to pre populate two custom fields username and email which are not in the original
Artists model.<commit_after>from django.shortcuts import render, get_object_or_404
from django.views.generic.edit import CreateView, UpdateView
from .form import CreateArtistForm, UpdateArtistForm, Artists, User
from django.contrib.auth.forms import AuthenticationForm
from django.contrib.auth import login, logout
from django.http import HttpResponseRedirect
class CreateArtistView(CreateView):
template_name = 'register.html'
form_class = CreateArtistForm
success_url = '/'
class UpdateArtistView(UpdateView):
template_name = 'update.html'
form_class = UpdateArtistForm
success_url = '/'
def get_initial(self):
initial = {}
user = User.objects.get(username=self.request.user.username)
initial['username'] = user.username
initial['email'] = user.email
return initial
def get_object(self):
return get_object_or_404(Artists, user=self.request.user)
def artist_login(request):
if request.method == 'POST':
form = AuthenticationForm(data=request.POST)
if form.is_valid():
form.clean()
login(request, form.user_cache)
return HttpResponseRedirect('/')
else:
form = AuthenticationForm()
return render(request, 'login.html', {'form' : form})
def artist_logout(request):
logout(request)
return HttpResponseRedirect('/') |
533d1462949ab451674d91dd7730957cb2252dde | susumutakuan.py | susumutakuan.py | import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"])
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN) | import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"], universal_newlines=True)
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN) | Add universal_newlines paramter to run call | Add universal_newlines paramter to run call
| Python | mit | gryffon/SusumuTakuan,gryffon/SusumuTakuan | import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"])
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN)Add universal_newlines paramter to run call | import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"], universal_newlines=True)
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN) | <commit_before>import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"])
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN)<commit_msg>Add universal_newlines paramter to run call<commit_after> | import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"], universal_newlines=True)
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN) | import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"])
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN)Add universal_newlines paramter to run callimport discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"], universal_newlines=True)
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN) | <commit_before>import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"])
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN)<commit_msg>Add universal_newlines paramter to run call<commit_after>import discord
import asyncio
import os
import signal
import sys
from subprocess import run
#Set up Client State
CLIENT_TOKEN=os.environ['TOKEN']
#Create Discord client
client = discord.Client()
#Handle shutdown gracefully
def sigterm_handler(signum, frame):
print('Logging out...', flush=True)
raise KeyboardInterrupt
print('Shutting down...')
sys.exit(0)
#Register SIGTERM Handler
signal.signal(signal.SIGTERM, sigterm_handler)
@client.event
async def on_ready():
print('Logged in as')
print(client.user.name)
print(client.user.id)
print('------')
@client.event
async def on_message(message):
#Look at DMs for special commands
if message.channel.type == discord.ChannelType.private:
if message.content.startswith('!update'):
tmp = await client.send_message(message.channel, 'Updating my code via git...')
process = run(["sh", "control.sh", "refresh"], universal_newlines=True)
tmp = await client.send_message(message.channel, process.stdout)
if message.content.startswith('!test'):
counter = 0
tmp = await client.send_message(message.channel, 'Calculating messages...')
async for log in client.logs_from(message.channel, limit=100):
if log.author == message.author:
counter += 1
await client.edit_message(tmp, 'You have {} messages.'.format(counter))
#Start event loop
client.run(CLIENT_TOKEN) |
f9835741804da062f1501b06560a2af75b199645 | scrapeOMDB.py | scrapeOMDB.py | #!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
| #!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
| Convert movie year to str | Convert movie year to str
| Python | mit | samcheck/PyMedia,samcheck/PyMedia,samcheck/PyMedia | #!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
Convert movie year to str | #!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
| <commit_before>#!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
<commit_msg>Convert movie year to str<commit_after> | #!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
| #!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
Convert movie year to str#!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
| <commit_before>#!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
<commit_msg>Convert movie year to str<commit_after>#!/usr/bin/python3
# scrapeOMDB.py - parses a movie and year from arguments and returns JSON
import json, requests
URL_BASE = 'http://www.omdbapi.com/?'
def OMDBmovie(mTitle, mYear):
# Craft the URL
url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
def OMDBtv(tvTitle, tvSeason, tvEpisode):
# Craft the URL
url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json'
# Try to get the url
response = requests.get(url)
response.raise_for_status()
theJSON = json.loads(response.text)
return(theJSON)
|
1d14d28d68278330855e585a859484019d8c3e43 | cacivicdata/manage.py | cacivicdata/manage.py | #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
sys.path.append(os.path.dirname(os.path.dirname(__file__)))
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| Change back to django default | Change back to django default
| Python | mit | california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website | #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
sys.path.append(os.path.dirname(os.path.dirname(__file__)))
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
Change back to django default | #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| <commit_before>#!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
sys.path.append(os.path.dirname(os.path.dirname(__file__)))
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
<commit_msg>Change back to django default<commit_after> | #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
sys.path.append(os.path.dirname(os.path.dirname(__file__)))
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
Change back to django default#!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
| <commit_before>#!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
sys.path.append(os.path.dirname(os.path.dirname(__file__)))
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
<commit_msg>Change back to django default<commit_after>#!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings")
from django.core.management import execute_from_command_line
execute_from_command_line(sys.argv)
|
5632daecf9c5f271eeba0f9948d88f44d6a070d0 | irclogview/models.py | irclogview/models.py | from django.db import models
from picklefield.fields import PickledObjectField
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
| from django.db import models
from picklefield.fields import PickledObjectField
from . import utils
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
def content_dict(self):
colors = utils.RainbowColor()
for data in self.content:
item = dict(zip(['time', 'type', 'name', 'text'], data))
item['name_color'] = colors.get_color(item['name'])
yield item
| Add function to get content in list of dicts format | Add function to get content in list of dicts format
| Python | agpl-3.0 | BlankOn/irclogview,fajran/irclogview,fajran/irclogview,BlankOn/irclogview | from django.db import models
from picklefield.fields import PickledObjectField
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
Add function to get content in list of dicts format | from django.db import models
from picklefield.fields import PickledObjectField
from . import utils
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
def content_dict(self):
colors = utils.RainbowColor()
for data in self.content:
item = dict(zip(['time', 'type', 'name', 'text'], data))
item['name_color'] = colors.get_color(item['name'])
yield item
| <commit_before>from django.db import models
from picklefield.fields import PickledObjectField
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
<commit_msg>Add function to get content in list of dicts format<commit_after> | from django.db import models
from picklefield.fields import PickledObjectField
from . import utils
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
def content_dict(self):
colors = utils.RainbowColor()
for data in self.content:
item = dict(zip(['time', 'type', 'name', 'text'], data))
item['name_color'] = colors.get_color(item['name'])
yield item
| from django.db import models
from picklefield.fields import PickledObjectField
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
Add function to get content in list of dicts formatfrom django.db import models
from picklefield.fields import PickledObjectField
from . import utils
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
def content_dict(self):
colors = utils.RainbowColor()
for data in self.content:
item = dict(zip(['time', 'type', 'name', 'text'], data))
item['name_color'] = colors.get_color(item['name'])
yield item
| <commit_before>from django.db import models
from picklefield.fields import PickledObjectField
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
<commit_msg>Add function to get content in list of dicts format<commit_after>from django.db import models
from picklefield.fields import PickledObjectField
from . import utils
class Channel(models.Model):
name = models.SlugField(max_length=50, unique=True)
updated = models.DateTimeField(auto_now=True)
def __unicode__(self):
return u'#%s' % self.name
class Log(models.Model):
channel = models.ForeignKey(Channel)
date = models.DateField()
mtime = models.DateTimeField()
updated = models.DateTimeField(auto_now=True)
content = PickledObjectField()
class Meta:
unique_together = ('channel', 'date')
def content_dict(self):
colors = utils.RainbowColor()
for data in self.content:
item = dict(zip(['time', 'type', 'name', 'text'], data))
item['name_color'] = colors.get_color(item['name'])
yield item
|
4228082c9c94b3e17e6b00fc1e380841d5389dc5 | crawler/models.py | crawler/models.py | from django.db import models
# Create your models here.
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
| from django.db import models
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
| Remove not needed comment line | Remove not needed comment line
| Python | mit | lucasgr7/silverplate,lucasgr7/silverplate,lucasgr7/silverplate | from django.db import models
# Create your models here.
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
Remove not needed comment line | from django.db import models
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
| <commit_before>from django.db import models
# Create your models here.
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
<commit_msg>Remove not needed comment line<commit_after> | from django.db import models
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
| from django.db import models
# Create your models here.
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
Remove not needed comment linefrom django.db import models
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
| <commit_before>from django.db import models
# Create your models here.
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
<commit_msg>Remove not needed comment line<commit_after>from django.db import models
class Data_Ingredient(models.Model):
""""Class used to Store Ingredients of the recipes found in the crawling process"""
Ingredient = models.CharField(max_length=1000)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Ingredientes')
def __str__(self):
return self.Ingredient
class Data_Way_Cooking(models.Model):
"""Class used to Store steps of the recipes found in the crawling process"""
Description = models.CharField(max_length=500)
Recipe = models.CharField(max_length=500)
Group = models.CharField(max_length=500, default='Modo de Fazer')
def __str__(self):
return self.Description
class Ingredient_Spec(models.Model):
"""Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients"""
Word = models.CharField(max_length=500)
Count = models.IntegerField()
Type = models.CharField(max_length=1)
class Ignore_Words(models.Model):
"""Model to store words to ignore from Ingredient Spec"""
Word = models.CharField(max_length=500)
|
1d77647efdb26b8282fc0624852d211fa9339539 | db/TableConfig.py | db/TableConfig.py | {
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'1.0'"}
]
}
| {
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'4.0'"}
]
}
| Introduce family id for all | Introduce family id for all
| Python | mit | eddiedb6/ej,eddiedb6/ej,eddiedb6/ej | {
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'1.0'"}
]
}
Introduce family id for all | {
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'4.0'"}
]
}
| <commit_before>{
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'1.0'"}
]
}
<commit_msg>Introduce family id for all<commit_after> | {
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'4.0'"}
]
}
| {
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'1.0'"}
]
}
Introduce family id for all{
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'4.0'"}
]
}
| <commit_before>{
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'1.0'"}
]
}
<commit_msg>Introduce family id for all<commit_after>{
PDBConst.Name: "config",
PDBConst.Columns: [
{
PDBConst.Name: "Name",
PDBConst.Attributes: ["varchar(128)", "not null", "primary key"]
},
{
PDBConst.Name: "Value",
PDBConst.Attributes: ["varchar(128)"]
}],
PDBConst.Initials: [
{"Name": "'version'", "Value": "'4.0'"}
]
}
|
22b6785695967a43ab9d187db60c201c3dc4a8e1 | peerinst/admin.py | peerinst/admin.py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL}
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
| Use radio buttons for the style and number of answers. | Use radio buttons for the style and number of answers.
| Python | agpl-3.0 | open-craft/dalite-ng,open-craft/dalite-ng,open-craft/dalite-ng | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
Use radio buttons for the style and number of answers. | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL}
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
| <commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
<commit_msg>Use radio buttons for the style and number of answers.<commit_after> | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL}
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
| # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
Use radio buttons for the style and number of answers.# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL}
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
| <commit_before># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
<commit_msg>Use radio buttons for the style and number of answers.<commit_after># -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.contrib import admin
from django.utils.translation import ugettext_lazy as _
from . import models
@admin.register(models.Question)
class QuestionAdmin(admin.ModelAdmin):
fieldsets = [
(None, {'fields': ['title']}),
(_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}),
(_('Secondary image or video'), {
'fields': ['secondary_image', 'secondary_video_url'],
'classes': ['collapse'],
'description': _(
'Choose either a video or image to include on the first page of the question, '
'where students select concept tags. This is only used if you want the question '
'to be hidden when students select concept tags; instead, a preliminary video or '
'image can be displayed. The main question image will be displayed on all '
'subsequent pages.'
),
}),
(_('Answers'), {'fields': [
'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer'
]}),
(None, {'fields': ['example_rationale']}),
]
radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL}
@admin.register(models.Assignment)
class AssignmentAdmin(admin.ModelAdmin):
filter_horizontal = ['questions']
|
b35ee625143fd57f5571f807d0cd4331be4e0947 | caprice/models.py | caprice/models.py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return "<{0}: '{1}'>".format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
| Use compatible(for Python2.6) format string. | Use compatible(for Python2.6) format string.
| Python | mit | FGtatsuro/Caprice,FGtatsuro/Caprice,FGtatsuro/Caprice | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
Use compatible(for Python2.6) format string. | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return "<{0}: '{1}'>".format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
<commit_msg>Use compatible(for Python2.6) format string.<commit_after> | #!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return "<{0}: '{1}'>".format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
| #!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
Use compatible(for Python2.6) format string.#!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return "<{0}: '{1}'>".format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
| <commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
<commit_msg>Use compatible(for Python2.6) format string.<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from sqlalchemy import Column, String
from .db import Base, Session
__all__ = ['Schema']
class Schema(Base):
__tablename__ = 'schemas'
# TODO: allow Only UUID? or user defined ID too?
id = Column(String, primary_key=True)
# TODO: JSON uniqueness is needed
body = Column(String)
# ID is generated in Python context(=in application)
def __init__(self, id=None, body=None):
self.id = id
self.body = body
def __repr__(self):
return "<{0}: '{1}'>".format(self.__class__.__name__, self.body)
# TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html
def save(self):
s = Session()
s.add(self)
try:
# TODO: logger
s.commit()
except:
s.rollback()
# TODO: Error message
raise RuntimeError('')
finally:
# TODO: logger
s.close()
|
8222bf717b92ab57b60b834b4496afb654b8a80b | krisk/connections.py | krisk/connections.py |
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
paths: {
echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min"
},
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES
|
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES | Change all js to point to krisk repo | Change all js to point to krisk repo
| Python | bsd-3-clause | napjon/krisk,napjon/krisk,napjon/krisk |
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
paths: {
echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min"
},
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES
Change all js to point to krisk repo |
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES | <commit_before>
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
paths: {
echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min"
},
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES
<commit_msg>Change all js to point to krisk repo<commit_after> |
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES |
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
paths: {
echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min"
},
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES
Change all js to point to krisk repo
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES | <commit_before>
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
paths: {
echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min"
},
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES
<commit_msg>Change all js to point to krisk repo<commit_after>
from collections import OrderedDict
from IPython.display import Javascript
import json
from krisk.util import join_current_dir
ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/'
ECHARTS_FILE = 'echarts.min'
d_paths = OrderedDict({})
THEMES = ['dark','vintage','roma','shine','infographic','macarons']
THEMES_URL='//echarts.baidu.com/asset/theme/'
PATH_LOCAL = join_current_dir('static')
# PATH_LOCAL = 'pandas-echarts/krisk/static'
#TODO FIX LOCAL PATH! NEED TO DO nbextension install
# def init_notebook():
# """Inject Javascript to notebook, default using local js.
# """
# return Javascript("""
# require.config({
# baseUrl : '%s',
# paths: {
# echarts: 'echarts.min'
# }
# });
# """ % PATH_LOCAL)
def init_notebook():
"""Inject Javascript to notebook, default using local js.
"""
return Javascript("""
require.config({
baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static",
waitSeconds: 15
});
""")
def get_paths():
return ['echarts'] + THEMES |
e299c07034e0ad1135bda999ad0c63f4b5a7fa40 | chaco/__init__.py | chaco/__init__.py | # Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0dev'
__requires__ = [
'enable',
]
| # Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0.dev'
__requires__ = [
'enable',
]
| Tweak the version number to match other ETS projects. | Tweak the version number to match other ETS projects.
| Python | bsd-3-clause | burnpanck/chaco,tommy-u/chaco,tommy-u/chaco,tommy-u/chaco,burnpanck/chaco,burnpanck/chaco | # Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0dev'
__requires__ = [
'enable',
]
Tweak the version number to match other ETS projects. | # Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0.dev'
__requires__ = [
'enable',
]
| <commit_before># Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0dev'
__requires__ = [
'enable',
]
<commit_msg>Tweak the version number to match other ETS projects.<commit_after> | # Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0.dev'
__requires__ = [
'enable',
]
| # Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0dev'
__requires__ = [
'enable',
]
Tweak the version number to match other ETS projects.# Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0.dev'
__requires__ = [
'enable',
]
| <commit_before># Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0dev'
__requires__ = [
'enable',
]
<commit_msg>Tweak the version number to match other ETS projects.<commit_after># Copyright (c) 2005-2013 by Enthought, Inc.
# All rights reserved.
""" Two-dimensional plotting application toolkit.
Part of the Chaco project of the Enthought Tool Suite.
"""
__version__ = '4.5.0.dev'
__requires__ = [
'enable',
]
|
0f53ec6ddeb236bee78794e8d1ed156ad182bc07 | projects/forms.py | projects/forms.py | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',
'status')
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| Add status to project form | Add status to project form
| Python | mit | Hackfmi/Diaphanum,Hackfmi/Diaphanum | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
Add status to project form | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',
'status')
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| <commit_before>from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
<commit_msg>Add status to project form<commit_after> | from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',
'status')
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
Add status to project formfrom django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',
'status')
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
| <commit_before>from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',)
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
<commit_msg>Add status to project form<commit_after>from django import forms
from .models import Project
class ProjectForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
self.user = kwargs.pop('user')
super(ProjectForm, self).__init__(*args, **kwargs)
def save(self, *args, **kwargs):
instance = super(ProjectForm, self).save(commit=False)
instance.user = self.user
instance.flp = self.user
instance.status = 'unrevised'
instance.save(*args, **kwargs)
self.save_m2m()
return instance
class Meta:
model = Project
fields = (
'name',
'team',
'description',
'targets',
'tasks',
'target_group',
'schedule',
'resources',
'finance_description',
'partners',
'status')
class RestrictedProjectForm(forms.ModelForm):
def save(self, *args, **kwargs):
instance = super(RestrictedProjectForm, self).save(commit=False)
return instance
class Meta:
model = Project
exclude = (
'name', 'team', 'description', 'targets', 'tasks', 'target_group',
'schedule', 'resources', 'finance_description', 'partners',
'flp', 'created_at', 'user',
)
fileds = (
'status',
'attitude', )
|
84e9532487615f684abbed17d6821ae7bc84c9be | virtualfish/loader/__init__.py | virtualfish/loader/__init__.py | from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=()):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
commands.append("emit virtualfish_did_setup_plugins")
return commands
| from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=(), full_install=True):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
if full_install:
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
else:
commands = []
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
if full_install:
commands.append("emit virtualfish_did_setup_plugins")
return commands
| Add kwarg to load function to distinguish from full install | Add kwarg to load function to distinguish from full install
The load function is used for a full install and thus always adds
general configuration lines to the loader file, but we don't want that
for plugin installation.
| Python | mit | adambrenecki/virtualfish,adambrenecki/virtualfish | from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=()):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
commands.append("emit virtualfish_did_setup_plugins")
return commands
Add kwarg to load function to distinguish from full install
The load function is used for a full install and thus always adds
general configuration lines to the loader file, but we don't want that
for plugin installation. | from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=(), full_install=True):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
if full_install:
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
else:
commands = []
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
if full_install:
commands.append("emit virtualfish_did_setup_plugins")
return commands
| <commit_before>from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=()):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
commands.append("emit virtualfish_did_setup_plugins")
return commands
<commit_msg>Add kwarg to load function to distinguish from full install
The load function is used for a full install and thus always adds
general configuration lines to the loader file, but we don't want that
for plugin installation.<commit_after> | from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=(), full_install=True):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
if full_install:
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
else:
commands = []
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
if full_install:
commands.append("emit virtualfish_did_setup_plugins")
return commands
| from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=()):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
commands.append("emit virtualfish_did_setup_plugins")
return commands
Add kwarg to load function to distinguish from full install
The load function is used for a full install and thus always adds
general configuration lines to the loader file, but we don't want that
for plugin installation.from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=(), full_install=True):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
if full_install:
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
else:
commands = []
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
if full_install:
commands.append("emit virtualfish_did_setup_plugins")
return commands
| <commit_before>from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=()):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
commands.append("emit virtualfish_did_setup_plugins")
return commands
<commit_msg>Add kwarg to load function to distinguish from full install
The load function is used for a full install and thus always adds
general configuration lines to the loader file, but we don't want that
for plugin installation.<commit_after>from __future__ import print_function
import os
import sys
import pkg_resources
def load(plugins=(), full_install=True):
try:
version = pkg_resources.get_distribution("virtualfish").version
commands = ["set -g VIRTUALFISH_VERSION {}".format(version)]
except pkg_resources.DistributionNotFound:
commands = []
base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
if full_install:
commands += [
"set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable),
"source {}".format(os.path.join(base_path, "virtual.fish")),
]
else:
commands = []
for plugin in plugins:
path = os.path.join(base_path, plugin + ".fish")
if os.path.exists(path):
commands.append("source {}".format(path))
else:
raise ValueError("Plugin does not exist: " + plugin)
if full_install:
commands.append("emit virtualfish_did_setup_plugins")
return commands
|
5b652fc1af9c72c195446aaaf3ff35a501766676 | tests/tests.py | tests/tests.py | # coding=UTF-8
import unittest
import treetojson
class TreeToJsonTests(unittest.TestCase):
def test_list(self):
result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \
"{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}"
sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'),
('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')]
output = treetojson.get_json(data=sentence)
self.assertEqual(output, result)
def main():
unittest.main()
if __name__ == '__main__':
main()
| Add test case for list containing words with tags | Add test case for list containing words with tags
| Python | mit | saadsahibjan/treetojson | Add test case for list containing words with tags | # coding=UTF-8
import unittest
import treetojson
class TreeToJsonTests(unittest.TestCase):
def test_list(self):
result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \
"{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}"
sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'),
('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')]
output = treetojson.get_json(data=sentence)
self.assertEqual(output, result)
def main():
unittest.main()
if __name__ == '__main__':
main()
| <commit_before><commit_msg>Add test case for list containing words with tags<commit_after> | # coding=UTF-8
import unittest
import treetojson
class TreeToJsonTests(unittest.TestCase):
def test_list(self):
result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \
"{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}"
sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'),
('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')]
output = treetojson.get_json(data=sentence)
self.assertEqual(output, result)
def main():
unittest.main()
if __name__ == '__main__':
main()
| Add test case for list containing words with tags# coding=UTF-8
import unittest
import treetojson
class TreeToJsonTests(unittest.TestCase):
def test_list(self):
result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \
"{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}"
sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'),
('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')]
output = treetojson.get_json(data=sentence)
self.assertEqual(output, result)
def main():
unittest.main()
if __name__ == '__main__':
main()
| <commit_before><commit_msg>Add test case for list containing words with tags<commit_after># coding=UTF-8
import unittest
import treetojson
class TreeToJsonTests(unittest.TestCase):
def test_list(self):
result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \
"{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}"
sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'),
('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')]
output = treetojson.get_json(data=sentence)
self.assertEqual(output, result)
def main():
unittest.main()
if __name__ == '__main__':
main()
| |
0db1575341ae37644f2ce43c0a89e4baf83f8d87 | filebrowser/urls.py | filebrowser/urls.py | from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
| from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
| Remove redirect again, it's somehow causing the JS issue and it won't work for other media types | Remove redirect again, it's somehow causing the JS issue and it won't work for other media types
| Python | bsd-3-clause | django-wodnas/django-filebrowser-no-grappelli,django-wodnas/django-filebrowser-no-grappelli,sandow-digital/django-filebrowser-no-grappelli-sandow,sandow-digital/django-filebrowser-no-grappelli-sandow,sandow-digital/django-filebrowser-no-grappelli-sandow,django-wodnas/django-filebrowser-no-grappelli | from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
Remove redirect again, it's somehow causing the JS issue and it won't work for other media types | from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
| <commit_before>from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
<commit_msg>Remove redirect again, it's somehow causing the JS issue and it won't work for other media types<commit_after> | from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
| from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
Remove redirect again, it's somehow causing the JS issue and it won't work for other media typesfrom django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
| <commit_before>from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
<commit_msg>Remove redirect again, it's somehow causing the JS issue and it won't work for other media types<commit_after>from django.conf.urls.defaults import *
from django.views.generic.simple import redirect_to
urlpatterns = patterns('',
# filebrowser urls
url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"),
url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"),
url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"),
url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"),
url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"),
url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"),
url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"),
url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"),
)
|
3039fec89f74618657db0509765dda48a090f0be | hetzner/__init__.py | hetzner/__init__.py | class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __repr__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
| class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __str__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
| Use __str__ for formatting the error. | RobotError: Use __str__ for formatting the error.
On a traceback, the __str__ value will be read instead of __repr__ so we
get the name of the exception class, but not the actual exception.
This is now no longer the case :-)
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>
| Python | bsd-3-clause | RedMoonStudios/hetzner | class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __repr__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
RobotError: Use __str__ for formatting the error.
On a traceback, the __str__ value will be read instead of __repr__ so we
get the name of the exception class, but not the actual exception.
This is now no longer the case :-)
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org> | class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __str__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
| <commit_before>class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __repr__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
<commit_msg>RobotError: Use __str__ for formatting the error.
On a traceback, the __str__ value will be read instead of __repr__ so we
get the name of the exception class, but not the actual exception.
This is now no longer the case :-)
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org><commit_after> | class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __str__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
| class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __repr__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
RobotError: Use __str__ for formatting the error.
On a traceback, the __str__ value will be read instead of __repr__ so we
get the name of the exception class, but not the actual exception.
This is now no longer the case :-)
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __str__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
| <commit_before>class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __repr__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
<commit_msg>RobotError: Use __str__ for formatting the error.
On a traceback, the __str__ value will be read instead of __repr__ so we
get the name of the exception class, but not the actual exception.
This is now no longer the case :-)
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org><commit_after>class RobotError(Exception):
def __init__(self, message, status=None):
self.message = message
self.status = status
def __str__(self):
if self.status is None:
return self.message
else:
return "{0} ({1})".format(self.message, self.status)
class ManualReboot(Exception):
pass
class ConnectError(Exception):
pass
class WebRobotError(RobotError):
pass
|
353098b81b0e281d5d78e820dd91c3f360d6e585 | ibmcnx/test/test.py | ibmcnx/test/test.py | import loadFunction.py
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
| import ibmcnx.test.loadFunction
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
| Customize scripts to work with menu | Customize scripts to work with menu
| Python | apache-2.0 | stoeps13/ibmcnx2,stoeps13/ibmcnx2 | import loadFunction.py
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
Customize scripts to work with menu | import ibmcnx.test.loadFunction
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
| <commit_before>import loadFunction.py
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
<commit_msg>Customize scripts to work with menu<commit_after> | import ibmcnx.test.loadFunction
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
| import loadFunction.py
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
Customize scripts to work with menuimport ibmcnx.test.loadFunction
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
| <commit_before>import loadFunction.py
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
<commit_msg>Customize scripts to work with menu<commit_after>import ibmcnx.test.loadFunction
loadFilesService()
FilesPolicyService.browse( "title", "true", 1, 25 )
|
e1721a515520a85fbbfae112eb63f877de33e7c9 | caffe2/python/test_util.py | caffe2/python/test_util.py | ## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
| ## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import core, workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
# clear the default engines settings to separate out its
# affect from the ops tests
core.SetEnginePref({}, {})
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
| Clear the operator default engines before running operator tests | Clear the operator default engines before running operator tests
Reviewed By: akyrola
Differential Revision: D5729024
fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75
| Python | apache-2.0 | sf-wind/caffe2,xzturn/caffe2,pietern/caffe2,davinwang/caffe2,pietern/caffe2,davinwang/caffe2,sf-wind/caffe2,Yangqing/caffe2,sf-wind/caffe2,Yangqing/caffe2,davinwang/caffe2,xzturn/caffe2,sf-wind/caffe2,sf-wind/caffe2,xzturn/caffe2,pietern/caffe2,pietern/caffe2,xzturn/caffe2,Yangqing/caffe2,Yangqing/caffe2,caffe2/caffe2,pietern/caffe2,Yangqing/caffe2,xzturn/caffe2,davinwang/caffe2,davinwang/caffe2 | ## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
Clear the operator default engines before running operator tests
Reviewed By: akyrola
Differential Revision: D5729024
fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75 | ## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import core, workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
# clear the default engines settings to separate out its
# affect from the ops tests
core.SetEnginePref({}, {})
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
| <commit_before>## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
<commit_msg>Clear the operator default engines before running operator tests
Reviewed By: akyrola
Differential Revision: D5729024
fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75<commit_after> | ## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import core, workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
# clear the default engines settings to separate out its
# affect from the ops tests
core.SetEnginePref({}, {})
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
| ## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
Clear the operator default engines before running operator tests
Reviewed By: akyrola
Differential Revision: D5729024
fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import core, workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
# clear the default engines settings to separate out its
# affect from the ops tests
core.SetEnginePref({}, {})
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
| <commit_before>## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
<commit_msg>Clear the operator default engines before running operator tests
Reviewed By: akyrola
Differential Revision: D5729024
fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75<commit_after>## @package test_util
# Module caffe2.python.test_util
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from caffe2.python import core, workspace
import unittest
def rand_array(*dims):
# np.random.rand() returns float instead of 0-dim array, that's why need to
# do some tricks
return np.array(np.random.rand(*dims) - 0.5).astype(np.float32)
class TestCase(unittest.TestCase):
@classmethod
def setUpClass(cls):
workspace.GlobalInit([
'caffe2',
'--caffe2_log_level=0',
])
# clear the default engines settings to separate out its
# affect from the ops tests
core.SetEnginePref({}, {})
def setUp(self):
self.ws = workspace.C.Workspace()
workspace.ResetWorkspace()
def tearDown(self):
workspace.ResetWorkspace()
|
bb24f9d650cc1e8ae4f7f3ffa53a662ff1788c89 | zuora/client.py | zuora/client.py | """
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def __str__(self):
return self.client.__str__()
| """
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def instanciate(self, instance_type_string):
"""
Create object for client.factory.
"""
return self.client.factory.create(instance_type_string)
def __str__(self):
"""
Display the client __str__ method.
"""
return self.client.__str__()
| Add instanciate method + docstring | Add instanciate method + docstring
| Python | bsd-3-clause | liberation/zuora-client | """
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def __str__(self):
return self.client.__str__()
Add instanciate method + docstring | """
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def instanciate(self, instance_type_string):
"""
Create object for client.factory.
"""
return self.client.factory.create(instance_type_string)
def __str__(self):
"""
Display the client __str__ method.
"""
return self.client.__str__()
| <commit_before>"""
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def __str__(self):
return self.client.__str__()
<commit_msg>Add instanciate method + docstring<commit_after> | """
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def instanciate(self, instance_type_string):
"""
Create object for client.factory.
"""
return self.client.factory.create(instance_type_string)
def __str__(self):
"""
Display the client __str__ method.
"""
return self.client.__str__()
| """
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def __str__(self):
return self.client.__str__()
Add instanciate method + docstring"""
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def instanciate(self, instance_type_string):
"""
Create object for client.factory.
"""
return self.client.factory.create(instance_type_string)
def __str__(self):
"""
Display the client __str__ method.
"""
return self.client.__str__()
| <commit_before>"""
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def __str__(self):
return self.client.__str__()
<commit_msg>Add instanciate method + docstring<commit_after>"""
Client for Zuora SOAP API
"""
# TODO:
# - Handle debug
# - Handle error
# - Session policy
import os
from suds.client import Client
from suds.sax.element import Element
from zuora.transport import HttpTransportWithKeepAlive
class ZuoraException(Exception):
"""
Base Zuora Exception.
"""
pass
class Zuora(object):
"""
SOAP Client based on Suds
"""
def __init__(self, wsdl, login, password):
self.wsdl = wsdl
self.login = login
self.password = password
self.session = None
self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl)
self.client = Client(
self.wsdl_path,
transport=HttpTransportWithKeepAlive())
def instanciate(self, instance_type_string):
"""
Create object for client.factory.
"""
return self.client.factory.create(instance_type_string)
def __str__(self):
"""
Display the client __str__ method.
"""
return self.client.__str__()
|
d37a05d305279d9d3bd74ebbdf500b56f83f4768 | salt/grains/extra.py | salt/grains/extra.py | # -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if not 'conf_file' in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
| # -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if 'conf_file' not in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
| Fix PEP8 E713 - test for membership should be "not in" | Fix PEP8 E713 - test for membership should be "not in"
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | # -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if not 'conf_file' in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
Fix PEP8 E713 - test for membership should be "not in" | # -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if 'conf_file' not in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
| <commit_before># -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if not 'conf_file' in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
<commit_msg>Fix PEP8 E713 - test for membership should be "not in"<commit_after> | # -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if 'conf_file' not in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
| # -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if not 'conf_file' in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
Fix PEP8 E713 - test for membership should be "not in"# -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if 'conf_file' not in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
| <commit_before># -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if not 'conf_file' in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
<commit_msg>Fix PEP8 E713 - test for membership should be "not in"<commit_after># -*- coding: utf-8 -*-
# Import python libs
import os
# Import third party libs
import yaml
import logging
# Import salt libs
import salt.utils
log = logging.getLogger(__name__)
def shell():
'''
Return the default shell to use on this system
'''
# Provides:
# shell
return {'shell': os.environ.get('SHELL', '/bin/sh')}
def config():
'''
Return the grains set in the grains file
'''
if 'conf_file' not in __opts__:
return {}
if os.path.isdir(__opts__['conf_file']):
gfn = os.path.join(
__opts__['conf_file'],
'grains'
)
else:
gfn = os.path.join(
os.path.dirname(__opts__['conf_file']),
'grains'
)
if os.path.isfile(gfn):
with salt.utils.fopen(gfn, 'rb') as fp_:
try:
return yaml.safe_load(fp_.read())
except Exception:
log.warn("Bad syntax in grains file! Skipping.")
return {}
return {}
|
7ec5786efbdb20b9cbcdf0b4f1b583a7e07e0644 | comrade/core/tests.py | comrade/core/tests.py | from nose.tools import ok_, eq_
import unittest
import models
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
| from nose.tools import ok_, eq_
import unittest
import models
def check_direct_to_template(prefix, pattern):
from django import test
from django.core.urlresolvers import reverse
client = test.Client()
response = client.get(reverse(prefix + ':' + pattern.name))
template_name = pattern.default_args['template']
template_names = [t.name for t in test.testcases.to_list(response.template)]
ok_(template_names)
ok_(template_name in template_names,
"Template '%s' was not a template used to render"
" the response. Actual template(s) used: %s" %
(template_name, u', '.join(template_names)))
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
| Add test helper method for checking direct_to_template views. | Add test helper method for checking direct_to_template views.
| Python | mit | bueda/django-comrade | from nose.tools import ok_, eq_
import unittest
import models
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
Add test helper method for checking direct_to_template views. | from nose.tools import ok_, eq_
import unittest
import models
def check_direct_to_template(prefix, pattern):
from django import test
from django.core.urlresolvers import reverse
client = test.Client()
response = client.get(reverse(prefix + ':' + pattern.name))
template_name = pattern.default_args['template']
template_names = [t.name for t in test.testcases.to_list(response.template)]
ok_(template_names)
ok_(template_name in template_names,
"Template '%s' was not a template used to render"
" the response. Actual template(s) used: %s" %
(template_name, u', '.join(template_names)))
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
| <commit_before>from nose.tools import ok_, eq_
import unittest
import models
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
<commit_msg>Add test helper method for checking direct_to_template views.<commit_after> | from nose.tools import ok_, eq_
import unittest
import models
def check_direct_to_template(prefix, pattern):
from django import test
from django.core.urlresolvers import reverse
client = test.Client()
response = client.get(reverse(prefix + ':' + pattern.name))
template_name = pattern.default_args['template']
template_names = [t.name for t in test.testcases.to_list(response.template)]
ok_(template_names)
ok_(template_name in template_names,
"Template '%s' was not a template used to render"
" the response. Actual template(s) used: %s" %
(template_name, u', '.join(template_names)))
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
| from nose.tools import ok_, eq_
import unittest
import models
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
Add test helper method for checking direct_to_template views.from nose.tools import ok_, eq_
import unittest
import models
def check_direct_to_template(prefix, pattern):
from django import test
from django.core.urlresolvers import reverse
client = test.Client()
response = client.get(reverse(prefix + ':' + pattern.name))
template_name = pattern.default_args['template']
template_names = [t.name for t in test.testcases.to_list(response.template)]
ok_(template_names)
ok_(template_name in template_names,
"Template '%s' was not a template used to render"
" the response. Actual template(s) used: %s" %
(template_name, u', '.join(template_names)))
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
| <commit_before>from nose.tools import ok_, eq_
import unittest
import models
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
<commit_msg>Add test helper method for checking direct_to_template views.<commit_after>from nose.tools import ok_, eq_
import unittest
import models
def check_direct_to_template(prefix, pattern):
from django import test
from django.core.urlresolvers import reverse
client = test.Client()
response = client.get(reverse(prefix + ':' + pattern.name))
template_name = pattern.default_args['template']
template_names = [t.name for t in test.testcases.to_list(response.template)]
ok_(template_names)
ok_(template_name in template_names,
"Template '%s' was not a template used to render"
" the response. Actual template(s) used: %s" %
(template_name, u', '.join(template_names)))
class SimpleModel(models.ComradeBaseModel):
def __unicode__(self):
return u'This is a unicode string'
class TestBaseModel(unittest.TestCase):
def setUp(self):
super(TestBaseModel, self).setUp()
self.obj = SimpleModel()
def test_repr(self):
ok_(isinstance(self.obj.__repr__(), str))
def test_str(self):
ok_(isinstance(self.obj.__str__(), str))
def test_unicode(self):
ok_(isinstance(self.obj.__unicode__(), unicode))
|
f30e441958b8354b189ee5b5ef1e7eb47ebb1b1a | nhs/gunicorn_conf.py | nhs/gunicorn_conf.py | bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 3
timeout = 60
| bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 6
timeout = 60
| Increase the number of Gunicorn workers | Increase the number of Gunicorn workers
| Python | agpl-3.0 | openhealthcare/open-prescribing,openhealthcare/open-prescribing,openhealthcare/open-prescribing | bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 3
timeout = 60
Increase the number of Gunicorn workers | bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 6
timeout = 60
| <commit_before>bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 3
timeout = 60
<commit_msg>Increase the number of Gunicorn workers<commit_after> | bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 6
timeout = 60
| bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 3
timeout = 60
Increase the number of Gunicorn workersbind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 6
timeout = 60
| <commit_before>bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 3
timeout = 60
<commit_msg>Increase the number of Gunicorn workers<commit_after>bind = "127.0.0.1:4567"
logfile = "/usr/local/ohc/log/op.gunicorn.log"
workers = 6
timeout = 60
|
4f3cfe6e990c932d7f86dbd0cf8ae762407278b0 | nucleus/base/urls.py | nucleus/base/urls.py | from django.conf.urls import url
from nucleus.base import views
urlpatterns = (
url(r'^/?$', views.home, name='base.home'),
)
| from django.conf.urls import url
from django.views.generic import RedirectView
urlpatterns = (
url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'),
)
| Change root URL to redirect to /rna/ | Change root URL to redirect to /rna/
| Python | mpl-2.0 | mozilla/nucleus,mozilla/nucleus,mozilla/nucleus,mozilla/nucleus | from django.conf.urls import url
from nucleus.base import views
urlpatterns = (
url(r'^/?$', views.home, name='base.home'),
)
Change root URL to redirect to /rna/ | from django.conf.urls import url
from django.views.generic import RedirectView
urlpatterns = (
url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'),
)
| <commit_before>from django.conf.urls import url
from nucleus.base import views
urlpatterns = (
url(r'^/?$', views.home, name='base.home'),
)
<commit_msg>Change root URL to redirect to /rna/<commit_after> | from django.conf.urls import url
from django.views.generic import RedirectView
urlpatterns = (
url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'),
)
| from django.conf.urls import url
from nucleus.base import views
urlpatterns = (
url(r'^/?$', views.home, name='base.home'),
)
Change root URL to redirect to /rna/from django.conf.urls import url
from django.views.generic import RedirectView
urlpatterns = (
url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'),
)
| <commit_before>from django.conf.urls import url
from nucleus.base import views
urlpatterns = (
url(r'^/?$', views.home, name='base.home'),
)
<commit_msg>Change root URL to redirect to /rna/<commit_after>from django.conf.urls import url
from django.views.generic import RedirectView
urlpatterns = (
url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'),
)
|
5d8ea747bd5f34b382cc9fef91105f3ed434c0db | pylearn2/datasets/hdf5.py | pylearn2/datasets/hdf5.py | """Objects for datasets serialized in HDF5 format (.h5)."""
import h5py
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
| """Objects for datasets serialized in HDF5 format (.h5)."""
import warnings
try:
import h5py
except ImportError:
warnings.warn("Could not import h5py")
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
| Fix import issue in h5py.py | Fix import issue in h5py.py
| Python | bsd-3-clause | fulmicoton/pylearn2,caidongyun/pylearn2,ddboline/pylearn2,chrish42/pylearn,CIFASIS/pylearn2,sandeepkbhat/pylearn2,aalmah/pylearn2,goodfeli/pylearn2,junbochen/pylearn2,jamessergeant/pylearn2,nouiz/pylearn2,mclaughlin6464/pylearn2,pombredanne/pylearn2,JesseLivezey/plankton,msingh172/pylearn2,w1kke/pylearn2,shiquanwang/pylearn2,mclaughlin6464/pylearn2,theoryno3/pylearn2,goodfeli/pylearn2,matrogers/pylearn2,CIFASIS/pylearn2,sandeepkbhat/pylearn2,JesseLivezey/pylearn2,mclaughlin6464/pylearn2,kastnerkyle/pylearn2,daemonmaker/pylearn2,aalmah/pylearn2,pombredanne/pylearn2,ashhher3/pylearn2,pkainz/pylearn2,ddboline/pylearn2,woozzu/pylearn2,Refefer/pylearn2,hyqneuron/pylearn2-maxsom,shiquanwang/pylearn2,skearnes/pylearn2,shiquanwang/pylearn2,nouiz/pylearn2,pkainz/pylearn2,Refefer/pylearn2,KennethPierce/pylearnk,jamessergeant/pylearn2,alexjc/pylearn2,lisa-lab/pylearn2,hyqneuron/pylearn2-maxsom,sandeepkbhat/pylearn2,abergeron/pylearn2,kose-y/pylearn2,lunyang/pylearn2,lunyang/pylearn2,lancezlin/pylearn2,KennethPierce/pylearnk,junbochen/pylearn2,JesseLivezey/plankton,ashhher3/pylearn2,se4u/pylearn2,se4u/pylearn2,matrogers/pylearn2,CIFASIS/pylearn2,fulmicoton/pylearn2,lisa-lab/pylearn2,mkraemer67/pylearn2,aalmah/pylearn2,junbochen/pylearn2,fishcorn/pylearn2,lancezlin/pylearn2,kose-y/pylearn2,bartvm/pylearn2,daemonmaker/pylearn2,ddboline/pylearn2,daemonmaker/pylearn2,fyffyt/pylearn2,jamessergeant/pylearn2,JesseLivezey/plankton,goodfeli/pylearn2,abergeron/pylearn2,msingh172/pylearn2,caidongyun/pylearn2,goodfeli/pylearn2,sandeepkbhat/pylearn2,se4u/pylearn2,hantek/pylearn2,kastnerkyle/pylearn2,skearnes/pylearn2,lamblin/pylearn2,mkraemer67/pylearn2,woozzu/pylearn2,mkraemer67/pylearn2,jeremyfix/pylearn2,bartvm/pylearn2,hyqneuron/pylearn2-maxsom,pombredanne/pylearn2,fyffyt/pylearn2,mkraemer67/pylearn2,jeremyfix/pylearn2,ashhher3/pylearn2,daemonmaker/pylearn2,pkainz/pylearn2,lunyang/pylearn2,bartvm/pylearn2,matrogers/pylearn2,chrish42/pylearn,fishcorn/pylearn2,caidongyun/pylearn2,skearnes/pylearn2,lancezlin/pylearn2,w1kke/pylearn2,w1kke/pylearn2,se4u/pylearn2,woozzu/pylearn2,lamblin/pylearn2,fyffyt/pylearn2,junbochen/pylearn2,alexjc/pylearn2,alexjc/pylearn2,hantek/pylearn2,ashhher3/pylearn2,msingh172/pylearn2,skearnes/pylearn2,fishcorn/pylearn2,TNick/pylearn2,chrish42/pylearn,cosmoharrigan/pylearn2,fyffyt/pylearn2,shiquanwang/pylearn2,jeremyfix/pylearn2,hyqneuron/pylearn2-maxsom,abergeron/pylearn2,lunyang/pylearn2,mclaughlin6464/pylearn2,abergeron/pylearn2,fishcorn/pylearn2,theoryno3/pylearn2,kastnerkyle/pylearn2,msingh172/pylearn2,pkainz/pylearn2,fulmicoton/pylearn2,chrish42/pylearn,nouiz/pylearn2,lamblin/pylearn2,jeremyfix/pylearn2,fulmicoton/pylearn2,theoryno3/pylearn2,theoryno3/pylearn2,cosmoharrigan/pylearn2,lisa-lab/pylearn2,lisa-lab/pylearn2,matrogers/pylearn2,pombredanne/pylearn2,JesseLivezey/pylearn2,TNick/pylearn2,KennethPierce/pylearnk,KennethPierce/pylearnk,JesseLivezey/plankton,Refefer/pylearn2,jamessergeant/pylearn2,alexjc/pylearn2,kose-y/pylearn2,hantek/pylearn2,aalmah/pylearn2,JesseLivezey/pylearn2,bartvm/pylearn2,lancezlin/pylearn2,TNick/pylearn2,lamblin/pylearn2,JesseLivezey/pylearn2,TNick/pylearn2,cosmoharrigan/pylearn2,woozzu/pylearn2,CIFASIS/pylearn2,nouiz/pylearn2,Refefer/pylearn2,ddboline/pylearn2,kose-y/pylearn2,w1kke/pylearn2,cosmoharrigan/pylearn2,kastnerkyle/pylearn2,hantek/pylearn2,caidongyun/pylearn2 | """Objects for datasets serialized in HDF5 format (.h5)."""
import h5py
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
Fix import issue in h5py.py | """Objects for datasets serialized in HDF5 format (.h5)."""
import warnings
try:
import h5py
except ImportError:
warnings.warn("Could not import h5py")
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
| <commit_before>"""Objects for datasets serialized in HDF5 format (.h5)."""
import h5py
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
<commit_msg>Fix import issue in h5py.py<commit_after> | """Objects for datasets serialized in HDF5 format (.h5)."""
import warnings
try:
import h5py
except ImportError:
warnings.warn("Could not import h5py")
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
| """Objects for datasets serialized in HDF5 format (.h5)."""
import h5py
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
Fix import issue in h5py.py"""Objects for datasets serialized in HDF5 format (.h5)."""
import warnings
try:
import h5py
except ImportError:
warnings.warn("Could not import h5py")
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
| <commit_before>"""Objects for datasets serialized in HDF5 format (.h5)."""
import h5py
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
<commit_msg>Fix import issue in h5py.py<commit_after>"""Objects for datasets serialized in HDF5 format (.h5)."""
import warnings
try:
import h5py
except ImportError:
warnings.warn("Could not import h5py")
from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix
class HDF5Dataset(DenseDesignMatrix):
"""Dense dataset loaded from an HDF5 file."""
def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs):
"""
Loads data and labels from HDF5 file.
Parameters
----------
filename: str
HDF5 file name.
X: str
Key into HDF5 file for dataset design matrix.
topo_view: str
Key into HDF5 file for topological view of dataset.
y: str
Key into HDF5 file for dataset targets.
kwargs: dict
Keyword arguments passed to `DenseDesignMatrix`.
"""
with h5py.File(filename) as f:
if X is not None:
X = f[X][:]
if topo_view is not None:
topo_view = f[topo_view][:]
if y is not None:
y = f[y][:]
super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y,
**kwargs)
|
9b8cbfcf33ba644670a42490db7de4249e5ff080 | invocations/docs.py | invocations/docs.py | import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs.body()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs.body()
| import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs()
| Leverage __call__ on task downstream | Leverage __call__ on task downstream
| Python | bsd-2-clause | mrjmad/invocations,alex/invocations,pyinvoke/invocations,singingwolfboy/invocations | import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs.body()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs.body()
Leverage __call__ on task downstream | import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs()
| <commit_before>import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs.body()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs.body()
<commit_msg>Leverage __call__ on task downstream<commit_after> | import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs()
| import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs.body()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs.body()
Leverage __call__ on task downstreamimport os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs()
| <commit_before>import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs.body()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs.body()
<commit_msg>Leverage __call__ on task downstream<commit_after>import os
from invoke.tasks import task
from invoke.runner import run
docs_dir = 'docs'
build = os.path.join(docs_dir, '_build')
@task
def clean_docs():
run("rm -rf %s" % build)
@task
def browse_docs():
run("open %s" % os.path.join(build, 'index.html'))
@task
def docs(clean=False, browse=False):
if clean:
clean_docs()
run("sphinx-build %s %s" % (docs_dir, build), pty=True)
if browse:
browse_docs()
|
153360072096d4a3cef783d371fbfabcd75bcf98 | script/lib/config.py | script/lib/config.py | #!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
| #!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
| Upgrade libchromiumcontent to loose iframe sandbox. | Upgrade libchromiumcontent to loose iframe sandbox.
| Python | mit | rreimann/electron,dongjoon-hyun/electron,brave/muon,trankmichael/electron,kokdemo/electron,robinvandernoord/electron,tonyganch/electron,tylergibson/electron,natgolov/electron,sircharleswatson/electron,matiasinsaurralde/electron,vaginessa/electron,fritx/electron,benweissmann/electron,xiruibing/electron,cos2004/electron,anko/electron,systembugtj/electron,matiasinsaurralde/electron,aichingm/electron,stevekinney/electron,astoilkov/electron,astoilkov/electron,chrisswk/electron,dkfiresky/electron,mattotodd/electron,oiledCode/electron,soulteary/electron,meowlab/electron,synaptek/electron,carsonmcdonald/electron,Ivshti/electron,digideskio/electron,faizalpribadi/electron,bpasero/electron,robinvandernoord/electron,mhkeller/electron,gerhardberger/electron,sky7sea/electron,anko/electron,mirrh/electron,jtburke/electron,aaron-goshine/electron,bruce/electron,nagyistoce/electron-atom-shell,tonyganch/electron,electron/electron,MaxGraey/electron,eric-seekas/electron,evgenyzinoviev/electron,thomsonreuters/electron,arturts/electron,matiasinsaurralde/electron,chriskdon/electron,xiruibing/electron,trankmichael/electron,etiktin/electron,gamedevsam/electron,Neron-X5/electron,pombredanne/electron,eriser/electron,trigrass2/electron,posix4e/electron,jacksondc/electron,RobertJGabriel/electron,John-Lin/electron,thompsonemerson/electron,christian-bromann/electron,bpasero/electron,subblue/electron,Jacobichou/electron,noikiy/electron,rreimann/electron,gabriel/electron,timruffles/electron,greyhwndz/electron,jhen0409/electron,Rokt33r/electron,abhishekgahlot/electron,meowlab/electron,gabrielPeart/electron,Andrey-Pavlov/electron,pirafrank/electron,brave/electron,jjz/electron,shockone/electron,dahal/electron,leolujuyi/electron,minggo/electron,thomsonreuters/electron,gabrielPeart/electron,IonicaBizauKitchen/electron,Floato/electron,setzer777/electron,Zagorakiss/electron,thompsonemerson/electron,jsutcodes/electron,kazupon/electron,webmechanicx/electron,meowlab/electron,gabriel/electron,thomsonreuters/electron,icattlecoder/electron,davazp/electron,iftekeriba/electron,subblue/electron,pandoraui/electron,gerhardberger/electron,synaptek/electron,ianscrivener/electron,wolfflow/electron,mirrh/electron,RobertJGabriel/electron,vHanda/electron,meowlab/electron,gbn972/electron,wan-qy/electron,smczk/electron,electron/electron,takashi/electron,GoooIce/electron,RobertJGabriel/electron,astoilkov/electron,SufianHassan/electron,Rokt33r/electron,webmechanicx/electron,systembugtj/electron,jlord/electron,vaginessa/electron,bpasero/electron,rsvip/electron,joaomoreno/atom-shell,smczk/electron,preco21/electron,eriser/electron,electron/electron,kenmozi/electron,joaomoreno/atom-shell,mjaniszew/electron,takashi/electron,Jacobichou/electron,gamedevsam/electron,brenca/electron,tincan24/electron,jonatasfreitasv/electron,roadev/electron,leftstick/electron,tonyganch/electron,mattotodd/electron,tonyganch/electron,pombredanne/electron,jaanus/electron,gamedevsam/electron,micalan/electron,twolfson/electron,rhencke/electron,gbn972/electron,smczk/electron,mattotodd/electron,Rokt33r/electron,Evercoder/electron,arturts/electron,Floato/electron,takashi/electron,darwin/electron,Ivshti/electron,medixdev/electron,brave/electron,jcblw/electron,mattdesl/electron,nagyistoce/electron-atom-shell,aliib/electron,tomashanacek/electron,davazp/electron,RIAEvangelist/electron,dkfiresky/electron,darwin/electron,rajatsingla28/electron,bright-sparks/electron,dongjoon-hyun/electron,ankitaggarwal011/electron,mrwizard82d1/electron,nicholasess/electron,digideskio/electron,icattlecoder/electron,darwin/electron,Neron-X5/electron,bitemyapp/electron,pandoraui/electron,aecca/electron,bwiggs/electron,John-Lin/electron,Gerhut/electron,coderhaoxin/electron,dahal/electron,mrwizard82d1/electron,brave/electron,voidbridge/electron,maxogden/atom-shell,mattotodd/electron,bobwol/electron,bpasero/electron,d-salas/electron,simongregory/electron,benweissmann/electron,shiftkey/electron,fireball-x/atom-shell,xfstudio/electron,Gerhut/electron,vipulroxx/electron,Rokt33r/electron,aliib/electron,rreimann/electron,IonicaBizauKitchen/electron,egoist/electron,twolfson/electron,John-Lin/electron,jsutcodes/electron,adamjgray/electron,vipulroxx/electron,jtburke/electron,biblerule/UMCTelnetHub,d-salas/electron,bruce/electron,nekuz0r/electron,Evercoder/electron,noikiy/electron,bruce/electron,gerhardberger/electron,kenmozi/electron,simonfork/electron,MaxWhere/electron,thingsinjars/electron,tinydew4/electron,kikong/electron,deepak1556/atom-shell,thompsonemerson/electron,miniak/electron,ianscrivener/electron,deepak1556/atom-shell,JesselJohn/electron,jhen0409/electron,hokein/atom-shell,fomojola/electron,JesselJohn/electron,robinvandernoord/electron,the-ress/electron,DivyaKMenon/electron,bitemyapp/electron,kokdemo/electron,MaxWhere/electron,Zagorakiss/electron,shennushi/electron,bbondy/electron,lrlna/electron,cqqccqc/electron,Faiz7412/electron,xfstudio/electron,beni55/electron,Jonekee/electron,hokein/atom-shell,baiwyc119/electron,robinvandernoord/electron,arusakov/electron,Andrey-Pavlov/electron,egoist/electron,simongregory/electron,tinydew4/electron,d-salas/electron,jannishuebl/electron,vaginessa/electron,jcblw/electron,tonyganch/electron,systembugtj/electron,vaginessa/electron,stevemao/electron,renaesop/electron,sky7sea/electron,digideskio/electron,yan-foto/electron,shaundunne/electron,farmisen/electron,bobwol/electron,jannishuebl/electron,nicholasess/electron,nicholasess/electron,bbondy/electron,leftstick/electron,fireball-x/atom-shell,simonfork/electron,ianscrivener/electron,renaesop/electron,faizalpribadi/electron,roadev/electron,natgolov/electron,xfstudio/electron,farmisen/electron,seanchas116/electron,cqqccqc/electron,joaomoreno/atom-shell,edulan/electron,miniak/electron,jsutcodes/electron,stevemao/electron,meowlab/electron,d-salas/electron,edulan/electron,jannishuebl/electron,rhencke/electron,sircharleswatson/electron,kazupon/electron,cqqccqc/electron,biblerule/UMCTelnetHub,dongjoon-hyun/electron,oiledCode/electron,jannishuebl/electron,evgenyzinoviev/electron,iftekeriba/electron,seanchas116/electron,Ivshti/electron,beni55/electron,preco21/electron,wan-qy/electron,iftekeriba/electron,RobertJGabriel/electron,ervinb/electron,eriser/electron,brenca/electron,tonyganch/electron,yalexx/electron,shiftkey/electron,thingsinjars/electron,michaelchiche/electron,leftstick/electron,systembugtj/electron,RIAEvangelist/electron,mhkeller/electron,rprichard/electron,astoilkov/electron,wolfflow/electron,shiftkey/electron,yalexx/electron,greyhwndz/electron,evgenyzinoviev/electron,brenca/electron,seanchas116/electron,JesselJohn/electron,brenca/electron,GoooIce/electron,saronwei/electron,timruffles/electron,nekuz0r/electron,cos2004/electron,jonatasfreitasv/electron,DivyaKMenon/electron,abhishekgahlot/electron,BionicClick/electron,eriser/electron,RobertJGabriel/electron,fomojola/electron,aecca/electron,davazp/electron,shennushi/electron,shaundunne/electron,destan/electron,lrlna/electron,thomsonreuters/electron,kokdemo/electron,neutrous/electron,trigrass2/electron,voidbridge/electron,Jonekee/electron,rsvip/electron,yalexx/electron,nicobot/electron,trigrass2/electron,bwiggs/electron,neutrous/electron,trankmichael/electron,trigrass2/electron,kostia/electron,kenmozi/electron,soulteary/electron,webmechanicx/electron,tylergibson/electron,twolfson/electron,sshiting/electron,electron/electron,bruce/electron,dkfiresky/electron,synaptek/electron,the-ress/electron,joaomoreno/atom-shell,hokein/atom-shell,deed02392/electron,rprichard/electron,darwin/electron,kazupon/electron,shennushi/electron,howmuchcomputer/electron,christian-bromann/electron,mattdesl/electron,oiledCode/electron,MaxWhere/electron,gabrielPeart/electron,adamjgray/electron,ianscrivener/electron,shiftkey/electron,voidbridge/electron,vHanda/electron,ankitaggarwal011/electron,leftstick/electron,nicobot/electron,fffej/electron,dahal/electron,trankmichael/electron,jaanus/electron,jonatasfreitasv/electron,arusakov/electron,kikong/electron,stevekinney/electron,gabriel/electron,tinydew4/electron,zhakui/electron,eriser/electron,coderhaoxin/electron,pombredanne/electron,minggo/electron,miniak/electron,gabrielPeart/electron,kostia/electron,Andrey-Pavlov/electron,mattdesl/electron,oiledCode/electron,nagyistoce/electron-atom-shell,etiktin/electron,beni55/electron,medixdev/electron,noikiy/electron,rhencke/electron,simongregory/electron,deepak1556/atom-shell,JesselJohn/electron,renaesop/electron,preco21/electron,jtburke/electron,SufianHassan/electron,gabriel/electron,aaron-goshine/electron,Faiz7412/electron,John-Lin/electron,wolfflow/electron,shaundunne/electron,pandoraui/electron,jannishuebl/electron,thompsonemerson/electron,soulteary/electron,smczk/electron,aliib/electron,neutrous/electron,eric-seekas/electron,joneit/electron,felixrieseberg/electron,mattdesl/electron,bright-sparks/electron,timruffles/electron,robinvandernoord/electron,subblue/electron,aecca/electron,oiledCode/electron,bright-sparks/electron,nicobot/electron,nekuz0r/electron,jiaz/electron,simonfork/electron,howmuchcomputer/electron,GoooIce/electron,etiktin/electron,tylergibson/electron,fffej/electron,bitemyapp/electron,shennushi/electron,chrisswk/electron,eric-seekas/electron,thompsonemerson/electron,icattlecoder/electron,setzer777/electron,chrisswk/electron,yan-foto/electron,preco21/electron,wan-qy/electron,chriskdon/electron,farmisen/electron,yan-foto/electron,pandoraui/electron,jacksondc/electron,aichingm/electron,adamjgray/electron,chrisswk/electron,vipulroxx/electron,JussMee15/electron,gbn972/electron,DivyaKMenon/electron,zhakui/electron,lrlna/electron,miniak/electron,lzpfmh/electron,howmuchcomputer/electron,Evercoder/electron,voidbridge/electron,fireball-x/atom-shell,shockone/electron,mjaniszew/electron,seanchas116/electron,mubassirhayat/electron,jjz/electron,carsonmcdonald/electron,seanchas116/electron,kokdemo/electron,bwiggs/electron,natgolov/electron,brave/muon,chriskdon/electron,faizalpribadi/electron,michaelchiche/electron,preco21/electron,darwin/electron,micalan/electron,jlord/electron,Gerhut/electron,ianscrivener/electron,fritx/electron,soulteary/electron,Jonekee/electron,Ivshti/electron,hokein/atom-shell,xiruibing/electron,biblerule/UMCTelnetHub,edulan/electron,jsutcodes/electron,nekuz0r/electron,dkfiresky/electron,Floato/electron,abhishekgahlot/electron,arusakov/electron,beni55/electron,digideskio/electron,mirrh/electron,pombredanne/electron,yalexx/electron,faizalpribadi/electron,eric-seekas/electron,icattlecoder/electron,bright-sparks/electron,LadyNaggaga/electron,saronwei/electron,christian-bromann/electron,rhencke/electron,shennushi/electron,JussMee15/electron,MaxGraey/electron,Zagorakiss/electron,joaomoreno/atom-shell,seanchas116/electron,kikong/electron,jjz/electron,John-Lin/electron,fireball-x/atom-shell,iftekeriba/electron,jtburke/electron,arusakov/electron,saronwei/electron,rajatsingla28/electron,mjaniszew/electron,jhen0409/electron,aliib/electron,icattlecoder/electron,systembugtj/electron,destan/electron,fritx/electron,simongregory/electron,MaxGraey/electron,trankmichael/electron,arturts/electron,jtburke/electron,felixrieseberg/electron,saronwei/electron,jlhbaseball15/electron,baiwyc119/electron,tomashanacek/electron,mrwizard82d1/electron,simongregory/electron,subblue/electron,oiledCode/electron,mrwizard82d1/electron,Zagorakiss/electron,bpasero/electron,wolfflow/electron,bobwol/electron,fffej/electron,yan-foto/electron,gamedevsam/electron,arturts/electron,subblue/electron,stevekinney/electron,SufianHassan/electron,mubassirhayat/electron,howmuchcomputer/electron,vHanda/electron,Neron-X5/electron,tincan24/electron,kostia/electron,mhkeller/electron,leethomas/electron,wan-qy/electron,Floato/electron,baiwyc119/electron,dkfiresky/electron,IonicaBizauKitchen/electron,leethomas/electron,pombredanne/electron,bruce/electron,digideskio/electron,mattdesl/electron,rreimann/electron,carsonmcdonald/electron,matiasinsaurralde/electron,bobwol/electron,brave/electron,saronwei/electron,ervinb/electron,GoooIce/electron,jlhbaseball15/electron,minggo/electron,bright-sparks/electron,jlhbaseball15/electron,adamjgray/electron,Andrey-Pavlov/electron,miniak/electron,leethomas/electron,etiktin/electron,fffej/electron,pirafrank/electron,gabrielPeart/electron,leethomas/electron,anko/electron,subblue/electron,lzpfmh/electron,synaptek/electron,ervinb/electron,egoist/electron,stevemao/electron,Rokt33r/electron,bwiggs/electron,timruffles/electron,destan/electron,nagyistoce/electron-atom-shell,Faiz7412/electron,fomojola/electron,kazupon/electron,mrwizard82d1/electron,adcentury/electron,evgenyzinoviev/electron,xfstudio/electron,kcrt/electron,John-Lin/electron,kenmozi/electron,Neron-X5/electron,joneit/electron,mattotodd/electron,trigrass2/electron,lrlna/electron,aaron-goshine/electron,gamedevsam/electron,RIAEvangelist/electron,tomashanacek/electron,bobwol/electron,egoist/electron,aecca/electron,jiaz/electron,deed02392/electron,Jonekee/electron,jlhbaseball15/electron,mirrh/electron,brave/muon,RobertJGabriel/electron,DivyaKMenon/electron,destan/electron,gbn972/electron,takashi/electron,fabien-d/electron,fffej/electron,rsvip/electron,BionicClick/electron,RIAEvangelist/electron,cqqccqc/electron,abhishekgahlot/electron,the-ress/electron,beni55/electron,shaundunne/electron,yan-foto/electron,faizalpribadi/electron,stevekinney/electron,mhkeller/electron,shockone/electron,michaelchiche/electron,rhencke/electron,brave/muon,micalan/electron,Floato/electron,howmuchcomputer/electron,zhakui/electron,natgolov/electron,renaesop/electron,mattdesl/electron,Gerhut/electron,carsonmcdonald/electron,ianscrivener/electron,natgolov/electron,gstack/infinium-shell,tomashanacek/electron,maxogden/atom-shell,vipulroxx/electron,MaxGraey/electron,roadev/electron,jcblw/electron,joneit/electron,dahal/electron,cos2004/electron,JussMee15/electron,BionicClick/electron,cos2004/electron,RIAEvangelist/electron,medixdev/electron,coderhaoxin/electron,noikiy/electron,bbondy/electron,mhkeller/electron,zhakui/electron,jlhbaseball15/electron,pandoraui/electron,thingsinjars/electron,rreimann/electron,anko/electron,coderhaoxin/electron,nekuz0r/electron,rsvip/electron,iftekeriba/electron,yan-foto/electron,tincan24/electron,leethomas/electron,sircharleswatson/electron,greyhwndz/electron,tylergibson/electron,arturts/electron,brenca/electron,brenca/electron,bitemyapp/electron,bwiggs/electron,kikong/electron,nicholasess/electron,synaptek/electron,simonfork/electron,adcentury/electron,tincan24/electron,dahal/electron,egoist/electron,carsonmcdonald/electron,farmisen/electron,kazupon/electron,jonatasfreitasv/electron,abhishekgahlot/electron,biblerule/UMCTelnetHub,jlord/electron,webmechanicx/electron,matiasinsaurralde/electron,thomsonreuters/electron,sshiting/electron,baiwyc119/electron,eriser/electron,baiwyc119/electron,the-ress/electron,electron/electron,SufianHassan/electron,minggo/electron,posix4e/electron,felixrieseberg/electron,saronwei/electron,kokdemo/electron,gerhardberger/electron,wolfflow/electron,kcrt/electron,shockone/electron,felixrieseberg/electron,wan-qy/electron,renaesop/electron,deed02392/electron,adamjgray/electron,jacksondc/electron,mjaniszew/electron,noikiy/electron,voidbridge/electron,aaron-goshine/electron,setzer777/electron,GoooIce/electron,rsvip/electron,neutrous/electron,thingsinjars/electron,Faiz7412/electron,ervinb/electron,fabien-d/electron,twolfson/electron,tylergibson/electron,fabien-d/electron,destan/electron,nicholasess/electron,miniak/electron,rajatsingla28/electron,mirrh/electron,christian-bromann/electron,JesselJohn/electron,aliib/electron,joneit/electron,astoilkov/electron,minggo/electron,kazupon/electron,hokein/atom-shell,kcrt/electron,simonfork/electron,fffej/electron,lzpfmh/electron,smczk/electron,carsonmcdonald/electron,jacksondc/electron,wolfflow/electron,dongjoon-hyun/electron,fireball-x/atom-shell,jsutcodes/electron,stevemao/electron,smczk/electron,fomojola/electron,edulan/electron,gerhardberger/electron,gstack/infinium-shell,adcentury/electron,xiruibing/electron,stevemao/electron,lzpfmh/electron,gabriel/electron,fritx/electron,twolfson/electron,shaundunne/electron,pirafrank/electron,kostia/electron,tinydew4/electron,adcentury/electron,renaesop/electron,JussMee15/electron,brave/electron,joneit/electron,Evercoder/electron,aaron-goshine/electron,kostia/electron,BionicClick/electron,rhencke/electron,posix4e/electron,d-salas/electron,davazp/electron,electron/electron,GoooIce/electron,jhen0409/electron,benweissmann/electron,simonfork/electron,vipulroxx/electron,farmisen/electron,shaundunne/electron,leolujuyi/electron,the-ress/electron,Faiz7412/electron,bright-sparks/electron,fomojola/electron,sircharleswatson/electron,thompsonemerson/electron,thingsinjars/electron,LadyNaggaga/electron,micalan/electron,roadev/electron,fabien-d/electron,jjz/electron,deed02392/electron,sshiting/electron,roadev/electron,Jacobichou/electron,trigrass2/electron,mrwizard82d1/electron,evgenyzinoviev/electron,bbondy/electron,bwiggs/electron,mjaniszew/electron,lrlna/electron,nicobot/electron,biblerule/UMCTelnetHub,trankmichael/electron,pirafrank/electron,jiaz/electron,shockone/electron,gabrielPeart/electron,Jacobichou/electron,dongjoon-hyun/electron,leolujuyi/electron,zhakui/electron,jaanus/electron,IonicaBizauKitchen/electron,synaptek/electron,davazp/electron,pombredanne/electron,bpasero/electron,gerhardberger/electron,mjaniszew/electron,webmechanicx/electron,nicobot/electron,JussMee15/electron,ankitaggarwal011/electron,jhen0409/electron,jcblw/electron,micalan/electron,mubassirhayat/electron,edulan/electron,deepak1556/atom-shell,kokdemo/electron,Floato/electron,mattotodd/electron,arturts/electron,vHanda/electron,chriskdon/electron,neutrous/electron,gstack/infinium-shell,pirafrank/electron,jonatasfreitasv/electron,posix4e/electron,tincan24/electron,soulteary/electron,twolfson/electron,LadyNaggaga/electron,sky7sea/electron,shockone/electron,greyhwndz/electron,sircharleswatson/electron,etiktin/electron,sky7sea/electron,eric-seekas/electron,electron/electron,coderhaoxin/electron,eric-seekas/electron,aliib/electron,aichingm/electron,adamjgray/electron,aichingm/electron,Jacobichou/electron,jiaz/electron,webmechanicx/electron,egoist/electron,the-ress/electron,nicobot/electron,medixdev/electron,jjz/electron,MaxGraey/electron,d-salas/electron,Evercoder/electron,coderhaoxin/electron,thingsinjars/electron,neutrous/electron,abhishekgahlot/electron,natgolov/electron,beni55/electron,vaginessa/electron,BionicClick/electron,anko/electron,bruce/electron,leolujuyi/electron,aaron-goshine/electron,jhen0409/electron,minggo/electron,mubassirhayat/electron,stevemao/electron,MaxWhere/electron,jiaz/electron,adcentury/electron,MaxWhere/electron,pirafrank/electron,jonatasfreitasv/electron,JussMee15/electron,Jonekee/electron,benweissmann/electron,Andrey-Pavlov/electron,LadyNaggaga/electron,shiftkey/electron,sshiting/electron,nagyistoce/electron-atom-shell,dahal/electron,davazp/electron,chrisswk/electron,Andrey-Pavlov/electron,gstack/infinium-shell,rajatsingla28/electron,shennushi/electron,setzer777/electron,robinvandernoord/electron,rajatsingla28/electron,brave/muon,digideskio/electron,christian-bromann/electron,kcrt/electron,tomashanacek/electron,Rokt33r/electron,wan-qy/electron,BionicClick/electron,michaelchiche/electron,takashi/electron,sshiting/electron,posix4e/electron,JesselJohn/electron,rreimann/electron,fritx/electron,adcentury/electron,LadyNaggaga/electron,benweissmann/electron,faizalpribadi/electron,farmisen/electron,arusakov/electron,nicholasess/electron,gerhardberger/electron,aecca/electron,jcblw/electron,leethomas/electron,gbn972/electron,bbondy/electron,greyhwndz/electron,noikiy/electron,thomsonreuters/electron,takashi/electron,voidbridge/electron,jcblw/electron,setzer777/electron,DivyaKMenon/electron,maxogden/atom-shell,bbondy/electron,kcrt/electron,stevekinney/electron,yalexx/electron,christian-bromann/electron,jlord/electron,vaginessa/electron,xfstudio/electron,mirrh/electron,ervinb/electron,meowlab/electron,SufianHassan/electron,etiktin/electron,leftstick/electron,sky7sea/electron,xiruibing/electron,bitemyapp/electron,joneit/electron,Gerhut/electron,matiasinsaurralde/electron,gabriel/electron,micalan/electron,Jonekee/electron,jannishuebl/electron,destan/electron,Neron-X5/electron,SufianHassan/electron,maxogden/atom-shell,jaanus/electron,Zagorakiss/electron,yalexx/electron,gbn972/electron,kenmozi/electron,joaomoreno/atom-shell,gstack/infinium-shell,timruffles/electron,jtburke/electron,leolujuyi/electron,ankitaggarwal011/electron,Neron-X5/electron,systembugtj/electron,jlhbaseball15/electron,kenmozi/electron,nekuz0r/electron,RIAEvangelist/electron,rajatsingla28/electron,shiftkey/electron,setzer777/electron,tinydew4/electron,tincan24/electron,fomojola/electron,maxogden/atom-shell,bobwol/electron,ervinb/electron,tylergibson/electron,deepak1556/atom-shell,evgenyzinoviev/electron,jiaz/electron,sircharleswatson/electron,jjz/electron,xfstudio/electron,felixrieseberg/electron,stevekinney/electron,roadev/electron,jlord/electron,michaelchiche/electron,mubassirhayat/electron,simongregory/electron,cos2004/electron,MaxWhere/electron,soulteary/electron,cqqccqc/electron,michaelchiche/electron,vipulroxx/electron,lzpfmh/electron,IonicaBizauKitchen/electron,tomashanacek/electron,biblerule/UMCTelnetHub,aecca/electron,fritx/electron,pandoraui/electron,preco21/electron,zhakui/electron,edulan/electron,LadyNaggaga/electron,Zagorakiss/electron,kostia/electron,mhkeller/electron,Evercoder/electron,sky7sea/electron,bitemyapp/electron,felixrieseberg/electron,deed02392/electron,cos2004/electron,jaanus/electron,chriskdon/electron,deed02392/electron,jacksondc/electron,lzpfmh/electron,posix4e/electron,leftstick/electron,ankitaggarwal011/electron,jacksondc/electron,leolujuyi/electron,cqqccqc/electron,gamedevsam/electron,ankitaggarwal011/electron,brave/muon,howmuchcomputer/electron,dkfiresky/electron,iftekeriba/electron,Ivshti/electron,rprichard/electron,benweissmann/electron,icattlecoder/electron,bpasero/electron,Jacobichou/electron,jsutcodes/electron,astoilkov/electron,brave/electron,lrlna/electron,kikong/electron,kcrt/electron,Gerhut/electron,sshiting/electron,baiwyc119/electron,aichingm/electron,DivyaKMenon/electron,vHanda/electron,jaanus/electron,dongjoon-hyun/electron,rprichard/electron,vHanda/electron,greyhwndz/electron,medixdev/electron,aichingm/electron,chriskdon/electron,IonicaBizauKitchen/electron,fabien-d/electron,anko/electron,medixdev/electron,arusakov/electron,tinydew4/electron,xiruibing/electron,the-ress/electron | #!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
Upgrade libchromiumcontent to loose iframe sandbox. | #!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
| <commit_before>#!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
<commit_msg>Upgrade libchromiumcontent to loose iframe sandbox.<commit_after> | #!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
| #!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
Upgrade libchromiumcontent to loose iframe sandbox.#!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
| <commit_before>#!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
<commit_msg>Upgrade libchromiumcontent to loose iframe sandbox.<commit_after>#!/usr/bin/env python
import platform
import sys
NODE_VERSION = 'v0.11.13'
BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent'
LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5'
ARCH = {
'cygwin': '32bit',
'darwin': '64bit',
'linux2': platform.architecture()[0],
'win32': '32bit',
}[sys.platform]
DIST_ARCH = {
'32bit': 'ia32',
'64bit': 'x64',
}[ARCH]
TARGET_PLATFORM = {
'cygwin': 'win32',
'darwin': 'darwin',
'linux2': 'linux',
'win32': 'win32',
}[sys.platform]
|
5c0ace537a073f3d851ad4e490a7f2b5a0062c62 | tfr/features.py | tfr/features.py | import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
if __name__ == '__main__':
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
| """
Example usage:
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
"""
import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
| Put the code into a comment as an example usage. | Put the code into a comment as an example usage.
| Python | mit | bzamecnik/tfr,bzamecnik/tfr | import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
if __name__ == '__main__':
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
Put the code into a comment as an example usage. | """
Example usage:
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
"""
import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
| <commit_before>import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
if __name__ == '__main__':
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
<commit_msg>Put the code into a comment as an example usage.<commit_after> | """
Example usage:
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
"""
import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
| import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
if __name__ == '__main__':
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
Put the code into a comment as an example usage."""
Example usage:
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
"""
import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
| <commit_before>import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
if __name__ == '__main__':
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
<commit_msg>Put the code into a comment as an example usage.<commit_after>"""
Example usage:
import matplotlib.pyplot as plt
from files import load_wav
from analysis import split_to_blocks
def analyze_mean_energy(file, block_size=1024):
x, fs = load_wav(file)
blocks, t = split_to_blocks(x, block_size)
y = mean_energy(blocks)
plt.semilogy(t, y)
plt.ylim(0, 1)
"""
import numpy as np
def mean_power(x_blocks):
return np.sqrt(np.mean(x_blocks**2, axis=-1))
def power(x_blocks):
return np.sqrt(np.sum(x_blocks**2, axis=-1))
def mean_energy(x_blocks):
return np.mean(x_blocks**2, axis=-1)
def energy(x_blocks):
return np.sum(x_blocks**2, axis=-1)
|
744c91ca30379d6cca7f7f9fc2b014e0f29e55e4 | keepaneyeon/http.py | keepaneyeon/http.py | import requests
class HttpDownloader():
def __init__(self, opts={}):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
| import requests
class HttpDownloader():
def __init__(self, **opts):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
| Make HTTPDownloader work with YAML config | Make HTTPDownloader work with YAML config
| Python | mit | mmcloughlin/keepaneyeon | import requests
class HttpDownloader():
def __init__(self, opts={}):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
Make HTTPDownloader work with YAML config | import requests
class HttpDownloader():
def __init__(self, **opts):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
| <commit_before>import requests
class HttpDownloader():
def __init__(self, opts={}):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
<commit_msg>Make HTTPDownloader work with YAML config<commit_after> | import requests
class HttpDownloader():
def __init__(self, **opts):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
| import requests
class HttpDownloader():
def __init__(self, opts={}):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
Make HTTPDownloader work with YAML configimport requests
class HttpDownloader():
def __init__(self, **opts):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
| <commit_before>import requests
class HttpDownloader():
def __init__(self, opts={}):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
<commit_msg>Make HTTPDownloader work with YAML config<commit_after>import requests
class HttpDownloader():
def __init__(self, **opts):
self.base = opts
def build_request_options(self, opts):
options = {'method': 'get'}
options.update(self.base)
options.update(opts)
options.update({'stream': True})
return options
def download(self, opts, out):
options = self.build_request_options(opts)
r = requests.request(**options)
r.raise_for_status()
for chunk in r.iter_content(1024):
out.write(chunk)
|
e3c53133b71d7426695fbf24cac5b8e82311c037 | seeker/middleware.py | seeker/middleware.py | from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
| from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
| Make signal dispatch_uid values more specific | Make signal dispatch_uid values more specific | Python | bsd-2-clause | imsweb/django-seeker,imsweb/django-seeker | from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
Make signal dispatch_uid values more specific | from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
| <commit_before>from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
<commit_msg>Make signal dispatch_uid values more specific<commit_after> | from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
| from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
Make signal dispatch_uid values more specificfrom .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
| <commit_before>from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
<commit_msg>Make signal dispatch_uid values more specific<commit_after>from .utils import index, delete
from django.db import models
import logging
logger = logging.getLogger(__name__)
class ModelIndexingMiddleware (object):
"""
Middleware class that automatically indexes any new or deleted model objects.
"""
def __init__(self):
models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save')
models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete')
def handle_save(self, sender, instance, **kwargs):
try:
index(instance)
except:
logger.exception('Error indexing %s instance: %s', sender, instance)
def handle_delete(self, sender, instance, **kwargs):
try:
delete(instance)
except:
logger.exception('Error deleting %s instance: %s', sender, instance)
def process_request(self, request):
# This is really just here so Django keeps the middleware installed.
pass
|
56d92af9ba0a9b81dd0e802d05717ec6e4f511d3 | seven23/api/views.py | seven23/api/views.py | """
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
| """
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d")
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
| Fix bug on API with date in Terms and Conditions not serializable | Fix bug on API with date in Terms and Conditions not serializable
| Python | mit | sebastienbarbier/723e_server,sebastienbarbier/723e,sebastienbarbier/723e_server,sebastienbarbier/723e | """
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
Fix bug on API with date in Terms and Conditions not serializable | """
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d")
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
| <commit_before>"""
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
<commit_msg>Fix bug on API with date in Terms and Conditions not serializable<commit_after> | """
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d")
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
| """
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
Fix bug on API with date in Terms and Conditions not serializable"""
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d")
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
| <commit_before>"""
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
<commit_msg>Fix bug on API with date in Terms and Conditions not serializable<commit_after>"""
Root views of api
"""
import json
import os
import markdown2
from django.http import HttpResponse
from django.db import models
from rest_framework.decorators import api_view
from seven23 import settings
from seven23.models.terms.models import TermsAndConditions
@api_view(["GET"])
def api_init(request):
"""
Return status on client initialisation
"""
result = {}
# Return API Version.
result['api_version'] = settings.API_VERSION
result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION
result['contact'] = settings.CONTACT_EMAIL
try:
terms = TermsAndConditions.objects.latest('date')
result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d")
result['terms_and_conditions'] = markdown2.markdown(terms.markdown)
except TermsAndConditions.DoesNotExist:
result['terms_and_conditions_date'] = None
result['terms_and_conditions'] = None
if request.user.is_authenticated():
result['is_authenticated'] = True
result['id'] = request.user.id
else:
result['is_authenticated'] = False
# Return json format string.
j = json.dumps(result, separators=(',', ':'))
return HttpResponse(j, content_type='application/json')
|
fecf53c0c4414f50a9c3937b05d27de8c1387c45 | src/hireme/tasks/task2.py | src/hireme/tasks/task2.py | # -*- coding: utf-8 -*-
from . import render_task
@render_task
def solve():
return dict(
solution='42',
title='task2'
) | # -*- coding: utf-8 -*-
import re
from flask import request
from werkzeug import exceptions
import numpy as np
from . import render_task
@render_task
def solve():
input_data = request.form.get('input')
method = request.method
title = 'task2'
if method == 'GET':
return dict(
title=title
)
lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')]
cases = int(0 if not lines else lines.pop(0))
counts = [1] * cases
for case in xrange(cases):
if not len(lines):
raise exceptions.BadRequest(
description='Specified %s cases, but only provided %s.' %
(cases, case)
)
dimension = int(lines.pop(0))
matrix = np.array([[int(n) for n in m] for m in lines[:dimension]])
lines = lines[dimension:]
if not matrix.shape == (dimension,) * 2:
raise exceptions.BadRequest(
description='Expected %s-dimensional matrix for case %s.' %
(dimension, case + 1)
)
def neighbours(idx):
def along_axis(axis):
for offset in (-1, 0, 1):
candidate = idx[axis] + offset
if candidate >= 0 and candidate < dimension:
yield candidate
hood = []
for x in along_axis(0):
for y in along_axis(1):
if (x, y) != idx and matrix[x, y] == 1:
hood.append((x, y))
return hood
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
while untouched:
def expand(resident):
matrix[resident] = counts[case]
hood = neighbours(resident)
if hood:
for neighbour in hood:
expand(neighbour)
counts[case] += 1
expand(untouched[0])
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
solution = '\n'.join([str(c - 1) for c in counts])
return dict(
input=input_data,
solution=solution,
title=title
)
| Implement rudimentary task 2 solution | Implement rudimentary task 2 solution
| Python | bsd-2-clause | cutoffthetop/hireme | # -*- coding: utf-8 -*-
from . import render_task
@render_task
def solve():
return dict(
solution='42',
title='task2'
)Implement rudimentary task 2 solution | # -*- coding: utf-8 -*-
import re
from flask import request
from werkzeug import exceptions
import numpy as np
from . import render_task
@render_task
def solve():
input_data = request.form.get('input')
method = request.method
title = 'task2'
if method == 'GET':
return dict(
title=title
)
lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')]
cases = int(0 if not lines else lines.pop(0))
counts = [1] * cases
for case in xrange(cases):
if not len(lines):
raise exceptions.BadRequest(
description='Specified %s cases, but only provided %s.' %
(cases, case)
)
dimension = int(lines.pop(0))
matrix = np.array([[int(n) for n in m] for m in lines[:dimension]])
lines = lines[dimension:]
if not matrix.shape == (dimension,) * 2:
raise exceptions.BadRequest(
description='Expected %s-dimensional matrix for case %s.' %
(dimension, case + 1)
)
def neighbours(idx):
def along_axis(axis):
for offset in (-1, 0, 1):
candidate = idx[axis] + offset
if candidate >= 0 and candidate < dimension:
yield candidate
hood = []
for x in along_axis(0):
for y in along_axis(1):
if (x, y) != idx and matrix[x, y] == 1:
hood.append((x, y))
return hood
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
while untouched:
def expand(resident):
matrix[resident] = counts[case]
hood = neighbours(resident)
if hood:
for neighbour in hood:
expand(neighbour)
counts[case] += 1
expand(untouched[0])
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
solution = '\n'.join([str(c - 1) for c in counts])
return dict(
input=input_data,
solution=solution,
title=title
)
| <commit_before># -*- coding: utf-8 -*-
from . import render_task
@render_task
def solve():
return dict(
solution='42',
title='task2'
)<commit_msg>Implement rudimentary task 2 solution<commit_after> | # -*- coding: utf-8 -*-
import re
from flask import request
from werkzeug import exceptions
import numpy as np
from . import render_task
@render_task
def solve():
input_data = request.form.get('input')
method = request.method
title = 'task2'
if method == 'GET':
return dict(
title=title
)
lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')]
cases = int(0 if not lines else lines.pop(0))
counts = [1] * cases
for case in xrange(cases):
if not len(lines):
raise exceptions.BadRequest(
description='Specified %s cases, but only provided %s.' %
(cases, case)
)
dimension = int(lines.pop(0))
matrix = np.array([[int(n) for n in m] for m in lines[:dimension]])
lines = lines[dimension:]
if not matrix.shape == (dimension,) * 2:
raise exceptions.BadRequest(
description='Expected %s-dimensional matrix for case %s.' %
(dimension, case + 1)
)
def neighbours(idx):
def along_axis(axis):
for offset in (-1, 0, 1):
candidate = idx[axis] + offset
if candidate >= 0 and candidate < dimension:
yield candidate
hood = []
for x in along_axis(0):
for y in along_axis(1):
if (x, y) != idx and matrix[x, y] == 1:
hood.append((x, y))
return hood
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
while untouched:
def expand(resident):
matrix[resident] = counts[case]
hood = neighbours(resident)
if hood:
for neighbour in hood:
expand(neighbour)
counts[case] += 1
expand(untouched[0])
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
solution = '\n'.join([str(c - 1) for c in counts])
return dict(
input=input_data,
solution=solution,
title=title
)
| # -*- coding: utf-8 -*-
from . import render_task
@render_task
def solve():
return dict(
solution='42',
title='task2'
)Implement rudimentary task 2 solution# -*- coding: utf-8 -*-
import re
from flask import request
from werkzeug import exceptions
import numpy as np
from . import render_task
@render_task
def solve():
input_data = request.form.get('input')
method = request.method
title = 'task2'
if method == 'GET':
return dict(
title=title
)
lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')]
cases = int(0 if not lines else lines.pop(0))
counts = [1] * cases
for case in xrange(cases):
if not len(lines):
raise exceptions.BadRequest(
description='Specified %s cases, but only provided %s.' %
(cases, case)
)
dimension = int(lines.pop(0))
matrix = np.array([[int(n) for n in m] for m in lines[:dimension]])
lines = lines[dimension:]
if not matrix.shape == (dimension,) * 2:
raise exceptions.BadRequest(
description='Expected %s-dimensional matrix for case %s.' %
(dimension, case + 1)
)
def neighbours(idx):
def along_axis(axis):
for offset in (-1, 0, 1):
candidate = idx[axis] + offset
if candidate >= 0 and candidate < dimension:
yield candidate
hood = []
for x in along_axis(0):
for y in along_axis(1):
if (x, y) != idx and matrix[x, y] == 1:
hood.append((x, y))
return hood
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
while untouched:
def expand(resident):
matrix[resident] = counts[case]
hood = neighbours(resident)
if hood:
for neighbour in hood:
expand(neighbour)
counts[case] += 1
expand(untouched[0])
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
solution = '\n'.join([str(c - 1) for c in counts])
return dict(
input=input_data,
solution=solution,
title=title
)
| <commit_before># -*- coding: utf-8 -*-
from . import render_task
@render_task
def solve():
return dict(
solution='42',
title='task2'
)<commit_msg>Implement rudimentary task 2 solution<commit_after># -*- coding: utf-8 -*-
import re
from flask import request
from werkzeug import exceptions
import numpy as np
from . import render_task
@render_task
def solve():
input_data = request.form.get('input')
method = request.method
title = 'task2'
if method == 'GET':
return dict(
title=title
)
lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')]
cases = int(0 if not lines else lines.pop(0))
counts = [1] * cases
for case in xrange(cases):
if not len(lines):
raise exceptions.BadRequest(
description='Specified %s cases, but only provided %s.' %
(cases, case)
)
dimension = int(lines.pop(0))
matrix = np.array([[int(n) for n in m] for m in lines[:dimension]])
lines = lines[dimension:]
if not matrix.shape == (dimension,) * 2:
raise exceptions.BadRequest(
description='Expected %s-dimensional matrix for case %s.' %
(dimension, case + 1)
)
def neighbours(idx):
def along_axis(axis):
for offset in (-1, 0, 1):
candidate = idx[axis] + offset
if candidate >= 0 and candidate < dimension:
yield candidate
hood = []
for x in along_axis(0):
for y in along_axis(1):
if (x, y) != idx and matrix[x, y] == 1:
hood.append((x, y))
return hood
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
while untouched:
def expand(resident):
matrix[resident] = counts[case]
hood = neighbours(resident)
if hood:
for neighbour in hood:
expand(neighbour)
counts[case] += 1
expand(untouched[0])
untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)])
solution = '\n'.join([str(c - 1) for c in counts])
return dict(
input=input_data,
solution=solution,
title=title
)
|
84041a2bb517841d725781bdd72b1daf4f8e603d | spacy/ja/__init__.py | spacy/ja/__init__.py | # encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class Japanese(Language):
lang = 'ja'
def make_doc(self, text):
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
words = [x.surface for x in Tokenizer().tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
| # encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language, BaseDefaults
from ..tokenizer import Tokenizer
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class JapaneseTokenizer(object):
def __init__(self, cls, nlp=None):
self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp)
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
self.tokenizer = Tokenizer()
def __call__(self, text):
words = [x.surface for x in self.tokenizer.tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
class JapaneseDefaults(BaseDefaults):
@classmethod
def create_tokenizer(cls, nlp=None):
return JapaneseTokenizer(cls, nlp)
class Japanese(Language):
lang = 'ja'
Defaults = JapaneseDefaults
def make_doc(self, text):
words = self.tokenizer(text)
return Doc(self.vocab, words=words, spaces=[False]*len(words))
| Make create_tokenizer work with Japanese | Make create_tokenizer work with Japanese
| Python | mit | spacy-io/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,raphael0202/spaCy,aikramer2/spaCy,honnibal/spaCy,spacy-io/spaCy,raphael0202/spaCy,raphael0202/spaCy,raphael0202/spaCy,spacy-io/spaCy,spacy-io/spaCy,explosion/spaCy,raphael0202/spaCy,honnibal/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy,raphael0202/spaCy,recognai/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,spacy-io/spaCy,aikramer2/spaCy,honnibal/spaCy | # encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class Japanese(Language):
lang = 'ja'
def make_doc(self, text):
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
words = [x.surface for x in Tokenizer().tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
Make create_tokenizer work with Japanese | # encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language, BaseDefaults
from ..tokenizer import Tokenizer
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class JapaneseTokenizer(object):
def __init__(self, cls, nlp=None):
self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp)
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
self.tokenizer = Tokenizer()
def __call__(self, text):
words = [x.surface for x in self.tokenizer.tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
class JapaneseDefaults(BaseDefaults):
@classmethod
def create_tokenizer(cls, nlp=None):
return JapaneseTokenizer(cls, nlp)
class Japanese(Language):
lang = 'ja'
Defaults = JapaneseDefaults
def make_doc(self, text):
words = self.tokenizer(text)
return Doc(self.vocab, words=words, spaces=[False]*len(words))
| <commit_before># encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class Japanese(Language):
lang = 'ja'
def make_doc(self, text):
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
words = [x.surface for x in Tokenizer().tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
<commit_msg>Make create_tokenizer work with Japanese<commit_after> | # encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language, BaseDefaults
from ..tokenizer import Tokenizer
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class JapaneseTokenizer(object):
def __init__(self, cls, nlp=None):
self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp)
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
self.tokenizer = Tokenizer()
def __call__(self, text):
words = [x.surface for x in self.tokenizer.tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
class JapaneseDefaults(BaseDefaults):
@classmethod
def create_tokenizer(cls, nlp=None):
return JapaneseTokenizer(cls, nlp)
class Japanese(Language):
lang = 'ja'
Defaults = JapaneseDefaults
def make_doc(self, text):
words = self.tokenizer(text)
return Doc(self.vocab, words=words, spaces=[False]*len(words))
| # encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class Japanese(Language):
lang = 'ja'
def make_doc(self, text):
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
words = [x.surface for x in Tokenizer().tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
Make create_tokenizer work with Japanese# encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language, BaseDefaults
from ..tokenizer import Tokenizer
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class JapaneseTokenizer(object):
def __init__(self, cls, nlp=None):
self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp)
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
self.tokenizer = Tokenizer()
def __call__(self, text):
words = [x.surface for x in self.tokenizer.tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
class JapaneseDefaults(BaseDefaults):
@classmethod
def create_tokenizer(cls, nlp=None):
return JapaneseTokenizer(cls, nlp)
class Japanese(Language):
lang = 'ja'
Defaults = JapaneseDefaults
def make_doc(self, text):
words = self.tokenizer(text)
return Doc(self.vocab, words=words, spaces=[False]*len(words))
| <commit_before># encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class Japanese(Language):
lang = 'ja'
def make_doc(self, text):
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
words = [x.surface for x in Tokenizer().tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
<commit_msg>Make create_tokenizer work with Japanese<commit_after># encoding: utf8
from __future__ import unicode_literals, print_function
from os import path
from ..language import Language, BaseDefaults
from ..tokenizer import Tokenizer
from ..attrs import LANG
from ..tokens import Doc
from .language_data import *
class JapaneseTokenizer(object):
def __init__(self, cls, nlp=None):
self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp)
try:
from janome.tokenizer import Tokenizer
except ImportError:
raise ImportError("The Japanese tokenizer requires the Janome library: "
"https://github.com/mocobeta/janome")
self.tokenizer = Tokenizer()
def __call__(self, text):
words = [x.surface for x in self.tokenizer.tokenize(text)]
return Doc(self.vocab, words=words, spaces=[False]*len(words))
class JapaneseDefaults(BaseDefaults):
@classmethod
def create_tokenizer(cls, nlp=None):
return JapaneseTokenizer(cls, nlp)
class Japanese(Language):
lang = 'ja'
Defaults = JapaneseDefaults
def make_doc(self, text):
words = self.tokenizer(text)
return Doc(self.vocab, words=words, spaces=[False]*len(words))
|
cec423c4a1e633193ef3e639a1cb202bddc27e18 | api/base/content_negotiation.py | api/base/content_negotiation.py |
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
Returns a tuple (renderer, media_type).
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
|
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
Returns appropriate tuple (renderer, media type).
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
| Add one-line summary to docstring. | Add one-line summary to docstring.
| Python | apache-2.0 | hmoco/osf.io,doublebits/osf.io,leb2dg/osf.io,doublebits/osf.io,erinspace/osf.io,billyhunt/osf.io,jnayak1/osf.io,felliott/osf.io,sbt9uc/osf.io,MerlinZhang/osf.io,caneruguz/osf.io,ZobairAlijan/osf.io,RomanZWang/osf.io,pattisdr/osf.io,RomanZWang/osf.io,zachjanicki/osf.io,kwierman/osf.io,ticklemepierce/osf.io,icereval/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,CenterForOpenScience/osf.io,leb2dg/osf.io,erinspace/osf.io,ticklemepierce/osf.io,wearpants/osf.io,billyhunt/osf.io,binoculars/osf.io,Nesiehr/osf.io,ZobairAlijan/osf.io,HalcyonChimera/osf.io,SSJohns/osf.io,kch8qx/osf.io,acshi/osf.io,brandonPurvis/osf.io,TomHeatwole/osf.io,zachjanicki/osf.io,hmoco/osf.io,DanielSBrown/osf.io,acshi/osf.io,MerlinZhang/osf.io,chennan47/osf.io,KAsante95/osf.io,alexschiller/osf.io,samanehsan/osf.io,mfraezz/osf.io,Ghalko/osf.io,chennan47/osf.io,sbt9uc/osf.io,zamattiac/osf.io,rdhyee/osf.io,cosenal/osf.io,caseyrygt/osf.io,caseyrollins/osf.io,abought/osf.io,jnayak1/osf.io,amyshi188/osf.io,acshi/osf.io,TomBaxter/osf.io,mluo613/osf.io,binoculars/osf.io,doublebits/osf.io,Nesiehr/osf.io,danielneis/osf.io,GageGaskins/osf.io,DanielSBrown/osf.io,cslzchen/osf.io,erinspace/osf.io,cslzchen/osf.io,Ghalko/osf.io,danielneis/osf.io,SSJohns/osf.io,cslzchen/osf.io,asanfilippo7/osf.io,saradbowman/osf.io,kwierman/osf.io,alexschiller/osf.io,icereval/osf.io,laurenrevere/osf.io,samanehsan/osf.io,abought/osf.io,mluke93/osf.io,ckc6cz/osf.io,monikagrabowska/osf.io,RomanZWang/osf.io,chrisseto/osf.io,zamattiac/osf.io,mfraezz/osf.io,caseyrollins/osf.io,samchrisinger/osf.io,petermalcolm/osf.io,arpitar/osf.io,icereval/osf.io,adlius/osf.io,ckc6cz/osf.io,mluo613/osf.io,amyshi188/osf.io,abought/osf.io,samchrisinger/osf.io,sloria/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,cslzchen/osf.io,cosenal/osf.io,samanehsan/osf.io,jnayak1/osf.io,SSJohns/osf.io,sbt9uc/osf.io,aaxelb/osf.io,felliott/osf.io,SSJohns/osf.io,GageGaskins/osf.io,binoculars/osf.io,kch8qx/osf.io,jmcarp/osf.io,sloria/osf.io,chrisseto/osf.io,monikagrabowska/osf.io,monikagrabowska/osf.io,KAsante95/osf.io,chrisseto/osf.io,Ghalko/osf.io,CenterForOpenScience/osf.io,GageGaskins/osf.io,haoyuchen1992/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,amyshi188/osf.io,brianjgeiger/osf.io,caseyrygt/osf.io,wearpants/osf.io,leb2dg/osf.io,cwisecarver/osf.io,arpitar/osf.io,danielneis/osf.io,doublebits/osf.io,monikagrabowska/osf.io,pattisdr/osf.io,wearpants/osf.io,brianjgeiger/osf.io,DanielSBrown/osf.io,emetsger/osf.io,njantrania/osf.io,aaxelb/osf.io,zachjanicki/osf.io,caneruguz/osf.io,mattclark/osf.io,chennan47/osf.io,aaxelb/osf.io,saradbowman/osf.io,mfraezz/osf.io,GageGaskins/osf.io,samchrisinger/osf.io,ticklemepierce/osf.io,caneruguz/osf.io,petermalcolm/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,alexschiller/osf.io,njantrania/osf.io,CenterForOpenScience/osf.io,rdhyee/osf.io,baylee-d/osf.io,cosenal/osf.io,alexschiller/osf.io,caseyrygt/osf.io,acshi/osf.io,GageGaskins/osf.io,ckc6cz/osf.io,samchrisinger/osf.io,danielneis/osf.io,adlius/osf.io,asanfilippo7/osf.io,felliott/osf.io,jmcarp/osf.io,amyshi188/osf.io,zamattiac/osf.io,adlius/osf.io,arpitar/osf.io,alexschiller/osf.io,rdhyee/osf.io,KAsante95/osf.io,ZobairAlijan/osf.io,Johnetordoff/osf.io,Nesiehr/osf.io,billyhunt/osf.io,sbt9uc/osf.io,wearpants/osf.io,emetsger/osf.io,brianjgeiger/osf.io,kch8qx/osf.io,rdhyee/osf.io,caseyrygt/osf.io,billyhunt/osf.io,felliott/osf.io,mattclark/osf.io,kwierman/osf.io,Nesiehr/osf.io,KAsante95/osf.io,laurenrevere/osf.io,ckc6cz/osf.io,DanielSBrown/osf.io,brandonPurvis/osf.io,baylee-d/osf.io,hmoco/osf.io,haoyuchen1992/osf.io,RomanZWang/osf.io,samanehsan/osf.io,petermalcolm/osf.io,mluke93/osf.io,hmoco/osf.io,cwisecarver/osf.io,caneruguz/osf.io,ticklemepierce/osf.io,Johnetordoff/osf.io,jnayak1/osf.io,haoyuchen1992/osf.io,CenterForOpenScience/osf.io,kwierman/osf.io,cwisecarver/osf.io,doublebits/osf.io,sloria/osf.io,mattclark/osf.io,Johnetordoff/osf.io,mluo613/osf.io,MerlinZhang/osf.io,brandonPurvis/osf.io,KAsante95/osf.io,petermalcolm/osf.io,MerlinZhang/osf.io,ZobairAlijan/osf.io,jmcarp/osf.io,asanfilippo7/osf.io,brianjgeiger/osf.io,cwisecarver/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,mluke93/osf.io,njantrania/osf.io,brandonPurvis/osf.io,cosenal/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,Johnetordoff/osf.io,mluo613/osf.io,pattisdr/osf.io,njantrania/osf.io,billyhunt/osf.io,kch8qx/osf.io,mluo613/osf.io,mluke93/osf.io,TomHeatwole/osf.io,Ghalko/osf.io,caseyrollins/osf.io,arpitar/osf.io,jmcarp/osf.io,emetsger/osf.io,zamattiac/osf.io,adlius/osf.io,kch8qx/osf.io,TomBaxter/osf.io,baylee-d/osf.io,abought/osf.io,leb2dg/osf.io,acshi/osf.io,mfraezz/osf.io,emetsger/osf.io,aaxelb/osf.io,RomanZWang/osf.io,crcresearch/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io |
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
Returns a tuple (renderer, media_type).
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
Add one-line summary to docstring. |
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
Returns appropriate tuple (renderer, media type).
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
| <commit_before>
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
Returns a tuple (renderer, media_type).
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
<commit_msg>Add one-line summary to docstring.<commit_after> |
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
Returns appropriate tuple (renderer, media type).
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
|
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
Returns a tuple (renderer, media_type).
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
Add one-line summary to docstring.
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
Returns appropriate tuple (renderer, media type).
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
| <commit_before>
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
Returns a tuple (renderer, media_type).
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
<commit_msg>Add one-line summary to docstring.<commit_after>
from rest_framework.negotiation import DefaultContentNegotiation
class JSONAPIContentNegotiation(DefaultContentNegotiation):
def select_renderer(self, request, renderers, format_suffix=None):
"""
Returns appropriate tuple (renderer, media type).
If 'application/json' in acceptable media types, use the first renderer in
DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'.
Media_type "application/vnd.api+json". Otherwise, use default select_renderer.
"""
accepts = self.get_accept_list(request)
if 'application/json' in accepts:
return (renderers[0], renderers[0].media_type)
return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
|
2604d759bfd9a18e5e594cfa5b50e83c73fbc2d8 | devito/interfaces.py | devito/interfaces.py | import numpy as np
from sympy import IndexedBase
class MatrixData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(MatrixData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
MatrixData._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
| import numpy as np
from sympy import IndexedBase
class DenseData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(DenseData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
super(TimeData, self)._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
| Change name from MatrixData to DenseData | Change name from MatrixData to DenseData
| Python | mit | opesci/devito,opesci/devito | import numpy as np
from sympy import IndexedBase
class MatrixData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(MatrixData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
MatrixData._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
Change name from MatrixData to DenseData | import numpy as np
from sympy import IndexedBase
class DenseData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(DenseData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
super(TimeData, self)._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
| <commit_before>import numpy as np
from sympy import IndexedBase
class MatrixData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(MatrixData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
MatrixData._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
<commit_msg>Change name from MatrixData to DenseData<commit_after> | import numpy as np
from sympy import IndexedBase
class DenseData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(DenseData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
super(TimeData, self)._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
| import numpy as np
from sympy import IndexedBase
class MatrixData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(MatrixData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
MatrixData._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
Change name from MatrixData to DenseDataimport numpy as np
from sympy import IndexedBase
class DenseData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(DenseData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
super(TimeData, self)._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
| <commit_before>import numpy as np
from sympy import IndexedBase
class MatrixData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(MatrixData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
MatrixData._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
<commit_msg>Change name from MatrixData to DenseData<commit_after>import numpy as np
from sympy import IndexedBase
class DenseData(IndexedBase):
def __init__(self, name, shape, dtype):
self.name = name
self.shape = shape
self.dtype = dtype
self.pointer = None
self.initializer = None
def set_initializer(self, lambda_initializer):
assert(callable(lambda_initializer))
self.initializer = lambda_initializer
def initialize(self):
assert(self.initializer is not None)
self.initializer(self.data)
def _allocate_memory(self):
self.pointer = np.zeros(self.shape, self.dtype, order='C')
@property
def data(self):
if self.pointer is None:
self._allocate_memory()
return self.pointer
class TimeData(DenseData):
def __init__(self, name, spc_shape, time_dim, time_order, save, dtype):
if save:
time_dim = time_dim + time_order
else:
time_dim = time_order + 1
shape = tuple((time_dim,) + spc_shape)
super(TimeData, self).__init__(name, shape, dtype)
self.save = save
self.time_order = time_order
def _allocate_memory(self):
super(TimeData, self)._allocate_memory(self)
if self.pad_time:
self.pointer = self.pointer[self.time_order]
|
840d80c543c4688ebd1bda41b8689cf404bf755c | edit_spectide.py | edit_spectide.py | """
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
| """
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
Lee's changeNC file should live in the SVN repository.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
| Update help to indicate where the necessary script lives | Update help to indicate where the necessary script lives
| Python | mit | pwcazenave/PyFVCOM | """
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
Update help to indicate where the necessary script lives | """
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
Lee's changeNC file should live in the SVN repository.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
| <commit_before>"""
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
<commit_msg>Update help to indicate where the necessary script lives<commit_after> | """
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
Lee's changeNC file should live in the SVN repository.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
| """
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
Update help to indicate where the necessary script lives"""
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
Lee's changeNC file should live in the SVN repository.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
| <commit_before>"""
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
<commit_msg>Update help to indicate where the necessary script lives<commit_after>"""
Edits the spectide amplitude values to some factor of their original value.
WARNING: When using this on FVCOM input files, it will change the format of
the variables. changeNC presumes each variable has a value and unit associated
with it, whereas some of the variables in the FVCOM inputs are in fact not
that sort of data, and so have different structures. Probably best to use the
combination of ncdump and ncgen to edit the values as text and generate a new
NetCDF file from that edited text.
Lee's changeNC file should live in the SVN repository.
"""
from changeNC import *
infile = './co2_spectide.nc'
outfile = './co2_spectide_scaled.nc'
scaleFact = 0.75
av = AutoVivification()
av['tide_Eamp']['convert'] = lambda x:x*scaleFact
changeNC(infile, outfile, av)
|
127b90c88d1362e7b10e7bf36dff56b96a5c4f0b | simpegEM/FDEM/__init__.py | simpegEM/FDEM/__init__.py | from SurveyFDEM import *
from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h
| from SurveyFDEM import *
from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
| Add more files to export on the init. | Add more files to export on the init.
| Python | mit | simpeg/discretize,lheagy/simpegem,simpeg/discretize,simpeg/discretize,simpeg/simpeg,simpeg/simpegem | from SurveyFDEM import *
from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h
Add more files to export on the init. | from SurveyFDEM import *
from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
| <commit_before>from SurveyFDEM import *
from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h
<commit_msg>Add more files to export on the init.<commit_after> | from SurveyFDEM import *
from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
| from SurveyFDEM import *
from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h
Add more files to export on the init.from SurveyFDEM import *
from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
| <commit_before>from SurveyFDEM import *
from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h
<commit_msg>Add more files to export on the init.<commit_after>from SurveyFDEM import *
from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
|
8ef4ca2166167f6370dd8c2f724e752210adf067 | sirius/SI_V07/__init__.py | sirius/SI_V07/__init__.py | from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| Fix bug when family_data.py was deleted | Fix bug when family_data.py was deleted
| Python | mit | lnls-fac/sirius | from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
Fix bug when family_data.py was deleted | from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| <commit_before>from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
<commit_msg>Fix bug when family_data.py was deleted<commit_after> | from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
Fix bug when family_data.py was deletedfrom . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
| <commit_before>from . import lattice as _lattice
from . import accelerator as _accelerator
from . import family_data as _family_data
from . import record_names
create_accelerator = _accelerator.create_accelerator
get_family_data = _family_data.get_family_data
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
<commit_msg>Fix bug when family_data.py was deleted<commit_after>from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = _accelerator.create_accelerator
# -- default accelerator values for SI_V07 --
energy = _lattice._energy
harmonic_number = _lattice._harmonic_number
default_cavity_on = _accelerator._default_cavity_on
default_radiation_on = _accelerator._default_cavity_on
default_vchamber_on = _accelerator._default_vchamber_on
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'SI_V07'
lattice_symmetry = _lattice._lattice_symmetry
family_data = _lattice._family_data
family_mapping = _lattice._family_mapping
global_coupling = 0.01 # expected corrected value
average_pressure = 1.333e-9 # average pressure [mbar]
|
55072134b8053ac126213e580fcc59977cfb7a02 | scikits/image/setup.py | scikits/image/setup.py | import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
| import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
config.add_subpackage('draw')
config.add_subpackage('feature')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
| Add 'draw' and 'feature' sub-modules. | BUG: Add 'draw' and 'feature' sub-modules.
| Python | bsd-3-clause | paalge/scikit-image,michaelaye/scikit-image,bennlich/scikit-image,chintak/scikit-image,paalge/scikit-image,ClinicalGraphics/scikit-image,warmspringwinds/scikit-image,michaelaye/scikit-image,chriscrosscutler/scikit-image,ClinicalGraphics/scikit-image,WarrenWeckesser/scikits-image,emmanuelle/scikits.image,Midafi/scikit-image,emmanuelle/scikits.image,WarrenWeckesser/scikits-image,bennlich/scikit-image,GaZ3ll3/scikit-image,michaelpacer/scikit-image,SamHames/scikit-image,vighneshbirodkar/scikit-image,almarklein/scikit-image,rjeli/scikit-image,blink1073/scikit-image,jwiggins/scikit-image,emmanuelle/scikits.image,emon10005/scikit-image,SamHames/scikit-image,warmspringwinds/scikit-image,almarklein/scikit-image,almarklein/scikit-image,robintw/scikit-image,keflavich/scikit-image,Midafi/scikit-image,ajaybhat/scikit-image,ajaybhat/scikit-image,dpshelio/scikit-image,emmanuelle/scikits.image,youprofit/scikit-image,jwiggins/scikit-image,SamHames/scikit-image,emon10005/scikit-image,ofgulban/scikit-image,paalge/scikit-image,chintak/scikit-image,newville/scikit-image,juliusbierk/scikit-image,keflavich/scikit-image,Hiyorimi/scikit-image,Hiyorimi/scikit-image,chintak/scikit-image,bsipocz/scikit-image,newville/scikit-image,robintw/scikit-image,vighneshbirodkar/scikit-image,Britefury/scikit-image,youprofit/scikit-image,chintak/scikit-image,oew1v07/scikit-image,blink1073/scikit-image,ofgulban/scikit-image,dpshelio/scikit-image,pratapvardhan/scikit-image,michaelpacer/scikit-image,pratapvardhan/scikit-image,chriscrosscutler/scikit-image,vighneshbirodkar/scikit-image,bsipocz/scikit-image,ofgulban/scikit-image,juliusbierk/scikit-image,Britefury/scikit-image,rjeli/scikit-image,GaZ3ll3/scikit-image,rjeli/scikit-image,SamHames/scikit-image,almarklein/scikit-image,oew1v07/scikit-image | import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
BUG: Add 'draw' and 'feature' sub-modules. | import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
config.add_subpackage('draw')
config.add_subpackage('feature')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
| <commit_before>import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
<commit_msg>BUG: Add 'draw' and 'feature' sub-modules.<commit_after> | import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
config.add_subpackage('draw')
config.add_subpackage('feature')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
| import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
BUG: Add 'draw' and 'feature' sub-modules.import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
config.add_subpackage('draw')
config.add_subpackage('feature')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
| <commit_before>import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
<commit_msg>BUG: Add 'draw' and 'feature' sub-modules.<commit_after>import os
def configuration(parent_package='', top_path=None):
from numpy.distutils.misc_util import Configuration
config = Configuration('image', parent_package, top_path)
config.add_subpackage('opencv')
config.add_subpackage('graph')
config.add_subpackage('io')
config.add_subpackage('morphology')
config.add_subpackage('filter')
config.add_subpackage('transform')
config.add_subpackage('data')
config.add_subpackage('util')
config.add_subpackage('color')
config.add_subpackage('draw')
config.add_subpackage('feature')
def add_test_directories(arg, dirname, fnames):
if dirname.split(os.path.sep)[-1] == 'tests':
config.add_data_dir(dirname)
# Add test directories
from os.path import isdir, dirname, join, abspath
rel_isdir = lambda d: isdir(join(curpath, d))
curpath = join(dirname(__file__), './')
subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)]
subdirs = [d for d in subdirs if rel_isdir(d)]
for test_dir in subdirs:
config.add_data_dir(test_dir)
return config
if __name__ == "__main__":
from numpy.distutils.core import setup
config = Configuration(top_path='').todict()
setup(**config)
|
7eadc9e514b1311409356f4c6c40ef8cdb2de809 | manager/__init__.py | manager/__init__.py | import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
output='css_all.css'
)
)
| import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
'fontawesome/css/font-awesome.min.css',
'admin-lte/dist/css/AdminLTE.css',
'admin-lte/dist/css/skins/skin-black.min.css',
output='css_all.css'
)
)
| Add new stuff to the css bundle | Add new stuff to the css bundle
| Python | mit | hreeder/ignition,hreeder/ignition,hreeder/ignition | import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
output='css_all.css'
)
)
Add new stuff to the css bundle | import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
'fontawesome/css/font-awesome.min.css',
'admin-lte/dist/css/AdminLTE.css',
'admin-lte/dist/css/skins/skin-black.min.css',
output='css_all.css'
)
)
| <commit_before>import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
output='css_all.css'
)
)
<commit_msg>Add new stuff to the css bundle<commit_after> | import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
'fontawesome/css/font-awesome.min.css',
'admin-lte/dist/css/AdminLTE.css',
'admin-lte/dist/css/skins/skin-black.min.css',
output='css_all.css'
)
)
| import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
output='css_all.css'
)
)
Add new stuff to the css bundleimport os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
'fontawesome/css/font-awesome.min.css',
'admin-lte/dist/css/AdminLTE.css',
'admin-lte/dist/css/skins/skin-black.min.css',
output='css_all.css'
)
)
| <commit_before>import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
output='css_all.css'
)
)
<commit_msg>Add new stuff to the css bundle<commit_after>import os
from flask import Flask
from flask.ext.assets import Bundle, Environment
from flask.ext.bcrypt import Bcrypt
from flask.ext.login import LoginManager, current_user
from flask.ext.migrate import Migrate
from flask.ext.sqlalchemy import SQLAlchemy
app = Flask(__name__)
# Load the app config
app.config.from_object("config.Config")
assets = Environment(app)
bcrypt = Bcrypt(app)
db = SQLAlchemy(app)
login_manager = LoginManager(app)
migrate = Migrate(app, db)
# Load Blueprints
from manager.core import core
from manager.dns import dns
app.register_blueprint(core)
app.register_blueprint(dns, url_prefix="/dns")
# Configure flask-login
login_manager.login_view = "core.login"
# Asset Management
assets.load_path = [
os.path.join(os.path.dirname(__file__), 'static'),
os.path.join(os.path.dirname(__file__), 'static', 'bower')
]
assets.register(
'js_all',
Bundle(
'jquery/dist/jquery.min.js',
'bootstrap/dist/js/bootstrap.min.js',
output='js_all.js'
)
)
assets.register(
'css_all',
Bundle(
'bootstrap/dist/css/bootstrap.css',
'fontawesome/css/font-awesome.min.css',
'admin-lte/dist/css/AdminLTE.css',
'admin-lte/dist/css/skins/skin-black.min.css',
output='css_all.css'
)
)
|
29ac3073b747d5bafaec240df25844d6d27c049a | marshmallow/base.py | marshmallow/base.py | # -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj):
raise NotImplementedError
def _deserialize(self, value, attr, data):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
| # -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj, **kwargs):
raise NotImplementedError
def _deserialize(self, value, attr, data, **kwargs):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
| Update signatures of FieldABC methods | Update signatures of FieldABC methods
| Python | mit | marshmallow-code/marshmallow,mwstobo/marshmallow | # -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj):
raise NotImplementedError
def _deserialize(self, value, attr, data):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
Update signatures of FieldABC methods | # -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj, **kwargs):
raise NotImplementedError
def _deserialize(self, value, attr, data, **kwargs):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
| <commit_before># -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj):
raise NotImplementedError
def _deserialize(self, value, attr, data):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
<commit_msg>Update signatures of FieldABC methods<commit_after> | # -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj, **kwargs):
raise NotImplementedError
def _deserialize(self, value, attr, data, **kwargs):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
| # -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj):
raise NotImplementedError
def _deserialize(self, value, attr, data):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
Update signatures of FieldABC methods# -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj, **kwargs):
raise NotImplementedError
def _deserialize(self, value, attr, data, **kwargs):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
| <commit_before># -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj):
raise NotImplementedError
def _deserialize(self, value, attr, data):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
<commit_msg>Update signatures of FieldABC methods<commit_after># -*- coding: utf-8 -*-
"""Abstract base classes.
These are necessary to avoid circular imports between core.py and fields.py.
"""
import copy
class FieldABC(object):
"""Abstract base class from which all Field classes inherit.
"""
parent = None
name = None
def serialize(self, attr, obj, accessor=None):
raise NotImplementedError
def deserialize(self, value):
raise NotImplementedError
def _serialize(self, value, attr, obj, **kwargs):
raise NotImplementedError
def _deserialize(self, value, attr, data, **kwargs):
raise NotImplementedError
def __deepcopy__(self, memo):
ret = copy.copy(self)
return ret
class SchemaABC(object):
"""Abstract base class from which all Schemas inherit."""
def dump(self, obj):
raise NotImplementedError
def dumps(self, obj, *args, **kwargs):
raise NotImplementedError
def load(self, data):
raise NotImplementedError
def loads(self, data):
raise NotImplementedError
|
9da01f39c8d9b73025d85be72b71399b6930b6fb | src/encoded/cache.py | src/encoded/cache.py | from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return
cached = cache.get(key)
if cached is not None:
return cached[1]
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
| from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return default
try:
return cache[key]
except KeyError:
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
| Use LRUCache correctly (minimal improvement) | Use LRUCache correctly (minimal improvement)
| Python | mit | ENCODE-DCC/encoded,kidaa/encoded,hms-dbmi/fourfront,philiptzou/clincoded,ENCODE-DCC/snovault,ENCODE-DCC/snovault,hms-dbmi/fourfront,philiptzou/clincoded,ENCODE-DCC/snovault,4dn-dcic/fourfront,ClinGen/clincoded,ClinGen/clincoded,4dn-dcic/fourfront,ENCODE-DCC/snovault,4dn-dcic/fourfront,ENCODE-DCC/encoded,T2DREAM/t2dream-portal,4dn-dcic/fourfront,philiptzou/clincoded,kidaa/encoded,hms-dbmi/fourfront,ENCODE-DCC/snovault,ENCODE-DCC/encoded,kidaa/encoded,ENCODE-DCC/encoded,hms-dbmi/fourfront,T2DREAM/t2dream-portal,T2DREAM/t2dream-portal,hms-dbmi/fourfront,ClinGen/clincoded,kidaa/encoded,kidaa/encoded,ClinGen/clincoded,ClinGen/clincoded,T2DREAM/t2dream-portal,philiptzou/clincoded,philiptzou/clincoded | from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return
cached = cache.get(key)
if cached is not None:
return cached[1]
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
Use LRUCache correctly (minimal improvement) | from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return default
try:
return cache[key]
except KeyError:
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
| <commit_before>from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return
cached = cache.get(key)
if cached is not None:
return cached[1]
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
<commit_msg>Use LRUCache correctly (minimal improvement)<commit_after> | from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return default
try:
return cache[key]
except KeyError:
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
| from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return
cached = cache.get(key)
if cached is not None:
return cached[1]
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
Use LRUCache correctly (minimal improvement)from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return default
try:
return cache[key]
except KeyError:
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
| <commit_before>from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return
cached = cache.get(key)
if cached is not None:
return cached[1]
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
<commit_msg>Use LRUCache correctly (minimal improvement)<commit_after>from pyramid.threadlocal import manager
from sqlalchemy.util import LRUCache
class ManagerLRUCache(object):
""" Override capacity in settings.
"""
def __init__(self, name, default_capacity=100, threshold=.5):
self.name = name
self.default_capacity = default_capacity
self.threshold = threshold
@property
def cache(self):
if not manager.stack:
return None
threadlocals = manager.stack[0]
if self.name not in threadlocals:
registry = threadlocals['registry']
capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity))
threadlocals[self.name] = LRUCache(capacity, self.threshold)
return threadlocals[self.name]
def get(self, key, default=None):
cache = self.cache
if cache is None:
return default
try:
return cache[key]
except KeyError:
return default
def __contains__(self, key):
cache = self.cache
if cache is None:
return False
return key in cache
def __setitem__(self, key, value):
cache = self.cache
if cache is None:
return
self.cache[key] = value
|
a6b39dde09777ff162fbf83976934cbf2ec14056 | app.py | app.py | from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:15]:
node = "{}".format(get_hash(item))
G.add_node(node)
commit = repo.commit(get_hash(item))
for parent in commit.parents:
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
| from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:20]:
node = "{}".format(get_hash(item))
commit = repo.commit(get_hash(item))
G.add_node(node, message=commit.message.split("\n")[0])
for parent in commit.parents:
G.add_node(parent.hexsha, message=parent.message.split("\n")[0])
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
| Add commit message for commits and for parents | Add commit message for commits and for parents
| Python | bsd-3-clause | kdheepak89/c3.py,kdheepak89/c3.py | from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:15]:
node = "{}".format(get_hash(item))
G.add_node(node)
commit = repo.commit(get_hash(item))
for parent in commit.parents:
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
Add commit message for commits and for parents | from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:20]:
node = "{}".format(get_hash(item))
commit = repo.commit(get_hash(item))
G.add_node(node, message=commit.message.split("\n")[0])
for parent in commit.parents:
G.add_node(parent.hexsha, message=parent.message.split("\n")[0])
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
| <commit_before>from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:15]:
node = "{}".format(get_hash(item))
G.add_node(node)
commit = repo.commit(get_hash(item))
for parent in commit.parents:
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
<commit_msg>Add commit message for commits and for parents<commit_after> | from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:20]:
node = "{}".format(get_hash(item))
commit = repo.commit(get_hash(item))
G.add_node(node, message=commit.message.split("\n")[0])
for parent in commit.parents:
G.add_node(parent.hexsha, message=parent.message.split("\n")[0])
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
| from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:15]:
node = "{}".format(get_hash(item))
G.add_node(node)
commit = repo.commit(get_hash(item))
for parent in commit.parents:
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
Add commit message for commits and for parentsfrom flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:20]:
node = "{}".format(get_hash(item))
commit = repo.commit(get_hash(item))
G.add_node(node, message=commit.message.split("\n")[0])
for parent in commit.parents:
G.add_node(parent.hexsha, message=parent.message.split("\n")[0])
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
| <commit_before>from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:15]:
node = "{}".format(get_hash(item))
G.add_node(node)
commit = repo.commit(get_hash(item))
for parent in commit.parents:
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
<commit_msg>Add commit message for commits and for parents<commit_after>from flask import Flask
from flask import render_template
app = Flask(__name__)
app.config['DEBUG'] = True
repo_path = '../ames-py'
@app.route("/")
def main():
return render_template("index.html")
@app.route("/data")
def data():
import json
import git
import networkx as nx
G = nx.DiGraph()
parent = None
g = git.Git(repo_path)
repo = git.Repo(repo_path)
loginfo = g.log()
def get_hash(commit):
return commit.split("\n")[0]
commits = loginfo.split("\n\ncommit ")
commits[0] = commits[0].replace("commit ", '')
for item in commits[:20]:
node = "{}".format(get_hash(item))
commit = repo.commit(get_hash(item))
G.add_node(node, message=commit.message.split("\n")[0])
for parent in commit.parents:
G.add_node(parent.hexsha, message=parent.message.split("\n")[0])
G.add_edge(node, parent.hexsha)
pos=nx.graphviz_layout(G, prog='dot')
from networkx.readwrite import json_graph
data = json_graph.node_link_data(G)
for node in data['nodes']:
node['pos'] = pos[node['id']]
j = json.dumps(data)
return(j)
if __name__ == "__main__":
app.run()
|
169ca5581c6c35d07dd772baf3119f45ba1c1e2e | app.py | app.py | from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
app.run(debug=True)
| from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
logging.getLogger().setLevel(logging.DEBUG)
app.run(debug=True)
| Make all log messages show by default | Make all log messages show by default
| Python | mit | karlalopez/Authentise-Store,addendumauto/Authentise-Store,addendumauto/Authentise-Store,addendumauto/Authentise-Store | from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
app.run(debug=True)
Make all log messages show by default | from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
logging.getLogger().setLevel(logging.DEBUG)
app.run(debug=True)
| <commit_before>from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
app.run(debug=True)
<commit_msg>Make all log messages show by default<commit_after> | from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
logging.getLogger().setLevel(logging.DEBUG)
app.run(debug=True)
| from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
app.run(debug=True)
Make all log messages show by defaultfrom flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
logging.getLogger().setLevel(logging.DEBUG)
app.run(debug=True)
| <commit_before>from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
app.run(debug=True)
<commit_msg>Make all log messages show by default<commit_after>from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from werkzeug import secure_filename
import os
import logging
import stripe
APP_ROOT = os.path.dirname(os.path.abspath(__file__))
UPLOAD_FOLDER = os.path.join('static/uploads')
MODELS_FOLDER = os.path.join('models')
ALLOWED_EXTENSIONS = set(['stl'])
stripe_keys = {
'secret_key': os.environ['SECRET_KEY'],
'publishable_key': os.environ['PUBLISHABLE_KEY']
}
stripe.api_key = stripe_keys['secret_key']
shop_name = "Shop name"
shop_tagline = "Best shop tagline ever"
app = Flask(__name__)
app.secret_key = 'thisisasecret' #You need to set up an app secret key.
app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER
app.config['MODELS_FOLDER'] = MODELS_FOLDER
# Set up the SQLAlchemy Database to be a local file 'store.db'
app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store'
db = SQLAlchemy(app)
if __name__ == "__main__":
from views import *
del session
logging.basicConfig()
logging.getLogger().setLevel(logging.DEBUG)
app.run(debug=True)
|
d560a809c4d0fd78e1ce0454ea5406e81f356906 | server_app/__main__.py | server_app/__main__.py | import sys
import os
import logging
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
| import sys
import os
import logging
import time
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
| Make logger sort by date | Make logger sort by date
| Python | bsd-3-clause | jos0003/Chat,jos0003/Chat,jos0003/Chat,jos0003/Chat,jos0003/Chat | import sys
import os
import logging
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
Make logger sort by date | import sys
import os
import logging
import time
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
| <commit_before>import sys
import os
import logging
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
<commit_msg>Make logger sort by date<commit_after> | import sys
import os
import logging
import time
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
| import sys
import os
import logging
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
Make logger sort by dateimport sys
import os
import logging
import time
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
| <commit_before>import sys
import os
import logging
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
<commit_msg>Make logger sort by date<commit_after>import sys
import os
import logging
import time
if not os.path.exists(os.path.expanduser("~/.chatserver")):
os.makedirs(os.path.expanduser("~/.chatserver"))
logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG)
sys.stderr.close()
sys.stdout.close()
sys.stdin.close()
from app import app, db, main, socketio
db.create_all()
app.register_blueprint(main)
port = app.config['PORT']
if len(sys.argv) == 2:
port = int(sys.argv[1])
logging.info("Chat server is now running on 0.0.0.0:%r" % port)
socketio.run(app, host="0.0.0.0", port=port)
|
0438825672a407eb30bff49e03dac89a0534f28a | minimax.py | minimax.py | class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
pass
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
| class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
raise NotImplementedError('Dont override this class')
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
def calculate_min_or_max(self, vector_values, color):
if color == self.me.color:
return self.me.eval(vector_values)
else:
return self.challenger.eval(vector_values)
| Create in MinMax the calculate_min_max | Create in MinMax the calculate_min_max
| Python | apache-2.0 | frila/agente-minimax | class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
pass
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
Create in MinMax the calculate_min_max | class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
raise NotImplementedError('Dont override this class')
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
def calculate_min_or_max(self, vector_values, color):
if color == self.me.color:
return self.me.eval(vector_values)
else:
return self.challenger.eval(vector_values)
| <commit_before>class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
pass
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
<commit_msg>Create in MinMax the calculate_min_max<commit_after> | class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
raise NotImplementedError('Dont override this class')
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
def calculate_min_or_max(self, vector_values, color):
if color == self.me.color:
return self.me.eval(vector_values)
else:
return self.challenger.eval(vector_values)
| class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
pass
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
Create in MinMax the calculate_min_maxclass Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
raise NotImplementedError('Dont override this class')
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
def calculate_min_or_max(self, vector_values, color):
if color == self.me.color:
return self.me.eval(vector_values)
else:
return self.challenger.eval(vector_values)
| <commit_before>class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
pass
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
<commit_msg>Create in MinMax the calculate_min_max<commit_after>class Heuristic:
def __init__(self, color):
self.color = color
def heuristic(self, board, color):
raise NotImplementedError('Dont override this class')
def eval(self, vector):
raise NotImplementedError('Dont override this class')
class Minimax:
def __init__(self, me, challenger):
self.me = me
self.challenger = challenger
def heuristic(self, board, color):
if color == self.color_me:
return self.me.heuristic(board, color)
else
return self.challenger.heuristic(board, color)
def calculate_min_or_max(self, vector_values, color):
if color == self.me.color:
return self.me.eval(vector_values)
else:
return self.challenger.eval(vector_values)
|
3b3a8dc6aa0b38cfbb68105eb5ef31e8e73ff3a4 | gcm_flask/application/models.py | gcm_flask/application/models.py | """
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty()
sent_at = db.DateTimeProperty(auto_now_add=True) | """
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty(auto_current_user=True)
sent_at = db.DateTimeProperty(auto_now_add=True) | Update user who sent the message | Update user who sent the message
| Python | apache-2.0 | BarcampBangalore/Barcamp-Bangalore-Android-App,BarcampBangalore/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App,BarcampBangalore/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App,BarcampBangalore/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App | """
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty()
sent_at = db.DateTimeProperty(auto_now_add=True)Update user who sent the message | """
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty(auto_current_user=True)
sent_at = db.DateTimeProperty(auto_now_add=True) | <commit_before>"""
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty()
sent_at = db.DateTimeProperty(auto_now_add=True)<commit_msg>Update user who sent the message<commit_after> | """
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty(auto_current_user=True)
sent_at = db.DateTimeProperty(auto_now_add=True) | """
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty()
sent_at = db.DateTimeProperty(auto_now_add=True)Update user who sent the message"""
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty(auto_current_user=True)
sent_at = db.DateTimeProperty(auto_now_add=True) | <commit_before>"""
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty()
sent_at = db.DateTimeProperty(auto_now_add=True)<commit_msg>Update user who sent the message<commit_after>"""
models.py
App Engine datastore models
"""
from google.appengine.ext import db
class ExampleModel(db.Model):
"""Example Model"""
example_name = db.StringProperty(required=True)
example_description = db.TextProperty(required=True)
added_by = db.UserProperty()
timestamp = db.DateTimeProperty(auto_now_add=True)
class RegIDModel(db.Model):
"""Regl IDs Model"""
regID = db.StringProperty(required=True)
class MessagesModel(db.Model):
"""Model for storing messages sent"""
message = db.StringProperty(required=True)
messagetype = db.StringProperty(required=True)
added_by = db.UserProperty(auto_current_user=True)
sent_at = db.DateTimeProperty(auto_now_add=True) |
c04b8932ec65480ba90dd4578d5f6bb8c3baa690 | demosys/project/default.py | demosys/project/default.py | from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
self.effect = self.create_effect('default', cls.__name__)
| from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
if self.effect_class_name:
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
else:
effect_package = effects.get_package(self.effect_package_name)
runnable_effects = effect_package.runnable_effects()
if not runnable_effects:
raise ValueError("No runnable effects found in effect package", self.effect_package_name)
cls = runnable_effects[-1]
self.effect = self.create_effect('default', cls.__name__)
| Improve errors when effect packages or effects are not found | Improve errors when effect packages or effects are not found
| Python | isc | Contraz/demosys-py | from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
self.effect = self.create_effect('default', cls.__name__)
Improve errors when effect packages or effects are not found | from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
if self.effect_class_name:
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
else:
effect_package = effects.get_package(self.effect_package_name)
runnable_effects = effect_package.runnable_effects()
if not runnable_effects:
raise ValueError("No runnable effects found in effect package", self.effect_package_name)
cls = runnable_effects[-1]
self.effect = self.create_effect('default', cls.__name__)
| <commit_before>from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
self.effect = self.create_effect('default', cls.__name__)
<commit_msg>Improve errors when effect packages or effects are not found<commit_after> | from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
if self.effect_class_name:
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
else:
effect_package = effects.get_package(self.effect_package_name)
runnable_effects = effect_package.runnable_effects()
if not runnable_effects:
raise ValueError("No runnable effects found in effect package", self.effect_package_name)
cls = runnable_effects[-1]
self.effect = self.create_effect('default', cls.__name__)
| from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
self.effect = self.create_effect('default', cls.__name__)
Improve errors when effect packages or effects are not foundfrom demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
if self.effect_class_name:
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
else:
effect_package = effects.get_package(self.effect_package_name)
runnable_effects = effect_package.runnable_effects()
if not runnable_effects:
raise ValueError("No runnable effects found in effect package", self.effect_package_name)
cls = runnable_effects[-1]
self.effect = self.create_effect('default', cls.__name__)
| <commit_before>from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
self.effect = self.create_effect('default', cls.__name__)
<commit_msg>Improve errors when effect packages or effects are not found<commit_after>from demosys.project.base import BaseProject
from demosys.effects.registry import effects, parse_package_string
class Project(BaseProject):
"""
The project what will be assigned when no project are specified.
This is mainly used when the ``runeffect`` command is used
"""
def __init__(self, effect_package):
super().__init__()
self.path = effect_package
self.effect_package_name, self.effect_class_name = parse_package_string(effect_package)
self.effect_packages = [self.effect_package_name]
self.effect = None
def get_default_effect(self):
return self.effect
def create_resources(self):
pass
def create_effect_instances(self):
if self.effect_class_name:
cls = effects.find_effect_class(self.path)
if not cls.runnable:
raise ValueError("Effect doesn't have the runnable flag set:", self.path)
else:
effect_package = effects.get_package(self.effect_package_name)
runnable_effects = effect_package.runnable_effects()
if not runnable_effects:
raise ValueError("No runnable effects found in effect package", self.effect_package_name)
cls = runnable_effects[-1]
self.effect = self.create_effect('default', cls.__name__)
|
891e8afe5deff5fe7d620abfe8189689d47ec4f8 | djangocms_inherit/forms.py | djangocms_inherit/forms.py | from django import forms
from django.forms.models import ModelForm
from django.forms.utils import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
| from django import forms
from django.forms.models import ModelForm
try:
from django.forms.utils import ErrorList
except ImportError:
# Django<1.7 (deprecated in Django 1.8, removed in 1.9)
from django.forms.util import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
| Make import backward compatible (Django<1.7) | Make import backward compatible (Django<1.7)
| Python | bsd-3-clause | bittner/djangocms-inherit,bittner/djangocms-inherit,divio/djangocms-inherit,divio/djangocms-inherit,divio/djangocms-inherit | from django import forms
from django.forms.models import ModelForm
from django.forms.utils import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
Make import backward compatible (Django<1.7) | from django import forms
from django.forms.models import ModelForm
try:
from django.forms.utils import ErrorList
except ImportError:
# Django<1.7 (deprecated in Django 1.8, removed in 1.9)
from django.forms.util import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
| <commit_before>from django import forms
from django.forms.models import ModelForm
from django.forms.utils import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
<commit_msg>Make import backward compatible (Django<1.7)<commit_after> | from django import forms
from django.forms.models import ModelForm
try:
from django.forms.utils import ErrorList
except ImportError:
# Django<1.7 (deprecated in Django 1.8, removed in 1.9)
from django.forms.util import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
| from django import forms
from django.forms.models import ModelForm
from django.forms.utils import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
Make import backward compatible (Django<1.7)from django import forms
from django.forms.models import ModelForm
try:
from django.forms.utils import ErrorList
except ImportError:
# Django<1.7 (deprecated in Django 1.8, removed in 1.9)
from django.forms.util import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
| <commit_before>from django import forms
from django.forms.models import ModelForm
from django.forms.utils import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
<commit_msg>Make import backward compatible (Django<1.7)<commit_after>from django import forms
from django.forms.models import ModelForm
try:
from django.forms.utils import ErrorList
except ImportError:
# Django<1.7 (deprecated in Django 1.8, removed in 1.9)
from django.forms.util import ErrorList
from django.utils.translation import ugettext_lazy as _
from cms.models import Page
from .models import InheritPagePlaceholder
class InheritForm(ModelForm):
from_page = forms.ModelChoiceField(
label=_("page"), queryset=Page.objects.drafts(), required=False)
class Meta:
model = InheritPagePlaceholder
exclude = ('page', 'position', 'placeholder', 'language',
'plugin_type')
def for_site(self, site):
# override the page_link fields queryset to containt just pages for
# current site
self.fields['from_page'].queryset = Page.objects.drafts().on_site(site)
def clean(self):
cleaned_data = super(InheritForm, self).clean()
if not cleaned_data['from_page'] and not cleaned_data['from_language']:
self._errors['from_page'] = ErrorList(
[_("Language or Page must be filled out")])
return cleaned_data
|
db59332e3d522c68c3eeef77fe4393fe137e5059 | inspectors/registration/util.py | inspectors/registration/util.py | import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result is not None
| import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result
| Fix logic bug for API result | Fix logic bug for API result
| Python | bsd-3-clause | codeforamerica/mdc-inspectors,codeforamerica/mdc-inspectors,codeforamerica/mdc-inspectors | import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result is not None
Fix logic bug for API result | import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result
| <commit_before>import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result is not None
<commit_msg>Fix logic bug for API result<commit_after> | import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result
| import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result is not None
Fix logic bug for API resultimport requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result
| <commit_before>import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result is not None
<commit_msg>Fix logic bug for API result<commit_after>import requests
API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json'
def is_valid_permit(id):
# checks if the ID is a valid Miami-Dade Permit or Process Number
API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27'
response = requests.get(API)
json_result = response.json()
return json_result
|
09fa23adfb76f052473ee38de94ce4bdfdcc48e1 | src/nodeconductor_assembly_waldur/packages/perms.py | src/nodeconductor_assembly_waldur/packages/perms.py | from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)),
)
| from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic
from nodeconductor.structure import models as structure_models
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic(
collaborators_query=[
'tenant__service_project_link__service__customer__roles__permission_group__user',
],
collaborators_filter=[
{'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER},
],
any_permission=True,
)),
)
| Allow customer owner to create packages | Allow customer owner to create packages
- wal-26
| Python | mit | opennode/nodeconductor-assembly-waldur,opennode/nodeconductor-assembly-waldur,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/nodeconductor-assembly-waldur | from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)),
)
Allow customer owner to create packages
- wal-26 | from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic
from nodeconductor.structure import models as structure_models
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic(
collaborators_query=[
'tenant__service_project_link__service__customer__roles__permission_group__user',
],
collaborators_filter=[
{'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER},
],
any_permission=True,
)),
)
| <commit_before>from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)),
)
<commit_msg>Allow customer owner to create packages
- wal-26<commit_after> | from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic
from nodeconductor.structure import models as structure_models
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic(
collaborators_query=[
'tenant__service_project_link__service__customer__roles__permission_group__user',
],
collaborators_filter=[
{'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER},
],
any_permission=True,
)),
)
| from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)),
)
Allow customer owner to create packages
- wal-26from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic
from nodeconductor.structure import models as structure_models
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic(
collaborators_query=[
'tenant__service_project_link__service__customer__roles__permission_group__user',
],
collaborators_filter=[
{'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER},
],
any_permission=True,
)),
)
| <commit_before>from nodeconductor.core.permissions import StaffPermissionLogic
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)),
)
<commit_msg>Allow customer owner to create packages
- wal-26<commit_after>from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic
from nodeconductor.structure import models as structure_models
PERMISSION_LOGICS = (
('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)),
('packages.PackageComponent', StaffPermissionLogic(any_permission=True)),
('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic(
collaborators_query=[
'tenant__service_project_link__service__customer__roles__permission_group__user',
],
collaborators_filter=[
{'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER},
],
any_permission=True,
)),
)
|
75d1241c5d62def89a7377e506afdacfa83cbdda | js2xml/parser.py | js2xml/parser.py | import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=False, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
write_tables=False,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
| import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=True, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
| Write tab but don't warn | Write tab but don't warn
| Python | mit | redapple/js2xml,redapple/js2xml,redapple/js2xml,redapple/js2xml | import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=False, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
write_tables=False,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
Write tab but don't warn | import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=True, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
| <commit_before>import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=False, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
write_tables=False,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
<commit_msg>Write tab but don't warn<commit_after> | import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=True, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
| import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=False, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
write_tables=False,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
Write tab but don't warnimport ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=True, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
| <commit_before>import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=False, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
write_tables=False,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
<commit_msg>Write tab but don't warn<commit_after>import ply.yacc
from slimit.parser import Parser
from js2xml.lexer import CustomLexer as Lexer
from js2xml.log import logger
lextab, yacctab = 'lextab', 'yacctab'
class CustomParser(Parser):
def __init__(self, lex_optimize=True, lextab=lextab,
yacc_optimize=True, yacctab=yacctab, yacc_debug=False,
logger=logger):
self.lex_optimize = lex_optimize
self.lextab = lextab
self.yacc_optimize = yacc_optimize
self.yacctab = yacctab
self.yacc_debug = yacc_debug
self.lexer = Lexer()
self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger)
self.tokens = self.lexer.tokens
self.parser = ply.yacc.yacc(
module=self,
optimize=yacc_optimize,
debug=yacc_debug,
tabmodule=yacctab,
start='program',
errorlog=logger)
self._error_tokens = {}
def parse(self, text, debug=False):
result = super(CustomParser, self).parse(text, debug=debug)
self._error_tokens = {}
return result
|
cc2b579377abde262d76e2484a6488e254b315fc | judge/caching.py | judge/caching.py | from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete('sub_stats_table')
cache.delete('sub_stats_data')
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
cache.delete('user_complete:%d' % sub.user_id)
if hasattr(sub, 'contest'):
participation = sub.contest.participation
cache.delete('contest_complete:%d' % participation.id)
cache.delete(make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id))) | from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete_many(('sub_stats_table', 'sub_stats_data'))
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
keys = ['user_complete:%d' % sub.user_id]
if hasattr(sub, 'contest'):
participation = sub.contest.participation
keys += ['contest_complete:%d' % participation.id,
make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id))]
cache.delete_many(keys)
| Delete many to reduce round trips to the cache. | Delete many to reduce round trips to the cache.
| Python | agpl-3.0 | Minkov/site,Minkov/site,DMOJ/site,Minkov/site,DMOJ/site,DMOJ/site,Phoenix1369/site,Phoenix1369/site,Minkov/site,monouno/site,Phoenix1369/site,monouno/site,monouno/site,monouno/site,monouno/site,DMOJ/site,Phoenix1369/site | from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete('sub_stats_table')
cache.delete('sub_stats_data')
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
cache.delete('user_complete:%d' % sub.user_id)
if hasattr(sub, 'contest'):
participation = sub.contest.participation
cache.delete('contest_complete:%d' % participation.id)
cache.delete(make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id)))Delete many to reduce round trips to the cache. | from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete_many(('sub_stats_table', 'sub_stats_data'))
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
keys = ['user_complete:%d' % sub.user_id]
if hasattr(sub, 'contest'):
participation = sub.contest.participation
keys += ['contest_complete:%d' % participation.id,
make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id))]
cache.delete_many(keys)
| <commit_before>from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete('sub_stats_table')
cache.delete('sub_stats_data')
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
cache.delete('user_complete:%d' % sub.user_id)
if hasattr(sub, 'contest'):
participation = sub.contest.participation
cache.delete('contest_complete:%d' % participation.id)
cache.delete(make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id)))<commit_msg>Delete many to reduce round trips to the cache.<commit_after> | from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete_many(('sub_stats_table', 'sub_stats_data'))
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
keys = ['user_complete:%d' % sub.user_id]
if hasattr(sub, 'contest'):
participation = sub.contest.participation
keys += ['contest_complete:%d' % participation.id,
make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id))]
cache.delete_many(keys)
| from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete('sub_stats_table')
cache.delete('sub_stats_data')
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
cache.delete('user_complete:%d' % sub.user_id)
if hasattr(sub, 'contest'):
participation = sub.contest.participation
cache.delete('contest_complete:%d' % participation.id)
cache.delete(make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id)))Delete many to reduce round trips to the cache.from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete_many(('sub_stats_table', 'sub_stats_data'))
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
keys = ['user_complete:%d' % sub.user_id]
if hasattr(sub, 'contest'):
participation = sub.contest.participation
keys += ['contest_complete:%d' % participation.id,
make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id))]
cache.delete_many(keys)
| <commit_before>from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete('sub_stats_table')
cache.delete('sub_stats_data')
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
cache.delete('user_complete:%d' % sub.user_id)
if hasattr(sub, 'contest'):
participation = sub.contest.participation
cache.delete('contest_complete:%d' % participation.id)
cache.delete(make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id)))<commit_msg>Delete many to reduce round trips to the cache.<commit_after>from django.core.cache import cache
from django.core.cache.utils import make_template_fragment_key
def update_submission(id):
key = 'version:submission-%d' % id
cache.add(key, 0, None)
cache.incr(key)
def update_stats():
cache.delete_many(('sub_stats_table', 'sub_stats_data'))
def point_update(profile):
cache.delete(make_template_fragment_key('global_user'))
def finished_submission(sub):
keys = ['user_complete:%d' % sub.user_id]
if hasattr(sub, 'contest'):
participation = sub.contest.participation
keys += ['contest_complete:%d' % participation.id,
make_template_fragment_key('conrank_user_prob',
(participation.profile.user_id,
participation.contest_id))]
cache.delete_many(keys)
|
05adb44cdec74256fa44ce3a3df61c6525ce7fac | dryscrape/xvfb.py | dryscrape/xvfb.py | import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
| import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
| Remove removal of DISPLAY environment variable | Remove removal of DISPLAY environment variable
The issue has to do with the two lines:
` if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]`
This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError. | Python | mit | niklasb/dryscrape | import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
Remove removal of DISPLAY environment variable
The issue has to do with the two lines:
` if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]`
This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError. | import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
| <commit_before>import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
<commit_msg>Remove removal of DISPLAY environment variable
The issue has to do with the two lines:
` if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]`
This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.<commit_after> | import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
| import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
Remove removal of DISPLAY environment variable
The issue has to do with the two lines:
` if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]`
This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
| <commit_before>import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
<commit_msg>Remove removal of DISPLAY environment variable
The issue has to do with the two lines:
` if "DISPLAY" in os.environ:
del os.environ["DISPLAY"]`
This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.<commit_after>import atexit
import os
_xvfb = None
def start_xvfb():
from xvfbwrapper import Xvfb
global _xvfb
_xvfb = Xvfb()
_xvfb.start()
atexit.register(_xvfb.stop)
def stop_xvfb():
global _xvfb
_xvfb.stop()
|
382c46366c4ae29456aae35420990ce25b61ce76 | studygroups/tasks.py | studygroups/tasks.py | from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
| from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
| Add check for failed reminders so that it doesn't case an Exception | Add check for failed reminders so that it doesn't case an Exception
| Python | mit | p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles | from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
Add check for failed reminders so that it doesn't case an Exception | from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
| <commit_before>from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
<commit_msg>Add check for failed reminders so that it doesn't case an Exception<commit_after> | from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
| from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
Add check for failed reminders so that it doesn't case an Exceptionfrom __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
| <commit_before>from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
<commit_msg>Add check for failed reminders so that it doesn't case an Exception<commit_after>from __future__ import absolute_import
from celery import shared_task
from django.utils import timezone
from django.conf import settings
from studygroups.models import StudyGroup
from studygroups.models import Reminder
from studygroups.models import generate_reminder
from studygroups.models import send_reminder
from studygroups.models import send_weekly_update
from django.utils import translation
import datetime
@shared_task
def send_reminders():
now = timezone.now()
translation.activate(settings.LANGUAGE_CODE)
# TODO - should this be set here or closer to where the language matters?
for reminder in Reminder.objects.filter(sent_at__isnull=True):
if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2):
send_reminder(reminder)
@shared_task
def gen_reminders():
for study_group in StudyGroup.objects.all():
translation.activate(settings.LANGUAGE_CODE)
generate_reminder(study_group)
@shared_task
def weekly_update():
# Create a report for the previous week
send_weekly_update()
|
4613daea5d9d603b5f092005627fabd805de8a45 | example/app/utils.py | example/app/utils.py | from django.contrib.auth import get_user_model
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
| from django.contrib.auth import get_user_model
from django.db.utils import ProgrammingError
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
try:
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
except ProgrammingError:
# auth_user doesn't exist, this allows the migrations to run properly.
user = None
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
| Make initial ./manage.py migrate work in example | Make initial ./manage.py migrate work in example
| Python | bsd-3-clause | zostera/django-modeltrans,zostera/django-modeltrans | from django.contrib.auth import get_user_model
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
Make initial ./manage.py migrate work in example | from django.contrib.auth import get_user_model
from django.db.utils import ProgrammingError
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
try:
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
except ProgrammingError:
# auth_user doesn't exist, this allows the migrations to run properly.
user = None
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
| <commit_before>from django.contrib.auth import get_user_model
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
<commit_msg>Make initial ./manage.py migrate work in example<commit_after> | from django.contrib.auth import get_user_model
from django.db.utils import ProgrammingError
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
try:
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
except ProgrammingError:
# auth_user doesn't exist, this allows the migrations to run properly.
user = None
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
| from django.contrib.auth import get_user_model
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
Make initial ./manage.py migrate work in examplefrom django.contrib.auth import get_user_model
from django.db.utils import ProgrammingError
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
try:
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
except ProgrammingError:
# auth_user doesn't exist, this allows the migrations to run properly.
user = None
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
| <commit_before>from django.contrib.auth import get_user_model
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
<commit_msg>Make initial ./manage.py migrate work in example<commit_after>from django.contrib.auth import get_user_model
from django.db.utils import ProgrammingError
def disable_admin_login():
"""
Disable admin login, but allow editing.
amended from: https://stackoverflow.com/a/40008282/517560
"""
User = get_user_model()
try:
user, created = User.objects.update_or_create(
id=1,
defaults=dict(
first_name="Default Admin",
last_name="User",
is_superuser=True,
is_active=True,
is_staff=True,
),
)
except ProgrammingError:
# auth_user doesn't exist, this allows the migrations to run properly.
user = None
def no_login_has_permission(request):
setattr(request, "user", user)
return True
return no_login_has_permission
|
187447322d74fc3070c9a75415a55a56ba840ef9 | extruct/jsonld.py | extruct/jsonld.py | # -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
# now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded
# TODO: `strict=False` can be configurable if needed
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
| # -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
try:
# TODO: `strict=False` can be configurable if needed
data = json.loads(script, strict=False)
except ValueError:
# sometimes JSON-decoding errors are due to leading HTML or JavaScript comments
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
| Make comment removal a fallback when failed. | Mod: Make comment removal a fallback when failed.
| Python | bsd-3-clause | scrapinghub/extruct | # -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
# now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded
# TODO: `strict=False` can be configurable if needed
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
Mod: Make comment removal a fallback when failed. | # -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
try:
# TODO: `strict=False` can be configurable if needed
data = json.loads(script, strict=False)
except ValueError:
# sometimes JSON-decoding errors are due to leading HTML or JavaScript comments
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
| <commit_before># -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
# now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded
# TODO: `strict=False` can be configurable if needed
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
<commit_msg>Mod: Make comment removal a fallback when failed.<commit_after> | # -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
try:
# TODO: `strict=False` can be configurable if needed
data = json.loads(script, strict=False)
except ValueError:
# sometimes JSON-decoding errors are due to leading HTML or JavaScript comments
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
| # -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
# now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded
# TODO: `strict=False` can be configurable if needed
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
Mod: Make comment removal a fallback when failed.# -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
try:
# TODO: `strict=False` can be configurable if needed
data = json.loads(script, strict=False)
except ValueError:
# sometimes JSON-decoding errors are due to leading HTML or JavaScript comments
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
| <commit_before># -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
# now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded
# TODO: `strict=False` can be configurable if needed
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
<commit_msg>Mod: Make comment removal a fallback when failed.<commit_after># -*- coding: utf-8 -*-
"""
JSON-LD extractor
"""
import json
import re
import lxml.etree
import lxml.html
HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)')
class JsonLdExtractor(object):
_xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]')
def extract(self, htmlstring, base_url=None, encoding="UTF-8"):
parser = lxml.html.HTMLParser(encoding=encoding)
lxmldoc = lxml.html.fromstring(htmlstring, parser=parser)
return self.extract_items(lxmldoc, base_url=base_url)
def extract_items(self, document, base_url=None):
return [item for items in map(self._extract_items,
self._xp_jsonld(document))
for item in items
if item]
def _extract_items(self, node):
script = node.xpath('string()')
try:
# TODO: `strict=False` can be configurable if needed
data = json.loads(script, strict=False)
except ValueError:
# sometimes JSON-decoding errors are due to leading HTML or JavaScript comments
data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False)
if isinstance(data, list):
return data
elif isinstance(data, dict):
return [data]
|
118aa612ef088dba90328f1775d8603ee12e5e5b | main.py | main.py | import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = map(lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness)
total = sum(normalized)
return map(lambda x: x/total, normalized)
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=1)
evolve()
| import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = list(
map(
lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness
)
)
total = sum(normalized)
return list(map(lambda x: x/total, normalized))
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=20)
evolve()
| Set log level to info, fixed bug with map object | Set log level to info, fixed bug with map object
| Python | mit | ray-dino/robby-genetic-algorithm | import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = map(lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness)
total = sum(normalized)
return map(lambda x: x/total, normalized)
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=1)
evolve()
Set log level to info, fixed bug with map object | import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = list(
map(
lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness
)
)
total = sum(normalized)
return list(map(lambda x: x/total, normalized))
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=20)
evolve()
| <commit_before>import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = map(lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness)
total = sum(normalized)
return map(lambda x: x/total, normalized)
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=1)
evolve()
<commit_msg>Set log level to info, fixed bug with map object<commit_after> | import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = list(
map(
lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness
)
)
total = sum(normalized)
return list(map(lambda x: x/total, normalized))
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=20)
evolve()
| import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = map(lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness)
total = sum(normalized)
return map(lambda x: x/total, normalized)
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=1)
evolve()
Set log level to info, fixed bug with map objectimport logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = list(
map(
lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness
)
)
total = sum(normalized)
return list(map(lambda x: x/total, normalized))
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=20)
evolve()
| <commit_before>import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = map(lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness)
total = sum(normalized)
return map(lambda x: x/total, normalized)
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=1)
evolve()
<commit_msg>Set log level to info, fixed bug with map object<commit_after>import logging
import numpy as np
import settings
from models import Robby
def evolve():
population = np.array([Robby() for i in range(0, settings.POPULATION)])
for gen in range(0, settings.GENERATIONS):
for individual in population:
individual.live()
new_population = list()
while len(new_population)<settings.POPULATION:
father, mother = np.random.choice(
population,
size=2,
p=get_relative_probabilities(population))
child1, child2 = father.mate(mother)
new_population.append(child1)
new_population.append(child2)
population = new_population
logging.info("Generation {}: {}".format(
gen,
max([r.get_fitness() for r in population])))
def get_relative_probabilities(population):
pop_fitness = [r.get_fitness() for r in population]
min_fitness = min(pop_fitness)
max_fitness = max(pop_fitness)
normalized = list(
map(
lambda x: normalize(x, min_fitness, max_fitness),
pop_fitness
)
)
total = sum(normalized)
return list(map(lambda x: x/total, normalized))
def normalize(x, minf, maxf):
return (x - minf) / (maxf - minf)
if __name__=='__main__':
logging.basicConfig(level=20)
evolve()
|
351bfe236f183c069314f5df7d3c4b8f9d8699b4 | final/problem6.py | final/problem6.py | # Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Person.say(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
| # Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Lecturer.lecture(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
| Modify lecture method in ArrogantProfessor class using inheritance | Modify lecture method in ArrogantProfessor class using inheritance
| Python | mit | Kunal57/MIT_6.00.1x | # Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Person.say(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
Modify lecture method in ArrogantProfessor class using inheritance | # Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Lecturer.lecture(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
| <commit_before># Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Person.say(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
<commit_msg>Modify lecture method in ArrogantProfessor class using inheritance<commit_after> | # Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Lecturer.lecture(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
| # Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Person.say(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
Modify lecture method in ArrogantProfessor class using inheritance# Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Lecturer.lecture(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
| <commit_before># Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Person.say(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
<commit_msg>Modify lecture method in ArrogantProfessor class using inheritance<commit_after># Problem 6-1
# 10.0 points possible (graded)
class Person(object):
def __init__(self, name):
self.name = name
def say(self, stuff):
return self.name + ' says: ' + stuff
def __str__(self):
return self.name
class Lecturer(Person):
def lecture(self, stuff):
return 'I believe that ' + Person.say(self, stuff)
class Professor(Lecturer):
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
class ArrogantProfessor(Professor):
def lecture(self, stuff):
return 'It is obvious that ' + Lecturer.lecture(self, stuff)
def say(self, stuff):
return self.name + ' says: ' + self.lecture(stuff)
e = Person('eric')
le = Lecturer('eric')
pe = Professor('eric')
ae = ArrogantProfessor('eric')
e.say('the sky is blue')
le.say('the sky is blue')
le.lecture('the sky is blue')
pe.say('the sky is blue')
pe.lecture('the sky is blue')
print(ae.say('the sky is blue'))
print(ae.lecture('the sky is blue'))
|
3fdb40934319d667ae9e8c550a0404cdd6a8cb64 | grum/api/models/message.py | grum/api/models/message.py | from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Timestamp)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text) | from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Integer)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text) | Change from timestamp to Integer | Change from timestamp to Integer
| Python | mit | Grum-Hackdee/grum-web,Grum-Hackdee/grum-web,Grum-Hackdee/grum-web,Grum-Hackdee/grum-web | from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Timestamp)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text)Change from timestamp to Integer | from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Integer)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text) | <commit_before>from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Timestamp)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text)<commit_msg>Change from timestamp to Integer<commit_after> | from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Integer)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text) | from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Timestamp)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text)Change from timestamp to Integerfrom grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Integer)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text) | <commit_before>from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Timestamp)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text)<commit_msg>Change from timestamp to Integer<commit_after>from grum import db
class Message(db.Model):
id = db.Column(db.String(128), primary_key=True)
recipient = db.Column(db.String(128))
recipient_nice = db.Column(db.String(128))
sender = db.Column(db.String(128))
sender_nice = db.Column(db.String(128))
sent_at = db.Column(db.Integer)
html = db.Column(db.Text)
html_stripped = db.Column(db.Text)
plaintext = db.Column(db.Text)
plaintext_stripped = db.Column(db.Text)
plaintext_stripped_signature = db.Column(db.Text) |
ff8aa2725001dbd1281357ccd5e0877257b5975d | hackernews_scrapy/items.py | hackernews_scrapy/items.py | # -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
| # -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
crawled_at = scrapy.Field()
| Add crawled_at field to HackernewsScrapyItem | Add crawled_at field to HackernewsScrapyItem
| Python | mit | mdsrosa/hackernews_scrapy | # -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
Add crawled_at field to HackernewsScrapyItem | # -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
crawled_at = scrapy.Field()
| <commit_before># -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
<commit_msg>Add crawled_at field to HackernewsScrapyItem<commit_after> | # -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
crawled_at = scrapy.Field()
| # -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
Add crawled_at field to HackernewsScrapyItem# -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
crawled_at = scrapy.Field()
| <commit_before># -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
<commit_msg>Add crawled_at field to HackernewsScrapyItem<commit_after># -*- coding: utf-8 -*-
import scrapy
class HackernewsScrapyItem(scrapy.Item):
title = scrapy.Field()
crawled_at = scrapy.Field()
|
4c31f637d2b7f75c35debc51498913139b5634c0 | pushhub/__init__.py | pushhub/__init__.py | from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='suscribe')
return config.make_wsgi_app()
| from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='subscribe')
return config.make_wsgi_app()
| Fix typo, make routes a little more legible. | Fix typo, make routes a little more legible.
| Python | bsd-3-clause | ucla/PushHubCore | from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='suscribe')
return config.make_wsgi_app()
Fix typo, make routes a little more legible. | from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='subscribe')
return config.make_wsgi_app()
| <commit_before>from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='suscribe')
return config.make_wsgi_app()
<commit_msg>Fix typo, make routes a little more legible.<commit_after> | from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='subscribe')
return config.make_wsgi_app()
| from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='suscribe')
return config.make_wsgi_app()
Fix typo, make routes a little more legible.from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='subscribe')
return config.make_wsgi_app()
| <commit_before>from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='suscribe')
return config.make_wsgi_app()
<commit_msg>Fix typo, make routes a little more legible.<commit_after>from pyramid.config import Configurator
from pyramid_zodbconn import get_connection
from .models import appmaker
from .views import publish, subscribe
def root_factory(request):
conn = get_connection(request)
return appmaker(conn.root())
def main(global_config, **settings):
""" This function returns a Pyramid WSGI application.
"""
config = Configurator(root_factory=root_factory, settings=settings)
config.add_static_view('static', 'static', cache_max_age=3600)
config.add_route('publish', '/publish')
config.add_view(publish, route_name='publish')
config.add_route('subscribe', '/subscribe')
config.add_view(subscribe, route_name='subscribe')
return config.make_wsgi_app()
|
c147629b4a0a5b405f7568b9278f288fa09fd97b | tests/aggregation/models.py | tests/aggregation/models.py | # coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
def __unicode__(self):
return self.name
| # coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
has_coffee = models.BooleanField()
def __unicode__(self):
return self.name
| Add a boolean field to Store model (store.has_coffee) | Add a boolean field to Store model (store.has_coffee)
| Python | mit | henriquebastos/django-aggregate-if | # coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
def __unicode__(self):
return self.name
Add a boolean field to Store model (store.has_coffee) | # coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
has_coffee = models.BooleanField()
def __unicode__(self):
return self.name
| <commit_before># coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
def __unicode__(self):
return self.name
<commit_msg>Add a boolean field to Store model (store.has_coffee)<commit_after> | # coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
has_coffee = models.BooleanField()
def __unicode__(self):
return self.name
| # coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
def __unicode__(self):
return self.name
Add a boolean field to Store model (store.has_coffee)# coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
has_coffee = models.BooleanField()
def __unicode__(self):
return self.name
| <commit_before># coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
def __unicode__(self):
return self.name
<commit_msg>Add a boolean field to Store model (store.has_coffee)<commit_after># coding: utf-8
from django.db import models
class Author(models.Model):
name = models.CharField(max_length=100)
age = models.IntegerField()
friends = models.ManyToManyField('self', blank=True)
def __unicode__(self):
return self.name
class Publisher(models.Model):
name = models.CharField(max_length=255)
num_awards = models.IntegerField()
def __unicode__(self):
return self.name
class Book(models.Model):
isbn = models.CharField(max_length=9)
name = models.CharField(max_length=255)
pages = models.IntegerField()
rating = models.FloatField()
price = models.DecimalField(decimal_places=2, max_digits=6)
authors = models.ManyToManyField(Author)
contact = models.ForeignKey(Author, related_name='book_contact_set')
publisher = models.ForeignKey(Publisher)
pubdate = models.DateField()
def __unicode__(self):
return self.name
class Store(models.Model):
name = models.CharField(max_length=255)
books = models.ManyToManyField(Book)
original_opening = models.DateTimeField()
friday_night_closing = models.TimeField()
has_coffee = models.BooleanField()
def __unicode__(self):
return self.name
|
0143e790245d19528af56df5428dc990d0689637 | node/multiply.py | node/multiply.py | #!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return a*b | #!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return[a*b] | Multiply now handles lists correctly | Multiply now handles lists correctly
| Python | mit | muddyfish/PYKE,muddyfish/PYKE | #!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return a*bMultiply now handles lists correctly | #!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return[a*b] | <commit_before>#!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return a*b<commit_msg>Multiply now handles lists correctly<commit_after> | #!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return[a*b] | #!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return a*bMultiply now handles lists correctly#!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return[a*b] | <commit_before>#!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return a*b<commit_msg>Multiply now handles lists correctly<commit_after>#!/usr/bin/env python
from nodes import Node
class Multiply(Node):
char = "*"
args = 2
results = 1
@Node.test_func([4,5], [20])
def func(self, a,b):
"""a*b"""
return[a*b] |
9f6ade7fab83f15b49e37e28ac2d044a41846809 | tests/test_create.py | tests/test_create.py | import globals as gbl
from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
def test_add_to_git(project_repo):
pass
def test_create_ticket(project_repo):
CreateTicket(ticket='test-ticket')
def test_create_package(project_repo):
CreatePackage(package='test-package')
| from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
from pathlib import Path
def test_add_to_git(project_repo):
pass
def test_create_ticket(session, project_repo):
test_ticket = 'test-ticket'
CreateTicket(ticket=test_ticket)
ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket)
deploy_file = Path(ticket_folder, 'deploy.py')
repo = Repo(str(project_repo))
last_commit = repo.get_object(repo.head())
commit_message = last_commit.message
assert ticket_folder.exists()
assert deploy_file.exists()
expected_message = bytes(
'Create ticket %s\n' % test_ticket, encoding='UTF-8')
assert commit_message == expected_message
def test_create_package(session, project_repo):
test_package = 'test-package'
CreatePackage(package=test_package)
package_folder = Path(project_repo, 'deploy', 'packages', test_package)
package_file = Path(package_folder, 'tickets.yml')
remove_file = Path(package_folder, 'remove.py')
assert package_folder.exists()
assert package_file.exists()
assert remove_file.exists()
| Add test for git commit | Add test for git commit
| Python | mit | Empiria/matador | import globals as gbl
from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
def test_add_to_git(project_repo):
pass
def test_create_ticket(project_repo):
CreateTicket(ticket='test-ticket')
def test_create_package(project_repo):
CreatePackage(package='test-package')
Add test for git commit | from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
from pathlib import Path
def test_add_to_git(project_repo):
pass
def test_create_ticket(session, project_repo):
test_ticket = 'test-ticket'
CreateTicket(ticket=test_ticket)
ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket)
deploy_file = Path(ticket_folder, 'deploy.py')
repo = Repo(str(project_repo))
last_commit = repo.get_object(repo.head())
commit_message = last_commit.message
assert ticket_folder.exists()
assert deploy_file.exists()
expected_message = bytes(
'Create ticket %s\n' % test_ticket, encoding='UTF-8')
assert commit_message == expected_message
def test_create_package(session, project_repo):
test_package = 'test-package'
CreatePackage(package=test_package)
package_folder = Path(project_repo, 'deploy', 'packages', test_package)
package_file = Path(package_folder, 'tickets.yml')
remove_file = Path(package_folder, 'remove.py')
assert package_folder.exists()
assert package_file.exists()
assert remove_file.exists()
| <commit_before>import globals as gbl
from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
def test_add_to_git(project_repo):
pass
def test_create_ticket(project_repo):
CreateTicket(ticket='test-ticket')
def test_create_package(project_repo):
CreatePackage(package='test-package')
<commit_msg>Add test for git commit<commit_after> | from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
from pathlib import Path
def test_add_to_git(project_repo):
pass
def test_create_ticket(session, project_repo):
test_ticket = 'test-ticket'
CreateTicket(ticket=test_ticket)
ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket)
deploy_file = Path(ticket_folder, 'deploy.py')
repo = Repo(str(project_repo))
last_commit = repo.get_object(repo.head())
commit_message = last_commit.message
assert ticket_folder.exists()
assert deploy_file.exists()
expected_message = bytes(
'Create ticket %s\n' % test_ticket, encoding='UTF-8')
assert commit_message == expected_message
def test_create_package(session, project_repo):
test_package = 'test-package'
CreatePackage(package=test_package)
package_folder = Path(project_repo, 'deploy', 'packages', test_package)
package_file = Path(package_folder, 'tickets.yml')
remove_file = Path(package_folder, 'remove.py')
assert package_folder.exists()
assert package_file.exists()
assert remove_file.exists()
| import globals as gbl
from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
def test_add_to_git(project_repo):
pass
def test_create_ticket(project_repo):
CreateTicket(ticket='test-ticket')
def test_create_package(project_repo):
CreatePackage(package='test-package')
Add test for git commitfrom matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
from pathlib import Path
def test_add_to_git(project_repo):
pass
def test_create_ticket(session, project_repo):
test_ticket = 'test-ticket'
CreateTicket(ticket=test_ticket)
ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket)
deploy_file = Path(ticket_folder, 'deploy.py')
repo = Repo(str(project_repo))
last_commit = repo.get_object(repo.head())
commit_message = last_commit.message
assert ticket_folder.exists()
assert deploy_file.exists()
expected_message = bytes(
'Create ticket %s\n' % test_ticket, encoding='UTF-8')
assert commit_message == expected_message
def test_create_package(session, project_repo):
test_package = 'test-package'
CreatePackage(package=test_package)
package_folder = Path(project_repo, 'deploy', 'packages', test_package)
package_file = Path(package_folder, 'tickets.yml')
remove_file = Path(package_folder, 'remove.py')
assert package_folder.exists()
assert package_file.exists()
assert remove_file.exists()
| <commit_before>import globals as gbl
from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
def test_add_to_git(project_repo):
pass
def test_create_ticket(project_repo):
CreateTicket(ticket='test-ticket')
def test_create_package(project_repo):
CreatePackage(package='test-package')
<commit_msg>Add test for git commit<commit_after>from matador.commands import CreateTicket, CreatePackage
from dulwich.repo import Repo
from pathlib import Path
def test_add_to_git(project_repo):
pass
def test_create_ticket(session, project_repo):
test_ticket = 'test-ticket'
CreateTicket(ticket=test_ticket)
ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket)
deploy_file = Path(ticket_folder, 'deploy.py')
repo = Repo(str(project_repo))
last_commit = repo.get_object(repo.head())
commit_message = last_commit.message
assert ticket_folder.exists()
assert deploy_file.exists()
expected_message = bytes(
'Create ticket %s\n' % test_ticket, encoding='UTF-8')
assert commit_message == expected_message
def test_create_package(session, project_repo):
test_package = 'test-package'
CreatePackage(package=test_package)
package_folder = Path(project_repo, 'deploy', 'packages', test_package)
package_file = Path(package_folder, 'tickets.yml')
remove_file = Path(package_folder, 'remove.py')
assert package_folder.exists()
assert package_file.exists()
assert remove_file.exists()
|
179df740725c0d3c9e256629e4718afcfa3b0cec | terminal_notifier.py | terminal_notifier.py | # This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
# 10-04-2015
# Version 1.0.0: initial release
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(signal_data))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
| # This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
#
# Version 1.0.0: initial release
# Version 1.0.1: fix escape characters which broke terminal-notifier
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
message = signal_data
if message[0] is "[":
message = "\\%s" % message
elif message[0] is "-":
message = "\\%s" % message
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(message))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
| Fix characters which break terminal-notifier | Fix characters which break terminal-notifier
If your message starts with either a [ or - (and probably more I haven't
found yet) terminal-notifier blows up because of the way it parses its
arguments
| Python | mit | keith/terminal-notifier-weechat | # This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
# 10-04-2015
# Version 1.0.0: initial release
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(signal_data))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
Fix characters which break terminal-notifier
If your message starts with either a [ or - (and probably more I haven't
found yet) terminal-notifier blows up because of the way it parses its
arguments | # This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
#
# Version 1.0.0: initial release
# Version 1.0.1: fix escape characters which broke terminal-notifier
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
message = signal_data
if message[0] is "[":
message = "\\%s" % message
elif message[0] is "-":
message = "\\%s" % message
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(message))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
| <commit_before># This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
# 10-04-2015
# Version 1.0.0: initial release
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(signal_data))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
<commit_msg>Fix characters which break terminal-notifier
If your message starts with either a [ or - (and probably more I haven't
found yet) terminal-notifier blows up because of the way it parses its
arguments<commit_after> | # This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
#
# Version 1.0.0: initial release
# Version 1.0.1: fix escape characters which broke terminal-notifier
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
message = signal_data
if message[0] is "[":
message = "\\%s" % message
elif message[0] is "-":
message = "\\%s" % message
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(message))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
| # This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
# 10-04-2015
# Version 1.0.0: initial release
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(signal_data))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
Fix characters which break terminal-notifier
If your message starts with either a [ or - (and probably more I haven't
found yet) terminal-notifier blows up because of the way it parses its
arguments# This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
#
# Version 1.0.0: initial release
# Version 1.0.1: fix escape characters which broke terminal-notifier
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
message = signal_data
if message[0] is "[":
message = "\\%s" % message
elif message[0] is "-":
message = "\\%s" % message
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(message))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
| <commit_before># This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
# 10-04-2015
# Version 1.0.0: initial release
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(signal_data))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
<commit_msg>Fix characters which break terminal-notifier
If your message starts with either a [ or - (and probably more I haven't
found yet) terminal-notifier blows up because of the way it parses its
arguments<commit_after># This weechat plugin sends OS X notifications for weechat messages
#
# Install terminal-notifier, no other configuration needed.
#
# History:
#
# Version 1.0.0: initial release
# Version 1.0.1: fix escape characters which broke terminal-notifier
import distutils.spawn
import os
import pipes
import weechat
def notify(data, signal, signal_data):
message = signal_data
if message[0] is "[":
message = "\\%s" % message
elif message[0] is "-":
message = "\\%s" % message
command = ("terminal-notifier -message %s -title WeeChat -sound Hero"
% pipes.quote(message))
exit_code = os.system(command)
if exit_code == 0:
return weechat.WEECHAT_RC_ERROR
else:
return weechat.WEECHAT_RC_OK
def main():
if distutils.spawn.find_executable("terminal-notifier") is None:
return weechat.WEECHAT_RC_ERROR
if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT",
"Get OS X notifications for messages", "", ""):
return weechat.WEECHAT_RC_ERROR
weechat.hook_signal("weechat_pv", "notify", "")
weechat.hook_signal("weechat_highlight", "notify", "")
return weechat.WEECHAT_RC_OK
if __name__ == "__main__":
main()
|
b2d9234ff6353191afc434556f9cfdea2448f726 | test/test_regexes.py | test/test_regexes.py | from findspam import FindSpam
import pytest
@pytest.mark.parametrize("text, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True),
('bagprada', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False),
('HOW DO YOU SOLVE THIS PROBLEM?', True),
])
def test_regexes(text, match):
result = FindSpam.testpost(text, "", "")
print text
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam | from findspam import FindSpam
import pytest
@pytest.mark.parametrize("title, username, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True),
('', 'bagprada', True),
('HOW DO YOU SOLVE THIS PROBLEM?', '', True),
('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True),
('support for yahoo mail 18669786819 @call for helpline number', '', True),
('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False),
('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False),
('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False),
('What is the proper way to say "queryer"', 'jedwards', False),
('What\'s a real-world example of "overfitting"?', 'user3851283', False),
('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False),
('Is it unfair to regrade prior work after detecting cheating?', 'Village', False),
])
def test_regexes(title, username, match):
result = FindSpam.testpost(title, username, "")
print title
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam | Update tests + add username field | Update tests + add username field
| Python | apache-2.0 | Charcoal-SE/SmokeDetector,NickVolynkin/SmokeDetector,ArtOfCode-/SmokeDetector,ArtOfCode-/SmokeDetector,Charcoal-SE/SmokeDetector,NickVolynkin/SmokeDetector | from findspam import FindSpam
import pytest
@pytest.mark.parametrize("text, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True),
('bagprada', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False),
('HOW DO YOU SOLVE THIS PROBLEM?', True),
])
def test_regexes(text, match):
result = FindSpam.testpost(text, "", "")
print text
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspamUpdate tests + add username field | from findspam import FindSpam
import pytest
@pytest.mark.parametrize("title, username, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True),
('', 'bagprada', True),
('HOW DO YOU SOLVE THIS PROBLEM?', '', True),
('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True),
('support for yahoo mail 18669786819 @call for helpline number', '', True),
('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False),
('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False),
('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False),
('What is the proper way to say "queryer"', 'jedwards', False),
('What\'s a real-world example of "overfitting"?', 'user3851283', False),
('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False),
('Is it unfair to regrade prior work after detecting cheating?', 'Village', False),
])
def test_regexes(title, username, match):
result = FindSpam.testpost(title, username, "")
print title
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam | <commit_before>from findspam import FindSpam
import pytest
@pytest.mark.parametrize("text, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True),
('bagprada', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False),
('HOW DO YOU SOLVE THIS PROBLEM?', True),
])
def test_regexes(text, match):
result = FindSpam.testpost(text, "", "")
print text
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam<commit_msg>Update tests + add username field<commit_after> | from findspam import FindSpam
import pytest
@pytest.mark.parametrize("title, username, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True),
('', 'bagprada', True),
('HOW DO YOU SOLVE THIS PROBLEM?', '', True),
('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True),
('support for yahoo mail 18669786819 @call for helpline number', '', True),
('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False),
('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False),
('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False),
('What is the proper way to say "queryer"', 'jedwards', False),
('What\'s a real-world example of "overfitting"?', 'user3851283', False),
('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False),
('Is it unfair to regrade prior work after detecting cheating?', 'Village', False),
])
def test_regexes(title, username, match):
result = FindSpam.testpost(title, username, "")
print title
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam | from findspam import FindSpam
import pytest
@pytest.mark.parametrize("text, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True),
('bagprada', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False),
('HOW DO YOU SOLVE THIS PROBLEM?', True),
])
def test_regexes(text, match):
result = FindSpam.testpost(text, "", "")
print text
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspamUpdate tests + add username fieldfrom findspam import FindSpam
import pytest
@pytest.mark.parametrize("title, username, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True),
('', 'bagprada', True),
('HOW DO YOU SOLVE THIS PROBLEM?', '', True),
('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True),
('support for yahoo mail 18669786819 @call for helpline number', '', True),
('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False),
('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False),
('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False),
('What is the proper way to say "queryer"', 'jedwards', False),
('What\'s a real-world example of "overfitting"?', 'user3851283', False),
('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False),
('Is it unfair to regrade prior work after detecting cheating?', 'Village', False),
])
def test_regexes(title, username, match):
result = FindSpam.testpost(title, username, "")
print title
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam | <commit_before>from findspam import FindSpam
import pytest
@pytest.mark.parametrize("text, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True),
('bagprada', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False),
('HOW DO YOU SOLVE THIS PROBLEM?', True),
])
def test_regexes(text, match):
result = FindSpam.testpost(text, "", "")
print text
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam<commit_msg>Update tests + add username field<commit_after>from findspam import FindSpam
import pytest
@pytest.mark.parametrize("title, username, match", [
('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True),
('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True),
('', 'bagprada', True),
('HOW DO YOU SOLVE THIS PROBLEM?', '', True),
('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True),
('support for yahoo mail 18669786819 @call for helpline number', '', True),
('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True),
('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False),
('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False),
('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False),
('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False),
('What is the proper way to say "queryer"', 'jedwards', False),
('What\'s a real-world example of "overfitting"?', 'user3851283', False),
('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False),
('Is it unfair to regrade prior work after detecting cheating?', 'Village', False),
])
def test_regexes(title, username, match):
result = FindSpam.testpost(title, username, "")
print title
print result
isspam = False
if (len(result) > 0):
isspam = True
assert match == isspam |
23c3b63e9f336ad70d481c0355d2f7f1544b9d54 | lattice_length.py | lattice_length.py | # Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
myLattice = ap.machines.getLattice()
length = 0
for key in range(myLattice.size()):
length += myLattice[key].length
print "The length of the lattice is {}.".format(length)
| # Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
my_lattice = ap.machines.getLattice()
length = 0
for key in range(my_lattice.size()):
length += my_lattice[key].length
print "The length of the lattice is {}.".format(length)
| Change variable name to more suitable ones | Change variable name to more suitable ones
| Python | apache-2.0 | razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects | # Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
myLattice = ap.machines.getLattice()
length = 0
for key in range(myLattice.size()):
length += myLattice[key].length
print "The length of the lattice is {}.".format(length)
Change variable name to more suitable ones | # Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
my_lattice = ap.machines.getLattice()
length = 0
for key in range(my_lattice.size()):
length += my_lattice[key].length
print "The length of the lattice is {}.".format(length)
| <commit_before># Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
myLattice = ap.machines.getLattice()
length = 0
for key in range(myLattice.size()):
length += myLattice[key].length
print "The length of the lattice is {}.".format(length)
<commit_msg>Change variable name to more suitable ones<commit_after> | # Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
my_lattice = ap.machines.getLattice()
length = 0
for key in range(my_lattice.size()):
length += my_lattice[key].length
print "The length of the lattice is {}.".format(length)
| # Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
myLattice = ap.machines.getLattice()
length = 0
for key in range(myLattice.size()):
length += myLattice[key].length
print "The length of the lattice is {}.".format(length)
Change variable name to more suitable ones# Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
my_lattice = ap.machines.getLattice()
length = 0
for key in range(my_lattice.size()):
length += my_lattice[key].length
print "The length of the lattice is {}.".format(length)
| <commit_before># Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
myLattice = ap.machines.getLattice()
length = 0
for key in range(myLattice.size()):
length += myLattice[key].length
print "The length of the lattice is {}.".format(length)
<commit_msg>Change variable name to more suitable ones<commit_after># Load the machine
import pkg_resources
pkg_resources.require('aphla')
import aphla as ap
# Load the machine
ap.machines.load('SRI21')
my_lattice = ap.machines.getLattice()
length = 0
for key in range(my_lattice.size()):
length += my_lattice[key].length
print "The length of the lattice is {}.".format(length)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.