commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
a2ee6106a6c98dae102cf14902c6b82f480e6cbe
python/main.py
python/main.py
import sys from enum import Enum class Furniture(Enum): bed = 1 couce = 2 desk = 3 chair = 4 tv = 5 table = 6 rug = 7 shelf = 8 f = open(sys.argv[1], 'r') print(f.read())
import sys from enum import Enum from furniture import * #class Furniture(Enum): # bed = 1 # couce = 2 # desk = 3 # chair = 4 # tv = 5 # table = 6 # rug = 7 # shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) placeDesksAndChairs() placeCouchesTablesAndTv() placeBeds() placeShelves() placeRugs()
Add calls to furniture placement functions
Add calls to furniture placement functions
Python
apache-2.0
TheZoq2/VRHack,TheZoq2/VRHack,TheZoq2/VRHack,TheZoq2/VRHack
import sys from enum import Enum class Furniture(Enum): bed = 1 couce = 2 desk = 3 chair = 4 tv = 5 table = 6 rug = 7 shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) Add calls to furniture placement functions
import sys from enum import Enum from furniture import * #class Furniture(Enum): # bed = 1 # couce = 2 # desk = 3 # chair = 4 # tv = 5 # table = 6 # rug = 7 # shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) placeDesksAndChairs() placeCouchesTablesAndTv() placeBeds() placeShelves() placeRugs()
<commit_before>import sys from enum import Enum class Furniture(Enum): bed = 1 couce = 2 desk = 3 chair = 4 tv = 5 table = 6 rug = 7 shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) <commit_msg>Add calls to furniture placement functions<commit_after>
import sys from enum import Enum from furniture import * #class Furniture(Enum): # bed = 1 # couce = 2 # desk = 3 # chair = 4 # tv = 5 # table = 6 # rug = 7 # shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) placeDesksAndChairs() placeCouchesTablesAndTv() placeBeds() placeShelves() placeRugs()
import sys from enum import Enum class Furniture(Enum): bed = 1 couce = 2 desk = 3 chair = 4 tv = 5 table = 6 rug = 7 shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) Add calls to furniture placement functionsimport sys from enum import Enum from furniture import * #class Furniture(Enum): # bed = 1 # couce = 2 # desk = 3 # chair = 4 # tv = 5 # table = 6 # rug = 7 # shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) placeDesksAndChairs() placeCouchesTablesAndTv() placeBeds() placeShelves() placeRugs()
<commit_before>import sys from enum import Enum class Furniture(Enum): bed = 1 couce = 2 desk = 3 chair = 4 tv = 5 table = 6 rug = 7 shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) <commit_msg>Add calls to furniture placement functions<commit_after>import sys from enum import Enum from furniture import * #class Furniture(Enum): # bed = 1 # couce = 2 # desk = 3 # chair = 4 # tv = 5 # table = 6 # rug = 7 # shelf = 8 f = open(sys.argv[1], 'r') print(f.read()) placeDesksAndChairs() placeCouchesTablesAndTv() placeBeds() placeShelves() placeRugs()
c7ed3e2a39c7de1120a33cd0253d9ac3bd9e7984
redcliff/cli.py
redcliff/cli.py
from sys import exit import argparse from .commands import dispatch from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url') parser.add_argument('-k', '--api-key') parser.add_argument('-C', '--config-file') parser.add_argument('cmd') parser.add_argument('args', nargs=argparse.REMAINDER) args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main())
from sys import exit import argparse from .commands import dispatch, choices from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url', metavar='https://redmine.example.com', help='Base URL of your Redmine installation.') parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify', action='store_const', const=False) parser.add_argument('-k', '--api-key', help='Your Redmine API key.') parser.add_argument('-C', '--config-file', help='Override default config path.') parser.add_argument('cmd', choices=choices, help='Command to execute.') parser.add_argument('args', nargs=argparse.REMAINDER, help='Arguments to command. Use --help to get ' 'command-specific help.') args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main())
Update main arguments parser config
Update main arguments parser config
Python
mit
dmedvinsky/redcliff
from sys import exit import argparse from .commands import dispatch from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url') parser.add_argument('-k', '--api-key') parser.add_argument('-C', '--config-file') parser.add_argument('cmd') parser.add_argument('args', nargs=argparse.REMAINDER) args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main()) Update main arguments parser config
from sys import exit import argparse from .commands import dispatch, choices from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url', metavar='https://redmine.example.com', help='Base URL of your Redmine installation.') parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify', action='store_const', const=False) parser.add_argument('-k', '--api-key', help='Your Redmine API key.') parser.add_argument('-C', '--config-file', help='Override default config path.') parser.add_argument('cmd', choices=choices, help='Command to execute.') parser.add_argument('args', nargs=argparse.REMAINDER, help='Arguments to command. Use --help to get ' 'command-specific help.') args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main())
<commit_before>from sys import exit import argparse from .commands import dispatch from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url') parser.add_argument('-k', '--api-key') parser.add_argument('-C', '--config-file') parser.add_argument('cmd') parser.add_argument('args', nargs=argparse.REMAINDER) args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main()) <commit_msg>Update main arguments parser config<commit_after>
from sys import exit import argparse from .commands import dispatch, choices from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url', metavar='https://redmine.example.com', help='Base URL of your Redmine installation.') parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify', action='store_const', const=False) parser.add_argument('-k', '--api-key', help='Your Redmine API key.') parser.add_argument('-C', '--config-file', help='Override default config path.') parser.add_argument('cmd', choices=choices, help='Command to execute.') parser.add_argument('args', nargs=argparse.REMAINDER, help='Arguments to command. Use --help to get ' 'command-specific help.') args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main())
from sys import exit import argparse from .commands import dispatch from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url') parser.add_argument('-k', '--api-key') parser.add_argument('-C', '--config-file') parser.add_argument('cmd') parser.add_argument('args', nargs=argparse.REMAINDER) args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main()) Update main arguments parser configfrom sys import exit import argparse from .commands import dispatch, choices from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url', metavar='https://redmine.example.com', help='Base URL of your Redmine installation.') parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify', action='store_const', const=False) parser.add_argument('-k', '--api-key', help='Your Redmine API key.') parser.add_argument('-C', '--config-file', help='Override default config path.') parser.add_argument('cmd', choices=choices, help='Command to execute.') parser.add_argument('args', nargs=argparse.REMAINDER, help='Arguments to command. Use --help to get ' 'command-specific help.') args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main())
<commit_before>from sys import exit import argparse from .commands import dispatch from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url') parser.add_argument('-k', '--api-key') parser.add_argument('-C', '--config-file') parser.add_argument('cmd') parser.add_argument('args', nargs=argparse.REMAINDER) args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main()) <commit_msg>Update main arguments parser config<commit_after>from sys import exit import argparse from .commands import dispatch, choices from .config import get_config from .utils import merge def main(): parser = argparse.ArgumentParser() parser.add_argument('-u', '--base-url', metavar='https://redmine.example.com', help='Base URL of your Redmine installation.') parser.add_argument('-S', '--no-ssl-verify', dest='ssl_verify', action='store_const', const=False) parser.add_argument('-k', '--api-key', help='Your Redmine API key.') parser.add_argument('-C', '--config-file', help='Override default config path.') parser.add_argument('cmd', choices=choices, help='Command to execute.') parser.add_argument('args', nargs=argparse.REMAINDER, help='Arguments to command. Use --help to get ' 'command-specific help.') args = vars(parser.parse_args()) conf = get_config(args.pop('config_file')) cmd = args.pop('cmd') cmd_args = args.pop('args') merged_conf = merge(conf, args) return dispatch(cmd, cmd_args, merged_conf) if __name__ == '__main__': exit(main())
02ca88129430044f4202991358939d87f8c6da0b
simple-cipher/simple_cipher.py
simple-cipher/simple_cipher.py
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
Refactor to reuse the encode method for decoding
Refactor to reuse the encode method for decoding
Python
agpl-3.0
CubicComet/exercism-python-solutions
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") Refactor to reuse the encode method for decoding
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
<commit_before>import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") <commit_msg>Refactor to reuse the encode method for decoding<commit_after>
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") Refactor to reuse the encode method for decodingimport math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
<commit_before>import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") <commit_msg>Refactor to reuse the encode method for decoding<commit_after>import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
3a86ea268b7cb9f00968e7dcb228d6821dafda99
simple-cipher/simple_cipher.py
simple-cipher/simple_cipher.py
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
Refactor to reuse the encode method for decoding
Refactor to reuse the encode method for decoding
Python
agpl-3.0
CubicComet/exercism-python-solutions
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") Refactor to reuse the encode method for decoding
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
<commit_before>import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") <commit_msg>Refactor to reuse the encode method for decoding<commit_after>
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") Refactor to reuse the encode method for decodingimport math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
<commit_before>import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, k) for c, k in zip(chars, key)) def decode(self, s): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, -k) for c, k in zip(chars, key)) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d") <commit_msg>Refactor to reuse the encode method for decoding<commit_after>import math import secrets from string import ascii_lowercase class Cipher(object): def __init__(self, key=None): if not key: key = self._random_key() if not key.isalpha() or not key.islower(): raise ValueError("Key must consist only of lowercase letters") self.key = key self._key = [ord(k)-97 for k in key] def encode(self, s, dirn=1): key = self._key * math.ceil(len(s)/len(self._key)) chars = [c for c in s.lower() if c in ascii_lowercase] return "".join(self._shift(c, dirn*k) for c, k in zip(chars, key)) def decode(self, s): return self.encode(s, dirn=-1) @staticmethod def _shift(char, key): return chr(97 + ((ord(char) - 97 + key) % 26)) @staticmethod def _random_key(length=256): return "".join(secrets.choice(ascii_lowercase) for _ in range(length)) class Caesar(Cipher): def __init__(self): super().__init__("d")
16c9563a75792aba7ccc0d979f579d64dc0140c1
common_rg_bar.py
common_rg_bar.py
#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) esc = chr(27) print (''.join(( esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console ''' % dict( prog_name=sys.argv[0], ))
#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width 3. (optional) Text to display (without color) shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) if len(sys.argv) >= 4: start_text = sys.argv[3] + ' ' else: start_text = '' esc = chr(27) print (''.join(( start_text, esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code) - len(start_text)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console 3. (optional) Text to display ''' % dict( prog_name=sys.argv[0], ))
Add optional text to display
Add optional text to display
Python
mit
kwadrat/rgb_tdd
#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) esc = chr(27) print (''.join(( esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console ''' % dict( prog_name=sys.argv[0], )) Add optional text to display
#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width 3. (optional) Text to display (without color) shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) if len(sys.argv) >= 4: start_text = sys.argv[3] + ' ' else: start_text = '' esc = chr(27) print (''.join(( start_text, esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code) - len(start_text)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console 3. (optional) Text to display ''' % dict( prog_name=sys.argv[0], ))
<commit_before>#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) esc = chr(27) print (''.join(( esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console ''' % dict( prog_name=sys.argv[0], )) <commit_msg>Add optional text to display<commit_after>
#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width 3. (optional) Text to display (without color) shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) if len(sys.argv) >= 4: start_text = sys.argv[3] + ' ' else: start_text = '' esc = chr(27) print (''.join(( start_text, esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code) - len(start_text)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console 3. (optional) Text to display ''' % dict( prog_name=sys.argv[0], ))
#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) esc = chr(27) print (''.join(( esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console ''' % dict( prog_name=sys.argv[0], )) Add optional text to display#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width 3. (optional) Text to display (without color) shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) if len(sys.argv) >= 4: start_text = sys.argv[3] + ' ' else: start_text = '' esc = chr(27) print (''.join(( start_text, esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code) - len(start_text)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console 3. (optional) Text to display ''' % dict( prog_name=sys.argv[0], ))
<commit_before>#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) esc = chr(27) print (''.join(( esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console ''' % dict( prog_name=sys.argv[0], )) <commit_msg>Add optional text to display<commit_after>#!/usr/bin/env python3 ''' Given: 1. status code: (0 - OK, other value - BAD) 2. terminal window width 3. (optional) Text to display (without color) shows red/green bar to visualize return code of previous command ''' import sys def main(): if len(sys.argv) >= 2: code = sys.argv[1] if code == 'x': col_char = '3' cols_limit = 78 code = '' # No code provided - only yellow bar else: if code == 'y': col_char = '3' else: value = int(code) if value: col_char = '1' else: col_char = '2' cols_limit = int(sys.argv[2]) if len(sys.argv) >= 4: start_text = sys.argv[3] + ' ' else: start_text = '' esc = chr(27) print (''.join(( start_text, esc, '[4', col_char, 'm', ' ' * (cols_limit - 2 - len(code) - len(start_text)), code, esc, '[0m', ))) else: print (''' Usage: %(prog_name)s status_code number_of_columns 1. status code: 0 - OK (green color), other values - BAD (red color) 2. number of columns: the width of text console 3. (optional) Text to display ''' % dict( prog_name=sys.argv[0], ))
ff391fc302b6d4e9fab0653522fa2fe47d8e8faa
beavy_modules/url_extractor/lib.py
beavy_modules/url_extractor/lib.py
import lassie from pyembed.core import PyEmbed from beavy.app import cache pyembed = PyEmbed() @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs)
from pyembed.core import PyEmbed from beavy.app import cache from lassie import Lassie import re # lassie by default isn't extensive enough for us # configure it so that it is. from lassie.filters import FILTER_MAPS FILTER_MAPS['meta']['open_graph']['map'].update({ # general "og:type": "type", "og:site_name": "site_name", }) FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I) FILTER_MAPS['meta']['generic']['map'].update({ # articles "article:published_time": "published_time", "article:modified_time": "modified_time", "article:expiration_time": "expiration_time", "article:section": "section", "article:section_url": "section_url", # music "music:duration": "duration", "music:release_date": "release_date", # video "video:duration": "duration", "video:release_date": "release_date", # author "author": "author", # book "book:author": "author", "book:isbn": "isbn", "book:release_date": "release_date", }) # general configuration pyembed = PyEmbed() lassie = Lassie() lassie.request_opts = { 'headers':{ # tell Lassie to tell others it is facebook 'User-Agent': 'facebookexternalhit/1.1' } } @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed(url, **kwargs)
Configure Lassie for more information
Configure Lassie for more information
Python
mpl-2.0
beavyHQ/beavy,beavyHQ/beavy,beavyHQ/beavy,beavyHQ/beavy
import lassie from pyembed.core import PyEmbed from beavy.app import cache pyembed = PyEmbed() @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs) Configure Lassie for more information
from pyembed.core import PyEmbed from beavy.app import cache from lassie import Lassie import re # lassie by default isn't extensive enough for us # configure it so that it is. from lassie.filters import FILTER_MAPS FILTER_MAPS['meta']['open_graph']['map'].update({ # general "og:type": "type", "og:site_name": "site_name", }) FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I) FILTER_MAPS['meta']['generic']['map'].update({ # articles "article:published_time": "published_time", "article:modified_time": "modified_time", "article:expiration_time": "expiration_time", "article:section": "section", "article:section_url": "section_url", # music "music:duration": "duration", "music:release_date": "release_date", # video "video:duration": "duration", "video:release_date": "release_date", # author "author": "author", # book "book:author": "author", "book:isbn": "isbn", "book:release_date": "release_date", }) # general configuration pyembed = PyEmbed() lassie = Lassie() lassie.request_opts = { 'headers':{ # tell Lassie to tell others it is facebook 'User-Agent': 'facebookexternalhit/1.1' } } @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed(url, **kwargs)
<commit_before>import lassie from pyembed.core import PyEmbed from beavy.app import cache pyembed = PyEmbed() @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs) <commit_msg>Configure Lassie for more information<commit_after>
from pyembed.core import PyEmbed from beavy.app import cache from lassie import Lassie import re # lassie by default isn't extensive enough for us # configure it so that it is. from lassie.filters import FILTER_MAPS FILTER_MAPS['meta']['open_graph']['map'].update({ # general "og:type": "type", "og:site_name": "site_name", }) FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I) FILTER_MAPS['meta']['generic']['map'].update({ # articles "article:published_time": "published_time", "article:modified_time": "modified_time", "article:expiration_time": "expiration_time", "article:section": "section", "article:section_url": "section_url", # music "music:duration": "duration", "music:release_date": "release_date", # video "video:duration": "duration", "video:release_date": "release_date", # author "author": "author", # book "book:author": "author", "book:isbn": "isbn", "book:release_date": "release_date", }) # general configuration pyembed = PyEmbed() lassie = Lassie() lassie.request_opts = { 'headers':{ # tell Lassie to tell others it is facebook 'User-Agent': 'facebookexternalhit/1.1' } } @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed(url, **kwargs)
import lassie from pyembed.core import PyEmbed from beavy.app import cache pyembed = PyEmbed() @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs) Configure Lassie for more information from pyembed.core import PyEmbed from beavy.app import cache from lassie import Lassie import re # lassie by default isn't extensive enough for us # configure it so that it is. from lassie.filters import FILTER_MAPS FILTER_MAPS['meta']['open_graph']['map'].update({ # general "og:type": "type", "og:site_name": "site_name", }) FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I) FILTER_MAPS['meta']['generic']['map'].update({ # articles "article:published_time": "published_time", "article:modified_time": "modified_time", "article:expiration_time": "expiration_time", "article:section": "section", "article:section_url": "section_url", # music "music:duration": "duration", "music:release_date": "release_date", # video "video:duration": "duration", "video:release_date": "release_date", # author "author": "author", # book "book:author": "author", "book:isbn": "isbn", "book:release_date": "release_date", }) # general configuration pyembed = PyEmbed() lassie = Lassie() lassie.request_opts = { 'headers':{ # tell Lassie to tell others it is facebook 'User-Agent': 'facebookexternalhit/1.1' } } @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed(url, **kwargs)
<commit_before>import lassie from pyembed.core import PyEmbed from beavy.app import cache pyembed = PyEmbed() @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed('http://www.youtube.com/watch?v=_PEdPBEpQfY', **kwargs) <commit_msg>Configure Lassie for more information<commit_after> from pyembed.core import PyEmbed from beavy.app import cache from lassie import Lassie import re # lassie by default isn't extensive enough for us # configure it so that it is. from lassie.filters import FILTER_MAPS FILTER_MAPS['meta']['open_graph']['map'].update({ # general "og:type": "type", "og:site_name": "site_name", }) FILTER_MAPS['meta']['generic']['pattern'] = re.compile(r"^(description|keywords|title|author|article:|music:|video:|book:)", re.I) FILTER_MAPS['meta']['generic']['map'].update({ # articles "article:published_time": "published_time", "article:modified_time": "modified_time", "article:expiration_time": "expiration_time", "article:section": "section", "article:section_url": "section_url", # music "music:duration": "duration", "music:release_date": "release_date", # video "video:duration": "duration", "video:release_date": "release_date", # author "author": "author", # book "book:author": "author", "book:isbn": "isbn", "book:release_date": "release_date", }) # general configuration pyembed = PyEmbed() lassie = Lassie() lassie.request_opts = { 'headers':{ # tell Lassie to tell others it is facebook 'User-Agent': 'facebookexternalhit/1.1' } } @cache.memoize() def extract_info(url): return lassie.fetch(url) @cache.memoize() def extract_oembed(url, **kwargs): return pyembed.embed(url, **kwargs)
c58b2dd49ad5c73d49b496025d13116da30b3b9a
examples/qm7_long.py
examples/qm7_long.py
import numpy from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] train_feats = [] test_feats = [] for tf in feats: X_train = tf.fit_transform(Xin_train) X_test = tf.transform(Xin_test) train_feats.append(X_train) test_feats.append(X_test) X_train = numpy.hstack(train_feats) X_test = numpy.hstack(test_feats) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print
from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity, MultiFeature from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] full_feat = MultiFeature(feats) X_train = full_feat.fit_transform(Xin_train) X_test = full_feat.transform(Xin_test) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print
Change qm7 example to use MultiFeature
Change qm7 example to use MultiFeature
Python
mit
crcollins/molml
import numpy from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] train_feats = [] test_feats = [] for tf in feats: X_train = tf.fit_transform(Xin_train) X_test = tf.transform(Xin_test) train_feats.append(X_train) test_feats.append(X_test) X_train = numpy.hstack(train_feats) X_test = numpy.hstack(test_feats) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print Change qm7 example to use MultiFeature
from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity, MultiFeature from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] full_feat = MultiFeature(feats) X_train = full_feat.fit_transform(Xin_train) X_test = full_feat.transform(Xin_test) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print
<commit_before>import numpy from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] train_feats = [] test_feats = [] for tf in feats: X_train = tf.fit_transform(Xin_train) X_test = tf.transform(Xin_test) train_feats.append(X_train) test_feats.append(X_test) X_train = numpy.hstack(train_feats) X_test = numpy.hstack(test_feats) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print <commit_msg>Change qm7 example to use MultiFeature<commit_after>
from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity, MultiFeature from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] full_feat = MultiFeature(feats) X_train = full_feat.fit_transform(Xin_train) X_test = full_feat.transform(Xin_test) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print
import numpy from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] train_feats = [] test_feats = [] for tf in feats: X_train = tf.fit_transform(Xin_train) X_test = tf.transform(Xin_test) train_feats.append(X_train) test_feats.append(X_test) X_train = numpy.hstack(train_feats) X_test = numpy.hstack(test_feats) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print Change qm7 example to use MultiFeaturefrom sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity, MultiFeature from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] full_feat = MultiFeature(feats) X_train = full_feat.fit_transform(Xin_train) X_test = full_feat.transform(Xin_test) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print
<commit_before>import numpy from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] train_feats = [] test_feats = [] for tf in feats: X_train = tf.fit_transform(Xin_train) X_test = tf.transform(Xin_test) train_feats.append(X_train) test_feats.append(X_test) X_train = numpy.hstack(train_feats) X_test = numpy.hstack(test_feats) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print <commit_msg>Change qm7 example to use MultiFeature<commit_after>from sklearn.linear_model import Ridge from sklearn.kernel_ridge import KernelRidge from sklearn.metrics import mean_absolute_error as MAE from molml.features import EncodedBond, Connectivity, MultiFeature from utils import load_qm7 if __name__ == "__main__": # This is just boiler plate code to load the data Xin_train, Xin_test, y_train, y_test = load_qm7() feats = [ EncodedBond(n_jobs=-1, max_depth=3), Connectivity(depth=1, n_jobs=-1), Connectivity(depth=3, use_coordination=True, n_jobs=-1), ] full_feat = MultiFeature(feats) X_train = full_feat.fit_transform(Xin_train) X_test = full_feat.transform(Xin_test) clfs = [ Ridge(alpha=0.01), KernelRidge(alpha=1e-9, gamma=1e-5, kernel="rbf"), ] for clf in clfs: print clf clf.fit(X_train, y_train) train_error = MAE(clf.predict(X_train), y_train) test_error = MAE(clf.predict(X_test), y_test) print "Train MAE: %.4f Test MAE: %.4f" % (train_error, test_error) print
67c444fb3603c234916b790d3dded3625f0512e5
pivot/test/test_utils.py
pivot/test/test_utils.py
""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_google_analytics_processor(self): self.assertEquals(get_latest_term(), 'au12')
""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_get_latest_term(self): self.assertEquals(get_latest_term(), 'au12')
Rename test to something more descriptive.
Rename test to something more descriptive.
Python
apache-2.0
uw-it-aca/pivot,uw-it-aca/pivot,uw-it-aca/pivot,uw-it-aca/pivot,uw-it-aca/pivot
""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_google_analytics_processor(self): self.assertEquals(get_latest_term(), 'au12') Rename test to something more descriptive.
""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_get_latest_term(self): self.assertEquals(get_latest_term(), 'au12')
<commit_before>""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_google_analytics_processor(self): self.assertEquals(get_latest_term(), 'au12') <commit_msg>Rename test to something more descriptive.<commit_after>
""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_get_latest_term(self): self.assertEquals(get_latest_term(), 'au12')
""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_google_analytics_processor(self): self.assertEquals(get_latest_term(), 'au12') Rename test to something more descriptive.""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_get_latest_term(self): self.assertEquals(get_latest_term(), 'au12')
<commit_before>""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_google_analytics_processor(self): self.assertEquals(get_latest_term(), 'au12') <commit_msg>Rename test to something more descriptive.<commit_after>""" Tests utility scripts """ import os from django.test import TestCase, RequestFactory from django.test.utils import override_settings import pivot from pivot.utils import get_latest_term TEST_CSV_PATH = os.path.join(os.path.dirname(pivot.__file__), 'test', 'test_resources', 'csvfiles/',) class UtilsTest(TestCase): @override_settings(CSV_ROOT=TEST_CSV_PATH) def test_get_latest_term(self): self.assertEquals(get_latest_term(), 'au12')
f7c03daa9ce803ec10e1c7cd9319840045f47663
ddsc_core/management/commands/export_pi_xml.py
ddsc_core/management/commands/export_pi_xml.py
import sys from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = "help" def handle(self, *args, **options): try: source = args[0] except IndexError: self.stdout.write(self.help) return reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(sys.stdout)
from optparse import make_option from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = ( "Create pi.xml from a template. A template is a valid pi.xml file " + "without events (they are ignored if present) and per series a " + "`comment` element containing a ddsc timeseries uuid." ) option_list = BaseCommand.option_list + ( make_option( '-f', '--file', dest='file', help='write to file instead of stdout' ), ) def handle(self, *args, **options): # source is a pi.xml file that serves as a template: its headers are # used, but any events are ignored. The `comment` element of each # series is expected to contain a ddsc uuid. try: source = args[0] except IndexError: self.stderr.write(self.help) return # destination is the resulting pi.xml file. If no destination is # given, output is written to stdout. try: destination = open(options.get('file'), 'w') except TypeError: destination = self.stdout reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(destination)
Improve management command for exporting pi-xml
Improve management command for exporting pi-xml
Python
mit
ddsc/ddsc-core,ddsc/ddsc-core
import sys from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = "help" def handle(self, *args, **options): try: source = args[0] except IndexError: self.stdout.write(self.help) return reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(sys.stdout) Improve management command for exporting pi-xml
from optparse import make_option from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = ( "Create pi.xml from a template. A template is a valid pi.xml file " + "without events (they are ignored if present) and per series a " + "`comment` element containing a ddsc timeseries uuid." ) option_list = BaseCommand.option_list + ( make_option( '-f', '--file', dest='file', help='write to file instead of stdout' ), ) def handle(self, *args, **options): # source is a pi.xml file that serves as a template: its headers are # used, but any events are ignored. The `comment` element of each # series is expected to contain a ddsc uuid. try: source = args[0] except IndexError: self.stderr.write(self.help) return # destination is the resulting pi.xml file. If no destination is # given, output is written to stdout. try: destination = open(options.get('file'), 'w') except TypeError: destination = self.stdout reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(destination)
<commit_before>import sys from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = "help" def handle(self, *args, **options): try: source = args[0] except IndexError: self.stdout.write(self.help) return reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(sys.stdout) <commit_msg>Improve management command for exporting pi-xml<commit_after>
from optparse import make_option from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = ( "Create pi.xml from a template. A template is a valid pi.xml file " + "without events (they are ignored if present) and per series a " + "`comment` element containing a ddsc timeseries uuid." ) option_list = BaseCommand.option_list + ( make_option( '-f', '--file', dest='file', help='write to file instead of stdout' ), ) def handle(self, *args, **options): # source is a pi.xml file that serves as a template: its headers are # used, but any events are ignored. The `comment` element of each # series is expected to contain a ddsc uuid. try: source = args[0] except IndexError: self.stderr.write(self.help) return # destination is the resulting pi.xml file. If no destination is # given, output is written to stdout. try: destination = open(options.get('file'), 'w') except TypeError: destination = self.stdout reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(destination)
import sys from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = "help" def handle(self, *args, **options): try: source = args[0] except IndexError: self.stdout.write(self.help) return reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(sys.stdout) Improve management command for exporting pi-xmlfrom optparse import make_option from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = ( "Create pi.xml from a template. A template is a valid pi.xml file " + "without events (they are ignored if present) and per series a " + "`comment` element containing a ddsc timeseries uuid." ) option_list = BaseCommand.option_list + ( make_option( '-f', '--file', dest='file', help='write to file instead of stdout' ), ) def handle(self, *args, **options): # source is a pi.xml file that serves as a template: its headers are # used, but any events are ignored. The `comment` element of each # series is expected to contain a ddsc uuid. try: source = args[0] except IndexError: self.stderr.write(self.help) return # destination is the resulting pi.xml file. If no destination is # given, output is written to stdout. try: destination = open(options.get('file'), 'w') except TypeError: destination = self.stdout reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(destination)
<commit_before>import sys from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = "help" def handle(self, *args, **options): try: source = args[0] except IndexError: self.stdout.write(self.help) return reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(sys.stdout) <commit_msg>Improve management command for exporting pi-xml<commit_after>from optparse import make_option from django.core.management.base import BaseCommand import pandas as pd from tslib.readers import PiXmlReader from tslib.writers import PiXmlWriter from ddsc_core.models import Timeseries class Command(BaseCommand): args = "<pi.xml>" help = ( "Create pi.xml from a template. A template is a valid pi.xml file " + "without events (they are ignored if present) and per series a " + "`comment` element containing a ddsc timeseries uuid." ) option_list = BaseCommand.option_list + ( make_option( '-f', '--file', dest='file', help='write to file instead of stdout' ), ) def handle(self, *args, **options): # source is a pi.xml file that serves as a template: its headers are # used, but any events are ignored. The `comment` element of each # series is expected to contain a ddsc uuid. try: source = args[0] except IndexError: self.stderr.write(self.help) return # destination is the resulting pi.xml file. If no destination is # given, output is written to stdout. try: destination = open(options.get('file'), 'w') except TypeError: destination = self.stdout reader = PiXmlReader(source) writer = PiXmlWriter(reader.get_tz()) for md, _ in reader.get_series(): try: uuid = md.pop('comment') timeseries = Timeseries.objects.get(uuid=uuid) df = timeseries.get_events() writer.set_series(md, df) except Timeseries.DoesNotExist: df = pd.DataFrame() writer.set_series(md, df) except: pass writer.write(destination)
afb398094e207fdd338a492dbbe9fca3f041e2c7
tests/test_postgres_processor.py
tests/test_postgres_processor.py
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from django.test import TestCase from scrapi.processing.postgres import PostgresProcessor, Document from . import utils from scrapi.linter.document import RawDocument test_db = PostgresProcessor() # NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) class DocumentTestCase(TestCase): @pytest.mark.django_db def test_Documents_can_speak(self): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
Make this test a django test case
Make this test a django test case
Python
apache-2.0
erinspace/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,erinspace/scrapi,felliott/scrapi,mehanig/scrapi,fabianvf/scrapi,fabianvf/scrapi,felliott/scrapi,mehanig/scrapi
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] Make this test a django test case
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from django.test import TestCase from scrapi.processing.postgres import PostgresProcessor, Document from . import utils from scrapi.linter.document import RawDocument test_db = PostgresProcessor() # NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) class DocumentTestCase(TestCase): @pytest.mark.django_db def test_Documents_can_speak(self): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
<commit_before>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] <commit_msg>Make this test a django test case<commit_after>
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from django.test import TestCase from scrapi.processing.postgres import PostgresProcessor, Document from . import utils from scrapi.linter.document import RawDocument test_db = PostgresProcessor() # NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) class DocumentTestCase(TestCase): @pytest.mark.django_db def test_Documents_can_speak(self): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] Make this test a django test caseimport pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from django.test import TestCase from scrapi.processing.postgres import PostgresProcessor, Document from . import utils from scrapi.linter.document import RawDocument test_db = PostgresProcessor() # NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) class DocumentTestCase(TestCase): @pytest.mark.django_db def test_Documents_can_speak(self): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
<commit_before>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from . import utils from scrapi.linter.document import NormalizedDocument, RawDocument from scrapi.processing.postgres import PostgresProcessor, Document test_db = PostgresProcessor() NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) @pytest.mark.django_db def test_process_raw(): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID'] <commit_msg>Make this test a django test case<commit_after>import pytest import os os.environ.setdefault("DJANGO_SETTINGS_MODULE", "api.api.settings") from django.test import TestCase from scrapi.processing.postgres import PostgresProcessor, Document from . import utils from scrapi.linter.document import RawDocument test_db = PostgresProcessor() # NORMALIZED = NormalizedDocument(utils.RECORD) RAW = RawDocument(utils.RAW_DOC) class DocumentTestCase(TestCase): @pytest.mark.django_db def test_Documents_can_speak(self): test_db.process_raw(RAW) queryset = Document(docID='someID', source=RAW['source']) assert queryset.docID == RAW.attributes['docID']
5d59f800da9fb737cd87d47301793f750ca1cbdd
pysnow/exceptions.py
pysnow/exceptions.py
# -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass
# -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class UnexpectedResponseFormat(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass
Add missing UnexpectedResponseFormat for backward compatability
Add missing UnexpectedResponseFormat for backward compatability Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com>
Python
mit
rbw0/pysnow
# -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass Add missing UnexpectedResponseFormat for backward compatability Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com>
# -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class UnexpectedResponseFormat(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass
<commit_before># -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass <commit_msg>Add missing UnexpectedResponseFormat for backward compatability Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com><commit_after>
# -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class UnexpectedResponseFormat(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass
# -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass Add missing UnexpectedResponseFormat for backward compatability Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com># -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class UnexpectedResponseFormat(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass
<commit_before># -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass <commit_msg>Add missing UnexpectedResponseFormat for backward compatability Signed-off-by: Abhijeet Kasurde <6334fd0c217b1f2a15926284df229acde5b4fc3a@redhat.com><commit_after># -*- coding: utf-8 -*- class PysnowException(Exception): pass class InvalidUsage(PysnowException): pass class UnexpectedResponseFormat(PysnowException): pass class ResponseError(PysnowException): message = "<empty>" detail = "<empty>" def __init__(self, error): if "message" in error: self.message = error["message"] or self.message if "detail" in error: self.detail = error["detail"] or self.detail def __str__(self): return "Error in response. Message: %s, Details: %s" % ( self.message, self.detail, ) class MissingResult(PysnowException): pass class NoResults(PysnowException): pass class EmptyContent(PysnowException): pass class MultipleResults(PysnowException): pass class MissingToken(PysnowException): pass class TokenCreateError(PysnowException): def __init__(self, error, description, status_code): self.error = error self.description = description self.snow_status_code = status_code class QueryTypeError(PysnowException): pass class QueryMissingField(PysnowException): pass class QueryEmpty(PysnowException): pass class QueryExpressionError(PysnowException): pass class QueryMultipleExpressions(PysnowException): pass
2e812af6b937091d65a0b83ead936894a2789d02
rdflib/serializer.py
rdflib/serializer.py
""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "UTF-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri
""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "utf-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri
Change to preferred encoding style.
Change to preferred encoding style. UTF-8 -> utf-8
Python
bsd-3-clause
RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib,RDFLib/rdflib
""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "UTF-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri Change to preferred encoding style. UTF-8 -> utf-8
""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "utf-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri
<commit_before>""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "UTF-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri <commit_msg>Change to preferred encoding style. UTF-8 -> utf-8<commit_after>
""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "utf-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri
""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "UTF-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri Change to preferred encoding style. UTF-8 -> utf-8""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "utf-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri
<commit_before>""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "UTF-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri <commit_msg>Change to preferred encoding style. UTF-8 -> utf-8<commit_after>""" Serializer plugin interface. This module is useful for those wanting to write a serializer that can plugin to rdflib. If you are wanting to invoke a serializer you likely want to do so through the Graph class serialize method. TODO: info for how to write a serializer that can plugin to rdflib. See also rdflib.plugin """ from typing import IO, TYPE_CHECKING, Optional from rdflib.term import URIRef if TYPE_CHECKING: from rdflib.graph import Graph __all__ = ["Serializer"] class Serializer: def __init__(self, store: "Graph"): self.store: "Graph" = store self.encoding: str = "utf-8" self.base: Optional[str] = None def serialize( self, stream: IO[bytes], base: Optional[str] = None, encoding: Optional[str] = None, **args ) -> None: """Abstract method""" def relativize(self, uri: str): base = self.base if base is not None and uri.startswith(base): uri = URIRef(uri.replace(base, "", 1)) return uri
6e32cfd9b2640b4f119a3a8e4138c883fd4bcef0
_tests/test_scikit_ci_addons.py
_tests/test_scikit_ci_addons.py
import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = ci_addons.home() + '/' + addon if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci/noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) )
import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = os.path.join(ci_addons.home(), addon) if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci' + os.path.sep + 'noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) )
Fix failing tests on appveyor
ci: Fix failing tests on appveyor
Python
apache-2.0
scikit-build/scikit-ci-addons,scikit-build/scikit-ci-addons
import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = ci_addons.home() + '/' + addon if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci/noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) ) ci: Fix failing tests on appveyor
import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = os.path.join(ci_addons.home(), addon) if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci' + os.path.sep + 'noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) )
<commit_before> import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = ci_addons.home() + '/' + addon if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci/noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) ) <commit_msg>ci: Fix failing tests on appveyor<commit_after>
import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = os.path.join(ci_addons.home(), addon) if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci' + os.path.sep + 'noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) )
import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = ci_addons.home() + '/' + addon if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci/noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) ) ci: Fix failing tests on appveyor import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = os.path.join(ci_addons.home(), addon) if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci' + os.path.sep + 'noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) )
<commit_before> import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = ci_addons.home() + '/' + addon if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci/noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert ci_addons.home() + '/anyci/noop.py foo bar' in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) ) <commit_msg>ci: Fix failing tests on appveyor<commit_after> import ci_addons import os import pytest import subprocess from . import captured_lines def test_home(): expected_home = os.path.abspath(os.path.dirname(__file__) + '/..') assert ci_addons.home() == expected_home @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_path(addon): expected_path = os.path.join(ci_addons.home(), addon) if not addon.endswith('.py'): expected_path += '.py' assert ci_addons.path(addon) == expected_path def test_list(capsys): ci_addons.list_addons() output_lines, _ = captured_lines(capsys) assert 'anyci' + os.path.sep + 'noop.py' in output_lines @pytest.mark.parametrize("addon", ['anyci/noop', 'anyci/noop.py']) def test_execute(addon, capfd): ci_addons.execute(addon, ['foo', 'bar']) output_lines, _ = captured_lines(capfd) assert os.path.join(ci_addons.home(), 'anyci/noop.py foo bar') in output_lines def test_cli(): root = os.path.abspath(os.path.join(os.path.dirname(__file__), '..')) environment = dict(os.environ) environment['PYTHONPATH'] = root subprocess.check_call( "python -m ci_addons", shell=True, env=environment, stderr=subprocess.STDOUT, cwd=str(root) )
43ab1500719665b44e3b4eca4def9002711c2ee8
githublist/parser.py
githublist/parser.py
import requests import collections API_URL = 'https://api.github.com/users/{}/repos' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u))
import requests import collections API_URL = 'https://api.github.com/users/{}/repos?per_page=100' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u))
Update api url for recent 100 instead of default 30
Update api url for recent 100 instead of default 30
Python
mit
kshvmdn/github-list,kshvmdn/github-list,kshvmdn/github-list
import requests import collections API_URL = 'https://api.github.com/users/{}/repos' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u)) Update api url for recent 100 instead of default 30
import requests import collections API_URL = 'https://api.github.com/users/{}/repos?per_page=100' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u))
<commit_before>import requests import collections API_URL = 'https://api.github.com/users/{}/repos' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u)) <commit_msg>Update api url for recent 100 instead of default 30<commit_after>
import requests import collections API_URL = 'https://api.github.com/users/{}/repos?per_page=100' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u))
import requests import collections API_URL = 'https://api.github.com/users/{}/repos' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u)) Update api url for recent 100 instead of default 30import requests import collections API_URL = 'https://api.github.com/users/{}/repos?per_page=100' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u))
<commit_before>import requests import collections API_URL = 'https://api.github.com/users/{}/repos' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u)) <commit_msg>Update api url for recent 100 instead of default 30<commit_after>import requests import collections API_URL = 'https://api.github.com/users/{}/repos?per_page=100' def main(user): return parse(request(user)) def request(user): return requests.get(url=API_URL.format(user)) def parse(response): repos = response.json() data = [] if repos is None: return None for repo in repos: if 'name' in repo and not repo['fork']: data.append( collections.OrderedDict([('name', repo['name']), ('desc', repo['description']), ('lang', repo['language']), ('stars', repo['stargazers_count'])])) return data if __name__ == '__main__': import pprint u = 'kshvmdn' pprint.pprint(main(u))
fdf0daefac50de71a8c4f80a9ef877669ebea48b
byceps/services/tourney/transfer/models.py
byceps/services/tourney/transfer/models.py
""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID
""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
Change tourney match transfer model from `attrs` to `dataclass`
Change tourney match transfer model from `attrs` to `dataclass`
Python
bsd-3-clause
homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps
""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID Change tourney match transfer model from `attrs` to `dataclass`
""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
<commit_before>""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID <commit_msg>Change tourney match transfer model from `attrs` to `dataclass`<commit_after>
""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID Change tourney match transfer model from `attrs` to `dataclass`""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
<commit_before>""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from typing import NewType from uuid import UUID from attr import attrs TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @attrs(auto_attribs=True, frozen=True, slots=True) class Match: id: MatchID <commit_msg>Change tourney match transfer model from `attrs` to `dataclass`<commit_after>""" byceps.services.tourney.transfer.models ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2019 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from dataclasses import dataclass from typing import NewType from uuid import UUID TourneyCategoryID = NewType('TourneyCategoryID', UUID) TourneyID = NewType('TourneyID', UUID) MatchID = NewType('MatchID', UUID) MatchCommentID = NewType('MatchCommentID', UUID) ParticipantID = NewType('ParticipantID', UUID) @dataclass(frozen=True) class Match: id: MatchID
50519406ac64766874ce9edf5cea69233461ffb2
tests/test_config.py
tests/test_config.py
# -*- coding: utf-8 -*- import pytest import uuid from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.fixture def dummyvalue(): return str(uuid.uuid4()) def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam, dummyvalue): config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue
# -*- coding: utf-8 -*- import pytest import uuid import tempfile import os from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.yield_fixture def dummy_config_file(): filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4())) yield filename if os.path.isfile(filename): os.remove(filename) @pytest.fixture def custom_config_file(dummy_config_file, scope='module'): return S3Keyring(profile_name='test', config_file=dummy_config_file).config def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam): dummyvalue = str(uuid.uuid4()) config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue def test_read_custom_config_file(custom_config_file, dummy_config_file): """Reads a parameter from a custom config file""" profile_name = custom_config_file.get('default', 'profile') assert profile_name == 'default' assert custom_config_file.config_file == dummy_config_file assert os.path.isfile(dummy_config_file) def test_write_config_in_custom_config_file(custom_config_file, dummyparam, config): dummyvalue = str(uuid.uuid4()) custom_config_file.set('default', dummyparam, dummyvalue) custom_config_file.save() custom_config_file.load() assert custom_config_file.get('default', dummyparam) == dummyvalue assert config.config_file != custom_config_file.config_file config.load() assert config.get('default', dummyparam) != dummyvalue
Test custom configuration file feature
Test custom configuration file feature
Python
mit
InnovativeTravel/s3-keyring
# -*- coding: utf-8 -*- import pytest import uuid from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.fixture def dummyvalue(): return str(uuid.uuid4()) def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam, dummyvalue): config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue Test custom configuration file feature
# -*- coding: utf-8 -*- import pytest import uuid import tempfile import os from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.yield_fixture def dummy_config_file(): filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4())) yield filename if os.path.isfile(filename): os.remove(filename) @pytest.fixture def custom_config_file(dummy_config_file, scope='module'): return S3Keyring(profile_name='test', config_file=dummy_config_file).config def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam): dummyvalue = str(uuid.uuid4()) config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue def test_read_custom_config_file(custom_config_file, dummy_config_file): """Reads a parameter from a custom config file""" profile_name = custom_config_file.get('default', 'profile') assert profile_name == 'default' assert custom_config_file.config_file == dummy_config_file assert os.path.isfile(dummy_config_file) def test_write_config_in_custom_config_file(custom_config_file, dummyparam, config): dummyvalue = str(uuid.uuid4()) custom_config_file.set('default', dummyparam, dummyvalue) custom_config_file.save() custom_config_file.load() assert custom_config_file.get('default', dummyparam) == dummyvalue assert config.config_file != custom_config_file.config_file config.load() assert config.get('default', dummyparam) != dummyvalue
<commit_before># -*- coding: utf-8 -*- import pytest import uuid from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.fixture def dummyvalue(): return str(uuid.uuid4()) def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam, dummyvalue): config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue <commit_msg>Test custom configuration file feature<commit_after>
# -*- coding: utf-8 -*- import pytest import uuid import tempfile import os from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.yield_fixture def dummy_config_file(): filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4())) yield filename if os.path.isfile(filename): os.remove(filename) @pytest.fixture def custom_config_file(dummy_config_file, scope='module'): return S3Keyring(profile_name='test', config_file=dummy_config_file).config def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam): dummyvalue = str(uuid.uuid4()) config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue def test_read_custom_config_file(custom_config_file, dummy_config_file): """Reads a parameter from a custom config file""" profile_name = custom_config_file.get('default', 'profile') assert profile_name == 'default' assert custom_config_file.config_file == dummy_config_file assert os.path.isfile(dummy_config_file) def test_write_config_in_custom_config_file(custom_config_file, dummyparam, config): dummyvalue = str(uuid.uuid4()) custom_config_file.set('default', dummyparam, dummyvalue) custom_config_file.save() custom_config_file.load() assert custom_config_file.get('default', dummyparam) == dummyvalue assert config.config_file != custom_config_file.config_file config.load() assert config.get('default', dummyparam) != dummyvalue
# -*- coding: utf-8 -*- import pytest import uuid from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.fixture def dummyvalue(): return str(uuid.uuid4()) def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam, dummyvalue): config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue Test custom configuration file feature# -*- coding: utf-8 -*- import pytest import uuid import tempfile import os from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.yield_fixture def dummy_config_file(): filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4())) yield filename if os.path.isfile(filename): os.remove(filename) @pytest.fixture def custom_config_file(dummy_config_file, scope='module'): return S3Keyring(profile_name='test', config_file=dummy_config_file).config def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam): dummyvalue = str(uuid.uuid4()) config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue def test_read_custom_config_file(custom_config_file, dummy_config_file): """Reads a parameter from a custom config file""" profile_name = custom_config_file.get('default', 'profile') assert profile_name == 'default' assert custom_config_file.config_file == dummy_config_file assert os.path.isfile(dummy_config_file) def test_write_config_in_custom_config_file(custom_config_file, dummyparam, config): dummyvalue = str(uuid.uuid4()) custom_config_file.set('default', dummyparam, dummyvalue) custom_config_file.save() custom_config_file.load() assert custom_config_file.get('default', dummyparam) == dummyvalue assert config.config_file != custom_config_file.config_file config.load() assert config.get('default', dummyparam) != dummyvalue
<commit_before># -*- coding: utf-8 -*- import pytest import uuid from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.fixture def dummyvalue(): return str(uuid.uuid4()) def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam, dummyvalue): config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue <commit_msg>Test custom configuration file feature<commit_after># -*- coding: utf-8 -*- import pytest import uuid import tempfile import os from s3keyring.s3 import S3Keyring @pytest.fixture def config(scope='module'): return S3Keyring(profile_name='test').config @pytest.yield_fixture def dummyparam(config, scope='module'): yield 'dummyparam' config.config.remove_option('default', 'dummyparam') @pytest.yield_fixture def dummy_config_file(): filename = os.path.join(tempfile.gettempdir(), str(uuid.uuid4())) yield filename if os.path.isfile(filename): os.remove(filename) @pytest.fixture def custom_config_file(dummy_config_file, scope='module'): return S3Keyring(profile_name='test', config_file=dummy_config_file).config def test_read_config(config): """Sets value for an existing configuration option""" profile_name = config.get('default', 'profile') assert profile_name == 'default' def test_write_config(config, dummyparam): dummyvalue = str(uuid.uuid4()) config.set('default', dummyparam, dummyvalue) config.save() config.load() assert config.get('default', dummyparam) == dummyvalue def test_read_custom_config_file(custom_config_file, dummy_config_file): """Reads a parameter from a custom config file""" profile_name = custom_config_file.get('default', 'profile') assert profile_name == 'default' assert custom_config_file.config_file == dummy_config_file assert os.path.isfile(dummy_config_file) def test_write_config_in_custom_config_file(custom_config_file, dummyparam, config): dummyvalue = str(uuid.uuid4()) custom_config_file.set('default', dummyparam, dummyvalue) custom_config_file.save() custom_config_file.load() assert custom_config_file.get('default', dummyparam) == dummyvalue assert config.config_file != custom_config_file.config_file config.load() assert config.get('default', dummyparam) != dummyvalue
025927fa19bb96095a2ea1c53524945f1f9590ce
spur/results.py
spur/results.py
def result(return_code, output, stderr_output, allow_error=False): if allow_error or return_code == 0: return ExecutionResult(return_code, output, stderr_output) else: raise RunProcessError(return_code, output, stderr_output) class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output
def result(return_code, output, stderr_output, allow_error=False): result = ExecutionResult(return_code, output, stderr_output) if allow_error or return_code == 0: return result else: raise result.to_error() class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output def to_error(self): return RunProcessError( self.return_code, self.output, self.stderr_output )
Move logic for creating RunProcessError to ExecutionResult.to_error
Move logic for creating RunProcessError to ExecutionResult.to_error
Python
bsd-2-clause
mwilliamson/spur.py
def result(return_code, output, stderr_output, allow_error=False): if allow_error or return_code == 0: return ExecutionResult(return_code, output, stderr_output) else: raise RunProcessError(return_code, output, stderr_output) class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output Move logic for creating RunProcessError to ExecutionResult.to_error
def result(return_code, output, stderr_output, allow_error=False): result = ExecutionResult(return_code, output, stderr_output) if allow_error or return_code == 0: return result else: raise result.to_error() class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output def to_error(self): return RunProcessError( self.return_code, self.output, self.stderr_output )
<commit_before>def result(return_code, output, stderr_output, allow_error=False): if allow_error or return_code == 0: return ExecutionResult(return_code, output, stderr_output) else: raise RunProcessError(return_code, output, stderr_output) class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output <commit_msg>Move logic for creating RunProcessError to ExecutionResult.to_error<commit_after>
def result(return_code, output, stderr_output, allow_error=False): result = ExecutionResult(return_code, output, stderr_output) if allow_error or return_code == 0: return result else: raise result.to_error() class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output def to_error(self): return RunProcessError( self.return_code, self.output, self.stderr_output )
def result(return_code, output, stderr_output, allow_error=False): if allow_error or return_code == 0: return ExecutionResult(return_code, output, stderr_output) else: raise RunProcessError(return_code, output, stderr_output) class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output Move logic for creating RunProcessError to ExecutionResult.to_errordef result(return_code, output, stderr_output, allow_error=False): result = ExecutionResult(return_code, output, stderr_output) if allow_error or return_code == 0: return result else: raise result.to_error() class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output def to_error(self): return RunProcessError( self.return_code, self.output, self.stderr_output )
<commit_before>def result(return_code, output, stderr_output, allow_error=False): if allow_error or return_code == 0: return ExecutionResult(return_code, output, stderr_output) else: raise RunProcessError(return_code, output, stderr_output) class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output <commit_msg>Move logic for creating RunProcessError to ExecutionResult.to_error<commit_after>def result(return_code, output, stderr_output, allow_error=False): result = ExecutionResult(return_code, output, stderr_output) if allow_error or return_code == 0: return result else: raise result.to_error() class RunProcessError(RuntimeError): def __init__(self, return_code, output, stderr_output): message = "return code: {0}\noutput: {1}\nstderr output: {2}".format( return_code, output, stderr_output) super(type(self), self).__init__(message) self.return_code = return_code self.output = output self.stderr_output = stderr_output class ExecutionResult(object): def __init__(self, return_code, output, stderr_output): self.return_code = return_code self.output = output self.stderr_output = stderr_output def to_error(self): return RunProcessError( self.return_code, self.output, self.stderr_output )
936db17eed36284917395a6a8272351dabbc8168
numpy/_array_api/_dtypes.py
numpy/_array_api/_dtypes.py
from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64 # Note: This name is changed from .. import bool_ as bool _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
import numpy as np # Note: we use dtype objects instead of dtype classes. The spec does not # require any behavior on dtypes other than equality. int8 = np.dtype('int8') int16 = np.dtype('int16') int32 = np.dtype('int32') int64 = np.dtype('int64') uint8 = np.dtype('uint8') uint16 = np.dtype('uint16') uint32 = np.dtype('uint32') uint64 = np.dtype('uint64') float32 = np.dtype('float32') float64 = np.dtype('float64') # Note: This name is changed bool = np.dtype('bool') _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
Use dtype objects instead of classes in the array API
Use dtype objects instead of classes in the array API The array API does not require any methods or behaviors on dtype objects, other than that they be literals that can be compared for equality and passed to dtype keywords in functions. Since dtype objects are already used by the dtype attribute of ndarray, this makes it consistent, so that func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the same thing as numpy._array_api.<dtype>. This also fixes an issue in the array API test suite due to the fact that dtype classes and objects are not equal as dictionary keys.
Python
mit
cupy/cupy,cupy/cupy,cupy/cupy,cupy/cupy
from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64 # Note: This name is changed from .. import bool_ as bool _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64] Use dtype objects instead of classes in the array API The array API does not require any methods or behaviors on dtype objects, other than that they be literals that can be compared for equality and passed to dtype keywords in functions. Since dtype objects are already used by the dtype attribute of ndarray, this makes it consistent, so that func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the same thing as numpy._array_api.<dtype>. This also fixes an issue in the array API test suite due to the fact that dtype classes and objects are not equal as dictionary keys.
import numpy as np # Note: we use dtype objects instead of dtype classes. The spec does not # require any behavior on dtypes other than equality. int8 = np.dtype('int8') int16 = np.dtype('int16') int32 = np.dtype('int32') int64 = np.dtype('int64') uint8 = np.dtype('uint8') uint16 = np.dtype('uint16') uint32 = np.dtype('uint32') uint64 = np.dtype('uint64') float32 = np.dtype('float32') float64 = np.dtype('float64') # Note: This name is changed bool = np.dtype('bool') _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
<commit_before>from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64 # Note: This name is changed from .. import bool_ as bool _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64] <commit_msg>Use dtype objects instead of classes in the array API The array API does not require any methods or behaviors on dtype objects, other than that they be literals that can be compared for equality and passed to dtype keywords in functions. Since dtype objects are already used by the dtype attribute of ndarray, this makes it consistent, so that func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the same thing as numpy._array_api.<dtype>. This also fixes an issue in the array API test suite due to the fact that dtype classes and objects are not equal as dictionary keys.<commit_after>
import numpy as np # Note: we use dtype objects instead of dtype classes. The spec does not # require any behavior on dtypes other than equality. int8 = np.dtype('int8') int16 = np.dtype('int16') int32 = np.dtype('int32') int64 = np.dtype('int64') uint8 = np.dtype('uint8') uint16 = np.dtype('uint16') uint32 = np.dtype('uint32') uint64 = np.dtype('uint64') float32 = np.dtype('float32') float64 = np.dtype('float64') # Note: This name is changed bool = np.dtype('bool') _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64 # Note: This name is changed from .. import bool_ as bool _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64] Use dtype objects instead of classes in the array API The array API does not require any methods or behaviors on dtype objects, other than that they be literals that can be compared for equality and passed to dtype keywords in functions. Since dtype objects are already used by the dtype attribute of ndarray, this makes it consistent, so that func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the same thing as numpy._array_api.<dtype>. This also fixes an issue in the array API test suite due to the fact that dtype classes and objects are not equal as dictionary keys.import numpy as np # Note: we use dtype objects instead of dtype classes. The spec does not # require any behavior on dtypes other than equality. int8 = np.dtype('int8') int16 = np.dtype('int16') int32 = np.dtype('int32') int64 = np.dtype('int64') uint8 = np.dtype('uint8') uint16 = np.dtype('uint16') uint32 = np.dtype('uint32') uint64 = np.dtype('uint64') float32 = np.dtype('float32') float64 = np.dtype('float64') # Note: This name is changed bool = np.dtype('bool') _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
<commit_before>from .. import int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64 # Note: This name is changed from .. import bool_ as bool _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64] <commit_msg>Use dtype objects instead of classes in the array API The array API does not require any methods or behaviors on dtype objects, other than that they be literals that can be compared for equality and passed to dtype keywords in functions. Since dtype objects are already used by the dtype attribute of ndarray, this makes it consistent, so that func(dtype=<dtype>).dtype will give exactly <dtype> back, which will be the same thing as numpy._array_api.<dtype>. This also fixes an issue in the array API test suite due to the fact that dtype classes and objects are not equal as dictionary keys.<commit_after>import numpy as np # Note: we use dtype objects instead of dtype classes. The spec does not # require any behavior on dtypes other than equality. int8 = np.dtype('int8') int16 = np.dtype('int16') int32 = np.dtype('int32') int64 = np.dtype('int64') uint8 = np.dtype('uint8') uint16 = np.dtype('uint16') uint32 = np.dtype('uint32') uint64 = np.dtype('uint64') float32 = np.dtype('float32') float64 = np.dtype('float64') # Note: This name is changed bool = np.dtype('bool') _all_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64, bool] _boolean_dtypes = [bool] _floating_dtypes = [float32, float64] _integer_dtypes = [int8, int16, int32, int64, uint8, uint16, uint32, uint64] _integer_or_boolean_dtypes = [bool, int8, int16, int32, int64, uint8, uint16, uint32, uint64] _numeric_dtypes = [float32, float64, int8, int16, int32, int64, uint8, uint16, uint32, uint64]
ffd4c52155acd7d04939e766ebe63171b580a2fa
src/__init__.py
src/__init__.py
import os import logging from kaa.base import ipc from client import * from server import * __all__ = [ 'connect' ] # connected client object _client = None def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') if epgdb.find(':') >= 0: # epg is remote: host:port # TODO: create socket, pass it to client _client = GuideClient("epg") else: # epg is local _client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg") return _client
import os import logging from socket import gethostbyname, gethostname from kaa.base import ipc from client import * from server import * __all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ] # connected client object _client = None def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client if _client: return _client if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \ address != gethostbyname(gethostname()): # epg is remote: host:port if address.find(':') >= 0: host, port = address.split(':', 1) else: host = address port = DEFAULT_EPG_PORT # create socket, pass it to client _client = GuideClient((host, port)) else: # EPG is local, only use unix socket # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') _client = ipc.launch([server, logfile, str(loglevel), epgdb, address], 2, GuideClient, "epg") return _client
Add the ability to use inet socket as well.
Add the ability to use inet socket as well. git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e
Python
lgpl-2.1
freevo/kaa-epg
import os import logging from kaa.base import ipc from client import * from server import * __all__ = [ 'connect' ] # connected client object _client = None def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') if epgdb.find(':') >= 0: # epg is remote: host:port # TODO: create socket, pass it to client _client = GuideClient("epg") else: # epg is local _client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg") return _client Add the ability to use inet socket as well. git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e
import os import logging from socket import gethostbyname, gethostname from kaa.base import ipc from client import * from server import * __all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ] # connected client object _client = None def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client if _client: return _client if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \ address != gethostbyname(gethostname()): # epg is remote: host:port if address.find(':') >= 0: host, port = address.split(':', 1) else: host = address port = DEFAULT_EPG_PORT # create socket, pass it to client _client = GuideClient((host, port)) else: # EPG is local, only use unix socket # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') _client = ipc.launch([server, logfile, str(loglevel), epgdb, address], 2, GuideClient, "epg") return _client
<commit_before>import os import logging from kaa.base import ipc from client import * from server import * __all__ = [ 'connect' ] # connected client object _client = None def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') if epgdb.find(':') >= 0: # epg is remote: host:port # TODO: create socket, pass it to client _client = GuideClient("epg") else: # epg is local _client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg") return _client <commit_msg>Add the ability to use inet socket as well. git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e<commit_after>
import os import logging from socket import gethostbyname, gethostname from kaa.base import ipc from client import * from server import * __all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ] # connected client object _client = None def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client if _client: return _client if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \ address != gethostbyname(gethostname()): # epg is remote: host:port if address.find(':') >= 0: host, port = address.split(':', 1) else: host = address port = DEFAULT_EPG_PORT # create socket, pass it to client _client = GuideClient((host, port)) else: # EPG is local, only use unix socket # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') _client = ipc.launch([server, logfile, str(loglevel), epgdb, address], 2, GuideClient, "epg") return _client
import os import logging from kaa.base import ipc from client import * from server import * __all__ = [ 'connect' ] # connected client object _client = None def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') if epgdb.find(':') >= 0: # epg is remote: host:port # TODO: create socket, pass it to client _client = GuideClient("epg") else: # epg is local _client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg") return _client Add the ability to use inet socket as well. git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8eimport os import logging from socket import gethostbyname, gethostname from kaa.base import ipc from client import * from server import * __all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ] # connected client object _client = None def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client if _client: return _client if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \ address != gethostbyname(gethostname()): # epg is remote: host:port if address.find(':') >= 0: host, port = address.split(':', 1) else: host = address port = DEFAULT_EPG_PORT # create socket, pass it to client _client = GuideClient((host, port)) else: # EPG is local, only use unix socket # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') _client = ipc.launch([server, logfile, str(loglevel), epgdb, address], 2, GuideClient, "epg") return _client
<commit_before>import os import logging from kaa.base import ipc from client import * from server import * __all__ = [ 'connect' ] # connected client object _client = None def connect(epgdb, logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') if epgdb.find(':') >= 0: # epg is remote: host:port # TODO: create socket, pass it to client _client = GuideClient("epg") else: # epg is local _client = ipc.launch([server, logfile, str(loglevel), epgdb], 2, GuideClient, "epg") return _client <commit_msg>Add the ability to use inet socket as well. git-svn-id: ffaf500d3baede20d2f41eac1d275ef07405e077@1236 a8f5125c-1e01-0410-8897-facf34644b8e<commit_after>import os import logging from socket import gethostbyname, gethostname from kaa.base import ipc from client import * from server import * __all__ = [ 'connect', 'DEFAULT_EPG_PORT', 'GuideClient', 'GuideServer' ] # connected client object _client = None def connect(epgdb, address='127.0.0.1', logfile='/tmp/kaa-epg.log', loglevel=logging.INFO): """ """ global _client if _client: return _client if address.split(':')[0] not in ['127.0.0.1', '0.0.0.0'] and \ address != gethostbyname(gethostname()): # epg is remote: host:port if address.find(':') >= 0: host, port = address.split(':', 1) else: host = address port = DEFAULT_EPG_PORT # create socket, pass it to client _client = GuideClient((host, port)) else: # EPG is local, only use unix socket # get server filename server = os.path.join(os.path.dirname(__file__), 'server.py') _client = ipc.launch([server, logfile, str(loglevel), epgdb, address], 2, GuideClient, "epg") return _client
9d960bfa74a09382839f9b671004bebaffe46611
reui/Screen.py
reui/Screen.py
""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display()
""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() def draw_pixel(self, box, x, y, on=True): bx, by = self._boxMap[box] self._bitmap.draw_pixel(bx + x, by + y, on) def draw_text(self, box, x, y, string, font=box.font): bx, by = self._boxMap[box] self._bitmap.draw_text(bx + x, by + y, string, font)
Support for Box direct drawing to screen bitmap
Support for Box direct drawing to screen bitmap
Python
mit
mharriger/reui
""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() Support for Box direct drawing to screen bitmap
""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() def draw_pixel(self, box, x, y, on=True): bx, by = self._boxMap[box] self._bitmap.draw_pixel(bx + x, by + y, on) def draw_text(self, box, x, y, string, font=box.font): bx, by = self._boxMap[box] self._bitmap.draw_text(bx + x, by + y, string, font)
<commit_before>""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() <commit_msg>Support for Box direct drawing to screen bitmap<commit_after>
""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() def draw_pixel(self, box, x, y, on=True): bx, by = self._boxMap[box] self._bitmap.draw_pixel(bx + x, by + y, on) def draw_text(self, box, x, y, string, font=box.font): bx, by = self._boxMap[box] self._bitmap.draw_text(bx + x, by + y, string, font)
""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() Support for Box direct drawing to screen bitmap""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() def draw_pixel(self, box, x, y, on=True): bx, by = self._boxMap[box] self._bitmap.draw_pixel(bx + x, by + y, on) def draw_text(self, box, x, y, string, font=box.font): bx, by = self._boxMap[box] self._bitmap.draw_text(bx + x, by + y, string, font)
<commit_before>""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() <commit_msg>Support for Box direct drawing to screen bitmap<commit_after>""" A screen object contains a collection of boxes to be displayed on a physical display device. """ from pydispatch import dispatcher from reui import SGL_BOX_UPDATE from gaugette import bitmap class Screen: _boxes = [] _boxMap = {} _bitmap = None def __init__(self, width, height, display): #self._bitmap = bitmap.Bitmap(width, height, 'y') self._display = display #self._display.bitmap = self._bitmap self._bitmap = self._display.bitmap def add_box(self, x, y, box): self._boxes.append((x, y, box)) self._boxMap[box] = (x, y) dispatcher.connect(self.on_box_update, signal=SGL_BOX_UPDATE, sender=box) def draw(self): for x, y, box in self._boxes: self._bitmap.replace_rect(x, y, box._bitmap) def on_box_update(self, **args): if 'sender' in args: sender = args['sender'] if sender in self._boxMap: (x, y) = self._boxMap[sender] self._bitmap.replace_rect(x, y, sender._bitmap) self._display.display() def draw_pixel(self, box, x, y, on=True): bx, by = self._boxMap[box] self._bitmap.draw_pixel(bx + x, by + y, on) def draw_text(self, box, x, y, string, font=box.font): bx, by = self._boxMap[box] self._bitmap.draw_text(bx + x, by + y, string, font)
f087ea792b1e093e6ed49e3dd3b647f2f8276f64
acme/_metadata.py
acme/_metadata.py
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '0' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '1' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
Update Acme version to 0.2.1.
Update Acme version to 0.2.1. PiperOrigin-RevId: 375471102 Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4
Python
apache-2.0
deepmind/acme,deepmind/acme
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '0' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION]) Update Acme version to 0.2.1. PiperOrigin-RevId: 375471102 Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '1' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
<commit_before># python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '0' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION]) <commit_msg>Update Acme version to 0.2.1. PiperOrigin-RevId: 375471102 Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4<commit_after>
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '1' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '0' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION]) Update Acme version to 0.2.1. PiperOrigin-RevId: 375471102 Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4# python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '1' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
<commit_before># python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '0' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION]) <commit_msg>Update Acme version to 0.2.1. PiperOrigin-RevId: 375471102 Change-Id: I9e134bfa61b07059eac564efd515ab788eb1e4f4<commit_after># python3 # Copyright 2018 DeepMind Technologies Limited. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package metadata for acme. This is kept in a separate module so that it can be imported from setup.py, at a time when acme's dependencies may not have been installed yet. """ # We follow Semantic Versioning (https://semver.org/) _MAJOR_VERSION = '0' _MINOR_VERSION = '2' _PATCH_VERSION = '1' # Example: '0.4.2' __version__ = '.'.join([_MAJOR_VERSION, _MINOR_VERSION, _PATCH_VERSION])
0cf0f3de5879795fcd01b8d88bf11efb3362f530
script/echo.py
script/echo.py
#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main()
#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(keepalive=True, post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main()
Make example bot use keepalive
[Instabot] Make example bot use keepalive
Python
mit
CylonicRaider/Instant,CylonicRaider/Instant,CylonicRaider/Instant,CylonicRaider/Instant,CylonicRaider/Instant
#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main() [Instabot] Make example bot use keepalive
#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(keepalive=True, post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main()
<commit_before>#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main() <commit_msg>[Instabot] Make example bot use keepalive<commit_after>
#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(keepalive=True, post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main()
#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main() [Instabot] Make example bot use keepalive#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(keepalive=True, post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main()
<commit_before>#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main() <commit_msg>[Instabot] Make example bot use keepalive<commit_after>#!/usr/bin/env python3 # -*- coding: ascii -*- # A small example bot for Instant. import sys import instabot NICKNAME = 'Echo' def post_cb(self, msg, meta): if msg['text'].startswith('!echo '): return msg['text'][6:] def main(): b = instabot.CmdlineBotBuilder(defnick=NICKNAME) b.make_parser(sys.argv[0], desc='An Instant bot bouncing back received messages.') b.parse(sys.argv[1:]) bot = b(keepalive=True, post_cb=post_cb) try: bot.run() except KeyboardInterrupt: sys.stderr.write('\n') finally: bot.close() if __name__ == '__main__': main()
d90d35063f1a79916c20d32d3634842dd59798f1
api/tests/conftest.py
api/tests/conftest.py
import pytest @pytest.fixture(scope='module') def app(): from api import app return app
import pytest @pytest.fixture(scope='module') def app(): from api import app, db app.config['TESTING'] = True db.create_all() return app
Fix default fixture to initialize database
Fix default fixture to initialize database
Python
mit
Demotivated/loadstone
import pytest @pytest.fixture(scope='module') def app(): from api import app return app Fix default fixture to initialize database
import pytest @pytest.fixture(scope='module') def app(): from api import app, db app.config['TESTING'] = True db.create_all() return app
<commit_before>import pytest @pytest.fixture(scope='module') def app(): from api import app return app <commit_msg>Fix default fixture to initialize database<commit_after>
import pytest @pytest.fixture(scope='module') def app(): from api import app, db app.config['TESTING'] = True db.create_all() return app
import pytest @pytest.fixture(scope='module') def app(): from api import app return app Fix default fixture to initialize databaseimport pytest @pytest.fixture(scope='module') def app(): from api import app, db app.config['TESTING'] = True db.create_all() return app
<commit_before>import pytest @pytest.fixture(scope='module') def app(): from api import app return app <commit_msg>Fix default fixture to initialize database<commit_after>import pytest @pytest.fixture(scope='module') def app(): from api import app, db app.config['TESTING'] = True db.create_all() return app
6decf1f48e56832b1d15d3fc26d92f9813d13353
coop_cms/moves.py
coop_cms/moves.py
# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
Fix HTMLParser compatibility in Python 3
Fix HTMLParser compatibility in Python 3
Python
bsd-3-clause
ljean/coop_cms,ljean/coop_cms,ljean/coop_cms
# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context Fix HTMLParser compatibility in Python 3
# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
<commit_before># -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context <commit_msg>Fix HTMLParser compatibility in Python 3<commit_after>
# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context Fix HTMLParser compatibility in Python 3# -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
<commit_before># -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from HTMLParser import HTMLParser from StringIO import StringIO else: # Python 3 from html.parser import HTMLParser from io import BytesIO as StringIO try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context <commit_msg>Fix HTMLParser compatibility in Python 3<commit_after># -*- coding: utf-8 -*- """ coop_cms manage compatibilty with django and python versions """ import sys from django import VERSION if sys.version_info[0] < 3: # Python 2 from StringIO import StringIO from HTMLParser import HTMLParser else: # Python 3 from io import BytesIO as StringIO from html.parser import HTMLParser as BaseHTMLParser class HTMLParser(BaseHTMLParser): def __init__(self): BaseHTMLParser.__init__(self, convert_charrefs=False) try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object if VERSION >= (1, 9, 0): from wsgiref.util import FileWrapper else: from django.core.servers.basehttp import FileWrapper if VERSION >= (1, 8, 0): from unittest import SkipTest else: # Deprecated in Django 1.9 from django.utils.unittest import SkipTest def make_context(request, context_dict): """""" if VERSION >= (1, 9, 0): context = dict(context_dict) if request: context['request'] = request else: from django.template import RequestContext, Context if request: context = RequestContext(request, context_dict) else: context = Context(context_dict) return context
9931bd1d5459a983717fb502826f3cca87225b96
src/qrl/services/grpcHelper.py
src/qrl/services/grpcHelper.py
# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f
# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: context.set_code(StatusCode.INVALID_ARGUMENT) self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f
Set code to Invalid argument for ValueErrors
Set code to Invalid argument for ValueErrors
Python
mit
jleni/QRL,cyyber/QRL,jleni/QRL,cyyber/QRL,theQRL/QRL,randomshinichi/QRL,theQRL/QRL,randomshinichi/QRL
# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f Set code to Invalid argument for ValueErrors
# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: context.set_code(StatusCode.INVALID_ARGUMENT) self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f
<commit_before># coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f <commit_msg>Set code to Invalid argument for ValueErrors<commit_after>
# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: context.set_code(StatusCode.INVALID_ARGUMENT) self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f
# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f Set code to Invalid argument for ValueErrors# coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: context.set_code(StatusCode.INVALID_ARGUMENT) self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f
<commit_before># coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f <commit_msg>Set code to Invalid argument for ValueErrors<commit_after># coding=utf-8 # Distributed under the MIT software license, see the accompanying # file LICENSE or http://www.opensource.org/licenses/mit-license.php. from grpc import StatusCode from qrl.core.misc import logger class GrpcExceptionWrapper(object): def __init__(self, response_type, state_code=StatusCode.UNKNOWN): self.response_type = response_type self.state_code = state_code def _set_context(self, context, exception): if context is not None: context.set_code(self.state_code) context.set_details(str(exception)) def __call__(self, f): def wrap_f(caller_self, request, context): try: return f(caller_self, request, context) except ValueError as e: context.set_code(StatusCode.INVALID_ARGUMENT) self._set_context(context, e) logger.info(str(e)) return self.response_type() except Exception as e: self._set_context(context, e) logger.exception(e) return self.response_type() return wrap_f
ce1fb05e825e9be7589fd12ab798cae760b605e6
sheldon/bot.py
sheldon/bot.py
# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self): """ Function for loading bot. :return: """ # Creating empty lists for plugins and adapters self.plugins = [] self.adapters = []
# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self, command_line_arguments): """ Function for loading bot. :param command_line_arguments: dict, arguments for start script :return: """ self._load_config(command_line_arguments) def _load_config(self, command_line_arguments): """ Сreate and load bot config. :param command_line_arguments: dict, arguments for creating config: config-prefix - prefix of environment variables. Default - 'SHELDON_' :return: """ # Config class is imported from sheldon.config if 'config-prefix' in command_line_arguments: self.config = Config(prefix=command_line_arguments['config-prefix']) else: self.config = Config()
Add basic load config function
Add basic load config function
Python
mit
lises/sheldon
# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self): """ Function for loading bot. :return: """ # Creating empty lists for plugins and adapters self.plugins = [] self.adapters = [] Add basic load config function
# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self, command_line_arguments): """ Function for loading bot. :param command_line_arguments: dict, arguments for start script :return: """ self._load_config(command_line_arguments) def _load_config(self, command_line_arguments): """ Сreate and load bot config. :param command_line_arguments: dict, arguments for creating config: config-prefix - prefix of environment variables. Default - 'SHELDON_' :return: """ # Config class is imported from sheldon.config if 'config-prefix' in command_line_arguments: self.config = Config(prefix=command_line_arguments['config-prefix']) else: self.config = Config()
<commit_before># -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self): """ Function for loading bot. :return: """ # Creating empty lists for plugins and adapters self.plugins = [] self.adapters = [] <commit_msg>Add basic load config function<commit_after>
# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self, command_line_arguments): """ Function for loading bot. :param command_line_arguments: dict, arguments for start script :return: """ self._load_config(command_line_arguments) def _load_config(self, command_line_arguments): """ Сreate and load bot config. :param command_line_arguments: dict, arguments for creating config: config-prefix - prefix of environment variables. Default - 'SHELDON_' :return: """ # Config class is imported from sheldon.config if 'config-prefix' in command_line_arguments: self.config = Config(prefix=command_line_arguments['config-prefix']) else: self.config = Config()
# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self): """ Function for loading bot. :return: """ # Creating empty lists for plugins and adapters self.plugins = [] self.adapters = [] Add basic load config function# -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self, command_line_arguments): """ Function for loading bot. :param command_line_arguments: dict, arguments for start script :return: """ self._load_config(command_line_arguments) def _load_config(self, command_line_arguments): """ Сreate and load bot config. :param command_line_arguments: dict, arguments for creating config: config-prefix - prefix of environment variables. Default - 'SHELDON_' :return: """ # Config class is imported from sheldon.config if 'config-prefix' in command_line_arguments: self.config = Config(prefix=command_line_arguments['config-prefix']) else: self.config = Config()
<commit_before># -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self): """ Function for loading bot. :return: """ # Creating empty lists for plugins and adapters self.plugins = [] self.adapters = [] <commit_msg>Add basic load config function<commit_after># -*- coding: utf-8 -*- """ @author: Seva Zhidkov @contact: zhidkovseva@gmail.com @license: The MIT license Copyright (C) 2015 """ from sheldon.exceptions import * from sheldon.manager import * from sheldon.config import * from sheldon.adapter import * from sheldon.storage import * class Sheldon: """ Main class of the bot. Run script creating new instance of this class and run it. """ def __init__(self, command_line_arguments): """ Function for loading bot. :param command_line_arguments: dict, arguments for start script :return: """ self._load_config(command_line_arguments) def _load_config(self, command_line_arguments): """ Сreate and load bot config. :param command_line_arguments: dict, arguments for creating config: config-prefix - prefix of environment variables. Default - 'SHELDON_' :return: """ # Config class is imported from sheldon.config if 'config-prefix' in command_line_arguments: self.config = Config(prefix=command_line_arguments['config-prefix']) else: self.config = Config()
74f82029223cc541beab98d7026abb1ec992be40
createTodoFile.py
createTodoFile.py
"""createTodoFile.py: Creates an todo file with title name as current date""" import time import os.path def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") else: print("ERROR: " + filename + " already exist! Exiting..") # TODO: To move files into archive if more than a week def archiveFiles(): pass def main(): createfile() if __name__ == '__main__': main()
"""createTodoFile.py: Creates an todo file with title name as current date""" import os.path import time def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") addfileToSublime(filename) else: print("ERROR: " + filename + " already exist! Exiting..") def addfileToSublime(file): os.system("subl --add " + file) def main(): createfile() if __name__ == '__main__': main()
Add created file to sublime
feat: Add created file to sublime
Python
mit
prajesh-ananthan/Tools
"""createTodoFile.py: Creates an todo file with title name as current date""" import time import os.path def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") else: print("ERROR: " + filename + " already exist! Exiting..") # TODO: To move files into archive if more than a week def archiveFiles(): pass def main(): createfile() if __name__ == '__main__': main() feat: Add created file to sublime
"""createTodoFile.py: Creates an todo file with title name as current date""" import os.path import time def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") addfileToSublime(filename) else: print("ERROR: " + filename + " already exist! Exiting..") def addfileToSublime(file): os.system("subl --add " + file) def main(): createfile() if __name__ == '__main__': main()
<commit_before>"""createTodoFile.py: Creates an todo file with title name as current date""" import time import os.path def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") else: print("ERROR: " + filename + " already exist! Exiting..") # TODO: To move files into archive if more than a week def archiveFiles(): pass def main(): createfile() if __name__ == '__main__': main() <commit_msg>feat: Add created file to sublime<commit_after>
"""createTodoFile.py: Creates an todo file with title name as current date""" import os.path import time def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") addfileToSublime(filename) else: print("ERROR: " + filename + " already exist! Exiting..") def addfileToSublime(file): os.system("subl --add " + file) def main(): createfile() if __name__ == '__main__': main()
"""createTodoFile.py: Creates an todo file with title name as current date""" import time import os.path def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") else: print("ERROR: " + filename + " already exist! Exiting..") # TODO: To move files into archive if more than a week def archiveFiles(): pass def main(): createfile() if __name__ == '__main__': main() feat: Add created file to sublime"""createTodoFile.py: Creates an todo file with title name as current date""" import os.path import time def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") addfileToSublime(filename) else: print("ERROR: " + filename + " already exist! Exiting..") def addfileToSublime(file): os.system("subl --add " + file) def main(): createfile() if __name__ == '__main__': main()
<commit_before>"""createTodoFile.py: Creates an todo file with title name as current date""" import time import os.path def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") else: print("ERROR: " + filename + " already exist! Exiting..") # TODO: To move files into archive if more than a week def archiveFiles(): pass def main(): createfile() if __name__ == '__main__': main() <commit_msg>feat: Add created file to sublime<commit_after>"""createTodoFile.py: Creates an todo file with title name as current date""" import os.path import time def createfile(): # My-File--2009-12-31--23-59-59.txt date = time.strftime("%d-%m-%Y") filename = "GOALS--" + date + ".todo" if not os.path.exists(filename): with open(filename, "a") as myfile: myfile.write("[RESULTS - {}]".format(date)) print("INFO: " + filename + " created!") addfileToSublime(filename) else: print("ERROR: " + filename + " already exist! Exiting..") def addfileToSublime(file): os.system("subl --add " + file) def main(): createfile() if __name__ == '__main__': main()
0f114a144268bb611ff00db9917756a8c02f84b9
project/api/signals.py
project/api/signals.py
# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff and not instance.person: activate_user(instance) return
# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import activate_user from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff and not instance.person: activate_user(instance) return
Connect person to user account
Connect person to user account
Python
bsd-2-clause
barberscore/barberscore-api,dbinetti/barberscore-django,barberscore/barberscore-api,barberscore/barberscore-api,dbinetti/barberscore-django,dbinetti/barberscore,dbinetti/barberscore,barberscore/barberscore-api
# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff and not instance.person: activate_user(instance) return Connect person to user account
# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import activate_user from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff and not instance.person: activate_user(instance) return
<commit_before># Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff and not instance.person: activate_user(instance) return <commit_msg>Connect person to user account<commit_after>
# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import activate_user from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff and not instance.person: activate_user(instance) return
# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff and not instance.person: activate_user(instance) return Connect person to user account# Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import activate_user from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff and not instance.person: activate_user(instance) return
<commit_before># Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): if settings.DJANGO_SETTINGS_MODULE == 'settings.prod': if not instance.is_staff and not instance.person: activate_user(instance) return <commit_msg>Connect person to user account<commit_after># Django # Third-Party from django.db.models.signals import pre_delete from django.db.models.signals import pre_save from django.dispatch import receiver from django.conf import settings # Local from .models import User from .tasks import activate_user from .tasks import delete_account @receiver(pre_delete, sender=User) def user_pre_delete(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff: delete_account(instance) return @receiver(pre_save, sender=User) def user_pre_save(sender, instance, **kwargs): allowed = any([ settings.DJANGO_SETTINGS_MODULE == 'settings.prod', settings.DJANGO_SETTINGS_MODULE == 'settings.dev', ]) if allowed: if not instance.is_staff and not instance.person: activate_user(instance) return
9568efceab48f87ed8302ec4f9bad4b15aac4c5a
tests/test_action.py
tests/test_action.py
import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com")
import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") def test_connection_closed_after_sending_mail(self, mock_smtp_class): mock_smtp = mock_smtp_class.return_value self.action.execute("MSFT has crossed $10 price level") mock_smtp.send_message.assert_called_with(mock.ANY) self.assertTrue(mock_smtp.quit.called) mock_smtp.assert_has_calls([ mock.call.send_message(mock.ANY), mock.call.quit() ])
Add test to check if connection is closed after email is sent.
Add test to check if connection is closed after email is sent.
Python
mit
bsmukasa/stock_alerter
import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") Add test to check if connection is closed after email is sent.
import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") def test_connection_closed_after_sending_mail(self, mock_smtp_class): mock_smtp = mock_smtp_class.return_value self.action.execute("MSFT has crossed $10 price level") mock_smtp.send_message.assert_called_with(mock.ANY) self.assertTrue(mock_smtp.quit.called) mock_smtp.assert_has_calls([ mock.call.send_message(mock.ANY), mock.call.quit() ])
<commit_before>import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") <commit_msg>Add test to check if connection is closed after email is sent.<commit_after>
import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") def test_connection_closed_after_sending_mail(self, mock_smtp_class): mock_smtp = mock_smtp_class.return_value self.action.execute("MSFT has crossed $10 price level") mock_smtp.send_message.assert_called_with(mock.ANY) self.assertTrue(mock_smtp.quit.called) mock_smtp.assert_has_calls([ mock.call.send_message(mock.ANY), mock.call.quit() ])
import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") Add test to check if connection is closed after email is sent.import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") def test_connection_closed_after_sending_mail(self, mock_smtp_class): mock_smtp = mock_smtp_class.return_value self.action.execute("MSFT has crossed $10 price level") mock_smtp.send_message.assert_called_with(mock.ANY) self.assertTrue(mock_smtp.quit.called) mock_smtp.assert_has_calls([ mock.call.send_message(mock.ANY), mock.call.quit() ])
<commit_before>import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") <commit_msg>Add test to check if connection is closed after email is sent.<commit_after>import smtplib import unittest from unittest import mock from action import PrintAction, EmailAction @mock.patch("builtins.print") class PrintActionTest(unittest.TestCase): def test_executing_action_prints_message(self, mock_print): action = PrintAction() action.execute("GOOG > $10") mock_print.assert_called_with("GOOG > $10") @mock.patch("smtplib.SMTP") class EmailActionTest(unittest.TestCase): def setUp(self): self.action = EmailAction(to="bsmukasa@gmail.com") def test_email_is_sent_to_the_right_server(self, mock_smtp_class): self.action.execute("MSFT has crossed $10 price level") mock_smtp_class.assert_called_with("email.stocks.com") def test_connection_closed_after_sending_mail(self, mock_smtp_class): mock_smtp = mock_smtp_class.return_value self.action.execute("MSFT has crossed $10 price level") mock_smtp.send_message.assert_called_with(mock.ANY) self.assertTrue(mock_smtp.quit.called) mock_smtp.assert_has_calls([ mock.call.send_message(mock.ANY), mock.call.quit() ])
5bc4aa60be988abc98ba76ca4b790b259d75af37
capstone/rl/policies/egreedy.py
capstone/rl/policies/egreedy.py
import random from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(vf, state, actions)
from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(state, actions, vf)
Fix EGreedy policy call order
Fix EGreedy policy call order
Python
mit
davidrobles/mlnd-capstone-code
import random from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(vf, state, actions) Fix EGreedy policy call order
from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(state, actions, vf)
<commit_before>import random from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(vf, state, actions) <commit_msg>Fix EGreedy policy call order<commit_after>
from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(state, actions, vf)
import random from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(vf, state, actions) Fix EGreedy policy call orderfrom .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(state, actions, vf)
<commit_before>import random from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(vf, state, actions) <commit_msg>Fix EGreedy policy call order<commit_after>from .greedy import Greedy from .random_policy import RandomPolicy from ..policy import Policy from ...utils import check_random_state class EGreedy(Policy): def __init__(self, e, random_state=None): self.e = e self.greedy = Greedy() self.rand = RandomPolicy() self.random_state = check_random_state(random_state) def action(self, state, actions=None, vf=None): policy = self.rand if self.random_state.rand() < self.e else self.greedy return policy.action(state, actions, vf)
547787272a6945bfefd086504e4c3dcaf483bc37
tests/test_logger.py
tests/test_logger.py
"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble'
"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble'
Remove one blank to be friends with the lama.
Remove one blank to be friends with the lama.
Python
mit
openspending/gobble
"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble' Remove one blank to be friends with the lama.
"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble'
<commit_before>"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble' <commit_msg>Remove one blank to be friends with the lama.<commit_after>
"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble'
"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble' Remove one blank to be friends with the lama."""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble'
<commit_before>"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble' <commit_msg>Remove one blank to be friends with the lama.<commit_after>"""Test the logger module""" from gobble.logger import log def test_gobble_logger_exists(): assert log.name == 'Gobble'
7be79e544eecf158a6ff1bde2f9f76f5145e4ae1
tests/tools_tests.py
tests/tools_tests.py
"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0'
"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from ifcfg.tools import exec_cmd from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' def test_command(self): output, __, __ = exec_cmd("echo -n 'this is a test'") self.assertEqual(output, "this is a test")
Add a test that calls exec_cmd
Add a test that calls exec_cmd
Python
bsd-3-clause
ftao/python-ifcfg
"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' Add a test that calls exec_cmd
"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from ifcfg.tools import exec_cmd from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' def test_command(self): output, __, __ = exec_cmd("echo -n 'this is a test'") self.assertEqual(output, "this is a test")
<commit_before>"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' <commit_msg>Add a test that calls exec_cmd<commit_after>
"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from ifcfg.tools import exec_cmd from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' def test_command(self): output, __, __ = exec_cmd("echo -n 'this is a test'") self.assertEqual(output, "this is a test")
"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' Add a test that calls exec_cmd"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from ifcfg.tools import exec_cmd from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' def test_command(self): output, __, __ = exec_cmd("echo -n 'this is a test'") self.assertEqual(output, "this is a test")
<commit_before>"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' <commit_msg>Add a test that calls exec_cmd<commit_after>"""Tests for ifcfg.tools.""" import logging import os import unittest import ifcfg from ifcfg.tools import exec_cmd from nose.tools import eq_ class IfcfgToolsTestCase(unittest.TestCase): def test_minimal_logger(self): os.environ['IFCFG_DEBUG'] = '1' log = ifcfg.tools.minimal_logger(__name__) eq_(log.level, logging.DEBUG) os.environ['IFCFG_DEBUG'] = '0' def test_command(self): output, __, __ = exec_cmd("echo -n 'this is a test'") self.assertEqual(output, "this is a test")
30dbda17bfa3b52dc2aace6eba6b8c1e4b3f7542
robot-name/robot_name.py
robot-name/robot_name.py
# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = ""
# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" def factory_name(self): char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2)) num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3)) self.name = char + num return self.name R1 = Robot() print(R1.factory_name())
Add methord to generate unique robot name
Add methord to generate unique robot name
Python
mit
amalshehu/exercism-python
# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" Add methord to generate unique robot name
# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" def factory_name(self): char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2)) num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3)) self.name = char + num return self.name R1 = Robot() print(R1.factory_name())
<commit_before># File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" <commit_msg>Add methord to generate unique robot name<commit_after>
# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" def factory_name(self): char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2)) num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3)) self.name = char + num return self.name R1 = Robot() print(R1.factory_name())
# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" Add methord to generate unique robot name# File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" def factory_name(self): char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2)) num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3)) self.name = char + num return self.name R1 = Robot() print(R1.factory_name())
<commit_before># File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" <commit_msg>Add methord to generate unique robot name<commit_after># File: robot_name.py # Purpose: Write a program that manages robot factory settings. # Programmer: Amal Shehu # Course: Exercism # Date: Friday 30 September 2016, 03:00 PM import string import random class Robot(): """Robot facory settings""" def __init__(self): self.name = "" def factory_name(self): char = ''.join(random.SystemRandom().choice(string.ascii_uppercase) for _ in range(2)) num = ''.join(random.SystemRandom().choice(string.digits) for _ in range(3)) self.name = char + num return self.name R1 = Robot() print(R1.factory_name())
3134af98d2fcf88752170d628400a7e863d4c959
was/artists/views.py
was/artists/views.py
from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'register.html' form_class = UpdateArtistForm success_url = '/' def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')
from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists, User from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'update.html' form_class = UpdateArtistForm success_url = '/' def get_initial(self): initial = {} user = User.objects.get(username=self.request.user.username) initial['username'] = user.username initial['email'] = user.email return initial def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')
Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model.
Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model.
Python
mit
KeserOner/where-artists-share,KeserOner/where-artists-share
from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'register.html' form_class = UpdateArtistForm success_url = '/' def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model.
from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists, User from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'update.html' form_class = UpdateArtistForm success_url = '/' def get_initial(self): initial = {} user = User.objects.get(username=self.request.user.username) initial['username'] = user.username initial['email'] = user.email return initial def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')
<commit_before>from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'register.html' form_class = UpdateArtistForm success_url = '/' def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')<commit_msg>Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model.<commit_after>
from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists, User from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'update.html' form_class = UpdateArtistForm success_url = '/' def get_initial(self): initial = {} user = User.objects.get(username=self.request.user.username) initial['username'] = user.username initial['email'] = user.email return initial def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')
from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'register.html' form_class = UpdateArtistForm success_url = '/' def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model.from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists, User from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'update.html' form_class = UpdateArtistForm success_url = '/' def get_initial(self): initial = {} user = User.objects.get(username=self.request.user.username) initial['username'] = user.username initial['email'] = user.email return initial def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')
<commit_before>from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'register.html' form_class = UpdateArtistForm success_url = '/' def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')<commit_msg>Create update view (extends UpdateView generic view). Define get_initail in order to pre populate two custom fields username and email which are not in the original Artists model.<commit_after>from django.shortcuts import render, get_object_or_404 from django.views.generic.edit import CreateView, UpdateView from .form import CreateArtistForm, UpdateArtistForm, Artists, User from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth import login, logout from django.http import HttpResponseRedirect class CreateArtistView(CreateView): template_name = 'register.html' form_class = CreateArtistForm success_url = '/' class UpdateArtistView(UpdateView): template_name = 'update.html' form_class = UpdateArtistForm success_url = '/' def get_initial(self): initial = {} user = User.objects.get(username=self.request.user.username) initial['username'] = user.username initial['email'] = user.email return initial def get_object(self): return get_object_or_404(Artists, user=self.request.user) def artist_login(request): if request.method == 'POST': form = AuthenticationForm(data=request.POST) if form.is_valid(): form.clean() login(request, form.user_cache) return HttpResponseRedirect('/') else: form = AuthenticationForm() return render(request, 'login.html', {'form' : form}) def artist_logout(request): logout(request) return HttpResponseRedirect('/')
533d1462949ab451674d91dd7730957cb2252dde
susumutakuan.py
susumutakuan.py
import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"]) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)
import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"], universal_newlines=True) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)
Add universal_newlines paramter to run call
Add universal_newlines paramter to run call
Python
mit
gryffon/SusumuTakuan,gryffon/SusumuTakuan
import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"]) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)Add universal_newlines paramter to run call
import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"], universal_newlines=True) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)
<commit_before>import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"]) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)<commit_msg>Add universal_newlines paramter to run call<commit_after>
import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"], universal_newlines=True) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)
import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"]) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)Add universal_newlines paramter to run callimport discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"], universal_newlines=True) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)
<commit_before>import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"]) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)<commit_msg>Add universal_newlines paramter to run call<commit_after>import discord import asyncio import os import signal import sys from subprocess import run #Set up Client State CLIENT_TOKEN=os.environ['TOKEN'] #Create Discord client client = discord.Client() #Handle shutdown gracefully def sigterm_handler(signum, frame): print('Logging out...', flush=True) raise KeyboardInterrupt print('Shutting down...') sys.exit(0) #Register SIGTERM Handler signal.signal(signal.SIGTERM, sigterm_handler) @client.event async def on_ready(): print('Logged in as') print(client.user.name) print(client.user.id) print('------') @client.event async def on_message(message): #Look at DMs for special commands if message.channel.type == discord.ChannelType.private: if message.content.startswith('!update'): tmp = await client.send_message(message.channel, 'Updating my code via git...') process = run(["sh", "control.sh", "refresh"], universal_newlines=True) tmp = await client.send_message(message.channel, process.stdout) if message.content.startswith('!test'): counter = 0 tmp = await client.send_message(message.channel, 'Calculating messages...') async for log in client.logs_from(message.channel, limit=100): if log.author == message.author: counter += 1 await client.edit_message(tmp, 'You have {} messages.'.format(counter)) #Start event loop client.run(CLIENT_TOKEN)
f9835741804da062f1501b06560a2af75b199645
scrapeOMDB.py
scrapeOMDB.py
#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON)
#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON)
Convert movie year to str
Convert movie year to str
Python
mit
samcheck/PyMedia,samcheck/PyMedia,samcheck/PyMedia
#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) Convert movie year to str
#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON)
<commit_before>#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) <commit_msg>Convert movie year to str<commit_after>
#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON)
#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) Convert movie year to str#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON)
<commit_before>#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + mYear + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) <commit_msg>Convert movie year to str<commit_after>#!/usr/bin/python3 # scrapeOMDB.py - parses a movie and year from arguments and returns JSON import json, requests URL_BASE = 'http://www.omdbapi.com/?' def OMDBmovie(mTitle, mYear): # Craft the URL url = URL_BASE + 't=' + mTitle + '&y=' + str(mYear) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON) def OMDBtv(tvTitle, tvSeason, tvEpisode): # Craft the URL url = URL_BASE + 't=' + tvTitle + '&Season=' + str(tvSeason) + '&Episode=' + str(tvEpisode) + '&plot=full&r=json' # Try to get the url response = requests.get(url) response.raise_for_status() theJSON = json.loads(response.text) return(theJSON)
1d14d28d68278330855e585a859484019d8c3e43
cacivicdata/manage.py
cacivicdata/manage.py
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings") sys.path.append(os.path.dirname(os.path.dirname(__file__))) from django.core.management import execute_from_command_line execute_from_command_line(sys.argv)
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings") from django.core.management import execute_from_command_line execute_from_command_line(sys.argv)
Change back to django default
Change back to django default
Python
mit
california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings") sys.path.append(os.path.dirname(os.path.dirname(__file__))) from django.core.management import execute_from_command_line execute_from_command_line(sys.argv) Change back to django default
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings") from django.core.management import execute_from_command_line execute_from_command_line(sys.argv)
<commit_before>#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings") sys.path.append(os.path.dirname(os.path.dirname(__file__))) from django.core.management import execute_from_command_line execute_from_command_line(sys.argv) <commit_msg>Change back to django default<commit_after>
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings") from django.core.management import execute_from_command_line execute_from_command_line(sys.argv)
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings") sys.path.append(os.path.dirname(os.path.dirname(__file__))) from django.core.management import execute_from_command_line execute_from_command_line(sys.argv) Change back to django default#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings") from django.core.management import execute_from_command_line execute_from_command_line(sys.argv)
<commit_before>#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings") sys.path.append(os.path.dirname(os.path.dirname(__file__))) from django.core.management import execute_from_command_line execute_from_command_line(sys.argv) <commit_msg>Change back to django default<commit_after>#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "cacivicdata.settings") from django.core.management import execute_from_command_line execute_from_command_line(sys.argv)
5632daecf9c5f271eeba0f9948d88f44d6a070d0
irclogview/models.py
irclogview/models.py
from django.db import models from picklefield.fields import PickledObjectField class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date')
from django.db import models from picklefield.fields import PickledObjectField from . import utils class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') def content_dict(self): colors = utils.RainbowColor() for data in self.content: item = dict(zip(['time', 'type', 'name', 'text'], data)) item['name_color'] = colors.get_color(item['name']) yield item
Add function to get content in list of dicts format
Add function to get content in list of dicts format
Python
agpl-3.0
BlankOn/irclogview,fajran/irclogview,fajran/irclogview,BlankOn/irclogview
from django.db import models from picklefield.fields import PickledObjectField class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') Add function to get content in list of dicts format
from django.db import models from picklefield.fields import PickledObjectField from . import utils class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') def content_dict(self): colors = utils.RainbowColor() for data in self.content: item = dict(zip(['time', 'type', 'name', 'text'], data)) item['name_color'] = colors.get_color(item['name']) yield item
<commit_before>from django.db import models from picklefield.fields import PickledObjectField class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') <commit_msg>Add function to get content in list of dicts format<commit_after>
from django.db import models from picklefield.fields import PickledObjectField from . import utils class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') def content_dict(self): colors = utils.RainbowColor() for data in self.content: item = dict(zip(['time', 'type', 'name', 'text'], data)) item['name_color'] = colors.get_color(item['name']) yield item
from django.db import models from picklefield.fields import PickledObjectField class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') Add function to get content in list of dicts formatfrom django.db import models from picklefield.fields import PickledObjectField from . import utils class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') def content_dict(self): colors = utils.RainbowColor() for data in self.content: item = dict(zip(['time', 'type', 'name', 'text'], data)) item['name_color'] = colors.get_color(item['name']) yield item
<commit_before>from django.db import models from picklefield.fields import PickledObjectField class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') <commit_msg>Add function to get content in list of dicts format<commit_after>from django.db import models from picklefield.fields import PickledObjectField from . import utils class Channel(models.Model): name = models.SlugField(max_length=50, unique=True) updated = models.DateTimeField(auto_now=True) def __unicode__(self): return u'#%s' % self.name class Log(models.Model): channel = models.ForeignKey(Channel) date = models.DateField() mtime = models.DateTimeField() updated = models.DateTimeField(auto_now=True) content = PickledObjectField() class Meta: unique_together = ('channel', 'date') def content_dict(self): colors = utils.RainbowColor() for data in self.content: item = dict(zip(['time', 'type', 'name', 'text'], data)) item['name_color'] = colors.get_color(item['name']) yield item
4228082c9c94b3e17e6b00fc1e380841d5389dc5
crawler/models.py
crawler/models.py
from django.db import models # Create your models here. class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500)
from django.db import models class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500)
Remove not needed comment line
Remove not needed comment line
Python
mit
lucasgr7/silverplate,lucasgr7/silverplate,lucasgr7/silverplate
from django.db import models # Create your models here. class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500) Remove not needed comment line
from django.db import models class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500)
<commit_before>from django.db import models # Create your models here. class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500) <commit_msg>Remove not needed comment line<commit_after>
from django.db import models class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500)
from django.db import models # Create your models here. class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500) Remove not needed comment linefrom django.db import models class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500)
<commit_before>from django.db import models # Create your models here. class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500) <commit_msg>Remove not needed comment line<commit_after>from django.db import models class Data_Ingredient(models.Model): """"Class used to Store Ingredients of the recipes found in the crawling process""" Ingredient = models.CharField(max_length=1000) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Ingredientes') def __str__(self): return self.Ingredient class Data_Way_Cooking(models.Model): """Class used to Store steps of the recipes found in the crawling process""" Description = models.CharField(max_length=500) Recipe = models.CharField(max_length=500) Group = models.CharField(max_length=500, default='Modo de Fazer') def __str__(self): return self.Description class Ingredient_Spec(models.Model): """Class used to manipulate Ingredients found and change data to data mining and found patterns of ingredients""" Word = models.CharField(max_length=500) Count = models.IntegerField() Type = models.CharField(max_length=1) class Ignore_Words(models.Model): """Model to store words to ignore from Ingredient Spec""" Word = models.CharField(max_length=500)
1d77647efdb26b8282fc0624852d211fa9339539
db/TableConfig.py
db/TableConfig.py
{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'1.0'"} ] }
{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'4.0'"} ] }
Introduce family id for all
Introduce family id for all
Python
mit
eddiedb6/ej,eddiedb6/ej,eddiedb6/ej
{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'1.0'"} ] } Introduce family id for all
{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'4.0'"} ] }
<commit_before>{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'1.0'"} ] } <commit_msg>Introduce family id for all<commit_after>
{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'4.0'"} ] }
{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'1.0'"} ] } Introduce family id for all{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'4.0'"} ] }
<commit_before>{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'1.0'"} ] } <commit_msg>Introduce family id for all<commit_after>{ PDBConst.Name: "config", PDBConst.Columns: [ { PDBConst.Name: "Name", PDBConst.Attributes: ["varchar(128)", "not null", "primary key"] }, { PDBConst.Name: "Value", PDBConst.Attributes: ["varchar(128)"] }], PDBConst.Initials: [ {"Name": "'version'", "Value": "'4.0'"} ] }
22b6785695967a43ab9d187db60c201c3dc4a8e1
peerinst/admin.py
peerinst/admin.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions']
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL} @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions']
Use radio buttons for the style and number of answers.
Use radio buttons for the style and number of answers.
Python
agpl-3.0
open-craft/dalite-ng,open-craft/dalite-ng,open-craft/dalite-ng
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions'] Use radio buttons for the style and number of answers.
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL} @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions']
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions'] <commit_msg>Use radio buttons for the style and number of answers.<commit_after>
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL} @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions']
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions'] Use radio buttons for the style and number of answers.# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL} @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions']
<commit_before># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions'] <commit_msg>Use radio buttons for the style and number of answers.<commit_after># -*- coding: utf-8 -*- from __future__ import unicode_literals from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from . import models @admin.register(models.Question) class QuestionAdmin(admin.ModelAdmin): fieldsets = [ (None, {'fields': ['title']}), (_('Main image or video'), {'fields': ['primary_image', 'primary_video_url']}), (_('Secondary image or video'), { 'fields': ['secondary_image', 'secondary_video_url'], 'classes': ['collapse'], 'description': _( 'Choose either a video or image to include on the first page of the question, ' 'where students select concept tags. This is only used if you want the question ' 'to be hidden when students select concept tags; instead, a preliminary video or ' 'image can be displayed. The main question image will be displayed on all ' 'subsequent pages.' ), }), (_('Answers'), {'fields': [ 'answer_style', 'answer_num_choices', 'correct_answer', 'second_best_answer' ]}), (None, {'fields': ['example_rationale']}), ] radio_fields = {'answer_style': admin.HORIZONTAL, 'answer_num_choices': admin.HORIZONTAL} @admin.register(models.Assignment) class AssignmentAdmin(admin.ModelAdmin): filter_horizontal = ['questions']
b35ee625143fd57f5571f807d0cd4331be4e0947
caprice/models.py
caprice/models.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close()
#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return "<{0}: '{1}'>".format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close()
Use compatible(for Python2.6) format string.
Use compatible(for Python2.6) format string.
Python
mit
FGtatsuro/Caprice,FGtatsuro/Caprice,FGtatsuro/Caprice
#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close() Use compatible(for Python2.6) format string.
#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return "<{0}: '{1}'>".format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close() <commit_msg>Use compatible(for Python2.6) format string.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return "<{0}: '{1}'>".format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close()
#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close() Use compatible(for Python2.6) format string.#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return "<{0}: '{1}'>".format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return '<{!s}: {!r}>'.format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close() <commit_msg>Use compatible(for Python2.6) format string.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from sqlalchemy import Column, String from .db import Base, Session __all__ = ['Schema'] class Schema(Base): __tablename__ = 'schemas' # TODO: allow Only UUID? or user defined ID too? id = Column(String, primary_key=True) # TODO: JSON uniqueness is needed body = Column(String) # ID is generated in Python context(=in application) def __init__(self, id=None, body=None): self.id = id self.body = body def __repr__(self): return "<{0}: '{1}'>".format(self.__class__.__name__, self.body) # TODO: Use contextmanager. Ref. http://docs.sqlalchemy.org/en/rel_1_0/orm/session_basics.html def save(self): s = Session() s.add(self) try: # TODO: logger s.commit() except: s.rollback() # TODO: Error message raise RuntimeError('') finally: # TODO: logger s.close()
8222bf717b92ab57b60b834b4496afb654b8a80b
krisk/connections.py
krisk/connections.py
from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", paths: { echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min" }, waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES
from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES
Change all js to point to krisk repo
Change all js to point to krisk repo
Python
bsd-3-clause
napjon/krisk,napjon/krisk,napjon/krisk
from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", paths: { echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min" }, waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES Change all js to point to krisk repo
from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES
<commit_before> from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", paths: { echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min" }, waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES <commit_msg>Change all js to point to krisk repo<commit_after>
from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES
from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", paths: { echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min" }, waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES Change all js to point to krisk repo from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES
<commit_before> from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", paths: { echarts: "//cdnjs.cloudflare.com/ajax/libs/echarts/3.2.1/echarts.min" }, waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES <commit_msg>Change all js to point to krisk repo<commit_after> from collections import OrderedDict from IPython.display import Javascript import json from krisk.util import join_current_dir ECHARTS_URL = 'https://cdnjs.cloudflare.com/ajax/libs/echarts/3.2.0/' ECHARTS_FILE = 'echarts.min' d_paths = OrderedDict({}) THEMES = ['dark','vintage','roma','shine','infographic','macarons'] THEMES_URL='//echarts.baidu.com/asset/theme/' PATH_LOCAL = join_current_dir('static') # PATH_LOCAL = 'pandas-echarts/krisk/static' #TODO FIX LOCAL PATH! NEED TO DO nbextension install # def init_notebook(): # """Inject Javascript to notebook, default using local js. # """ # return Javascript(""" # require.config({ # baseUrl : '%s', # paths: { # echarts: 'echarts.min' # } # }); # """ % PATH_LOCAL) def init_notebook(): """Inject Javascript to notebook, default using local js. """ return Javascript(""" require.config({ baseUrl : "//rawgit.com/napjon/krisk/master/krisk/static", waitSeconds: 15 }); """) def get_paths(): return ['echarts'] + THEMES
e299c07034e0ad1135bda999ad0c63f4b5a7fa40
chaco/__init__.py
chaco/__init__.py
# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0dev' __requires__ = [ 'enable', ]
# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0.dev' __requires__ = [ 'enable', ]
Tweak the version number to match other ETS projects.
Tweak the version number to match other ETS projects.
Python
bsd-3-clause
burnpanck/chaco,tommy-u/chaco,tommy-u/chaco,tommy-u/chaco,burnpanck/chaco,burnpanck/chaco
# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0dev' __requires__ = [ 'enable', ] Tweak the version number to match other ETS projects.
# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0.dev' __requires__ = [ 'enable', ]
<commit_before># Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0dev' __requires__ = [ 'enable', ] <commit_msg>Tweak the version number to match other ETS projects.<commit_after>
# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0.dev' __requires__ = [ 'enable', ]
# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0dev' __requires__ = [ 'enable', ] Tweak the version number to match other ETS projects.# Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0.dev' __requires__ = [ 'enable', ]
<commit_before># Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0dev' __requires__ = [ 'enable', ] <commit_msg>Tweak the version number to match other ETS projects.<commit_after># Copyright (c) 2005-2013 by Enthought, Inc. # All rights reserved. """ Two-dimensional plotting application toolkit. Part of the Chaco project of the Enthought Tool Suite. """ __version__ = '4.5.0.dev' __requires__ = [ 'enable', ]
0f53ec6ddeb236bee78794e8d1ed156ad182bc07
projects/forms.py
projects/forms.py
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'status') class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
Add status to project form
Add status to project form
Python
mit
Hackfmi/Diaphanum,Hackfmi/Diaphanum
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', ) Add status to project form
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'status') class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
<commit_before>from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', ) <commit_msg>Add status to project form<commit_after>
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'status') class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', ) Add status to project formfrom django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'status') class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
<commit_before>from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners',) class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', ) <commit_msg>Add status to project form<commit_after>from django import forms from .models import Project class ProjectForm(forms.ModelForm): def __init__(self, *args, **kwargs): self.user = kwargs.pop('user') super(ProjectForm, self).__init__(*args, **kwargs) def save(self, *args, **kwargs): instance = super(ProjectForm, self).save(commit=False) instance.user = self.user instance.flp = self.user instance.status = 'unrevised' instance.save(*args, **kwargs) self.save_m2m() return instance class Meta: model = Project fields = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'status') class RestrictedProjectForm(forms.ModelForm): def save(self, *args, **kwargs): instance = super(RestrictedProjectForm, self).save(commit=False) return instance class Meta: model = Project exclude = ( 'name', 'team', 'description', 'targets', 'tasks', 'target_group', 'schedule', 'resources', 'finance_description', 'partners', 'flp', 'created_at', 'user', ) fileds = ( 'status', 'attitude', )
84e9532487615f684abbed17d6821ae7bc84c9be
virtualfish/loader/__init__.py
virtualfish/loader/__init__.py
from __future__ import print_function import os import sys import pkg_resources def load(plugins=()): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) commands.append("emit virtualfish_did_setup_plugins") return commands
from __future__ import print_function import os import sys import pkg_resources def load(plugins=(), full_install=True): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) if full_install: commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] else: commands = [] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) if full_install: commands.append("emit virtualfish_did_setup_plugins") return commands
Add kwarg to load function to distinguish from full install
Add kwarg to load function to distinguish from full install The load function is used for a full install and thus always adds general configuration lines to the loader file, but we don't want that for plugin installation.
Python
mit
adambrenecki/virtualfish,adambrenecki/virtualfish
from __future__ import print_function import os import sys import pkg_resources def load(plugins=()): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) commands.append("emit virtualfish_did_setup_plugins") return commands Add kwarg to load function to distinguish from full install The load function is used for a full install and thus always adds general configuration lines to the loader file, but we don't want that for plugin installation.
from __future__ import print_function import os import sys import pkg_resources def load(plugins=(), full_install=True): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) if full_install: commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] else: commands = [] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) if full_install: commands.append("emit virtualfish_did_setup_plugins") return commands
<commit_before>from __future__ import print_function import os import sys import pkg_resources def load(plugins=()): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) commands.append("emit virtualfish_did_setup_plugins") return commands <commit_msg>Add kwarg to load function to distinguish from full install The load function is used for a full install and thus always adds general configuration lines to the loader file, but we don't want that for plugin installation.<commit_after>
from __future__ import print_function import os import sys import pkg_resources def load(plugins=(), full_install=True): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) if full_install: commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] else: commands = [] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) if full_install: commands.append("emit virtualfish_did_setup_plugins") return commands
from __future__ import print_function import os import sys import pkg_resources def load(plugins=()): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) commands.append("emit virtualfish_did_setup_plugins") return commands Add kwarg to load function to distinguish from full install The load function is used for a full install and thus always adds general configuration lines to the loader file, but we don't want that for plugin installation.from __future__ import print_function import os import sys import pkg_resources def load(plugins=(), full_install=True): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) if full_install: commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] else: commands = [] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) if full_install: commands.append("emit virtualfish_did_setup_plugins") return commands
<commit_before>from __future__ import print_function import os import sys import pkg_resources def load(plugins=()): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) commands.append("emit virtualfish_did_setup_plugins") return commands <commit_msg>Add kwarg to load function to distinguish from full install The load function is used for a full install and thus always adds general configuration lines to the loader file, but we don't want that for plugin installation.<commit_after>from __future__ import print_function import os import sys import pkg_resources def load(plugins=(), full_install=True): try: version = pkg_resources.get_distribution("virtualfish").version commands = ["set -g VIRTUALFISH_VERSION {}".format(version)] except pkg_resources.DistributionNotFound: commands = [] base_path = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) if full_install: commands += [ "set -g VIRTUALFISH_PYTHON_EXEC {}".format(sys.executable), "source {}".format(os.path.join(base_path, "virtual.fish")), ] else: commands = [] for plugin in plugins: path = os.path.join(base_path, plugin + ".fish") if os.path.exists(path): commands.append("source {}".format(path)) else: raise ValueError("Plugin does not exist: " + plugin) if full_install: commands.append("emit virtualfish_did_setup_plugins") return commands
5b652fc1af9c72c195446aaaf3ff35a501766676
tests/tests.py
tests/tests.py
# coding=UTF-8 import unittest import treetojson class TreeToJsonTests(unittest.TestCase): def test_list(self): result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \ "{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}" sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'), ('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')] output = treetojson.get_json(data=sentence) self.assertEqual(output, result) def main(): unittest.main() if __name__ == '__main__': main()
Add test case for list containing words with tags
Add test case for list containing words with tags
Python
mit
saadsahibjan/treetojson
Add test case for list containing words with tags
# coding=UTF-8 import unittest import treetojson class TreeToJsonTests(unittest.TestCase): def test_list(self): result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \ "{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}" sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'), ('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')] output = treetojson.get_json(data=sentence) self.assertEqual(output, result) def main(): unittest.main() if __name__ == '__main__': main()
<commit_before><commit_msg>Add test case for list containing words with tags<commit_after>
# coding=UTF-8 import unittest import treetojson class TreeToJsonTests(unittest.TestCase): def test_list(self): result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \ "{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}" sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'), ('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')] output = treetojson.get_json(data=sentence) self.assertEqual(output, result) def main(): unittest.main() if __name__ == '__main__': main()
Add test case for list containing words with tags# coding=UTF-8 import unittest import treetojson class TreeToJsonTests(unittest.TestCase): def test_list(self): result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \ "{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}" sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'), ('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')] output = treetojson.get_json(data=sentence) self.assertEqual(output, result) def main(): unittest.main() if __name__ == '__main__': main()
<commit_before><commit_msg>Add test case for list containing words with tags<commit_after># coding=UTF-8 import unittest import treetojson class TreeToJsonTests(unittest.TestCase): def test_list(self): result = "{\"SENTENCE\":[{\"NN\":\"Everyone\"},{\"VBZ\":\"knows\"},{\"DT\":\"an\"},{\"NN\":\"Elephant\"}," \ "{\"VBZ\":\"is\"},{\"JJR\":\"larger\"},{\"IN\":\"than\"},{\"DT\":\"a\"},{\"NN\":\"Dog\"}]}" sentence = [('Everyone', 'NN'), ('knows', 'VBZ'), ('an', 'DT'), ('Elephant', 'NN'), ('is', 'VBZ'), ('larger', 'JJR'), ('than', 'IN'), ('a', 'DT'), ('Dog', 'NN')] output = treetojson.get_json(data=sentence) self.assertEqual(output, result) def main(): unittest.main() if __name__ == '__main__': main()
0db1575341ae37644f2ce43c0a89e4baf83f8d87
filebrowser/urls.py
filebrowser/urls.py
from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), )
from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), )
Remove redirect again, it's somehow causing the JS issue and it won't work for other media types
Remove redirect again, it's somehow causing the JS issue and it won't work for other media types
Python
bsd-3-clause
django-wodnas/django-filebrowser-no-grappelli,django-wodnas/django-filebrowser-no-grappelli,sandow-digital/django-filebrowser-no-grappelli-sandow,sandow-digital/django-filebrowser-no-grappelli-sandow,sandow-digital/django-filebrowser-no-grappelli-sandow,django-wodnas/django-filebrowser-no-grappelli
from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), ) Remove redirect again, it's somehow causing the JS issue and it won't work for other media types
from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), )
<commit_before>from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), ) <commit_msg>Remove redirect again, it's somehow causing the JS issue and it won't work for other media types<commit_after>
from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), )
from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), ) Remove redirect again, it's somehow causing the JS issue and it won't work for other media typesfrom django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), )
<commit_before>from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', redirect_to, {'url': '/admin/business/photo/?_popup=1', 'permanent': True}, name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), ) <commit_msg>Remove redirect again, it's somehow causing the JS issue and it won't work for other media types<commit_after>from django.conf.urls.defaults import * from django.views.generic.simple import redirect_to urlpatterns = patterns('', # filebrowser urls url(r'^browse/$', 'filebrowser.views.browse', name="fb_browse"), url(r'^mkdir/', 'filebrowser.views.mkdir', name="fb_mkdir"), url(r'^upload/', 'filebrowser.views.upload', name="fb_upload"), url(r'^rename/$', 'filebrowser.views.rename', name="fb_rename"), url(r'^delete/$', 'filebrowser.views.delete', name="fb_delete"), url(r'^versions/$', 'filebrowser.views.versions', name="fb_versions"), url(r'^check_file/$', 'filebrowser.views._check_file', name="fb_check"), url(r'^upload_file/$', 'filebrowser.views._upload_file', name="fb_do_upload"), )
3039fec89f74618657db0509765dda48a090f0be
hetzner/__init__.py
hetzner/__init__.py
class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __repr__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass
class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __str__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass
Use __str__ for formatting the error.
RobotError: Use __str__ for formatting the error. On a traceback, the __str__ value will be read instead of __repr__ so we get the name of the exception class, but not the actual exception. This is now no longer the case :-) Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>
Python
bsd-3-clause
RedMoonStudios/hetzner
class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __repr__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass RobotError: Use __str__ for formatting the error. On a traceback, the __str__ value will be read instead of __repr__ so we get the name of the exception class, but not the actual exception. This is now no longer the case :-) Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>
class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __str__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass
<commit_before>class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __repr__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass <commit_msg>RobotError: Use __str__ for formatting the error. On a traceback, the __str__ value will be read instead of __repr__ so we get the name of the exception class, but not the actual exception. This is now no longer the case :-) Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org><commit_after>
class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __str__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass
class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __repr__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass RobotError: Use __str__ for formatting the error. On a traceback, the __str__ value will be read instead of __repr__ so we get the name of the exception class, but not the actual exception. This is now no longer the case :-) Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __str__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass
<commit_before>class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __repr__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass <commit_msg>RobotError: Use __str__ for formatting the error. On a traceback, the __str__ value will be read instead of __repr__ so we get the name of the exception class, but not the actual exception. This is now no longer the case :-) Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org><commit_after>class RobotError(Exception): def __init__(self, message, status=None): self.message = message self.status = status def __str__(self): if self.status is None: return self.message else: return "{0} ({1})".format(self.message, self.status) class ManualReboot(Exception): pass class ConnectError(Exception): pass class WebRobotError(RobotError): pass
353098b81b0e281d5d78e820dd91c3f360d6e585
ibmcnx/test/test.py
ibmcnx/test/test.py
import loadFunction.py loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 )
import ibmcnx.test.loadFunction loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 )
Customize scripts to work with menu
Customize scripts to work with menu
Python
apache-2.0
stoeps13/ibmcnx2,stoeps13/ibmcnx2
import loadFunction.py loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 ) Customize scripts to work with menu
import ibmcnx.test.loadFunction loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 )
<commit_before>import loadFunction.py loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 ) <commit_msg>Customize scripts to work with menu<commit_after>
import ibmcnx.test.loadFunction loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 )
import loadFunction.py loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 ) Customize scripts to work with menuimport ibmcnx.test.loadFunction loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 )
<commit_before>import loadFunction.py loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 ) <commit_msg>Customize scripts to work with menu<commit_after>import ibmcnx.test.loadFunction loadFilesService() FilesPolicyService.browse( "title", "true", 1, 25 )
e1721a515520a85fbbfae112eb63f877de33e7c9
caffe2/python/test_util.py
caffe2/python/test_util.py
## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace()
## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import core, workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) # clear the default engines settings to separate out its # affect from the ops tests core.SetEnginePref({}, {}) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace()
Clear the operator default engines before running operator tests
Clear the operator default engines before running operator tests Reviewed By: akyrola Differential Revision: D5729024 fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75
Python
apache-2.0
sf-wind/caffe2,xzturn/caffe2,pietern/caffe2,davinwang/caffe2,pietern/caffe2,davinwang/caffe2,sf-wind/caffe2,Yangqing/caffe2,sf-wind/caffe2,Yangqing/caffe2,davinwang/caffe2,xzturn/caffe2,sf-wind/caffe2,sf-wind/caffe2,xzturn/caffe2,pietern/caffe2,pietern/caffe2,xzturn/caffe2,Yangqing/caffe2,Yangqing/caffe2,caffe2/caffe2,pietern/caffe2,Yangqing/caffe2,xzturn/caffe2,davinwang/caffe2,davinwang/caffe2
## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace() Clear the operator default engines before running operator tests Reviewed By: akyrola Differential Revision: D5729024 fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75
## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import core, workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) # clear the default engines settings to separate out its # affect from the ops tests core.SetEnginePref({}, {}) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace()
<commit_before>## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace() <commit_msg>Clear the operator default engines before running operator tests Reviewed By: akyrola Differential Revision: D5729024 fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75<commit_after>
## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import core, workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) # clear the default engines settings to separate out its # affect from the ops tests core.SetEnginePref({}, {}) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace()
## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace() Clear the operator default engines before running operator tests Reviewed By: akyrola Differential Revision: D5729024 fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import core, workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) # clear the default engines settings to separate out its # affect from the ops tests core.SetEnginePref({}, {}) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace()
<commit_before>## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace() <commit_msg>Clear the operator default engines before running operator tests Reviewed By: akyrola Differential Revision: D5729024 fbshipit-source-id: f2850d5cf53537b22298b39a07f64dfcc2753c75<commit_after>## @package test_util # Module caffe2.python.test_util from __future__ import absolute_import from __future__ import division from __future__ import print_function from __future__ import unicode_literals import numpy as np from caffe2.python import core, workspace import unittest def rand_array(*dims): # np.random.rand() returns float instead of 0-dim array, that's why need to # do some tricks return np.array(np.random.rand(*dims) - 0.5).astype(np.float32) class TestCase(unittest.TestCase): @classmethod def setUpClass(cls): workspace.GlobalInit([ 'caffe2', '--caffe2_log_level=0', ]) # clear the default engines settings to separate out its # affect from the ops tests core.SetEnginePref({}, {}) def setUp(self): self.ws = workspace.C.Workspace() workspace.ResetWorkspace() def tearDown(self): workspace.ResetWorkspace()
bb24f9d650cc1e8ae4f7f3ffa53a662ff1788c89
zuora/client.py
zuora/client.py
""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def __str__(self): return self.client.__str__()
""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def instanciate(self, instance_type_string): """ Create object for client.factory. """ return self.client.factory.create(instance_type_string) def __str__(self): """ Display the client __str__ method. """ return self.client.__str__()
Add instanciate method + docstring
Add instanciate method + docstring
Python
bsd-3-clause
liberation/zuora-client
""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def __str__(self): return self.client.__str__() Add instanciate method + docstring
""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def instanciate(self, instance_type_string): """ Create object for client.factory. """ return self.client.factory.create(instance_type_string) def __str__(self): """ Display the client __str__ method. """ return self.client.__str__()
<commit_before>""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def __str__(self): return self.client.__str__() <commit_msg>Add instanciate method + docstring<commit_after>
""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def instanciate(self, instance_type_string): """ Create object for client.factory. """ return self.client.factory.create(instance_type_string) def __str__(self): """ Display the client __str__ method. """ return self.client.__str__()
""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def __str__(self): return self.client.__str__() Add instanciate method + docstring""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def instanciate(self, instance_type_string): """ Create object for client.factory. """ return self.client.factory.create(instance_type_string) def __str__(self): """ Display the client __str__ method. """ return self.client.__str__()
<commit_before>""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def __str__(self): return self.client.__str__() <commit_msg>Add instanciate method + docstring<commit_after>""" Client for Zuora SOAP API """ # TODO: # - Handle debug # - Handle error # - Session policy import os from suds.client import Client from suds.sax.element import Element from zuora.transport import HttpTransportWithKeepAlive class ZuoraException(Exception): """ Base Zuora Exception. """ pass class Zuora(object): """ SOAP Client based on Suds """ def __init__(self, wsdl, login, password): self.wsdl = wsdl self.login = login self.password = password self.session = None self.wsdl_path = 'file://%s' % os.path.abspath(self.wsdl) self.client = Client( self.wsdl_path, transport=HttpTransportWithKeepAlive()) def instanciate(self, instance_type_string): """ Create object for client.factory. """ return self.client.factory.create(instance_type_string) def __str__(self): """ Display the client __str__ method. """ return self.client.__str__()
d37a05d305279d9d3bd74ebbdf500b56f83f4768
salt/grains/extra.py
salt/grains/extra.py
# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if not 'conf_file' in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {}
# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if 'conf_file' not in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {}
Fix PEP8 E713 - test for membership should be "not in"
Fix PEP8 E713 - test for membership should be "not in"
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if not 'conf_file' in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {} Fix PEP8 E713 - test for membership should be "not in"
# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if 'conf_file' not in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {}
<commit_before># -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if not 'conf_file' in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {} <commit_msg>Fix PEP8 E713 - test for membership should be "not in"<commit_after>
# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if 'conf_file' not in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {}
# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if not 'conf_file' in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {} Fix PEP8 E713 - test for membership should be "not in"# -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if 'conf_file' not in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {}
<commit_before># -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if not 'conf_file' in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {} <commit_msg>Fix PEP8 E713 - test for membership should be "not in"<commit_after># -*- coding: utf-8 -*- # Import python libs import os # Import third party libs import yaml import logging # Import salt libs import salt.utils log = logging.getLogger(__name__) def shell(): ''' Return the default shell to use on this system ''' # Provides: # shell return {'shell': os.environ.get('SHELL', '/bin/sh')} def config(): ''' Return the grains set in the grains file ''' if 'conf_file' not in __opts__: return {} if os.path.isdir(__opts__['conf_file']): gfn = os.path.join( __opts__['conf_file'], 'grains' ) else: gfn = os.path.join( os.path.dirname(__opts__['conf_file']), 'grains' ) if os.path.isfile(gfn): with salt.utils.fopen(gfn, 'rb') as fp_: try: return yaml.safe_load(fp_.read()) except Exception: log.warn("Bad syntax in grains file! Skipping.") return {} return {}
7ec5786efbdb20b9cbcdf0b4f1b583a7e07e0644
comrade/core/tests.py
comrade/core/tests.py
from nose.tools import ok_, eq_ import unittest import models class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode))
from nose.tools import ok_, eq_ import unittest import models def check_direct_to_template(prefix, pattern): from django import test from django.core.urlresolvers import reverse client = test.Client() response = client.get(reverse(prefix + ':' + pattern.name)) template_name = pattern.default_args['template'] template_names = [t.name for t in test.testcases.to_list(response.template)] ok_(template_names) ok_(template_name in template_names, "Template '%s' was not a template used to render" " the response. Actual template(s) used: %s" % (template_name, u', '.join(template_names))) class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode))
Add test helper method for checking direct_to_template views.
Add test helper method for checking direct_to_template views.
Python
mit
bueda/django-comrade
from nose.tools import ok_, eq_ import unittest import models class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode)) Add test helper method for checking direct_to_template views.
from nose.tools import ok_, eq_ import unittest import models def check_direct_to_template(prefix, pattern): from django import test from django.core.urlresolvers import reverse client = test.Client() response = client.get(reverse(prefix + ':' + pattern.name)) template_name = pattern.default_args['template'] template_names = [t.name for t in test.testcases.to_list(response.template)] ok_(template_names) ok_(template_name in template_names, "Template '%s' was not a template used to render" " the response. Actual template(s) used: %s" % (template_name, u', '.join(template_names))) class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode))
<commit_before>from nose.tools import ok_, eq_ import unittest import models class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode)) <commit_msg>Add test helper method for checking direct_to_template views.<commit_after>
from nose.tools import ok_, eq_ import unittest import models def check_direct_to_template(prefix, pattern): from django import test from django.core.urlresolvers import reverse client = test.Client() response = client.get(reverse(prefix + ':' + pattern.name)) template_name = pattern.default_args['template'] template_names = [t.name for t in test.testcases.to_list(response.template)] ok_(template_names) ok_(template_name in template_names, "Template '%s' was not a template used to render" " the response. Actual template(s) used: %s" % (template_name, u', '.join(template_names))) class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode))
from nose.tools import ok_, eq_ import unittest import models class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode)) Add test helper method for checking direct_to_template views.from nose.tools import ok_, eq_ import unittest import models def check_direct_to_template(prefix, pattern): from django import test from django.core.urlresolvers import reverse client = test.Client() response = client.get(reverse(prefix + ':' + pattern.name)) template_name = pattern.default_args['template'] template_names = [t.name for t in test.testcases.to_list(response.template)] ok_(template_names) ok_(template_name in template_names, "Template '%s' was not a template used to render" " the response. Actual template(s) used: %s" % (template_name, u', '.join(template_names))) class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode))
<commit_before>from nose.tools import ok_, eq_ import unittest import models class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode)) <commit_msg>Add test helper method for checking direct_to_template views.<commit_after>from nose.tools import ok_, eq_ import unittest import models def check_direct_to_template(prefix, pattern): from django import test from django.core.urlresolvers import reverse client = test.Client() response = client.get(reverse(prefix + ':' + pattern.name)) template_name = pattern.default_args['template'] template_names = [t.name for t in test.testcases.to_list(response.template)] ok_(template_names) ok_(template_name in template_names, "Template '%s' was not a template used to render" " the response. Actual template(s) used: %s" % (template_name, u', '.join(template_names))) class SimpleModel(models.ComradeBaseModel): def __unicode__(self): return u'This is a unicode string' class TestBaseModel(unittest.TestCase): def setUp(self): super(TestBaseModel, self).setUp() self.obj = SimpleModel() def test_repr(self): ok_(isinstance(self.obj.__repr__(), str)) def test_str(self): ok_(isinstance(self.obj.__str__(), str)) def test_unicode(self): ok_(isinstance(self.obj.__unicode__(), unicode))
f30e441958b8354b189ee5b5ef1e7eb47ebb1b1a
nhs/gunicorn_conf.py
nhs/gunicorn_conf.py
bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 3 timeout = 60
bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 6 timeout = 60
Increase the number of Gunicorn workers
Increase the number of Gunicorn workers
Python
agpl-3.0
openhealthcare/open-prescribing,openhealthcare/open-prescribing,openhealthcare/open-prescribing
bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 3 timeout = 60 Increase the number of Gunicorn workers
bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 6 timeout = 60
<commit_before>bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 3 timeout = 60 <commit_msg>Increase the number of Gunicorn workers<commit_after>
bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 6 timeout = 60
bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 3 timeout = 60 Increase the number of Gunicorn workersbind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 6 timeout = 60
<commit_before>bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 3 timeout = 60 <commit_msg>Increase the number of Gunicorn workers<commit_after>bind = "127.0.0.1:4567" logfile = "/usr/local/ohc/log/op.gunicorn.log" workers = 6 timeout = 60
4f3cfe6e990c932d7f86dbd0cf8ae762407278b0
nucleus/base/urls.py
nucleus/base/urls.py
from django.conf.urls import url from nucleus.base import views urlpatterns = ( url(r'^/?$', views.home, name='base.home'), )
from django.conf.urls import url from django.views.generic import RedirectView urlpatterns = ( url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'), )
Change root URL to redirect to /rna/
Change root URL to redirect to /rna/
Python
mpl-2.0
mozilla/nucleus,mozilla/nucleus,mozilla/nucleus,mozilla/nucleus
from django.conf.urls import url from nucleus.base import views urlpatterns = ( url(r'^/?$', views.home, name='base.home'), ) Change root URL to redirect to /rna/
from django.conf.urls import url from django.views.generic import RedirectView urlpatterns = ( url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'), )
<commit_before>from django.conf.urls import url from nucleus.base import views urlpatterns = ( url(r'^/?$', views.home, name='base.home'), ) <commit_msg>Change root URL to redirect to /rna/<commit_after>
from django.conf.urls import url from django.views.generic import RedirectView urlpatterns = ( url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'), )
from django.conf.urls import url from nucleus.base import views urlpatterns = ( url(r'^/?$', views.home, name='base.home'), ) Change root URL to redirect to /rna/from django.conf.urls import url from django.views.generic import RedirectView urlpatterns = ( url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'), )
<commit_before>from django.conf.urls import url from nucleus.base import views urlpatterns = ( url(r'^/?$', views.home, name='base.home'), ) <commit_msg>Change root URL to redirect to /rna/<commit_after>from django.conf.urls import url from django.views.generic import RedirectView urlpatterns = ( url(r'^/?$', RedirectView.as_view(url='/rna/', permanent=True), name='base.home'), )
5d8ea747bd5f34b382cc9fef91105f3ed434c0db
pylearn2/datasets/hdf5.py
pylearn2/datasets/hdf5.py
"""Objects for datasets serialized in HDF5 format (.h5).""" import h5py from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs)
"""Objects for datasets serialized in HDF5 format (.h5).""" import warnings try: import h5py except ImportError: warnings.warn("Could not import h5py") from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs)
Fix import issue in h5py.py
Fix import issue in h5py.py
Python
bsd-3-clause
fulmicoton/pylearn2,caidongyun/pylearn2,ddboline/pylearn2,chrish42/pylearn,CIFASIS/pylearn2,sandeepkbhat/pylearn2,aalmah/pylearn2,goodfeli/pylearn2,junbochen/pylearn2,jamessergeant/pylearn2,nouiz/pylearn2,mclaughlin6464/pylearn2,pombredanne/pylearn2,JesseLivezey/plankton,msingh172/pylearn2,w1kke/pylearn2,shiquanwang/pylearn2,mclaughlin6464/pylearn2,theoryno3/pylearn2,goodfeli/pylearn2,matrogers/pylearn2,CIFASIS/pylearn2,sandeepkbhat/pylearn2,JesseLivezey/pylearn2,mclaughlin6464/pylearn2,kastnerkyle/pylearn2,daemonmaker/pylearn2,aalmah/pylearn2,pombredanne/pylearn2,ashhher3/pylearn2,pkainz/pylearn2,ddboline/pylearn2,woozzu/pylearn2,Refefer/pylearn2,hyqneuron/pylearn2-maxsom,shiquanwang/pylearn2,skearnes/pylearn2,shiquanwang/pylearn2,nouiz/pylearn2,pkainz/pylearn2,Refefer/pylearn2,KennethPierce/pylearnk,jamessergeant/pylearn2,alexjc/pylearn2,lisa-lab/pylearn2,hyqneuron/pylearn2-maxsom,sandeepkbhat/pylearn2,abergeron/pylearn2,kose-y/pylearn2,lunyang/pylearn2,lunyang/pylearn2,lancezlin/pylearn2,KennethPierce/pylearnk,junbochen/pylearn2,JesseLivezey/plankton,ashhher3/pylearn2,se4u/pylearn2,se4u/pylearn2,matrogers/pylearn2,CIFASIS/pylearn2,fulmicoton/pylearn2,lisa-lab/pylearn2,mkraemer67/pylearn2,aalmah/pylearn2,junbochen/pylearn2,fishcorn/pylearn2,lancezlin/pylearn2,kose-y/pylearn2,bartvm/pylearn2,daemonmaker/pylearn2,ddboline/pylearn2,daemonmaker/pylearn2,fyffyt/pylearn2,jamessergeant/pylearn2,JesseLivezey/plankton,goodfeli/pylearn2,abergeron/pylearn2,msingh172/pylearn2,caidongyun/pylearn2,goodfeli/pylearn2,sandeepkbhat/pylearn2,se4u/pylearn2,hantek/pylearn2,kastnerkyle/pylearn2,skearnes/pylearn2,lamblin/pylearn2,mkraemer67/pylearn2,woozzu/pylearn2,mkraemer67/pylearn2,jeremyfix/pylearn2,bartvm/pylearn2,hyqneuron/pylearn2-maxsom,pombredanne/pylearn2,fyffyt/pylearn2,mkraemer67/pylearn2,jeremyfix/pylearn2,ashhher3/pylearn2,daemonmaker/pylearn2,pkainz/pylearn2,lunyang/pylearn2,bartvm/pylearn2,matrogers/pylearn2,chrish42/pylearn,fishcorn/pylearn2,caidongyun/pylearn2,skearnes/pylearn2,lancezlin/pylearn2,w1kke/pylearn2,w1kke/pylearn2,se4u/pylearn2,woozzu/pylearn2,lamblin/pylearn2,fyffyt/pylearn2,junbochen/pylearn2,alexjc/pylearn2,alexjc/pylearn2,hantek/pylearn2,ashhher3/pylearn2,msingh172/pylearn2,skearnes/pylearn2,fishcorn/pylearn2,TNick/pylearn2,chrish42/pylearn,cosmoharrigan/pylearn2,fyffyt/pylearn2,shiquanwang/pylearn2,jeremyfix/pylearn2,hyqneuron/pylearn2-maxsom,abergeron/pylearn2,lunyang/pylearn2,mclaughlin6464/pylearn2,abergeron/pylearn2,fishcorn/pylearn2,theoryno3/pylearn2,kastnerkyle/pylearn2,msingh172/pylearn2,pkainz/pylearn2,fulmicoton/pylearn2,chrish42/pylearn,nouiz/pylearn2,lamblin/pylearn2,jeremyfix/pylearn2,fulmicoton/pylearn2,theoryno3/pylearn2,theoryno3/pylearn2,cosmoharrigan/pylearn2,lisa-lab/pylearn2,lisa-lab/pylearn2,matrogers/pylearn2,pombredanne/pylearn2,JesseLivezey/pylearn2,TNick/pylearn2,KennethPierce/pylearnk,KennethPierce/pylearnk,JesseLivezey/plankton,Refefer/pylearn2,jamessergeant/pylearn2,alexjc/pylearn2,kose-y/pylearn2,hantek/pylearn2,aalmah/pylearn2,JesseLivezey/pylearn2,bartvm/pylearn2,lancezlin/pylearn2,TNick/pylearn2,lamblin/pylearn2,JesseLivezey/pylearn2,TNick/pylearn2,cosmoharrigan/pylearn2,woozzu/pylearn2,CIFASIS/pylearn2,nouiz/pylearn2,Refefer/pylearn2,ddboline/pylearn2,kose-y/pylearn2,w1kke/pylearn2,cosmoharrigan/pylearn2,kastnerkyle/pylearn2,hantek/pylearn2,caidongyun/pylearn2
"""Objects for datasets serialized in HDF5 format (.h5).""" import h5py from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs) Fix import issue in h5py.py
"""Objects for datasets serialized in HDF5 format (.h5).""" import warnings try: import h5py except ImportError: warnings.warn("Could not import h5py") from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs)
<commit_before>"""Objects for datasets serialized in HDF5 format (.h5).""" import h5py from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs) <commit_msg>Fix import issue in h5py.py<commit_after>
"""Objects for datasets serialized in HDF5 format (.h5).""" import warnings try: import h5py except ImportError: warnings.warn("Could not import h5py") from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs)
"""Objects for datasets serialized in HDF5 format (.h5).""" import h5py from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs) Fix import issue in h5py.py"""Objects for datasets serialized in HDF5 format (.h5).""" import warnings try: import h5py except ImportError: warnings.warn("Could not import h5py") from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs)
<commit_before>"""Objects for datasets serialized in HDF5 format (.h5).""" import h5py from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs) <commit_msg>Fix import issue in h5py.py<commit_after>"""Objects for datasets serialized in HDF5 format (.h5).""" import warnings try: import h5py except ImportError: warnings.warn("Could not import h5py") from pylearn2.datasets.dense_design_matrix import DenseDesignMatrix class HDF5Dataset(DenseDesignMatrix): """Dense dataset loaded from an HDF5 file.""" def __init__(self, filename, X=None, topo_view=None, y=None, **kwargs): """ Loads data and labels from HDF5 file. Parameters ---------- filename: str HDF5 file name. X: str Key into HDF5 file for dataset design matrix. topo_view: str Key into HDF5 file for topological view of dataset. y: str Key into HDF5 file for dataset targets. kwargs: dict Keyword arguments passed to `DenseDesignMatrix`. """ with h5py.File(filename) as f: if X is not None: X = f[X][:] if topo_view is not None: topo_view = f[topo_view][:] if y is not None: y = f[y][:] super(HDF5Dataset, self).__init__(X=X, topo_view=topo_view, y=y, **kwargs)
9b8cbfcf33ba644670a42490db7de4249e5ff080
invocations/docs.py
invocations/docs.py
import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs.body() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs.body()
import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs()
Leverage __call__ on task downstream
Leverage __call__ on task downstream
Python
bsd-2-clause
mrjmad/invocations,alex/invocations,pyinvoke/invocations,singingwolfboy/invocations
import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs.body() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs.body() Leverage __call__ on task downstream
import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs()
<commit_before>import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs.body() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs.body() <commit_msg>Leverage __call__ on task downstream<commit_after>
import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs()
import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs.body() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs.body() Leverage __call__ on task downstreamimport os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs()
<commit_before>import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs.body() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs.body() <commit_msg>Leverage __call__ on task downstream<commit_after>import os from invoke.tasks import task from invoke.runner import run docs_dir = 'docs' build = os.path.join(docs_dir, '_build') @task def clean_docs(): run("rm -rf %s" % build) @task def browse_docs(): run("open %s" % os.path.join(build, 'index.html')) @task def docs(clean=False, browse=False): if clean: clean_docs() run("sphinx-build %s %s" % (docs_dir, build), pty=True) if browse: browse_docs()
153360072096d4a3cef783d371fbfabcd75bcf98
script/lib/config.py
script/lib/config.py
#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
Upgrade libchromiumcontent to loose iframe sandbox.
Upgrade libchromiumcontent to loose iframe sandbox.
Python
mit
rreimann/electron,dongjoon-hyun/electron,brave/muon,trankmichael/electron,kokdemo/electron,robinvandernoord/electron,tonyganch/electron,tylergibson/electron,natgolov/electron,sircharleswatson/electron,matiasinsaurralde/electron,vaginessa/electron,fritx/electron,benweissmann/electron,xiruibing/electron,cos2004/electron,anko/electron,systembugtj/electron,matiasinsaurralde/electron,aichingm/electron,stevekinney/electron,astoilkov/electron,astoilkov/electron,chrisswk/electron,dkfiresky/electron,mattotodd/electron,oiledCode/electron,soulteary/electron,meowlab/electron,synaptek/electron,carsonmcdonald/electron,Ivshti/electron,digideskio/electron,faizalpribadi/electron,bpasero/electron,robinvandernoord/electron,mhkeller/electron,gerhardberger/electron,sky7sea/electron,anko/electron,mirrh/electron,jtburke/electron,aaron-goshine/electron,bruce/electron,nagyistoce/electron-atom-shell,tonyganch/electron,electron/electron,MaxGraey/electron,eric-seekas/electron,evgenyzinoviev/electron,thomsonreuters/electron,arturts/electron,matiasinsaurralde/electron,chriskdon/electron,xiruibing/electron,trankmichael/electron,etiktin/electron,gamedevsam/electron,Neron-X5/electron,pombredanne/electron,eriser/electron,trigrass2/electron,posix4e/electron,jacksondc/electron,RobertJGabriel/electron,John-Lin/electron,thompsonemerson/electron,christian-bromann/electron,bpasero/electron,subblue/electron,Jacobichou/electron,noikiy/electron,rreimann/electron,gabriel/electron,timruffles/electron,greyhwndz/electron,jhen0409/electron,Rokt33r/electron,abhishekgahlot/electron,meowlab/electron,gabrielPeart/electron,Andrey-Pavlov/electron,pirafrank/electron,brave/electron,jjz/electron,shockone/electron,dahal/electron,leolujuyi/electron,minggo/electron,thomsonreuters/electron,gabrielPeart/electron,IonicaBizauKitchen/electron,Floato/electron,setzer777/electron,Zagorakiss/electron,thompsonemerson/electron,jsutcodes/electron,kazupon/electron,webmechanicx/electron,meowlab/electron,gabriel/electron,thomsonreuters/electron,icattlecoder/electron,davazp/electron,iftekeriba/electron,subblue/electron,pandoraui/electron,gerhardberger/electron,synaptek/electron,ianscrivener/electron,wolfflow/electron,mirrh/electron,RobertJGabriel/electron,vHanda/electron,meowlab/electron,gbn972/electron,wan-qy/electron,smczk/electron,electron/electron,takashi/electron,GoooIce/electron,RobertJGabriel/electron,astoilkov/electron,SufianHassan/electron,Rokt33r/electron,webmechanicx/electron,systembugtj/electron,jlord/electron,vaginessa/electron,bpasero/electron,rsvip/electron,joaomoreno/atom-shell,smczk/electron,preco21/electron,eriser/electron,electron/electron,kenmozi/electron,joaomoreno/atom-shell,mjaniszew/electron,takashi/electron,Jacobichou/electron,gamedevsam/electron,brenca/electron,tincan24/electron,jonatasfreitasv/electron,roadev/electron,leftstick/electron,tonyganch/electron,mattotodd/electron,tonyganch/electron,pombredanne/electron,jaanus/electron,gamedevsam/electron,micalan/electron,twolfson/electron,rhencke/electron,gbn972/electron,smczk/electron,mattotodd/electron,Rokt33r/electron,Evercoder/electron,arturts/electron,Floato/electron,takashi/electron,darwin/electron,Ivshti/electron,medixdev/electron,brave/electron,jcblw/electron,mattdesl/electron,nagyistoce/electron-atom-shell,aliib/electron,tomashanacek/electron,davazp/electron,RIAEvangelist/electron,dkfiresky/electron,darwin/electron,rajatsingla28/electron,bright-sparks/electron,dongjoon-hyun/electron,ankitaggarwal011/electron,mrwizard82d1/electron,nicholasess/electron,digideskio/electron,icattlecoder/electron,darwin/electron,Neron-X5/electron,bitemyapp/electron,pandoraui/electron,aecca/electron,bwiggs/electron,John-Lin/electron,Gerhut/electron,coderhaoxin/electron,dahal/electron,mrwizard82d1/electron,brave/electron,voidbridge/electron,maxogden/atom-shell,mattotodd/electron,bobwol/electron,bpasero/electron,d-salas/electron,simongregory/electron,benweissmann/electron,shiftkey/electron,fireball-x/atom-shell,xfstudio/electron,Gerhut/electron,vipulroxx/electron,Rokt33r/electron,aliib/electron,rreimann/electron,IonicaBizauKitchen/electron,egoist/electron,twolfson/electron,John-Lin/electron,jsutcodes/electron,adamjgray/electron,vipulroxx/electron,jtburke/electron,biblerule/UMCTelnetHub,d-salas/electron,bruce/electron,nekuz0r/electron,Evercoder/electron,noikiy/electron,bruce/electron,gerhardberger/electron,kenmozi/electron,simonfork/electron,MaxWhere/electron,thingsinjars/electron,tinydew4/electron,kikong/electron,deepak1556/atom-shell,thompsonemerson/electron,miniak/electron,ianscrivener/electron,deepak1556/atom-shell,JesselJohn/electron,jhen0409/electron,hokein/atom-shell,fomojola/electron,JesselJohn/electron,robinvandernoord/electron,the-ress/electron,DivyaKMenon/electron,bitemyapp/electron,kokdemo/electron,MaxWhere/electron,Zagorakiss/electron,shennushi/electron,bbondy/electron,lrlna/electron,cqqccqc/electron,Faiz7412/electron,xfstudio/electron,beni55/electron,Jonekee/electron,hokein/atom-shell,baiwyc119/electron,robinvandernoord/electron,arusakov/electron,Andrey-Pavlov/electron,egoist/electron,simongregory/electron,tinydew4/electron,d-salas/electron,jannishuebl/electron,vaginessa/electron,jcblw/electron,tonyganch/electron,systembugtj/electron,vaginessa/electron,stevemao/electron,renaesop/electron,sky7sea/electron,digideskio/electron,yan-foto/electron,shaundunne/electron,farmisen/electron,bobwol/electron,jannishuebl/electron,nicholasess/electron,nicholasess/electron,bbondy/electron,leftstick/electron,fireball-x/atom-shell,simonfork/electron,ianscrivener/electron,renaesop/electron,faizalpribadi/electron,roadev/electron,natgolov/electron,xfstudio/electron,farmisen/electron,seanchas116/electron,cqqccqc/electron,joaomoreno/atom-shell,edulan/electron,miniak/electron,jsutcodes/electron,stevemao/electron,meowlab/electron,d-salas/electron,edulan/electron,jannishuebl/electron,rhencke/electron,sircharleswatson/electron,kazupon/electron,cqqccqc/electron,biblerule/UMCTelnetHub,dongjoon-hyun/electron,oiledCode/electron,jannishuebl/electron,evgenyzinoviev/electron,iftekeriba/electron,seanchas116/electron,Ivshti/electron,beni55/electron,preco21/electron,wan-qy/electron,iftekeriba/electron,RobertJGabriel/electron,ervinb/electron,eriser/electron,brenca/electron,tonyganch/electron,yalexx/electron,shiftkey/electron,thingsinjars/electron,michaelchiche/electron,leftstick/electron,systembugtj/electron,RIAEvangelist/electron,mhkeller/electron,rprichard/electron,astoilkov/electron,wolfflow/electron,shiftkey/electron,yalexx/electron,greyhwndz/electron,evgenyzinoviev/electron,brenca/electron,seanchas116/electron,JesselJohn/electron,brenca/electron,GoooIce/electron,saronwei/electron,timruffles/electron,nekuz0r/electron,cos2004/electron,jonatasfreitasv/electron,DivyaKMenon/electron,abhishekgahlot/electron,BionicClick/electron,eriser/electron,RobertJGabriel/electron,fomojola/electron,aecca/electron,davazp/electron,shennushi/electron,shaundunne/electron,destan/electron,lrlna/electron,thomsonreuters/electron,kokdemo/electron,neutrous/electron,trigrass2/electron,voidbridge/electron,Jonekee/electron,rsvip/electron,yalexx/electron,nicobot/electron,trigrass2/electron,bwiggs/electron,neutrous/electron,trankmichael/electron,trigrass2/electron,kostia/electron,kenmozi/electron,soulteary/electron,webmechanicx/electron,tylergibson/electron,twolfson/electron,sshiting/electron,electron/electron,bruce/electron,dkfiresky/electron,synaptek/electron,the-ress/electron,joaomoreno/atom-shell,hokein/atom-shell,deed02392/electron,rprichard/electron,darwin/electron,kazupon/electron,shennushi/electron,howmuchcomputer/electron,christian-bromann/electron,mattdesl/electron,oiledCode/electron,MaxWhere/electron,gabrielPeart/electron,adamjgray/electron,ianscrivener/electron,shiftkey/electron,voidbridge/electron,vHanda/electron,ankitaggarwal011/electron,leftstick/electron,nicobot/electron,fffej/electron,dahal/electron,trankmichael/electron,jaanus/electron,jonatasfreitasv/electron,arusakov/electron,kikong/electron,stevekinney/electron,gabriel/electron,tinydew4/electron,zhakui/electron,eriser/electron,coderhaoxin/electron,pombredanne/electron,minggo/electron,miniak/electron,gabrielPeart/electron,kostia/electron,Andrey-Pavlov/electron,mattdesl/electron,oiledCode/electron,nagyistoce/electron-atom-shell,etiktin/electron,beni55/electron,medixdev/electron,noikiy/electron,rhencke/electron,simongregory/electron,deepak1556/atom-shell,JesselJohn/electron,renaesop/electron,preco21/electron,jtburke/electron,SufianHassan/electron,gabriel/electron,aaron-goshine/electron,Faiz7412/electron,John-Lin/electron,wolfflow/electron,shaundunne/electron,pandoraui/electron,jannishuebl/electron,thompsonemerson/electron,soulteary/electron,smczk/electron,aliib/electron,neutrous/electron,eric-seekas/electron,joneit/electron,felixrieseberg/electron,mattdesl/electron,bright-sparks/electron,timruffles/electron,robinvandernoord/electron,subblue/electron,aecca/electron,oiledCode/electron,bright-sparks/electron,nicobot/electron,nekuz0r/electron,jiaz/electron,simonfork/electron,howmuchcomputer/electron,GoooIce/electron,etiktin/electron,tylergibson/electron,fffej/electron,bitemyapp/electron,shennushi/electron,chrisswk/electron,eric-seekas/electron,thompsonemerson/electron,icattlecoder/electron,setzer777/electron,chrisswk/electron,yan-foto/electron,preco21/electron,wan-qy/electron,chriskdon/electron,farmisen/electron,yan-foto/electron,pandoraui/electron,jacksondc/electron,aichingm/electron,adamjgray/electron,chrisswk/electron,vipulroxx/electron,JussMee15/electron,gbn972/electron,DivyaKMenon/electron,zhakui/electron,lrlna/electron,miniak/electron,lzpfmh/electron,howmuchcomputer/electron,Evercoder/electron,voidbridge/electron,fireball-x/atom-shell,shockone/electron,mjaniszew/electron,seanchas116/electron,mubassirhayat/electron,jjz/electron,carsonmcdonald/electron,seanchas116/electron,kokdemo/electron,bwiggs/electron,natgolov/electron,brave/muon,chriskdon/electron,faizalpribadi/electron,michaelchiche/electron,preco21/electron,darwin/electron,micalan/electron,jlord/electron,Gerhut/electron,ianscrivener/electron,fritx/electron,soulteary/electron,Jonekee/electron,Ivshti/electron,hokein/atom-shell,xiruibing/electron,biblerule/UMCTelnetHub,edulan/electron,jsutcodes/electron,nekuz0r/electron,dkfiresky/electron,Floato/electron,abhishekgahlot/electron,arusakov/electron,beni55/electron,digideskio/electron,mirrh/electron,pombredanne/electron,yalexx/electron,faizalpribadi/electron,eric-seekas/electron,icattlecoder/electron,bright-sparks/electron,LadyNaggaga/electron,saronwei/electron,christian-bromann/electron,rhencke/electron,shennushi/electron,JussMee15/electron,MaxGraey/electron,Zagorakiss/electron,joaomoreno/atom-shell,seanchas116/electron,kikong/electron,jjz/electron,John-Lin/electron,fireball-x/atom-shell,iftekeriba/electron,jtburke/electron,arusakov/electron,saronwei/electron,rajatsingla28/electron,mjaniszew/electron,jhen0409/electron,aliib/electron,icattlecoder/electron,systembugtj/electron,destan/electron,fritx/electron,simongregory/electron,MaxGraey/electron,trankmichael/electron,arturts/electron,jtburke/electron,felixrieseberg/electron,saronwei/electron,jlhbaseball15/electron,baiwyc119/electron,tomashanacek/electron,mrwizard82d1/electron,simongregory/electron,subblue/electron,oiledCode/electron,mrwizard82d1/electron,Zagorakiss/electron,bpasero/electron,wolfflow/electron,bobwol/electron,fffej/electron,yan-foto/electron,gamedevsam/electron,arturts/electron,subblue/electron,stevekinney/electron,SufianHassan/electron,mubassirhayat/electron,howmuchcomputer/electron,vHanda/electron,Neron-X5/electron,tincan24/electron,kostia/electron,mhkeller/electron,leethomas/electron,wan-qy/electron,Floato/electron,baiwyc119/electron,dkfiresky/electron,IonicaBizauKitchen/electron,leethomas/electron,pombredanne/electron,bruce/electron,digideskio/electron,mattdesl/electron,rreimann/electron,carsonmcdonald/electron,matiasinsaurralde/electron,bobwol/electron,brave/electron,saronwei/electron,ervinb/electron,GoooIce/electron,jlhbaseball15/electron,minggo/electron,bright-sparks/electron,jlhbaseball15/electron,adamjgray/electron,Andrey-Pavlov/electron,miniak/electron,leethomas/electron,etiktin/electron,fffej/electron,pirafrank/electron,gabrielPeart/electron,leethomas/electron,anko/electron,subblue/electron,lzpfmh/electron,synaptek/electron,ervinb/electron,egoist/electron,stevemao/electron,Rokt33r/electron,bwiggs/electron,timruffles/electron,destan/electron,nagyistoce/electron-atom-shell,Faiz7412/electron,fomojola/electron,kazupon/electron,mrwizard82d1/electron,adcentury/electron,evgenyzinoviev/electron,xfstudio/electron,kcrt/electron,John-Lin/electron,kenmozi/electron,Neron-X5/electron,joneit/electron,mattotodd/electron,trigrass2/electron,lrlna/electron,aaron-goshine/electron,gamedevsam/electron,RIAEvangelist/electron,tomashanacek/electron,bobwol/electron,egoist/electron,aecca/electron,jiaz/electron,deed02392/electron,Jonekee/electron,jlhbaseball15/electron,mirrh/electron,brave/muon,RobertJGabriel/electron,DivyaKMenon/electron,destan/electron,gbn972/electron,takashi/electron,fabien-d/electron,fffej/electron,rsvip/electron,BionicClick/electron,RIAEvangelist/electron,cqqccqc/electron,abhishekgahlot/electron,the-ress/electron,beni55/electron,shaundunne/electron,yan-foto/electron,faizalpribadi/electron,stevekinney/electron,mhkeller/electron,shockone/electron,michaelchiche/electron,rhencke/electron,brave/muon,micalan/electron,Floato/electron,howmuchcomputer/electron,zhakui/electron,natgolov/electron,renaesop/electron,mattdesl/electron,Gerhut/electron,carsonmcdonald/electron,ianscrivener/electron,natgolov/electron,gstack/infinium-shell,tomashanacek/electron,maxogden/atom-shell,vipulroxx/electron,MaxGraey/electron,roadev/electron,jcblw/electron,joneit/electron,dahal/electron,cos2004/electron,JussMee15/electron,BionicClick/electron,cos2004/electron,RIAEvangelist/electron,medixdev/electron,coderhaoxin/electron,noikiy/electron,bbondy/electron,mhkeller/electron,zhakui/electron,jlhbaseball15/electron,pandoraui/electron,thingsinjars/electron,rreimann/electron,anko/electron,coderhaoxin/electron,nekuz0r/electron,rsvip/electron,iftekeriba/electron,yan-foto/electron,tincan24/electron,leethomas/electron,sircharleswatson/electron,greyhwndz/electron,tylergibson/electron,arturts/electron,brenca/electron,brenca/electron,bitemyapp/electron,bwiggs/electron,kikong/electron,nicholasess/electron,synaptek/electron,simonfork/electron,adcentury/electron,tincan24/electron,dahal/electron,egoist/electron,carsonmcdonald/electron,farmisen/electron,kazupon/electron,jonatasfreitasv/electron,abhishekgahlot/electron,biblerule/UMCTelnetHub,jlord/electron,webmechanicx/electron,matiasinsaurralde/electron,thomsonreuters/electron,sshiting/electron,baiwyc119/electron,eriser/electron,baiwyc119/electron,the-ress/electron,electron/electron,SufianHassan/electron,minggo/electron,posix4e/electron,felixrieseberg/electron,saronwei/electron,kokdemo/electron,gerhardberger/electron,wolfflow/electron,kcrt/electron,shockone/electron,felixrieseberg/electron,wan-qy/electron,renaesop/electron,deed02392/electron,adamjgray/electron,jacksondc/electron,mjaniszew/electron,noikiy/electron,voidbridge/electron,aaron-goshine/electron,setzer777/electron,GoooIce/electron,rsvip/electron,neutrous/electron,thingsinjars/electron,Faiz7412/electron,ervinb/electron,fabien-d/electron,twolfson/electron,tylergibson/electron,fabien-d/electron,destan/electron,nicholasess/electron,miniak/electron,rajatsingla28/electron,mirrh/electron,christian-bromann/electron,JesselJohn/electron,aliib/electron,joneit/electron,astoilkov/electron,minggo/electron,kazupon/electron,hokein/atom-shell,kcrt/electron,simonfork/electron,fffej/electron,lzpfmh/electron,smczk/electron,carsonmcdonald/electron,jacksondc/electron,wolfflow/electron,dongjoon-hyun/electron,fireball-x/atom-shell,jsutcodes/electron,stevemao/electron,smczk/electron,fomojola/electron,edulan/electron,gerhardberger/electron,gstack/infinium-shell,adcentury/electron,xiruibing/electron,stevemao/electron,lzpfmh/electron,gabriel/electron,fritx/electron,twolfson/electron,shaundunne/electron,pirafrank/electron,kostia/electron,tinydew4/electron,adcentury/electron,renaesop/electron,JussMee15/electron,brave/electron,joneit/electron,Evercoder/electron,aaron-goshine/electron,kostia/electron,BionicClick/electron,rhencke/electron,posix4e/electron,d-salas/electron,davazp/electron,electron/electron,GoooIce/electron,jhen0409/electron,benweissmann/electron,simonfork/electron,vipulroxx/electron,farmisen/electron,shaundunne/electron,leolujuyi/electron,the-ress/electron,Faiz7412/electron,bright-sparks/electron,fomojola/electron,sircharleswatson/electron,thompsonemerson/electron,thingsinjars/electron,LadyNaggaga/electron,micalan/electron,roadev/electron,fabien-d/electron,jjz/electron,deed02392/electron,sshiting/electron,roadev/electron,Jacobichou/electron,trigrass2/electron,mrwizard82d1/electron,evgenyzinoviev/electron,bbondy/electron,bwiggs/electron,mjaniszew/electron,lrlna/electron,nicobot/electron,biblerule/UMCTelnetHub,trankmichael/electron,pirafrank/electron,jiaz/electron,shockone/electron,gabrielPeart/electron,Jacobichou/electron,dongjoon-hyun/electron,leolujuyi/electron,zhakui/electron,jaanus/electron,IonicaBizauKitchen/electron,synaptek/electron,davazp/electron,pombredanne/electron,bpasero/electron,gerhardberger/electron,mjaniszew/electron,webmechanicx/electron,nicobot/electron,JussMee15/electron,ankitaggarwal011/electron,jhen0409/electron,jcblw/electron,micalan/electron,mubassirhayat/electron,edulan/electron,deepak1556/atom-shell,kokdemo/electron,Floato/electron,mattotodd/electron,arturts/electron,vHanda/electron,chriskdon/electron,neutrous/electron,gstack/infinium-shell,pirafrank/electron,jonatasfreitasv/electron,posix4e/electron,tincan24/electron,soulteary/electron,twolfson/electron,LadyNaggaga/electron,sky7sea/electron,shockone/electron,greyhwndz/electron,sircharleswatson/electron,etiktin/electron,sky7sea/electron,eric-seekas/electron,electron/electron,coderhaoxin/electron,eric-seekas/electron,aliib/electron,aichingm/electron,adamjgray/electron,aichingm/electron,Jacobichou/electron,jiaz/electron,webmechanicx/electron,egoist/electron,the-ress/electron,nicobot/electron,medixdev/electron,jjz/electron,MaxGraey/electron,d-salas/electron,Evercoder/electron,coderhaoxin/electron,thingsinjars/electron,neutrous/electron,abhishekgahlot/electron,natgolov/electron,beni55/electron,vaginessa/electron,BionicClick/electron,anko/electron,bruce/electron,leolujuyi/electron,aaron-goshine/electron,jhen0409/electron,minggo/electron,mubassirhayat/electron,stevemao/electron,MaxWhere/electron,jiaz/electron,adcentury/electron,MaxWhere/electron,pirafrank/electron,jonatasfreitasv/electron,JussMee15/electron,Jonekee/electron,benweissmann/electron,Andrey-Pavlov/electron,LadyNaggaga/electron,shiftkey/electron,sshiting/electron,nagyistoce/electron-atom-shell,dahal/electron,davazp/electron,chrisswk/electron,Andrey-Pavlov/electron,gstack/infinium-shell,rajatsingla28/electron,shennushi/electron,setzer777/electron,robinvandernoord/electron,rajatsingla28/electron,brave/muon,digideskio/electron,christian-bromann/electron,kcrt/electron,tomashanacek/electron,Rokt33r/electron,wan-qy/electron,BionicClick/electron,michaelchiche/electron,takashi/electron,sshiting/electron,posix4e/electron,JesselJohn/electron,rreimann/electron,fritx/electron,adcentury/electron,LadyNaggaga/electron,benweissmann/electron,faizalpribadi/electron,farmisen/electron,arusakov/electron,nicholasess/electron,gerhardberger/electron,aecca/electron,jcblw/electron,leethomas/electron,gbn972/electron,bbondy/electron,greyhwndz/electron,noikiy/electron,thomsonreuters/electron,takashi/electron,voidbridge/electron,jcblw/electron,setzer777/electron,DivyaKMenon/electron,maxogden/atom-shell,bbondy/electron,kcrt/electron,stevekinney/electron,yalexx/electron,christian-bromann/electron,jlord/electron,vaginessa/electron,xfstudio/electron,mirrh/electron,ervinb/electron,meowlab/electron,SufianHassan/electron,etiktin/electron,leftstick/electron,sky7sea/electron,xiruibing/electron,bitemyapp/electron,joneit/electron,Gerhut/electron,matiasinsaurralde/electron,gabriel/electron,micalan/electron,Jonekee/electron,jannishuebl/electron,destan/electron,Neron-X5/electron,SufianHassan/electron,maxogden/atom-shell,jaanus/electron,Zagorakiss/electron,yalexx/electron,gbn972/electron,kenmozi/electron,joaomoreno/atom-shell,gstack/infinium-shell,timruffles/electron,jtburke/electron,leolujuyi/electron,ankitaggarwal011/electron,Neron-X5/electron,systembugtj/electron,jlhbaseball15/electron,kenmozi/electron,nekuz0r/electron,RIAEvangelist/electron,rajatsingla28/electron,shiftkey/electron,setzer777/electron,tinydew4/electron,tincan24/electron,fomojola/electron,maxogden/atom-shell,bobwol/electron,ervinb/electron,tylergibson/electron,deepak1556/atom-shell,evgenyzinoviev/electron,jiaz/electron,sircharleswatson/electron,jjz/electron,xfstudio/electron,felixrieseberg/electron,stevekinney/electron,roadev/electron,jlord/electron,michaelchiche/electron,mubassirhayat/electron,simongregory/electron,cos2004/electron,MaxWhere/electron,soulteary/electron,cqqccqc/electron,michaelchiche/electron,vipulroxx/electron,lzpfmh/electron,IonicaBizauKitchen/electron,tomashanacek/electron,biblerule/UMCTelnetHub,aecca/electron,fritx/electron,pandoraui/electron,preco21/electron,zhakui/electron,edulan/electron,LadyNaggaga/electron,Zagorakiss/electron,kostia/electron,mhkeller/electron,Evercoder/electron,sky7sea/electron,bitemyapp/electron,felixrieseberg/electron,deed02392/electron,cos2004/electron,jaanus/electron,chriskdon/electron,deed02392/electron,jacksondc/electron,lzpfmh/electron,posix4e/electron,leftstick/electron,ankitaggarwal011/electron,jacksondc/electron,leolujuyi/electron,cqqccqc/electron,gamedevsam/electron,ankitaggarwal011/electron,brave/muon,howmuchcomputer/electron,dkfiresky/electron,iftekeriba/electron,Ivshti/electron,rprichard/electron,benweissmann/electron,icattlecoder/electron,bpasero/electron,Jacobichou/electron,jsutcodes/electron,astoilkov/electron,brave/electron,lrlna/electron,kikong/electron,kcrt/electron,Gerhut/electron,sshiting/electron,baiwyc119/electron,aichingm/electron,DivyaKMenon/electron,vHanda/electron,jaanus/electron,dongjoon-hyun/electron,rprichard/electron,vHanda/electron,greyhwndz/electron,medixdev/electron,aichingm/electron,chriskdon/electron,IonicaBizauKitchen/electron,fabien-d/electron,anko/electron,medixdev/electron,arusakov/electron,tinydew4/electron,xiruibing/electron,the-ress/electron
#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] Upgrade libchromiumcontent to loose iframe sandbox.
#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
<commit_before>#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] <commit_msg>Upgrade libchromiumcontent to loose iframe sandbox.<commit_after>
#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] Upgrade libchromiumcontent to loose iframe sandbox.#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
<commit_before>#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = '432720d4613e3aac939f127fe55b9d44fea349e5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] <commit_msg>Upgrade libchromiumcontent to loose iframe sandbox.<commit_after>#!/usr/bin/env python import platform import sys NODE_VERSION = 'v0.11.13' BASE_URL = 'https://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'afb4570ceee2ad10f3caf5a81335a2ee11ec68a5' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform]
5c0ace537a073f3d851ad4e490a7f2b5a0062c62
tfr/features.py
tfr/features.py
import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1) if __name__ == '__main__': import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1)
""" Example usage: import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) """ import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1)
Put the code into a comment as an example usage.
Put the code into a comment as an example usage.
Python
mit
bzamecnik/tfr,bzamecnik/tfr
import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1) if __name__ == '__main__': import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) Put the code into a comment as an example usage.
""" Example usage: import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) """ import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1)
<commit_before>import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1) if __name__ == '__main__': import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) <commit_msg>Put the code into a comment as an example usage.<commit_after>
""" Example usage: import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) """ import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1)
import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1) if __name__ == '__main__': import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) Put the code into a comment as an example usage.""" Example usage: import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) """ import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1)
<commit_before>import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1) if __name__ == '__main__': import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) <commit_msg>Put the code into a comment as an example usage.<commit_after>""" Example usage: import matplotlib.pyplot as plt from files import load_wav from analysis import split_to_blocks def analyze_mean_energy(file, block_size=1024): x, fs = load_wav(file) blocks, t = split_to_blocks(x, block_size) y = mean_energy(blocks) plt.semilogy(t, y) plt.ylim(0, 1) """ import numpy as np def mean_power(x_blocks): return np.sqrt(np.mean(x_blocks**2, axis=-1)) def power(x_blocks): return np.sqrt(np.sum(x_blocks**2, axis=-1)) def mean_energy(x_blocks): return np.mean(x_blocks**2, axis=-1) def energy(x_blocks): return np.sum(x_blocks**2, axis=-1)
744c91ca30379d6cca7f7f9fc2b014e0f29e55e4
keepaneyeon/http.py
keepaneyeon/http.py
import requests class HttpDownloader(): def __init__(self, opts={}): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk)
import requests class HttpDownloader(): def __init__(self, **opts): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk)
Make HTTPDownloader work with YAML config
Make HTTPDownloader work with YAML config
Python
mit
mmcloughlin/keepaneyeon
import requests class HttpDownloader(): def __init__(self, opts={}): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk) Make HTTPDownloader work with YAML config
import requests class HttpDownloader(): def __init__(self, **opts): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk)
<commit_before>import requests class HttpDownloader(): def __init__(self, opts={}): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk) <commit_msg>Make HTTPDownloader work with YAML config<commit_after>
import requests class HttpDownloader(): def __init__(self, **opts): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk)
import requests class HttpDownloader(): def __init__(self, opts={}): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk) Make HTTPDownloader work with YAML configimport requests class HttpDownloader(): def __init__(self, **opts): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk)
<commit_before>import requests class HttpDownloader(): def __init__(self, opts={}): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk) <commit_msg>Make HTTPDownloader work with YAML config<commit_after>import requests class HttpDownloader(): def __init__(self, **opts): self.base = opts def build_request_options(self, opts): options = {'method': 'get'} options.update(self.base) options.update(opts) options.update({'stream': True}) return options def download(self, opts, out): options = self.build_request_options(opts) r = requests.request(**options) r.raise_for_status() for chunk in r.iter_content(1024): out.write(chunk)
e3c53133b71d7426695fbf24cac5b8e82311c037
seeker/middleware.py
seeker/middleware.py
from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass
from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass
Make signal dispatch_uid values more specific
Make signal dispatch_uid values more specific
Python
bsd-2-clause
imsweb/django-seeker,imsweb/django-seeker
from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass Make signal dispatch_uid values more specific
from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass
<commit_before>from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass <commit_msg>Make signal dispatch_uid values more specific<commit_after>
from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass
from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass Make signal dispatch_uid values more specificfrom .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass
<commit_before>from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass <commit_msg>Make signal dispatch_uid values more specific<commit_after>from .utils import index, delete from django.db import models import logging logger = logging.getLogger(__name__) class ModelIndexingMiddleware (object): """ Middleware class that automatically indexes any new or deleted model objects. """ def __init__(self): models.signals.post_save.connect(self.handle_save, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_save') models.signals.post_delete.connect(self.handle_delete, dispatch_uid='seeker.middleware.ModelIndexingMiddleware.handle_delete') def handle_save(self, sender, instance, **kwargs): try: index(instance) except: logger.exception('Error indexing %s instance: %s', sender, instance) def handle_delete(self, sender, instance, **kwargs): try: delete(instance) except: logger.exception('Error deleting %s instance: %s', sender, instance) def process_request(self, request): # This is really just here so Django keeps the middleware installed. pass
56d92af9ba0a9b81dd0e802d05717ec6e4f511d3
seven23/api/views.py
seven23/api/views.py
""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json')
""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d") result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json')
Fix bug on API with date in Terms and Conditions not serializable
Fix bug on API with date in Terms and Conditions not serializable
Python
mit
sebastienbarbier/723e_server,sebastienbarbier/723e,sebastienbarbier/723e_server,sebastienbarbier/723e
""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json') Fix bug on API with date in Terms and Conditions not serializable
""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d") result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json')
<commit_before>""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json') <commit_msg>Fix bug on API with date in Terms and Conditions not serializable<commit_after>
""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d") result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json')
""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json') Fix bug on API with date in Terms and Conditions not serializable""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d") result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json')
<commit_before>""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json') <commit_msg>Fix bug on API with date in Terms and Conditions not serializable<commit_after>""" Root views of api """ import json import os import markdown2 from django.http import HttpResponse from django.db import models from rest_framework.decorators import api_view from seven23 import settings from seven23.models.terms.models import TermsAndConditions @api_view(["GET"]) def api_init(request): """ Return status on client initialisation """ result = {} # Return API Version. result['api_version'] = settings.API_VERSION result['allow_account_creation'] = settings.ALLOW_ACCOUNT_CREATION result['contact'] = settings.CONTACT_EMAIL try: terms = TermsAndConditions.objects.latest('date') result['terms_and_conditions_date'] = terms.date.strftime("%Y-%m-%d") result['terms_and_conditions'] = markdown2.markdown(terms.markdown) except TermsAndConditions.DoesNotExist: result['terms_and_conditions_date'] = None result['terms_and_conditions'] = None if request.user.is_authenticated(): result['is_authenticated'] = True result['id'] = request.user.id else: result['is_authenticated'] = False # Return json format string. j = json.dumps(result, separators=(',', ':')) return HttpResponse(j, content_type='application/json')
fecf53c0c4414f50a9c3937b05d27de8c1387c45
src/hireme/tasks/task2.py
src/hireme/tasks/task2.py
# -*- coding: utf-8 -*- from . import render_task @render_task def solve(): return dict( solution='42', title='task2' )
# -*- coding: utf-8 -*- import re from flask import request from werkzeug import exceptions import numpy as np from . import render_task @render_task def solve(): input_data = request.form.get('input') method = request.method title = 'task2' if method == 'GET': return dict( title=title ) lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')] cases = int(0 if not lines else lines.pop(0)) counts = [1] * cases for case in xrange(cases): if not len(lines): raise exceptions.BadRequest( description='Specified %s cases, but only provided %s.' % (cases, case) ) dimension = int(lines.pop(0)) matrix = np.array([[int(n) for n in m] for m in lines[:dimension]]) lines = lines[dimension:] if not matrix.shape == (dimension,) * 2: raise exceptions.BadRequest( description='Expected %s-dimensional matrix for case %s.' % (dimension, case + 1) ) def neighbours(idx): def along_axis(axis): for offset in (-1, 0, 1): candidate = idx[axis] + offset if candidate >= 0 and candidate < dimension: yield candidate hood = [] for x in along_axis(0): for y in along_axis(1): if (x, y) != idx and matrix[x, y] == 1: hood.append((x, y)) return hood untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) while untouched: def expand(resident): matrix[resident] = counts[case] hood = neighbours(resident) if hood: for neighbour in hood: expand(neighbour) counts[case] += 1 expand(untouched[0]) untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) solution = '\n'.join([str(c - 1) for c in counts]) return dict( input=input_data, solution=solution, title=title )
Implement rudimentary task 2 solution
Implement rudimentary task 2 solution
Python
bsd-2-clause
cutoffthetop/hireme
# -*- coding: utf-8 -*- from . import render_task @render_task def solve(): return dict( solution='42', title='task2' )Implement rudimentary task 2 solution
# -*- coding: utf-8 -*- import re from flask import request from werkzeug import exceptions import numpy as np from . import render_task @render_task def solve(): input_data = request.form.get('input') method = request.method title = 'task2' if method == 'GET': return dict( title=title ) lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')] cases = int(0 if not lines else lines.pop(0)) counts = [1] * cases for case in xrange(cases): if not len(lines): raise exceptions.BadRequest( description='Specified %s cases, but only provided %s.' % (cases, case) ) dimension = int(lines.pop(0)) matrix = np.array([[int(n) for n in m] for m in lines[:dimension]]) lines = lines[dimension:] if not matrix.shape == (dimension,) * 2: raise exceptions.BadRequest( description='Expected %s-dimensional matrix for case %s.' % (dimension, case + 1) ) def neighbours(idx): def along_axis(axis): for offset in (-1, 0, 1): candidate = idx[axis] + offset if candidate >= 0 and candidate < dimension: yield candidate hood = [] for x in along_axis(0): for y in along_axis(1): if (x, y) != idx and matrix[x, y] == 1: hood.append((x, y)) return hood untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) while untouched: def expand(resident): matrix[resident] = counts[case] hood = neighbours(resident) if hood: for neighbour in hood: expand(neighbour) counts[case] += 1 expand(untouched[0]) untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) solution = '\n'.join([str(c - 1) for c in counts]) return dict( input=input_data, solution=solution, title=title )
<commit_before># -*- coding: utf-8 -*- from . import render_task @render_task def solve(): return dict( solution='42', title='task2' )<commit_msg>Implement rudimentary task 2 solution<commit_after>
# -*- coding: utf-8 -*- import re from flask import request from werkzeug import exceptions import numpy as np from . import render_task @render_task def solve(): input_data = request.form.get('input') method = request.method title = 'task2' if method == 'GET': return dict( title=title ) lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')] cases = int(0 if not lines else lines.pop(0)) counts = [1] * cases for case in xrange(cases): if not len(lines): raise exceptions.BadRequest( description='Specified %s cases, but only provided %s.' % (cases, case) ) dimension = int(lines.pop(0)) matrix = np.array([[int(n) for n in m] for m in lines[:dimension]]) lines = lines[dimension:] if not matrix.shape == (dimension,) * 2: raise exceptions.BadRequest( description='Expected %s-dimensional matrix for case %s.' % (dimension, case + 1) ) def neighbours(idx): def along_axis(axis): for offset in (-1, 0, 1): candidate = idx[axis] + offset if candidate >= 0 and candidate < dimension: yield candidate hood = [] for x in along_axis(0): for y in along_axis(1): if (x, y) != idx and matrix[x, y] == 1: hood.append((x, y)) return hood untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) while untouched: def expand(resident): matrix[resident] = counts[case] hood = neighbours(resident) if hood: for neighbour in hood: expand(neighbour) counts[case] += 1 expand(untouched[0]) untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) solution = '\n'.join([str(c - 1) for c in counts]) return dict( input=input_data, solution=solution, title=title )
# -*- coding: utf-8 -*- from . import render_task @render_task def solve(): return dict( solution='42', title='task2' )Implement rudimentary task 2 solution# -*- coding: utf-8 -*- import re from flask import request from werkzeug import exceptions import numpy as np from . import render_task @render_task def solve(): input_data = request.form.get('input') method = request.method title = 'task2' if method == 'GET': return dict( title=title ) lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')] cases = int(0 if not lines else lines.pop(0)) counts = [1] * cases for case in xrange(cases): if not len(lines): raise exceptions.BadRequest( description='Specified %s cases, but only provided %s.' % (cases, case) ) dimension = int(lines.pop(0)) matrix = np.array([[int(n) for n in m] for m in lines[:dimension]]) lines = lines[dimension:] if not matrix.shape == (dimension,) * 2: raise exceptions.BadRequest( description='Expected %s-dimensional matrix for case %s.' % (dimension, case + 1) ) def neighbours(idx): def along_axis(axis): for offset in (-1, 0, 1): candidate = idx[axis] + offset if candidate >= 0 and candidate < dimension: yield candidate hood = [] for x in along_axis(0): for y in along_axis(1): if (x, y) != idx and matrix[x, y] == 1: hood.append((x, y)) return hood untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) while untouched: def expand(resident): matrix[resident] = counts[case] hood = neighbours(resident) if hood: for neighbour in hood: expand(neighbour) counts[case] += 1 expand(untouched[0]) untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) solution = '\n'.join([str(c - 1) for c in counts]) return dict( input=input_data, solution=solution, title=title )
<commit_before># -*- coding: utf-8 -*- from . import render_task @render_task def solve(): return dict( solution='42', title='task2' )<commit_msg>Implement rudimentary task 2 solution<commit_after># -*- coding: utf-8 -*- import re from flask import request from werkzeug import exceptions import numpy as np from . import render_task @render_task def solve(): input_data = request.form.get('input') method = request.method title = 'task2' if method == 'GET': return dict( title=title ) lines = [re.sub('[^0-9]+', '', i) for i in input_data.split('\n')] cases = int(0 if not lines else lines.pop(0)) counts = [1] * cases for case in xrange(cases): if not len(lines): raise exceptions.BadRequest( description='Specified %s cases, but only provided %s.' % (cases, case) ) dimension = int(lines.pop(0)) matrix = np.array([[int(n) for n in m] for m in lines[:dimension]]) lines = lines[dimension:] if not matrix.shape == (dimension,) * 2: raise exceptions.BadRequest( description='Expected %s-dimensional matrix for case %s.' % (dimension, case + 1) ) def neighbours(idx): def along_axis(axis): for offset in (-1, 0, 1): candidate = idx[axis] + offset if candidate >= 0 and candidate < dimension: yield candidate hood = [] for x in along_axis(0): for y in along_axis(1): if (x, y) != idx and matrix[x, y] == 1: hood.append((x, y)) return hood untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) while untouched: def expand(resident): matrix[resident] = counts[case] hood = neighbours(resident) if hood: for neighbour in hood: expand(neighbour) counts[case] += 1 expand(untouched[0]) untouched = zip(*[idx.tolist() for idx in np.where(matrix == 1)]) solution = '\n'.join([str(c - 1) for c in counts]) return dict( input=input_data, solution=solution, title=title )
84041a2bb517841d725781bdd72b1daf4f8e603d
spacy/ja/__init__.py
spacy/ja/__init__.py
# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language from ..attrs import LANG from ..tokens import Doc from .language_data import * class Japanese(Language): lang = 'ja' def make_doc(self, text): try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") words = [x.surface for x in Tokenizer().tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words))
# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language, BaseDefaults from ..tokenizer import Tokenizer from ..attrs import LANG from ..tokens import Doc from .language_data import * class JapaneseTokenizer(object): def __init__(self, cls, nlp=None): self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp) try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") self.tokenizer = Tokenizer() def __call__(self, text): words = [x.surface for x in self.tokenizer.tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) class JapaneseDefaults(BaseDefaults): @classmethod def create_tokenizer(cls, nlp=None): return JapaneseTokenizer(cls, nlp) class Japanese(Language): lang = 'ja' Defaults = JapaneseDefaults def make_doc(self, text): words = self.tokenizer(text) return Doc(self.vocab, words=words, spaces=[False]*len(words))
Make create_tokenizer work with Japanese
Make create_tokenizer work with Japanese
Python
mit
spacy-io/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,raphael0202/spaCy,aikramer2/spaCy,honnibal/spaCy,spacy-io/spaCy,raphael0202/spaCy,raphael0202/spaCy,raphael0202/spaCy,spacy-io/spaCy,spacy-io/spaCy,explosion/spaCy,raphael0202/spaCy,honnibal/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy,raphael0202/spaCy,recognai/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,spacy-io/spaCy,aikramer2/spaCy,honnibal/spaCy
# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language from ..attrs import LANG from ..tokens import Doc from .language_data import * class Japanese(Language): lang = 'ja' def make_doc(self, text): try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") words = [x.surface for x in Tokenizer().tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) Make create_tokenizer work with Japanese
# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language, BaseDefaults from ..tokenizer import Tokenizer from ..attrs import LANG from ..tokens import Doc from .language_data import * class JapaneseTokenizer(object): def __init__(self, cls, nlp=None): self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp) try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") self.tokenizer = Tokenizer() def __call__(self, text): words = [x.surface for x in self.tokenizer.tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) class JapaneseDefaults(BaseDefaults): @classmethod def create_tokenizer(cls, nlp=None): return JapaneseTokenizer(cls, nlp) class Japanese(Language): lang = 'ja' Defaults = JapaneseDefaults def make_doc(self, text): words = self.tokenizer(text) return Doc(self.vocab, words=words, spaces=[False]*len(words))
<commit_before># encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language from ..attrs import LANG from ..tokens import Doc from .language_data import * class Japanese(Language): lang = 'ja' def make_doc(self, text): try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") words = [x.surface for x in Tokenizer().tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) <commit_msg>Make create_tokenizer work with Japanese<commit_after>
# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language, BaseDefaults from ..tokenizer import Tokenizer from ..attrs import LANG from ..tokens import Doc from .language_data import * class JapaneseTokenizer(object): def __init__(self, cls, nlp=None): self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp) try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") self.tokenizer = Tokenizer() def __call__(self, text): words = [x.surface for x in self.tokenizer.tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) class JapaneseDefaults(BaseDefaults): @classmethod def create_tokenizer(cls, nlp=None): return JapaneseTokenizer(cls, nlp) class Japanese(Language): lang = 'ja' Defaults = JapaneseDefaults def make_doc(self, text): words = self.tokenizer(text) return Doc(self.vocab, words=words, spaces=[False]*len(words))
# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language from ..attrs import LANG from ..tokens import Doc from .language_data import * class Japanese(Language): lang = 'ja' def make_doc(self, text): try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") words = [x.surface for x in Tokenizer().tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) Make create_tokenizer work with Japanese# encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language, BaseDefaults from ..tokenizer import Tokenizer from ..attrs import LANG from ..tokens import Doc from .language_data import * class JapaneseTokenizer(object): def __init__(self, cls, nlp=None): self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp) try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") self.tokenizer = Tokenizer() def __call__(self, text): words = [x.surface for x in self.tokenizer.tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) class JapaneseDefaults(BaseDefaults): @classmethod def create_tokenizer(cls, nlp=None): return JapaneseTokenizer(cls, nlp) class Japanese(Language): lang = 'ja' Defaults = JapaneseDefaults def make_doc(self, text): words = self.tokenizer(text) return Doc(self.vocab, words=words, spaces=[False]*len(words))
<commit_before># encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language from ..attrs import LANG from ..tokens import Doc from .language_data import * class Japanese(Language): lang = 'ja' def make_doc(self, text): try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") words = [x.surface for x in Tokenizer().tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) <commit_msg>Make create_tokenizer work with Japanese<commit_after># encoding: utf8 from __future__ import unicode_literals, print_function from os import path from ..language import Language, BaseDefaults from ..tokenizer import Tokenizer from ..attrs import LANG from ..tokens import Doc from .language_data import * class JapaneseTokenizer(object): def __init__(self, cls, nlp=None): self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp) try: from janome.tokenizer import Tokenizer except ImportError: raise ImportError("The Japanese tokenizer requires the Janome library: " "https://github.com/mocobeta/janome") self.tokenizer = Tokenizer() def __call__(self, text): words = [x.surface for x in self.tokenizer.tokenize(text)] return Doc(self.vocab, words=words, spaces=[False]*len(words)) class JapaneseDefaults(BaseDefaults): @classmethod def create_tokenizer(cls, nlp=None): return JapaneseTokenizer(cls, nlp) class Japanese(Language): lang = 'ja' Defaults = JapaneseDefaults def make_doc(self, text): words = self.tokenizer(text) return Doc(self.vocab, words=words, spaces=[False]*len(words))
cec423c4a1e633193ef3e639a1cb202bddc27e18
api/base/content_negotiation.py
api/base/content_negotiation.py
from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. Returns a tuple (renderer, media_type). """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ Returns appropriate tuple (renderer, media type). If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
Add one-line summary to docstring.
Add one-line summary to docstring.
Python
apache-2.0
hmoco/osf.io,doublebits/osf.io,leb2dg/osf.io,doublebits/osf.io,erinspace/osf.io,billyhunt/osf.io,jnayak1/osf.io,felliott/osf.io,sbt9uc/osf.io,MerlinZhang/osf.io,caneruguz/osf.io,ZobairAlijan/osf.io,RomanZWang/osf.io,pattisdr/osf.io,RomanZWang/osf.io,zachjanicki/osf.io,kwierman/osf.io,ticklemepierce/osf.io,icereval/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,CenterForOpenScience/osf.io,leb2dg/osf.io,erinspace/osf.io,ticklemepierce/osf.io,wearpants/osf.io,billyhunt/osf.io,binoculars/osf.io,Nesiehr/osf.io,ZobairAlijan/osf.io,HalcyonChimera/osf.io,SSJohns/osf.io,kch8qx/osf.io,acshi/osf.io,brandonPurvis/osf.io,TomHeatwole/osf.io,zachjanicki/osf.io,hmoco/osf.io,DanielSBrown/osf.io,acshi/osf.io,MerlinZhang/osf.io,chennan47/osf.io,KAsante95/osf.io,alexschiller/osf.io,samanehsan/osf.io,mfraezz/osf.io,Ghalko/osf.io,chennan47/osf.io,sbt9uc/osf.io,zamattiac/osf.io,rdhyee/osf.io,cosenal/osf.io,caseyrygt/osf.io,caseyrollins/osf.io,abought/osf.io,jnayak1/osf.io,amyshi188/osf.io,acshi/osf.io,TomBaxter/osf.io,mluo613/osf.io,binoculars/osf.io,doublebits/osf.io,Nesiehr/osf.io,danielneis/osf.io,GageGaskins/osf.io,DanielSBrown/osf.io,cslzchen/osf.io,erinspace/osf.io,cslzchen/osf.io,Ghalko/osf.io,danielneis/osf.io,SSJohns/osf.io,cslzchen/osf.io,asanfilippo7/osf.io,saradbowman/osf.io,kwierman/osf.io,alexschiller/osf.io,icereval/osf.io,laurenrevere/osf.io,samanehsan/osf.io,abought/osf.io,mluke93/osf.io,ckc6cz/osf.io,monikagrabowska/osf.io,RomanZWang/osf.io,chrisseto/osf.io,zamattiac/osf.io,mfraezz/osf.io,caseyrollins/osf.io,samchrisinger/osf.io,petermalcolm/osf.io,arpitar/osf.io,icereval/osf.io,adlius/osf.io,ckc6cz/osf.io,mluo613/osf.io,amyshi188/osf.io,abought/osf.io,samchrisinger/osf.io,sloria/osf.io,haoyuchen1992/osf.io,HalcyonChimera/osf.io,cslzchen/osf.io,cosenal/osf.io,samanehsan/osf.io,jnayak1/osf.io,SSJohns/osf.io,sbt9uc/osf.io,aaxelb/osf.io,felliott/osf.io,SSJohns/osf.io,GageGaskins/osf.io,binoculars/osf.io,kch8qx/osf.io,jmcarp/osf.io,sloria/osf.io,chrisseto/osf.io,monikagrabowska/osf.io,monikagrabowska/osf.io,KAsante95/osf.io,chrisseto/osf.io,Ghalko/osf.io,CenterForOpenScience/osf.io,GageGaskins/osf.io,haoyuchen1992/osf.io,asanfilippo7/osf.io,chrisseto/osf.io,amyshi188/osf.io,brianjgeiger/osf.io,caseyrygt/osf.io,wearpants/osf.io,leb2dg/osf.io,cwisecarver/osf.io,arpitar/osf.io,danielneis/osf.io,doublebits/osf.io,monikagrabowska/osf.io,pattisdr/osf.io,wearpants/osf.io,brianjgeiger/osf.io,DanielSBrown/osf.io,emetsger/osf.io,njantrania/osf.io,aaxelb/osf.io,zachjanicki/osf.io,caneruguz/osf.io,mattclark/osf.io,chennan47/osf.io,aaxelb/osf.io,saradbowman/osf.io,mfraezz/osf.io,GageGaskins/osf.io,samchrisinger/osf.io,ticklemepierce/osf.io,caneruguz/osf.io,petermalcolm/osf.io,TomHeatwole/osf.io,crcresearch/osf.io,alexschiller/osf.io,njantrania/osf.io,CenterForOpenScience/osf.io,rdhyee/osf.io,baylee-d/osf.io,cosenal/osf.io,alexschiller/osf.io,caseyrygt/osf.io,acshi/osf.io,GageGaskins/osf.io,ckc6cz/osf.io,samchrisinger/osf.io,danielneis/osf.io,adlius/osf.io,asanfilippo7/osf.io,felliott/osf.io,jmcarp/osf.io,amyshi188/osf.io,zamattiac/osf.io,adlius/osf.io,arpitar/osf.io,alexschiller/osf.io,rdhyee/osf.io,KAsante95/osf.io,ZobairAlijan/osf.io,Johnetordoff/osf.io,Nesiehr/osf.io,billyhunt/osf.io,sbt9uc/osf.io,wearpants/osf.io,emetsger/osf.io,brianjgeiger/osf.io,kch8qx/osf.io,rdhyee/osf.io,caseyrygt/osf.io,billyhunt/osf.io,felliott/osf.io,mattclark/osf.io,kwierman/osf.io,Nesiehr/osf.io,KAsante95/osf.io,laurenrevere/osf.io,ckc6cz/osf.io,DanielSBrown/osf.io,brandonPurvis/osf.io,baylee-d/osf.io,hmoco/osf.io,haoyuchen1992/osf.io,RomanZWang/osf.io,samanehsan/osf.io,petermalcolm/osf.io,mluke93/osf.io,hmoco/osf.io,cwisecarver/osf.io,caneruguz/osf.io,ticklemepierce/osf.io,Johnetordoff/osf.io,jnayak1/osf.io,haoyuchen1992/osf.io,CenterForOpenScience/osf.io,kwierman/osf.io,cwisecarver/osf.io,doublebits/osf.io,sloria/osf.io,mattclark/osf.io,Johnetordoff/osf.io,mluo613/osf.io,MerlinZhang/osf.io,brandonPurvis/osf.io,KAsante95/osf.io,petermalcolm/osf.io,MerlinZhang/osf.io,ZobairAlijan/osf.io,jmcarp/osf.io,asanfilippo7/osf.io,brianjgeiger/osf.io,cwisecarver/osf.io,monikagrabowska/osf.io,HalcyonChimera/osf.io,mluke93/osf.io,njantrania/osf.io,brandonPurvis/osf.io,cosenal/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,laurenrevere/osf.io,Johnetordoff/osf.io,mluo613/osf.io,pattisdr/osf.io,njantrania/osf.io,billyhunt/osf.io,kch8qx/osf.io,mluo613/osf.io,mluke93/osf.io,TomHeatwole/osf.io,Ghalko/osf.io,caseyrollins/osf.io,arpitar/osf.io,jmcarp/osf.io,emetsger/osf.io,zamattiac/osf.io,adlius/osf.io,kch8qx/osf.io,TomBaxter/osf.io,baylee-d/osf.io,abought/osf.io,leb2dg/osf.io,acshi/osf.io,mfraezz/osf.io,emetsger/osf.io,aaxelb/osf.io,RomanZWang/osf.io,crcresearch/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io
from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. Returns a tuple (renderer, media_type). """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers) Add one-line summary to docstring.
from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ Returns appropriate tuple (renderer, media type). If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
<commit_before> from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. Returns a tuple (renderer, media_type). """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers) <commit_msg>Add one-line summary to docstring.<commit_after>
from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ Returns appropriate tuple (renderer, media type). If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. Returns a tuple (renderer, media_type). """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers) Add one-line summary to docstring. from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ Returns appropriate tuple (renderer, media type). If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
<commit_before> from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. Returns a tuple (renderer, media_type). """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers) <commit_msg>Add one-line summary to docstring.<commit_after> from rest_framework.negotiation import DefaultContentNegotiation class JSONAPIContentNegotiation(DefaultContentNegotiation): def select_renderer(self, request, renderers, format_suffix=None): """ Returns appropriate tuple (renderer, media type). If 'application/json' in acceptable media types, use the first renderer in DEFAULT_RENDERER_CLASSES which should be 'api.base.renderers.JSONAPIRenderer'. Media_type "application/vnd.api+json". Otherwise, use default select_renderer. """ accepts = self.get_accept_list(request) if 'application/json' in accepts: return (renderers[0], renderers[0].media_type) return super(JSONAPIContentNegotiation, self).select_renderer(request, renderers)
2604d759bfd9a18e5e594cfa5b50e83c73fbc2d8
devito/interfaces.py
devito/interfaces.py
import numpy as np from sympy import IndexedBase class MatrixData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(MatrixData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): MatrixData._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order]
import numpy as np from sympy import IndexedBase class DenseData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(DenseData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): super(TimeData, self)._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order]
Change name from MatrixData to DenseData
Change name from MatrixData to DenseData
Python
mit
opesci/devito,opesci/devito
import numpy as np from sympy import IndexedBase class MatrixData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(MatrixData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): MatrixData._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order] Change name from MatrixData to DenseData
import numpy as np from sympy import IndexedBase class DenseData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(DenseData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): super(TimeData, self)._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order]
<commit_before>import numpy as np from sympy import IndexedBase class MatrixData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(MatrixData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): MatrixData._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order] <commit_msg>Change name from MatrixData to DenseData<commit_after>
import numpy as np from sympy import IndexedBase class DenseData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(DenseData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): super(TimeData, self)._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order]
import numpy as np from sympy import IndexedBase class MatrixData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(MatrixData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): MatrixData._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order] Change name from MatrixData to DenseDataimport numpy as np from sympy import IndexedBase class DenseData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(DenseData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): super(TimeData, self)._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order]
<commit_before>import numpy as np from sympy import IndexedBase class MatrixData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(MatrixData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): MatrixData._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order] <commit_msg>Change name from MatrixData to DenseData<commit_after>import numpy as np from sympy import IndexedBase class DenseData(IndexedBase): def __init__(self, name, shape, dtype): self.name = name self.shape = shape self.dtype = dtype self.pointer = None self.initializer = None def set_initializer(self, lambda_initializer): assert(callable(lambda_initializer)) self.initializer = lambda_initializer def initialize(self): assert(self.initializer is not None) self.initializer(self.data) def _allocate_memory(self): self.pointer = np.zeros(self.shape, self.dtype, order='C') @property def data(self): if self.pointer is None: self._allocate_memory() return self.pointer class TimeData(DenseData): def __init__(self, name, spc_shape, time_dim, time_order, save, dtype): if save: time_dim = time_dim + time_order else: time_dim = time_order + 1 shape = tuple((time_dim,) + spc_shape) super(TimeData, self).__init__(name, shape, dtype) self.save = save self.time_order = time_order def _allocate_memory(self): super(TimeData, self)._allocate_memory(self) if self.pad_time: self.pointer = self.pointer[self.time_order]
840d80c543c4688ebd1bda41b8689cf404bf755c
edit_spectide.py
edit_spectide.py
""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av)
""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. Lee's changeNC file should live in the SVN repository. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av)
Update help to indicate where the necessary script lives
Update help to indicate where the necessary script lives
Python
mit
pwcazenave/PyFVCOM
""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av) Update help to indicate where the necessary script lives
""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. Lee's changeNC file should live in the SVN repository. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av)
<commit_before>""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av) <commit_msg>Update help to indicate where the necessary script lives<commit_after>
""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. Lee's changeNC file should live in the SVN repository. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av)
""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av) Update help to indicate where the necessary script lives""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. Lee's changeNC file should live in the SVN repository. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av)
<commit_before>""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av) <commit_msg>Update help to indicate where the necessary script lives<commit_after>""" Edits the spectide amplitude values to some factor of their original value. WARNING: When using this on FVCOM input files, it will change the format of the variables. changeNC presumes each variable has a value and unit associated with it, whereas some of the variables in the FVCOM inputs are in fact not that sort of data, and so have different structures. Probably best to use the combination of ncdump and ncgen to edit the values as text and generate a new NetCDF file from that edited text. Lee's changeNC file should live in the SVN repository. """ from changeNC import * infile = './co2_spectide.nc' outfile = './co2_spectide_scaled.nc' scaleFact = 0.75 av = AutoVivification() av['tide_Eamp']['convert'] = lambda x:x*scaleFact changeNC(infile, outfile, av)
127b90c88d1362e7b10e7bf36dff56b96a5c4f0b
simpegEM/FDEM/__init__.py
simpegEM/FDEM/__init__.py
from SurveyFDEM import * from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h
from SurveyFDEM import * from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
Add more files to export on the init.
Add more files to export on the init.
Python
mit
simpeg/discretize,lheagy/simpegem,simpeg/discretize,simpeg/discretize,simpeg/simpeg,simpeg/simpegem
from SurveyFDEM import * from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h Add more files to export on the init.
from SurveyFDEM import * from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
<commit_before>from SurveyFDEM import * from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h <commit_msg>Add more files to export on the init.<commit_after>
from SurveyFDEM import * from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
from SurveyFDEM import * from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h Add more files to export on the init.from SurveyFDEM import * from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
<commit_before>from SurveyFDEM import * from FDEM import ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h <commit_msg>Add more files to export on the init.<commit_after>from SurveyFDEM import * from FDEM import BaseFDEMProblem, ProblemFDEM_e, ProblemFDEM_b, ProblemFDEM_j, ProblemFDEM_h, omega
8ef4ca2166167f6370dd8c2f724e752210adf067
sirius/SI_V07/__init__.py
sirius/SI_V07/__init__.py
from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
Fix bug when family_data.py was deleted
Fix bug when family_data.py was deleted
Python
mit
lnls-fac/sirius
from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar] Fix bug when family_data.py was deleted
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
<commit_before>from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar] <commit_msg>Fix bug when family_data.py was deleted<commit_after>
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar] Fix bug when family_data.py was deletedfrom . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
<commit_before>from . import lattice as _lattice from . import accelerator as _accelerator from . import family_data as _family_data from . import record_names create_accelerator = _accelerator.create_accelerator get_family_data = _family_data.get_family_data # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar] <commit_msg>Fix bug when family_data.py was deleted<commit_after>from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = _accelerator.create_accelerator # -- default accelerator values for SI_V07 -- energy = _lattice._energy harmonic_number = _lattice._harmonic_number default_cavity_on = _accelerator._default_cavity_on default_radiation_on = _accelerator._default_cavity_on default_vchamber_on = _accelerator._default_vchamber_on default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'SI_V07' lattice_symmetry = _lattice._lattice_symmetry family_data = _lattice._family_data family_mapping = _lattice._family_mapping global_coupling = 0.01 # expected corrected value average_pressure = 1.333e-9 # average pressure [mbar]
55072134b8053ac126213e580fcc59977cfb7a02
scikits/image/setup.py
scikits/image/setup.py
import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config)
import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') config.add_subpackage('draw') config.add_subpackage('feature') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config)
Add 'draw' and 'feature' sub-modules.
BUG: Add 'draw' and 'feature' sub-modules.
Python
bsd-3-clause
paalge/scikit-image,michaelaye/scikit-image,bennlich/scikit-image,chintak/scikit-image,paalge/scikit-image,ClinicalGraphics/scikit-image,warmspringwinds/scikit-image,michaelaye/scikit-image,chriscrosscutler/scikit-image,ClinicalGraphics/scikit-image,WarrenWeckesser/scikits-image,emmanuelle/scikits.image,Midafi/scikit-image,emmanuelle/scikits.image,WarrenWeckesser/scikits-image,bennlich/scikit-image,GaZ3ll3/scikit-image,michaelpacer/scikit-image,SamHames/scikit-image,vighneshbirodkar/scikit-image,almarklein/scikit-image,rjeli/scikit-image,blink1073/scikit-image,jwiggins/scikit-image,emmanuelle/scikits.image,emon10005/scikit-image,SamHames/scikit-image,warmspringwinds/scikit-image,almarklein/scikit-image,almarklein/scikit-image,robintw/scikit-image,keflavich/scikit-image,Midafi/scikit-image,ajaybhat/scikit-image,ajaybhat/scikit-image,dpshelio/scikit-image,emmanuelle/scikits.image,youprofit/scikit-image,jwiggins/scikit-image,SamHames/scikit-image,emon10005/scikit-image,ofgulban/scikit-image,paalge/scikit-image,chintak/scikit-image,newville/scikit-image,juliusbierk/scikit-image,keflavich/scikit-image,Hiyorimi/scikit-image,Hiyorimi/scikit-image,chintak/scikit-image,bsipocz/scikit-image,newville/scikit-image,robintw/scikit-image,vighneshbirodkar/scikit-image,Britefury/scikit-image,youprofit/scikit-image,chintak/scikit-image,oew1v07/scikit-image,blink1073/scikit-image,ofgulban/scikit-image,dpshelio/scikit-image,pratapvardhan/scikit-image,michaelpacer/scikit-image,pratapvardhan/scikit-image,chriscrosscutler/scikit-image,vighneshbirodkar/scikit-image,bsipocz/scikit-image,ofgulban/scikit-image,juliusbierk/scikit-image,Britefury/scikit-image,rjeli/scikit-image,GaZ3ll3/scikit-image,rjeli/scikit-image,SamHames/scikit-image,almarklein/scikit-image,oew1v07/scikit-image
import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config) BUG: Add 'draw' and 'feature' sub-modules.
import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') config.add_subpackage('draw') config.add_subpackage('feature') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config)
<commit_before>import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config) <commit_msg>BUG: Add 'draw' and 'feature' sub-modules.<commit_after>
import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') config.add_subpackage('draw') config.add_subpackage('feature') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config)
import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config) BUG: Add 'draw' and 'feature' sub-modules.import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') config.add_subpackage('draw') config.add_subpackage('feature') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config)
<commit_before>import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config) <commit_msg>BUG: Add 'draw' and 'feature' sub-modules.<commit_after>import os def configuration(parent_package='', top_path=None): from numpy.distutils.misc_util import Configuration config = Configuration('image', parent_package, top_path) config.add_subpackage('opencv') config.add_subpackage('graph') config.add_subpackage('io') config.add_subpackage('morphology') config.add_subpackage('filter') config.add_subpackage('transform') config.add_subpackage('data') config.add_subpackage('util') config.add_subpackage('color') config.add_subpackage('draw') config.add_subpackage('feature') def add_test_directories(arg, dirname, fnames): if dirname.split(os.path.sep)[-1] == 'tests': config.add_data_dir(dirname) # Add test directories from os.path import isdir, dirname, join, abspath rel_isdir = lambda d: isdir(join(curpath, d)) curpath = join(dirname(__file__), './') subdirs = [join(d, 'tests') for d in os.listdir(curpath) if rel_isdir(d)] subdirs = [d for d in subdirs if rel_isdir(d)] for test_dir in subdirs: config.add_data_dir(test_dir) return config if __name__ == "__main__": from numpy.distutils.core import setup config = Configuration(top_path='').todict() setup(**config)
7eadc9e514b1311409356f4c6c40ef8cdb2de809
manager/__init__.py
manager/__init__.py
import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', output='css_all.css' ) )
import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', 'fontawesome/css/font-awesome.min.css', 'admin-lte/dist/css/AdminLTE.css', 'admin-lte/dist/css/skins/skin-black.min.css', output='css_all.css' ) )
Add new stuff to the css bundle
Add new stuff to the css bundle
Python
mit
hreeder/ignition,hreeder/ignition,hreeder/ignition
import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', output='css_all.css' ) ) Add new stuff to the css bundle
import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', 'fontawesome/css/font-awesome.min.css', 'admin-lte/dist/css/AdminLTE.css', 'admin-lte/dist/css/skins/skin-black.min.css', output='css_all.css' ) )
<commit_before>import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', output='css_all.css' ) ) <commit_msg>Add new stuff to the css bundle<commit_after>
import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', 'fontawesome/css/font-awesome.min.css', 'admin-lte/dist/css/AdminLTE.css', 'admin-lte/dist/css/skins/skin-black.min.css', output='css_all.css' ) )
import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', output='css_all.css' ) ) Add new stuff to the css bundleimport os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', 'fontawesome/css/font-awesome.min.css', 'admin-lte/dist/css/AdminLTE.css', 'admin-lte/dist/css/skins/skin-black.min.css', output='css_all.css' ) )
<commit_before>import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', output='css_all.css' ) ) <commit_msg>Add new stuff to the css bundle<commit_after>import os from flask import Flask from flask.ext.assets import Bundle, Environment from flask.ext.bcrypt import Bcrypt from flask.ext.login import LoginManager, current_user from flask.ext.migrate import Migrate from flask.ext.sqlalchemy import SQLAlchemy app = Flask(__name__) # Load the app config app.config.from_object("config.Config") assets = Environment(app) bcrypt = Bcrypt(app) db = SQLAlchemy(app) login_manager = LoginManager(app) migrate = Migrate(app, db) # Load Blueprints from manager.core import core from manager.dns import dns app.register_blueprint(core) app.register_blueprint(dns, url_prefix="/dns") # Configure flask-login login_manager.login_view = "core.login" # Asset Management assets.load_path = [ os.path.join(os.path.dirname(__file__), 'static'), os.path.join(os.path.dirname(__file__), 'static', 'bower') ] assets.register( 'js_all', Bundle( 'jquery/dist/jquery.min.js', 'bootstrap/dist/js/bootstrap.min.js', output='js_all.js' ) ) assets.register( 'css_all', Bundle( 'bootstrap/dist/css/bootstrap.css', 'fontawesome/css/font-awesome.min.css', 'admin-lte/dist/css/AdminLTE.css', 'admin-lte/dist/css/skins/skin-black.min.css', output='css_all.css' ) )
29ac3073b747d5bafaec240df25844d6d27c049a
marshmallow/base.py
marshmallow/base.py
# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj): raise NotImplementedError def _deserialize(self, value, attr, data): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError
# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj, **kwargs): raise NotImplementedError def _deserialize(self, value, attr, data, **kwargs): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError
Update signatures of FieldABC methods
Update signatures of FieldABC methods
Python
mit
marshmallow-code/marshmallow,mwstobo/marshmallow
# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj): raise NotImplementedError def _deserialize(self, value, attr, data): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError Update signatures of FieldABC methods
# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj, **kwargs): raise NotImplementedError def _deserialize(self, value, attr, data, **kwargs): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError
<commit_before># -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj): raise NotImplementedError def _deserialize(self, value, attr, data): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError <commit_msg>Update signatures of FieldABC methods<commit_after>
# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj, **kwargs): raise NotImplementedError def _deserialize(self, value, attr, data, **kwargs): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError
# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj): raise NotImplementedError def _deserialize(self, value, attr, data): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError Update signatures of FieldABC methods# -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj, **kwargs): raise NotImplementedError def _deserialize(self, value, attr, data, **kwargs): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError
<commit_before># -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj): raise NotImplementedError def _deserialize(self, value, attr, data): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError <commit_msg>Update signatures of FieldABC methods<commit_after># -*- coding: utf-8 -*- """Abstract base classes. These are necessary to avoid circular imports between core.py and fields.py. """ import copy class FieldABC(object): """Abstract base class from which all Field classes inherit. """ parent = None name = None def serialize(self, attr, obj, accessor=None): raise NotImplementedError def deserialize(self, value): raise NotImplementedError def _serialize(self, value, attr, obj, **kwargs): raise NotImplementedError def _deserialize(self, value, attr, data, **kwargs): raise NotImplementedError def __deepcopy__(self, memo): ret = copy.copy(self) return ret class SchemaABC(object): """Abstract base class from which all Schemas inherit.""" def dump(self, obj): raise NotImplementedError def dumps(self, obj, *args, **kwargs): raise NotImplementedError def load(self, data): raise NotImplementedError def loads(self, data): raise NotImplementedError
9da01f39c8d9b73025d85be72b71399b6930b6fb
src/encoded/cache.py
src/encoded/cache.py
from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return cached = cache.get(key) if cached is not None: return cached[1] return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value
from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return default try: return cache[key] except KeyError: return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value
Use LRUCache correctly (minimal improvement)
Use LRUCache correctly (minimal improvement)
Python
mit
ENCODE-DCC/encoded,kidaa/encoded,hms-dbmi/fourfront,philiptzou/clincoded,ENCODE-DCC/snovault,ENCODE-DCC/snovault,hms-dbmi/fourfront,philiptzou/clincoded,ENCODE-DCC/snovault,4dn-dcic/fourfront,ClinGen/clincoded,ClinGen/clincoded,4dn-dcic/fourfront,ENCODE-DCC/snovault,4dn-dcic/fourfront,ENCODE-DCC/encoded,T2DREAM/t2dream-portal,4dn-dcic/fourfront,philiptzou/clincoded,kidaa/encoded,hms-dbmi/fourfront,ENCODE-DCC/snovault,ENCODE-DCC/encoded,kidaa/encoded,ENCODE-DCC/encoded,hms-dbmi/fourfront,T2DREAM/t2dream-portal,T2DREAM/t2dream-portal,hms-dbmi/fourfront,ClinGen/clincoded,kidaa/encoded,kidaa/encoded,ClinGen/clincoded,ClinGen/clincoded,T2DREAM/t2dream-portal,philiptzou/clincoded,philiptzou/clincoded
from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return cached = cache.get(key) if cached is not None: return cached[1] return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value Use LRUCache correctly (minimal improvement)
from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return default try: return cache[key] except KeyError: return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value
<commit_before>from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return cached = cache.get(key) if cached is not None: return cached[1] return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value <commit_msg>Use LRUCache correctly (minimal improvement)<commit_after>
from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return default try: return cache[key] except KeyError: return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value
from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return cached = cache.get(key) if cached is not None: return cached[1] return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value Use LRUCache correctly (minimal improvement)from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return default try: return cache[key] except KeyError: return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value
<commit_before>from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return cached = cache.get(key) if cached is not None: return cached[1] return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value <commit_msg>Use LRUCache correctly (minimal improvement)<commit_after>from pyramid.threadlocal import manager from sqlalchemy.util import LRUCache class ManagerLRUCache(object): """ Override capacity in settings. """ def __init__(self, name, default_capacity=100, threshold=.5): self.name = name self.default_capacity = default_capacity self.threshold = threshold @property def cache(self): if not manager.stack: return None threadlocals = manager.stack[0] if self.name not in threadlocals: registry = threadlocals['registry'] capacity = int(registry.settings.get(self.name + '.capacity', self.default_capacity)) threadlocals[self.name] = LRUCache(capacity, self.threshold) return threadlocals[self.name] def get(self, key, default=None): cache = self.cache if cache is None: return default try: return cache[key] except KeyError: return default def __contains__(self, key): cache = self.cache if cache is None: return False return key in cache def __setitem__(self, key, value): cache = self.cache if cache is None: return self.cache[key] = value
a6b39dde09777ff162fbf83976934cbf2ec14056
app.py
app.py
from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:15]: node = "{}".format(get_hash(item)) G.add_node(node) commit = repo.commit(get_hash(item)) for parent in commit.parents: G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run()
from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:20]: node = "{}".format(get_hash(item)) commit = repo.commit(get_hash(item)) G.add_node(node, message=commit.message.split("\n")[0]) for parent in commit.parents: G.add_node(parent.hexsha, message=parent.message.split("\n")[0]) G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run()
Add commit message for commits and for parents
Add commit message for commits and for parents
Python
bsd-3-clause
kdheepak89/c3.py,kdheepak89/c3.py
from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:15]: node = "{}".format(get_hash(item)) G.add_node(node) commit = repo.commit(get_hash(item)) for parent in commit.parents: G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run() Add commit message for commits and for parents
from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:20]: node = "{}".format(get_hash(item)) commit = repo.commit(get_hash(item)) G.add_node(node, message=commit.message.split("\n")[0]) for parent in commit.parents: G.add_node(parent.hexsha, message=parent.message.split("\n")[0]) G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run()
<commit_before>from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:15]: node = "{}".format(get_hash(item)) G.add_node(node) commit = repo.commit(get_hash(item)) for parent in commit.parents: G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run() <commit_msg>Add commit message for commits and for parents<commit_after>
from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:20]: node = "{}".format(get_hash(item)) commit = repo.commit(get_hash(item)) G.add_node(node, message=commit.message.split("\n")[0]) for parent in commit.parents: G.add_node(parent.hexsha, message=parent.message.split("\n")[0]) G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run()
from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:15]: node = "{}".format(get_hash(item)) G.add_node(node) commit = repo.commit(get_hash(item)) for parent in commit.parents: G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run() Add commit message for commits and for parentsfrom flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:20]: node = "{}".format(get_hash(item)) commit = repo.commit(get_hash(item)) G.add_node(node, message=commit.message.split("\n")[0]) for parent in commit.parents: G.add_node(parent.hexsha, message=parent.message.split("\n")[0]) G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run()
<commit_before>from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:15]: node = "{}".format(get_hash(item)) G.add_node(node) commit = repo.commit(get_hash(item)) for parent in commit.parents: G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run() <commit_msg>Add commit message for commits and for parents<commit_after>from flask import Flask from flask import render_template app = Flask(__name__) app.config['DEBUG'] = True repo_path = '../ames-py' @app.route("/") def main(): return render_template("index.html") @app.route("/data") def data(): import json import git import networkx as nx G = nx.DiGraph() parent = None g = git.Git(repo_path) repo = git.Repo(repo_path) loginfo = g.log() def get_hash(commit): return commit.split("\n")[0] commits = loginfo.split("\n\ncommit ") commits[0] = commits[0].replace("commit ", '') for item in commits[:20]: node = "{}".format(get_hash(item)) commit = repo.commit(get_hash(item)) G.add_node(node, message=commit.message.split("\n")[0]) for parent in commit.parents: G.add_node(parent.hexsha, message=parent.message.split("\n")[0]) G.add_edge(node, parent.hexsha) pos=nx.graphviz_layout(G, prog='dot') from networkx.readwrite import json_graph data = json_graph.node_link_data(G) for node in data['nodes']: node['pos'] = pos[node['id']] j = json.dumps(data) return(j) if __name__ == "__main__": app.run()
169ca5581c6c35d07dd772baf3119f45ba1c1e2e
app.py
app.py
from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() app.run(debug=True)
from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() logging.getLogger().setLevel(logging.DEBUG) app.run(debug=True)
Make all log messages show by default
Make all log messages show by default
Python
mit
karlalopez/Authentise-Store,addendumauto/Authentise-Store,addendumauto/Authentise-Store,addendumauto/Authentise-Store
from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() app.run(debug=True) Make all log messages show by default
from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() logging.getLogger().setLevel(logging.DEBUG) app.run(debug=True)
<commit_before>from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() app.run(debug=True) <commit_msg>Make all log messages show by default<commit_after>
from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() logging.getLogger().setLevel(logging.DEBUG) app.run(debug=True)
from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() app.run(debug=True) Make all log messages show by defaultfrom flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() logging.getLogger().setLevel(logging.DEBUG) app.run(debug=True)
<commit_before>from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() app.run(debug=True) <commit_msg>Make all log messages show by default<commit_after>from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from werkzeug import secure_filename import os import logging import stripe APP_ROOT = os.path.dirname(os.path.abspath(__file__)) UPLOAD_FOLDER = os.path.join('static/uploads') MODELS_FOLDER = os.path.join('models') ALLOWED_EXTENSIONS = set(['stl']) stripe_keys = { 'secret_key': os.environ['SECRET_KEY'], 'publishable_key': os.environ['PUBLISHABLE_KEY'] } stripe.api_key = stripe_keys['secret_key'] shop_name = "Shop name" shop_tagline = "Best shop tagline ever" app = Flask(__name__) app.secret_key = 'thisisasecret' #You need to set up an app secret key. app.config['UPLOAD_FOLDER'] = UPLOAD_FOLDER app.config['MODELS_FOLDER'] = MODELS_FOLDER # Set up the SQLAlchemy Database to be a local file 'store.db' app.config['SQLALCHEMY_DATABASE_URI'] = 'postgresql://localhost/store' db = SQLAlchemy(app) if __name__ == "__main__": from views import * del session logging.basicConfig() logging.getLogger().setLevel(logging.DEBUG) app.run(debug=True)
d560a809c4d0fd78e1ce0454ea5406e81f356906
server_app/__main__.py
server_app/__main__.py
import sys import os import logging if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port)
import sys import os import logging import time if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port)
Make logger sort by date
Make logger sort by date
Python
bsd-3-clause
jos0003/Chat,jos0003/Chat,jos0003/Chat,jos0003/Chat,jos0003/Chat
import sys import os import logging if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port) Make logger sort by date
import sys import os import logging import time if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port)
<commit_before>import sys import os import logging if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port) <commit_msg>Make logger sort by date<commit_after>
import sys import os import logging import time if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port)
import sys import os import logging if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port) Make logger sort by dateimport sys import os import logging import time if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port)
<commit_before>import sys import os import logging if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port) <commit_msg>Make logger sort by date<commit_after>import sys import os import logging import time if not os.path.exists(os.path.expanduser("~/.chatserver")): os.makedirs(os.path.expanduser("~/.chatserver")) logging.basicConfig(filename=os.path.expanduser("~/.chatserver/chat-"+time.strftime("%d-%m-%Y.log"), level=logging.DEBUG) sys.stderr.close() sys.stdout.close() sys.stdin.close() from app import app, db, main, socketio db.create_all() app.register_blueprint(main) port = app.config['PORT'] if len(sys.argv) == 2: port = int(sys.argv[1]) logging.info("Chat server is now running on 0.0.0.0:%r" % port) socketio.run(app, host="0.0.0.0", port=port)
0438825672a407eb30bff49e03dac89a0534f28a
minimax.py
minimax.py
class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): pass class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color)
class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): raise NotImplementedError('Dont override this class') class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) def calculate_min_or_max(self, vector_values, color): if color == self.me.color: return self.me.eval(vector_values) else: return self.challenger.eval(vector_values)
Create in MinMax the calculate_min_max
Create in MinMax the calculate_min_max
Python
apache-2.0
frila/agente-minimax
class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): pass class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) Create in MinMax the calculate_min_max
class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): raise NotImplementedError('Dont override this class') class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) def calculate_min_or_max(self, vector_values, color): if color == self.me.color: return self.me.eval(vector_values) else: return self.challenger.eval(vector_values)
<commit_before>class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): pass class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) <commit_msg>Create in MinMax the calculate_min_max<commit_after>
class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): raise NotImplementedError('Dont override this class') class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) def calculate_min_or_max(self, vector_values, color): if color == self.me.color: return self.me.eval(vector_values) else: return self.challenger.eval(vector_values)
class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): pass class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) Create in MinMax the calculate_min_maxclass Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): raise NotImplementedError('Dont override this class') class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) def calculate_min_or_max(self, vector_values, color): if color == self.me.color: return self.me.eval(vector_values) else: return self.challenger.eval(vector_values)
<commit_before>class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): pass class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) <commit_msg>Create in MinMax the calculate_min_max<commit_after>class Heuristic: def __init__(self, color): self.color = color def heuristic(self, board, color): raise NotImplementedError('Dont override this class') def eval(self, vector): raise NotImplementedError('Dont override this class') class Minimax: def __init__(self, me, challenger): self.me = me self.challenger = challenger def heuristic(self, board, color): if color == self.color_me: return self.me.heuristic(board, color) else return self.challenger.heuristic(board, color) def calculate_min_or_max(self, vector_values, color): if color == self.me.color: return self.me.eval(vector_values) else: return self.challenger.eval(vector_values)
3b3a8dc6aa0b38cfbb68105eb5ef31e8e73ff3a4
gcm_flask/application/models.py
gcm_flask/application/models.py
""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty() sent_at = db.DateTimeProperty(auto_now_add=True)
""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty(auto_current_user=True) sent_at = db.DateTimeProperty(auto_now_add=True)
Update user who sent the message
Update user who sent the message
Python
apache-2.0
BarcampBangalore/Barcamp-Bangalore-Android-App,BarcampBangalore/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App,BarcampBangalore/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App,BarcampBangalore/Barcamp-Bangalore-Android-App,rajeefmk/Barcamp-Bangalore-Android-App
""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty() sent_at = db.DateTimeProperty(auto_now_add=True)Update user who sent the message
""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty(auto_current_user=True) sent_at = db.DateTimeProperty(auto_now_add=True)
<commit_before>""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty() sent_at = db.DateTimeProperty(auto_now_add=True)<commit_msg>Update user who sent the message<commit_after>
""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty(auto_current_user=True) sent_at = db.DateTimeProperty(auto_now_add=True)
""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty() sent_at = db.DateTimeProperty(auto_now_add=True)Update user who sent the message""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty(auto_current_user=True) sent_at = db.DateTimeProperty(auto_now_add=True)
<commit_before>""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty() sent_at = db.DateTimeProperty(auto_now_add=True)<commit_msg>Update user who sent the message<commit_after>""" models.py App Engine datastore models """ from google.appengine.ext import db class ExampleModel(db.Model): """Example Model""" example_name = db.StringProperty(required=True) example_description = db.TextProperty(required=True) added_by = db.UserProperty() timestamp = db.DateTimeProperty(auto_now_add=True) class RegIDModel(db.Model): """Regl IDs Model""" regID = db.StringProperty(required=True) class MessagesModel(db.Model): """Model for storing messages sent""" message = db.StringProperty(required=True) messagetype = db.StringProperty(required=True) added_by = db.UserProperty(auto_current_user=True) sent_at = db.DateTimeProperty(auto_now_add=True)
c04b8932ec65480ba90dd4578d5f6bb8c3baa690
demosys/project/default.py
demosys/project/default.py
from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) self.effect = self.create_effect('default', cls.__name__)
from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): if self.effect_class_name: cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) else: effect_package = effects.get_package(self.effect_package_name) runnable_effects = effect_package.runnable_effects() if not runnable_effects: raise ValueError("No runnable effects found in effect package", self.effect_package_name) cls = runnable_effects[-1] self.effect = self.create_effect('default', cls.__name__)
Improve errors when effect packages or effects are not found
Improve errors when effect packages or effects are not found
Python
isc
Contraz/demosys-py
from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) self.effect = self.create_effect('default', cls.__name__) Improve errors when effect packages or effects are not found
from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): if self.effect_class_name: cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) else: effect_package = effects.get_package(self.effect_package_name) runnable_effects = effect_package.runnable_effects() if not runnable_effects: raise ValueError("No runnable effects found in effect package", self.effect_package_name) cls = runnable_effects[-1] self.effect = self.create_effect('default', cls.__name__)
<commit_before>from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) self.effect = self.create_effect('default', cls.__name__) <commit_msg>Improve errors when effect packages or effects are not found<commit_after>
from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): if self.effect_class_name: cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) else: effect_package = effects.get_package(self.effect_package_name) runnable_effects = effect_package.runnable_effects() if not runnable_effects: raise ValueError("No runnable effects found in effect package", self.effect_package_name) cls = runnable_effects[-1] self.effect = self.create_effect('default', cls.__name__)
from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) self.effect = self.create_effect('default', cls.__name__) Improve errors when effect packages or effects are not foundfrom demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): if self.effect_class_name: cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) else: effect_package = effects.get_package(self.effect_package_name) runnable_effects = effect_package.runnable_effects() if not runnable_effects: raise ValueError("No runnable effects found in effect package", self.effect_package_name) cls = runnable_effects[-1] self.effect = self.create_effect('default', cls.__name__)
<commit_before>from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) self.effect = self.create_effect('default', cls.__name__) <commit_msg>Improve errors when effect packages or effects are not found<commit_after>from demosys.project.base import BaseProject from demosys.effects.registry import effects, parse_package_string class Project(BaseProject): """ The project what will be assigned when no project are specified. This is mainly used when the ``runeffect`` command is used """ def __init__(self, effect_package): super().__init__() self.path = effect_package self.effect_package_name, self.effect_class_name = parse_package_string(effect_package) self.effect_packages = [self.effect_package_name] self.effect = None def get_default_effect(self): return self.effect def create_resources(self): pass def create_effect_instances(self): if self.effect_class_name: cls = effects.find_effect_class(self.path) if not cls.runnable: raise ValueError("Effect doesn't have the runnable flag set:", self.path) else: effect_package = effects.get_package(self.effect_package_name) runnable_effects = effect_package.runnable_effects() if not runnable_effects: raise ValueError("No runnable effects found in effect package", self.effect_package_name) cls = runnable_effects[-1] self.effect = self.create_effect('default', cls.__name__)
891e8afe5deff5fe7d620abfe8189689d47ec4f8
djangocms_inherit/forms.py
djangocms_inherit/forms.py
from django import forms from django.forms.models import ModelForm from django.forms.utils import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data
from django import forms from django.forms.models import ModelForm try: from django.forms.utils import ErrorList except ImportError: # Django<1.7 (deprecated in Django 1.8, removed in 1.9) from django.forms.util import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data
Make import backward compatible (Django<1.7)
Make import backward compatible (Django<1.7)
Python
bsd-3-clause
bittner/djangocms-inherit,bittner/djangocms-inherit,divio/djangocms-inherit,divio/djangocms-inherit,divio/djangocms-inherit
from django import forms from django.forms.models import ModelForm from django.forms.utils import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data Make import backward compatible (Django<1.7)
from django import forms from django.forms.models import ModelForm try: from django.forms.utils import ErrorList except ImportError: # Django<1.7 (deprecated in Django 1.8, removed in 1.9) from django.forms.util import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data
<commit_before>from django import forms from django.forms.models import ModelForm from django.forms.utils import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data <commit_msg>Make import backward compatible (Django<1.7)<commit_after>
from django import forms from django.forms.models import ModelForm try: from django.forms.utils import ErrorList except ImportError: # Django<1.7 (deprecated in Django 1.8, removed in 1.9) from django.forms.util import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data
from django import forms from django.forms.models import ModelForm from django.forms.utils import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data Make import backward compatible (Django<1.7)from django import forms from django.forms.models import ModelForm try: from django.forms.utils import ErrorList except ImportError: # Django<1.7 (deprecated in Django 1.8, removed in 1.9) from django.forms.util import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data
<commit_before>from django import forms from django.forms.models import ModelForm from django.forms.utils import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data <commit_msg>Make import backward compatible (Django<1.7)<commit_after>from django import forms from django.forms.models import ModelForm try: from django.forms.utils import ErrorList except ImportError: # Django<1.7 (deprecated in Django 1.8, removed in 1.9) from django.forms.util import ErrorList from django.utils.translation import ugettext_lazy as _ from cms.models import Page from .models import InheritPagePlaceholder class InheritForm(ModelForm): from_page = forms.ModelChoiceField( label=_("page"), queryset=Page.objects.drafts(), required=False) class Meta: model = InheritPagePlaceholder exclude = ('page', 'position', 'placeholder', 'language', 'plugin_type') def for_site(self, site): # override the page_link fields queryset to containt just pages for # current site self.fields['from_page'].queryset = Page.objects.drafts().on_site(site) def clean(self): cleaned_data = super(InheritForm, self).clean() if not cleaned_data['from_page'] and not cleaned_data['from_language']: self._errors['from_page'] = ErrorList( [_("Language or Page must be filled out")]) return cleaned_data
db59332e3d522c68c3eeef77fe4393fe137e5059
inspectors/registration/util.py
inspectors/registration/util.py
import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result is not None
import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result
Fix logic bug for API result
Fix logic bug for API result
Python
bsd-3-clause
codeforamerica/mdc-inspectors,codeforamerica/mdc-inspectors,codeforamerica/mdc-inspectors
import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result is not None Fix logic bug for API result
import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result
<commit_before>import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result is not None <commit_msg>Fix logic bug for API result<commit_after>
import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result
import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result is not None Fix logic bug for API resultimport requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result
<commit_before>import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result is not None <commit_msg>Fix logic bug for API result<commit_after>import requests API_URL = 'https://opendata.miamidade.gov/resource/vvjq-pfmc.json' def is_valid_permit(id): # checks if the ID is a valid Miami-Dade Permit or Process Number API = API_URL + '?$where=permit_number=%27' + id + '%27%20or%20process_number=%27' + id + '%27' response = requests.get(API) json_result = response.json() return json_result
09fa23adfb76f052473ee38de94ce4bdfdcc48e1
src/nodeconductor_assembly_waldur/packages/perms.py
src/nodeconductor_assembly_waldur/packages/perms.py
from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)), )
from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic from nodeconductor.structure import models as structure_models PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic( collaborators_query=[ 'tenant__service_project_link__service__customer__roles__permission_group__user', ], collaborators_filter=[ {'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER}, ], any_permission=True, )), )
Allow customer owner to create packages
Allow customer owner to create packages - wal-26
Python
mit
opennode/nodeconductor-assembly-waldur,opennode/nodeconductor-assembly-waldur,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/nodeconductor-assembly-waldur
from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)), ) Allow customer owner to create packages - wal-26
from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic from nodeconductor.structure import models as structure_models PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic( collaborators_query=[ 'tenant__service_project_link__service__customer__roles__permission_group__user', ], collaborators_filter=[ {'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER}, ], any_permission=True, )), )
<commit_before>from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)), ) <commit_msg>Allow customer owner to create packages - wal-26<commit_after>
from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic from nodeconductor.structure import models as structure_models PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic( collaborators_query=[ 'tenant__service_project_link__service__customer__roles__permission_group__user', ], collaborators_filter=[ {'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER}, ], any_permission=True, )), )
from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)), ) Allow customer owner to create packages - wal-26from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic from nodeconductor.structure import models as structure_models PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic( collaborators_query=[ 'tenant__service_project_link__service__customer__roles__permission_group__user', ], collaborators_filter=[ {'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER}, ], any_permission=True, )), )
<commit_before>from nodeconductor.core.permissions import StaffPermissionLogic PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', StaffPermissionLogic(any_permission=True)), ) <commit_msg>Allow customer owner to create packages - wal-26<commit_after>from nodeconductor.core.permissions import StaffPermissionLogic, FilteredCollaboratorsPermissionLogic from nodeconductor.structure import models as structure_models PERMISSION_LOGICS = ( ('packages.PackageTemplate', StaffPermissionLogic(any_permission=True)), ('packages.PackageComponent', StaffPermissionLogic(any_permission=True)), ('packages.OpenStackPackage', FilteredCollaboratorsPermissionLogic( collaborators_query=[ 'tenant__service_project_link__service__customer__roles__permission_group__user', ], collaborators_filter=[ {'tenant__service_project_link__service__customer__roles__role_type': structure_models.CustomerRole.OWNER}, ], any_permission=True, )), )
75d1241c5d62def89a7377e506afdacfa83cbdda
js2xml/parser.py
js2xml/parser.py
import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=False, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, write_tables=False, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result
import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=True, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result
Write tab but don't warn
Write tab but don't warn
Python
mit
redapple/js2xml,redapple/js2xml,redapple/js2xml,redapple/js2xml
import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=False, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, write_tables=False, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result Write tab but don't warn
import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=True, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result
<commit_before>import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=False, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, write_tables=False, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result <commit_msg>Write tab but don't warn<commit_after>
import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=True, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result
import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=False, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, write_tables=False, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result Write tab but don't warnimport ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=True, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result
<commit_before>import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=False, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, write_tables=False, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result <commit_msg>Write tab but don't warn<commit_after>import ply.yacc from slimit.parser import Parser from js2xml.lexer import CustomLexer as Lexer from js2xml.log import logger lextab, yacctab = 'lextab', 'yacctab' class CustomParser(Parser): def __init__(self, lex_optimize=True, lextab=lextab, yacc_optimize=True, yacctab=yacctab, yacc_debug=False, logger=logger): self.lex_optimize = lex_optimize self.lextab = lextab self.yacc_optimize = yacc_optimize self.yacctab = yacctab self.yacc_debug = yacc_debug self.lexer = Lexer() self.lexer.build(optimize=lex_optimize, lextab=lextab, errorlog=logger) self.tokens = self.lexer.tokens self.parser = ply.yacc.yacc( module=self, optimize=yacc_optimize, debug=yacc_debug, tabmodule=yacctab, start='program', errorlog=logger) self._error_tokens = {} def parse(self, text, debug=False): result = super(CustomParser, self).parse(text, debug=debug) self._error_tokens = {} return result
cc2b579377abde262d76e2484a6488e254b315fc
judge/caching.py
judge/caching.py
from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete('sub_stats_table') cache.delete('sub_stats_data') def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): cache.delete('user_complete:%d' % sub.user_id) if hasattr(sub, 'contest'): participation = sub.contest.participation cache.delete('contest_complete:%d' % participation.id) cache.delete(make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id)))
from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete_many(('sub_stats_table', 'sub_stats_data')) def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): keys = ['user_complete:%d' % sub.user_id] if hasattr(sub, 'contest'): participation = sub.contest.participation keys += ['contest_complete:%d' % participation.id, make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id))] cache.delete_many(keys)
Delete many to reduce round trips to the cache.
Delete many to reduce round trips to the cache.
Python
agpl-3.0
Minkov/site,Minkov/site,DMOJ/site,Minkov/site,DMOJ/site,DMOJ/site,Phoenix1369/site,Phoenix1369/site,Minkov/site,monouno/site,Phoenix1369/site,monouno/site,monouno/site,monouno/site,monouno/site,DMOJ/site,Phoenix1369/site
from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete('sub_stats_table') cache.delete('sub_stats_data') def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): cache.delete('user_complete:%d' % sub.user_id) if hasattr(sub, 'contest'): participation = sub.contest.participation cache.delete('contest_complete:%d' % participation.id) cache.delete(make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id)))Delete many to reduce round trips to the cache.
from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete_many(('sub_stats_table', 'sub_stats_data')) def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): keys = ['user_complete:%d' % sub.user_id] if hasattr(sub, 'contest'): participation = sub.contest.participation keys += ['contest_complete:%d' % participation.id, make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id))] cache.delete_many(keys)
<commit_before>from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete('sub_stats_table') cache.delete('sub_stats_data') def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): cache.delete('user_complete:%d' % sub.user_id) if hasattr(sub, 'contest'): participation = sub.contest.participation cache.delete('contest_complete:%d' % participation.id) cache.delete(make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id)))<commit_msg>Delete many to reduce round trips to the cache.<commit_after>
from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete_many(('sub_stats_table', 'sub_stats_data')) def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): keys = ['user_complete:%d' % sub.user_id] if hasattr(sub, 'contest'): participation = sub.contest.participation keys += ['contest_complete:%d' % participation.id, make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id))] cache.delete_many(keys)
from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete('sub_stats_table') cache.delete('sub_stats_data') def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): cache.delete('user_complete:%d' % sub.user_id) if hasattr(sub, 'contest'): participation = sub.contest.participation cache.delete('contest_complete:%d' % participation.id) cache.delete(make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id)))Delete many to reduce round trips to the cache.from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete_many(('sub_stats_table', 'sub_stats_data')) def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): keys = ['user_complete:%d' % sub.user_id] if hasattr(sub, 'contest'): participation = sub.contest.participation keys += ['contest_complete:%d' % participation.id, make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id))] cache.delete_many(keys)
<commit_before>from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete('sub_stats_table') cache.delete('sub_stats_data') def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): cache.delete('user_complete:%d' % sub.user_id) if hasattr(sub, 'contest'): participation = sub.contest.participation cache.delete('contest_complete:%d' % participation.id) cache.delete(make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id)))<commit_msg>Delete many to reduce round trips to the cache.<commit_after>from django.core.cache import cache from django.core.cache.utils import make_template_fragment_key def update_submission(id): key = 'version:submission-%d' % id cache.add(key, 0, None) cache.incr(key) def update_stats(): cache.delete_many(('sub_stats_table', 'sub_stats_data')) def point_update(profile): cache.delete(make_template_fragment_key('global_user')) def finished_submission(sub): keys = ['user_complete:%d' % sub.user_id] if hasattr(sub, 'contest'): participation = sub.contest.participation keys += ['contest_complete:%d' % participation.id, make_template_fragment_key('conrank_user_prob', (participation.profile.user_id, participation.contest_id))] cache.delete_many(keys)
05adb44cdec74256fa44ce3a3df61c6525ce7fac
dryscrape/xvfb.py
dryscrape/xvfb.py
import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb if "DISPLAY" in os.environ: del os.environ["DISPLAY"] _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop()
import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop()
Remove removal of DISPLAY environment variable
Remove removal of DISPLAY environment variable The issue has to do with the two lines: ` if "DISPLAY" in os.environ: del os.environ["DISPLAY"]` This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.
Python
mit
niklasb/dryscrape
import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb if "DISPLAY" in os.environ: del os.environ["DISPLAY"] _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop() Remove removal of DISPLAY environment variable The issue has to do with the two lines: ` if "DISPLAY" in os.environ: del os.environ["DISPLAY"]` This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.
import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop()
<commit_before>import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb if "DISPLAY" in os.environ: del os.environ["DISPLAY"] _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop() <commit_msg>Remove removal of DISPLAY environment variable The issue has to do with the two lines: ` if "DISPLAY" in os.environ: del os.environ["DISPLAY"]` This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.<commit_after>
import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop()
import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb if "DISPLAY" in os.environ: del os.environ["DISPLAY"] _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop() Remove removal of DISPLAY environment variable The issue has to do with the two lines: ` if "DISPLAY" in os.environ: del os.environ["DISPLAY"]` This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop()
<commit_before>import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb if "DISPLAY" in os.environ: del os.environ["DISPLAY"] _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop() <commit_msg>Remove removal of DISPLAY environment variable The issue has to do with the two lines: ` if "DISPLAY" in os.environ: del os.environ["DISPLAY"]` This seems to remove the DISPLAY environment variable unnecessarily, as on line 50 of xvfbwrapper.py, self.orig_display is set to the value of DISPLAY. self.orig_display is checked on line 83, which is where the error occurs. Because of xvfb.py removing the environment variable and self.orig_display being set to the original value, on line 84 when it tries to remove DISPLAY, it has already been removed by xvfb.py, so it throws a KeyError.<commit_after>import atexit import os _xvfb = None def start_xvfb(): from xvfbwrapper import Xvfb global _xvfb _xvfb = Xvfb() _xvfb.start() atexit.register(_xvfb.stop) def stop_xvfb(): global _xvfb _xvfb.stop()
382c46366c4ae29456aae35420990ce25b61ce76
studygroups/tasks.py
studygroups/tasks.py
from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update()
from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update()
Add check for failed reminders so that it doesn't case an Exception
Add check for failed reminders so that it doesn't case an Exception
Python
mit
p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles,p2pu/learning-circles
from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update() Add check for failed reminders so that it doesn't case an Exception
from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update()
<commit_before>from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update() <commit_msg>Add check for failed reminders so that it doesn't case an Exception<commit_after>
from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update()
from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update() Add check for failed reminders so that it doesn't case an Exceptionfrom __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update()
<commit_before>from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update() <commit_msg>Add check for failed reminders so that it doesn't case an Exception<commit_after>from __future__ import absolute_import from celery import shared_task from django.utils import timezone from django.conf import settings from studygroups.models import StudyGroup from studygroups.models import Reminder from studygroups.models import generate_reminder from studygroups.models import send_reminder from studygroups.models import send_weekly_update from django.utils import translation import datetime @shared_task def send_reminders(): now = timezone.now() translation.activate(settings.LANGUAGE_CODE) # TODO - should this be set here or closer to where the language matters? for reminder in Reminder.objects.filter(sent_at__isnull=True): if reminder.study_group_meeting and reminder.study_group_meeting.meeting_time - now < datetime.timedelta(days=2): send_reminder(reminder) @shared_task def gen_reminders(): for study_group in StudyGroup.objects.all(): translation.activate(settings.LANGUAGE_CODE) generate_reminder(study_group) @shared_task def weekly_update(): # Create a report for the previous week send_weekly_update()
4613daea5d9d603b5f092005627fabd805de8a45
example/app/utils.py
example/app/utils.py
from django.contrib.auth import get_user_model def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission
from django.contrib.auth import get_user_model from django.db.utils import ProgrammingError def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() try: user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) except ProgrammingError: # auth_user doesn't exist, this allows the migrations to run properly. user = None def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission
Make initial ./manage.py migrate work in example
Make initial ./manage.py migrate work in example
Python
bsd-3-clause
zostera/django-modeltrans,zostera/django-modeltrans
from django.contrib.auth import get_user_model def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission Make initial ./manage.py migrate work in example
from django.contrib.auth import get_user_model from django.db.utils import ProgrammingError def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() try: user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) except ProgrammingError: # auth_user doesn't exist, this allows the migrations to run properly. user = None def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission
<commit_before>from django.contrib.auth import get_user_model def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission <commit_msg>Make initial ./manage.py migrate work in example<commit_after>
from django.contrib.auth import get_user_model from django.db.utils import ProgrammingError def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() try: user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) except ProgrammingError: # auth_user doesn't exist, this allows the migrations to run properly. user = None def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission
from django.contrib.auth import get_user_model def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission Make initial ./manage.py migrate work in examplefrom django.contrib.auth import get_user_model from django.db.utils import ProgrammingError def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() try: user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) except ProgrammingError: # auth_user doesn't exist, this allows the migrations to run properly. user = None def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission
<commit_before>from django.contrib.auth import get_user_model def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission <commit_msg>Make initial ./manage.py migrate work in example<commit_after>from django.contrib.auth import get_user_model from django.db.utils import ProgrammingError def disable_admin_login(): """ Disable admin login, but allow editing. amended from: https://stackoverflow.com/a/40008282/517560 """ User = get_user_model() try: user, created = User.objects.update_or_create( id=1, defaults=dict( first_name="Default Admin", last_name="User", is_superuser=True, is_active=True, is_staff=True, ), ) except ProgrammingError: # auth_user doesn't exist, this allows the migrations to run properly. user = None def no_login_has_permission(request): setattr(request, "user", user) return True return no_login_has_permission
187447322d74fc3070c9a75415a55a56ba840ef9
extruct/jsonld.py
extruct/jsonld.py
# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') # now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded # TODO: `strict=False` can be configurable if needed data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data]
# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') try: # TODO: `strict=False` can be configurable if needed data = json.loads(script, strict=False) except ValueError: # sometimes JSON-decoding errors are due to leading HTML or JavaScript comments data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data]
Make comment removal a fallback when failed.
Mod: Make comment removal a fallback when failed.
Python
bsd-3-clause
scrapinghub/extruct
# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') # now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded # TODO: `strict=False` can be configurable if needed data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data] Mod: Make comment removal a fallback when failed.
# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') try: # TODO: `strict=False` can be configurable if needed data = json.loads(script, strict=False) except ValueError: # sometimes JSON-decoding errors are due to leading HTML or JavaScript comments data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data]
<commit_before># -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') # now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded # TODO: `strict=False` can be configurable if needed data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data] <commit_msg>Mod: Make comment removal a fallback when failed.<commit_after>
# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') try: # TODO: `strict=False` can be configurable if needed data = json.loads(script, strict=False) except ValueError: # sometimes JSON-decoding errors are due to leading HTML or JavaScript comments data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data]
# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') # now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded # TODO: `strict=False` can be configurable if needed data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data] Mod: Make comment removal a fallback when failed.# -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') try: # TODO: `strict=False` can be configurable if needed data = json.loads(script, strict=False) except ValueError: # sometimes JSON-decoding errors are due to leading HTML or JavaScript comments data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data]
<commit_before># -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') # now do remove possible leading HTML/JavaScript comment first, allow control characters to be loaded # TODO: `strict=False` can be configurable if needed data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data] <commit_msg>Mod: Make comment removal a fallback when failed.<commit_after># -*- coding: utf-8 -*- """ JSON-LD extractor """ import json import re import lxml.etree import lxml.html HTML_OR_JS_COMMENTLINE = re.compile('^\s*(//.*|<!--.*-->)') class JsonLdExtractor(object): _xp_jsonld = lxml.etree.XPath('descendant-or-self::script[@type="application/ld+json"]') def extract(self, htmlstring, base_url=None, encoding="UTF-8"): parser = lxml.html.HTMLParser(encoding=encoding) lxmldoc = lxml.html.fromstring(htmlstring, parser=parser) return self.extract_items(lxmldoc, base_url=base_url) def extract_items(self, document, base_url=None): return [item for items in map(self._extract_items, self._xp_jsonld(document)) for item in items if item] def _extract_items(self, node): script = node.xpath('string()') try: # TODO: `strict=False` can be configurable if needed data = json.loads(script, strict=False) except ValueError: # sometimes JSON-decoding errors are due to leading HTML or JavaScript comments data = json.loads(HTML_OR_JS_COMMENTLINE.sub('', script), strict=False) if isinstance(data, list): return data elif isinstance(data, dict): return [data]
118aa612ef088dba90328f1775d8603ee12e5e5b
main.py
main.py
import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = map(lambda x: normalize(x, min_fitness, max_fitness), pop_fitness) total = sum(normalized) return map(lambda x: x/total, normalized) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=1) evolve()
import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = list( map( lambda x: normalize(x, min_fitness, max_fitness), pop_fitness ) ) total = sum(normalized) return list(map(lambda x: x/total, normalized)) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=20) evolve()
Set log level to info, fixed bug with map object
Set log level to info, fixed bug with map object
Python
mit
ray-dino/robby-genetic-algorithm
import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = map(lambda x: normalize(x, min_fitness, max_fitness), pop_fitness) total = sum(normalized) return map(lambda x: x/total, normalized) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=1) evolve() Set log level to info, fixed bug with map object
import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = list( map( lambda x: normalize(x, min_fitness, max_fitness), pop_fitness ) ) total = sum(normalized) return list(map(lambda x: x/total, normalized)) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=20) evolve()
<commit_before>import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = map(lambda x: normalize(x, min_fitness, max_fitness), pop_fitness) total = sum(normalized) return map(lambda x: x/total, normalized) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=1) evolve() <commit_msg>Set log level to info, fixed bug with map object<commit_after>
import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = list( map( lambda x: normalize(x, min_fitness, max_fitness), pop_fitness ) ) total = sum(normalized) return list(map(lambda x: x/total, normalized)) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=20) evolve()
import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = map(lambda x: normalize(x, min_fitness, max_fitness), pop_fitness) total = sum(normalized) return map(lambda x: x/total, normalized) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=1) evolve() Set log level to info, fixed bug with map objectimport logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = list( map( lambda x: normalize(x, min_fitness, max_fitness), pop_fitness ) ) total = sum(normalized) return list(map(lambda x: x/total, normalized)) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=20) evolve()
<commit_before>import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = map(lambda x: normalize(x, min_fitness, max_fitness), pop_fitness) total = sum(normalized) return map(lambda x: x/total, normalized) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=1) evolve() <commit_msg>Set log level to info, fixed bug with map object<commit_after>import logging import numpy as np import settings from models import Robby def evolve(): population = np.array([Robby() for i in range(0, settings.POPULATION)]) for gen in range(0, settings.GENERATIONS): for individual in population: individual.live() new_population = list() while len(new_population)<settings.POPULATION: father, mother = np.random.choice( population, size=2, p=get_relative_probabilities(population)) child1, child2 = father.mate(mother) new_population.append(child1) new_population.append(child2) population = new_population logging.info("Generation {}: {}".format( gen, max([r.get_fitness() for r in population]))) def get_relative_probabilities(population): pop_fitness = [r.get_fitness() for r in population] min_fitness = min(pop_fitness) max_fitness = max(pop_fitness) normalized = list( map( lambda x: normalize(x, min_fitness, max_fitness), pop_fitness ) ) total = sum(normalized) return list(map(lambda x: x/total, normalized)) def normalize(x, minf, maxf): return (x - minf) / (maxf - minf) if __name__=='__main__': logging.basicConfig(level=20) evolve()
351bfe236f183c069314f5df7d3c4b8f9d8699b4
final/problem6.py
final/problem6.py
# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Person.say(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue'))
# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Lecturer.lecture(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue'))
Modify lecture method in ArrogantProfessor class using inheritance
Modify lecture method in ArrogantProfessor class using inheritance
Python
mit
Kunal57/MIT_6.00.1x
# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Person.say(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue')) Modify lecture method in ArrogantProfessor class using inheritance
# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Lecturer.lecture(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue'))
<commit_before># Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Person.say(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue')) <commit_msg>Modify lecture method in ArrogantProfessor class using inheritance<commit_after>
# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Lecturer.lecture(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue'))
# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Person.say(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue')) Modify lecture method in ArrogantProfessor class using inheritance# Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Lecturer.lecture(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue'))
<commit_before># Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Person.say(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue')) <commit_msg>Modify lecture method in ArrogantProfessor class using inheritance<commit_after># Problem 6-1 # 10.0 points possible (graded) class Person(object): def __init__(self, name): self.name = name def say(self, stuff): return self.name + ' says: ' + stuff def __str__(self): return self.name class Lecturer(Person): def lecture(self, stuff): return 'I believe that ' + Person.say(self, stuff) class Professor(Lecturer): def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) class ArrogantProfessor(Professor): def lecture(self, stuff): return 'It is obvious that ' + Lecturer.lecture(self, stuff) def say(self, stuff): return self.name + ' says: ' + self.lecture(stuff) e = Person('eric') le = Lecturer('eric') pe = Professor('eric') ae = ArrogantProfessor('eric') e.say('the sky is blue') le.say('the sky is blue') le.lecture('the sky is blue') pe.say('the sky is blue') pe.lecture('the sky is blue') print(ae.say('the sky is blue')) print(ae.lecture('the sky is blue'))
3fdb40934319d667ae9e8c550a0404cdd6a8cb64
grum/api/models/message.py
grum/api/models/message.py
from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Timestamp) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)
from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Integer) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)
Change from timestamp to Integer
Change from timestamp to Integer
Python
mit
Grum-Hackdee/grum-web,Grum-Hackdee/grum-web,Grum-Hackdee/grum-web,Grum-Hackdee/grum-web
from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Timestamp) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)Change from timestamp to Integer
from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Integer) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)
<commit_before>from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Timestamp) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)<commit_msg>Change from timestamp to Integer<commit_after>
from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Integer) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)
from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Timestamp) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)Change from timestamp to Integerfrom grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Integer) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)
<commit_before>from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Timestamp) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)<commit_msg>Change from timestamp to Integer<commit_after>from grum import db class Message(db.Model): id = db.Column(db.String(128), primary_key=True) recipient = db.Column(db.String(128)) recipient_nice = db.Column(db.String(128)) sender = db.Column(db.String(128)) sender_nice = db.Column(db.String(128)) sent_at = db.Column(db.Integer) html = db.Column(db.Text) html_stripped = db.Column(db.Text) plaintext = db.Column(db.Text) plaintext_stripped = db.Column(db.Text) plaintext_stripped_signature = db.Column(db.Text)
ff8aa2725001dbd1281357ccd5e0877257b5975d
hackernews_scrapy/items.py
hackernews_scrapy/items.py
# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field()
# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() crawled_at = scrapy.Field()
Add crawled_at field to HackernewsScrapyItem
Add crawled_at field to HackernewsScrapyItem
Python
mit
mdsrosa/hackernews_scrapy
# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() Add crawled_at field to HackernewsScrapyItem
# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() crawled_at = scrapy.Field()
<commit_before># -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() <commit_msg>Add crawled_at field to HackernewsScrapyItem<commit_after>
# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() crawled_at = scrapy.Field()
# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() Add crawled_at field to HackernewsScrapyItem# -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() crawled_at = scrapy.Field()
<commit_before># -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() <commit_msg>Add crawled_at field to HackernewsScrapyItem<commit_after># -*- coding: utf-8 -*- import scrapy class HackernewsScrapyItem(scrapy.Item): title = scrapy.Field() crawled_at = scrapy.Field()
4c31f637d2b7f75c35debc51498913139b5634c0
pushhub/__init__.py
pushhub/__init__.py
from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='suscribe') return config.make_wsgi_app()
from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='subscribe') return config.make_wsgi_app()
Fix typo, make routes a little more legible.
Fix typo, make routes a little more legible.
Python
bsd-3-clause
ucla/PushHubCore
from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='suscribe') return config.make_wsgi_app() Fix typo, make routes a little more legible.
from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='subscribe') return config.make_wsgi_app()
<commit_before>from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='suscribe') return config.make_wsgi_app() <commit_msg>Fix typo, make routes a little more legible.<commit_after>
from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='subscribe') return config.make_wsgi_app()
from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='suscribe') return config.make_wsgi_app() Fix typo, make routes a little more legible.from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='subscribe') return config.make_wsgi_app()
<commit_before>from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='suscribe') return config.make_wsgi_app() <commit_msg>Fix typo, make routes a little more legible.<commit_after>from pyramid.config import Configurator from pyramid_zodbconn import get_connection from .models import appmaker from .views import publish, subscribe def root_factory(request): conn = get_connection(request) return appmaker(conn.root()) def main(global_config, **settings): """ This function returns a Pyramid WSGI application. """ config = Configurator(root_factory=root_factory, settings=settings) config.add_static_view('static', 'static', cache_max_age=3600) config.add_route('publish', '/publish') config.add_view(publish, route_name='publish') config.add_route('subscribe', '/subscribe') config.add_view(subscribe, route_name='subscribe') return config.make_wsgi_app()
c147629b4a0a5b405f7568b9278f288fa09fd97b
tests/aggregation/models.py
tests/aggregation/models.py
# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() def __unicode__(self): return self.name
# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() has_coffee = models.BooleanField() def __unicode__(self): return self.name
Add a boolean field to Store model (store.has_coffee)
Add a boolean field to Store model (store.has_coffee)
Python
mit
henriquebastos/django-aggregate-if
# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() def __unicode__(self): return self.name Add a boolean field to Store model (store.has_coffee)
# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() has_coffee = models.BooleanField() def __unicode__(self): return self.name
<commit_before># coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() def __unicode__(self): return self.name <commit_msg>Add a boolean field to Store model (store.has_coffee)<commit_after>
# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() has_coffee = models.BooleanField() def __unicode__(self): return self.name
# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() def __unicode__(self): return self.name Add a boolean field to Store model (store.has_coffee)# coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() has_coffee = models.BooleanField() def __unicode__(self): return self.name
<commit_before># coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() def __unicode__(self): return self.name <commit_msg>Add a boolean field to Store model (store.has_coffee)<commit_after># coding: utf-8 from django.db import models class Author(models.Model): name = models.CharField(max_length=100) age = models.IntegerField() friends = models.ManyToManyField('self', blank=True) def __unicode__(self): return self.name class Publisher(models.Model): name = models.CharField(max_length=255) num_awards = models.IntegerField() def __unicode__(self): return self.name class Book(models.Model): isbn = models.CharField(max_length=9) name = models.CharField(max_length=255) pages = models.IntegerField() rating = models.FloatField() price = models.DecimalField(decimal_places=2, max_digits=6) authors = models.ManyToManyField(Author) contact = models.ForeignKey(Author, related_name='book_contact_set') publisher = models.ForeignKey(Publisher) pubdate = models.DateField() def __unicode__(self): return self.name class Store(models.Model): name = models.CharField(max_length=255) books = models.ManyToManyField(Book) original_opening = models.DateTimeField() friday_night_closing = models.TimeField() has_coffee = models.BooleanField() def __unicode__(self): return self.name
0143e790245d19528af56df5428dc990d0689637
node/multiply.py
node/multiply.py
#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return a*b
#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return[a*b]
Multiply now handles lists correctly
Multiply now handles lists correctly
Python
mit
muddyfish/PYKE,muddyfish/PYKE
#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return a*bMultiply now handles lists correctly
#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return[a*b]
<commit_before>#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return a*b<commit_msg>Multiply now handles lists correctly<commit_after>
#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return[a*b]
#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return a*bMultiply now handles lists correctly#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return[a*b]
<commit_before>#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return a*b<commit_msg>Multiply now handles lists correctly<commit_after>#!/usr/bin/env python from nodes import Node class Multiply(Node): char = "*" args = 2 results = 1 @Node.test_func([4,5], [20]) def func(self, a,b): """a*b""" return[a*b]
9f6ade7fab83f15b49e37e28ac2d044a41846809
tests/test_create.py
tests/test_create.py
import globals as gbl from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo def test_add_to_git(project_repo): pass def test_create_ticket(project_repo): CreateTicket(ticket='test-ticket') def test_create_package(project_repo): CreatePackage(package='test-package')
from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo from pathlib import Path def test_add_to_git(project_repo): pass def test_create_ticket(session, project_repo): test_ticket = 'test-ticket' CreateTicket(ticket=test_ticket) ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket) deploy_file = Path(ticket_folder, 'deploy.py') repo = Repo(str(project_repo)) last_commit = repo.get_object(repo.head()) commit_message = last_commit.message assert ticket_folder.exists() assert deploy_file.exists() expected_message = bytes( 'Create ticket %s\n' % test_ticket, encoding='UTF-8') assert commit_message == expected_message def test_create_package(session, project_repo): test_package = 'test-package' CreatePackage(package=test_package) package_folder = Path(project_repo, 'deploy', 'packages', test_package) package_file = Path(package_folder, 'tickets.yml') remove_file = Path(package_folder, 'remove.py') assert package_folder.exists() assert package_file.exists() assert remove_file.exists()
Add test for git commit
Add test for git commit
Python
mit
Empiria/matador
import globals as gbl from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo def test_add_to_git(project_repo): pass def test_create_ticket(project_repo): CreateTicket(ticket='test-ticket') def test_create_package(project_repo): CreatePackage(package='test-package') Add test for git commit
from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo from pathlib import Path def test_add_to_git(project_repo): pass def test_create_ticket(session, project_repo): test_ticket = 'test-ticket' CreateTicket(ticket=test_ticket) ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket) deploy_file = Path(ticket_folder, 'deploy.py') repo = Repo(str(project_repo)) last_commit = repo.get_object(repo.head()) commit_message = last_commit.message assert ticket_folder.exists() assert deploy_file.exists() expected_message = bytes( 'Create ticket %s\n' % test_ticket, encoding='UTF-8') assert commit_message == expected_message def test_create_package(session, project_repo): test_package = 'test-package' CreatePackage(package=test_package) package_folder = Path(project_repo, 'deploy', 'packages', test_package) package_file = Path(package_folder, 'tickets.yml') remove_file = Path(package_folder, 'remove.py') assert package_folder.exists() assert package_file.exists() assert remove_file.exists()
<commit_before>import globals as gbl from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo def test_add_to_git(project_repo): pass def test_create_ticket(project_repo): CreateTicket(ticket='test-ticket') def test_create_package(project_repo): CreatePackage(package='test-package') <commit_msg>Add test for git commit<commit_after>
from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo from pathlib import Path def test_add_to_git(project_repo): pass def test_create_ticket(session, project_repo): test_ticket = 'test-ticket' CreateTicket(ticket=test_ticket) ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket) deploy_file = Path(ticket_folder, 'deploy.py') repo = Repo(str(project_repo)) last_commit = repo.get_object(repo.head()) commit_message = last_commit.message assert ticket_folder.exists() assert deploy_file.exists() expected_message = bytes( 'Create ticket %s\n' % test_ticket, encoding='UTF-8') assert commit_message == expected_message def test_create_package(session, project_repo): test_package = 'test-package' CreatePackage(package=test_package) package_folder = Path(project_repo, 'deploy', 'packages', test_package) package_file = Path(package_folder, 'tickets.yml') remove_file = Path(package_folder, 'remove.py') assert package_folder.exists() assert package_file.exists() assert remove_file.exists()
import globals as gbl from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo def test_add_to_git(project_repo): pass def test_create_ticket(project_repo): CreateTicket(ticket='test-ticket') def test_create_package(project_repo): CreatePackage(package='test-package') Add test for git commitfrom matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo from pathlib import Path def test_add_to_git(project_repo): pass def test_create_ticket(session, project_repo): test_ticket = 'test-ticket' CreateTicket(ticket=test_ticket) ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket) deploy_file = Path(ticket_folder, 'deploy.py') repo = Repo(str(project_repo)) last_commit = repo.get_object(repo.head()) commit_message = last_commit.message assert ticket_folder.exists() assert deploy_file.exists() expected_message = bytes( 'Create ticket %s\n' % test_ticket, encoding='UTF-8') assert commit_message == expected_message def test_create_package(session, project_repo): test_package = 'test-package' CreatePackage(package=test_package) package_folder = Path(project_repo, 'deploy', 'packages', test_package) package_file = Path(package_folder, 'tickets.yml') remove_file = Path(package_folder, 'remove.py') assert package_folder.exists() assert package_file.exists() assert remove_file.exists()
<commit_before>import globals as gbl from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo def test_add_to_git(project_repo): pass def test_create_ticket(project_repo): CreateTicket(ticket='test-ticket') def test_create_package(project_repo): CreatePackage(package='test-package') <commit_msg>Add test for git commit<commit_after>from matador.commands import CreateTicket, CreatePackage from dulwich.repo import Repo from pathlib import Path def test_add_to_git(project_repo): pass def test_create_ticket(session, project_repo): test_ticket = 'test-ticket' CreateTicket(ticket=test_ticket) ticket_folder = Path(project_repo, 'deploy', 'tickets', test_ticket) deploy_file = Path(ticket_folder, 'deploy.py') repo = Repo(str(project_repo)) last_commit = repo.get_object(repo.head()) commit_message = last_commit.message assert ticket_folder.exists() assert deploy_file.exists() expected_message = bytes( 'Create ticket %s\n' % test_ticket, encoding='UTF-8') assert commit_message == expected_message def test_create_package(session, project_repo): test_package = 'test-package' CreatePackage(package=test_package) package_folder = Path(project_repo, 'deploy', 'packages', test_package) package_file = Path(package_folder, 'tickets.yml') remove_file = Path(package_folder, 'remove.py') assert package_folder.exists() assert package_file.exists() assert remove_file.exists()
179df740725c0d3c9e256629e4718afcfa3b0cec
terminal_notifier.py
terminal_notifier.py
# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # 10-04-2015 # Version 1.0.0: initial release import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(signal_data)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main()
# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # # Version 1.0.0: initial release # Version 1.0.1: fix escape characters which broke terminal-notifier import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): message = signal_data if message[0] is "[": message = "\\%s" % message elif message[0] is "-": message = "\\%s" % message command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(message)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main()
Fix characters which break terminal-notifier
Fix characters which break terminal-notifier If your message starts with either a [ or - (and probably more I haven't found yet) terminal-notifier blows up because of the way it parses its arguments
Python
mit
keith/terminal-notifier-weechat
# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # 10-04-2015 # Version 1.0.0: initial release import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(signal_data)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main() Fix characters which break terminal-notifier If your message starts with either a [ or - (and probably more I haven't found yet) terminal-notifier blows up because of the way it parses its arguments
# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # # Version 1.0.0: initial release # Version 1.0.1: fix escape characters which broke terminal-notifier import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): message = signal_data if message[0] is "[": message = "\\%s" % message elif message[0] is "-": message = "\\%s" % message command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(message)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main()
<commit_before># This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # 10-04-2015 # Version 1.0.0: initial release import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(signal_data)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main() <commit_msg>Fix characters which break terminal-notifier If your message starts with either a [ or - (and probably more I haven't found yet) terminal-notifier blows up because of the way it parses its arguments<commit_after>
# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # # Version 1.0.0: initial release # Version 1.0.1: fix escape characters which broke terminal-notifier import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): message = signal_data if message[0] is "[": message = "\\%s" % message elif message[0] is "-": message = "\\%s" % message command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(message)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main()
# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # 10-04-2015 # Version 1.0.0: initial release import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(signal_data)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main() Fix characters which break terminal-notifier If your message starts with either a [ or - (and probably more I haven't found yet) terminal-notifier blows up because of the way it parses its arguments# This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # # Version 1.0.0: initial release # Version 1.0.1: fix escape characters which broke terminal-notifier import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): message = signal_data if message[0] is "[": message = "\\%s" % message elif message[0] is "-": message = "\\%s" % message command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(message)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main()
<commit_before># This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # 10-04-2015 # Version 1.0.0: initial release import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(signal_data)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main() <commit_msg>Fix characters which break terminal-notifier If your message starts with either a [ or - (and probably more I haven't found yet) terminal-notifier blows up because of the way it parses its arguments<commit_after># This weechat plugin sends OS X notifications for weechat messages # # Install terminal-notifier, no other configuration needed. # # History: # # Version 1.0.0: initial release # Version 1.0.1: fix escape characters which broke terminal-notifier import distutils.spawn import os import pipes import weechat def notify(data, signal, signal_data): message = signal_data if message[0] is "[": message = "\\%s" % message elif message[0] is "-": message = "\\%s" % message command = ("terminal-notifier -message %s -title WeeChat -sound Hero" % pipes.quote(message)) exit_code = os.system(command) if exit_code == 0: return weechat.WEECHAT_RC_ERROR else: return weechat.WEECHAT_RC_OK def main(): if distutils.spawn.find_executable("terminal-notifier") is None: return weechat.WEECHAT_RC_ERROR if not weechat.register("terminal_notifier", "Keith Smiley", "1.0.0", "MIT", "Get OS X notifications for messages", "", ""): return weechat.WEECHAT_RC_ERROR weechat.hook_signal("weechat_pv", "notify", "") weechat.hook_signal("weechat_highlight", "notify", "") return weechat.WEECHAT_RC_OK if __name__ == "__main__": main()
b2d9234ff6353191afc434556f9cfdea2448f726
test/test_regexes.py
test/test_regexes.py
from findspam import FindSpam import pytest @pytest.mark.parametrize("text, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True), ('bagprada', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False), ('HOW DO YOU SOLVE THIS PROBLEM?', True), ]) def test_regexes(text, match): result = FindSpam.testpost(text, "", "") print text print result isspam = False if (len(result) > 0): isspam = True assert match == isspam
from findspam import FindSpam import pytest @pytest.mark.parametrize("title, username, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True), ('', 'bagprada', True), ('HOW DO YOU SOLVE THIS PROBLEM?', '', True), ('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True), ('support for yahoo mail 18669786819 @call for helpline number', '', True), ('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False), ('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False), ('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False), ('What is the proper way to say "queryer"', 'jedwards', False), ('What\'s a real-world example of "overfitting"?', 'user3851283', False), ('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False), ('Is it unfair to regrade prior work after detecting cheating?', 'Village', False), ]) def test_regexes(title, username, match): result = FindSpam.testpost(title, username, "") print title print result isspam = False if (len(result) > 0): isspam = True assert match == isspam
Update tests + add username field
Update tests + add username field
Python
apache-2.0
Charcoal-SE/SmokeDetector,NickVolynkin/SmokeDetector,ArtOfCode-/SmokeDetector,ArtOfCode-/SmokeDetector,Charcoal-SE/SmokeDetector,NickVolynkin/SmokeDetector
from findspam import FindSpam import pytest @pytest.mark.parametrize("text, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True), ('bagprada', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False), ('HOW DO YOU SOLVE THIS PROBLEM?', True), ]) def test_regexes(text, match): result = FindSpam.testpost(text, "", "") print text print result isspam = False if (len(result) > 0): isspam = True assert match == isspamUpdate tests + add username field
from findspam import FindSpam import pytest @pytest.mark.parametrize("title, username, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True), ('', 'bagprada', True), ('HOW DO YOU SOLVE THIS PROBLEM?', '', True), ('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True), ('support for yahoo mail 18669786819 @call for helpline number', '', True), ('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False), ('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False), ('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False), ('What is the proper way to say "queryer"', 'jedwards', False), ('What\'s a real-world example of "overfitting"?', 'user3851283', False), ('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False), ('Is it unfair to regrade prior work after detecting cheating?', 'Village', False), ]) def test_regexes(title, username, match): result = FindSpam.testpost(title, username, "") print title print result isspam = False if (len(result) > 0): isspam = True assert match == isspam
<commit_before>from findspam import FindSpam import pytest @pytest.mark.parametrize("text, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True), ('bagprada', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False), ('HOW DO YOU SOLVE THIS PROBLEM?', True), ]) def test_regexes(text, match): result = FindSpam.testpost(text, "", "") print text print result isspam = False if (len(result) > 0): isspam = True assert match == isspam<commit_msg>Update tests + add username field<commit_after>
from findspam import FindSpam import pytest @pytest.mark.parametrize("title, username, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True), ('', 'bagprada', True), ('HOW DO YOU SOLVE THIS PROBLEM?', '', True), ('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True), ('support for yahoo mail 18669786819 @call for helpline number', '', True), ('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False), ('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False), ('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False), ('What is the proper way to say "queryer"', 'jedwards', False), ('What\'s a real-world example of "overfitting"?', 'user3851283', False), ('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False), ('Is it unfair to regrade prior work after detecting cheating?', 'Village', False), ]) def test_regexes(title, username, match): result = FindSpam.testpost(title, username, "") print title print result isspam = False if (len(result) > 0): isspam = True assert match == isspam
from findspam import FindSpam import pytest @pytest.mark.parametrize("text, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True), ('bagprada', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False), ('HOW DO YOU SOLVE THIS PROBLEM?', True), ]) def test_regexes(text, match): result = FindSpam.testpost(text, "", "") print text print result isspam = False if (len(result) > 0): isspam = True assert match == isspamUpdate tests + add username fieldfrom findspam import FindSpam import pytest @pytest.mark.parametrize("title, username, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True), ('', 'bagprada', True), ('HOW DO YOU SOLVE THIS PROBLEM?', '', True), ('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True), ('support for yahoo mail 18669786819 @call for helpline number', '', True), ('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False), ('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False), ('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False), ('What is the proper way to say "queryer"', 'jedwards', False), ('What\'s a real-world example of "overfitting"?', 'user3851283', False), ('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False), ('Is it unfair to regrade prior work after detecting cheating?', 'Village', False), ]) def test_regexes(title, username, match): result = FindSpam.testpost(title, username, "") print title print result isspam = False if (len(result) > 0): isspam = True assert match == isspam
<commit_before>from findspam import FindSpam import pytest @pytest.mark.parametrize("text, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', True), ('bagprada', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', False), ('HOW DO YOU SOLVE THIS PROBLEM?', True), ]) def test_regexes(text, match): result = FindSpam.testpost(text, "", "") print text print result isspam = False if (len(result) > 0): isspam = True assert match == isspam<commit_msg>Update tests + add username field<commit_after>from findspam import FindSpam import pytest @pytest.mark.parametrize("title, username, match", [ ('18669786819 gmail customer service number 1866978-6819 gmail support number', '', True), ('Is there any http://www.hindawi.com/ template for Cloud-Oriented Data Center Networking?', '', True), ('', 'bagprada', True), ('HOW DO YOU SOLVE THIS PROBLEM?', '', True), ('12 Month Loans quick @ http://www.quick12monthpaydayloans.co.uk/Elimination of collateral pledging', '', True), ('support for yahoo mail 18669786819 @call for helpline number', '', True), ('yahoo email tech support 1 866 978 6819 Yahoo Customer Phone Number ,Shortest Wait', '', True), ('What is the value of MD5 checksums if the MD5 hash itself could potentially also have been manipulated?', '', False), ('Probability: 6 Dice are rolled. Which is more likely, that you get exactly one 6, or that you get 6 different numbers?', '', False), ('The Challenge of Controlling a Powerful AI', 'Serban Tanasa', False), ('Reproducing image of a spiral using TikZ', 'Kristoffer Ryhl', False), ('What is the proper way to say "queryer"', 'jedwards', False), ('What\'s a real-world example of "overfitting"?', 'user3851283', False), ('How to avoid objects when traveling at greater than .75 light speed. or How Not to Go SPLAT?', 'bowlturner', False), ('Is it unfair to regrade prior work after detecting cheating?', 'Village', False), ]) def test_regexes(title, username, match): result = FindSpam.testpost(title, username, "") print title print result isspam = False if (len(result) > 0): isspam = True assert match == isspam
23c3b63e9f336ad70d481c0355d2f7f1544b9d54
lattice_length.py
lattice_length.py
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') myLattice = ap.machines.getLattice() length = 0 for key in range(myLattice.size()): length += myLattice[key].length print "The length of the lattice is {}.".format(length)
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') my_lattice = ap.machines.getLattice() length = 0 for key in range(my_lattice.size()): length += my_lattice[key].length print "The length of the lattice is {}.".format(length)
Change variable name to more suitable ones
Change variable name to more suitable ones
Python
apache-2.0
razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') myLattice = ap.machines.getLattice() length = 0 for key in range(myLattice.size()): length += myLattice[key].length print "The length of the lattice is {}.".format(length) Change variable name to more suitable ones
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') my_lattice = ap.machines.getLattice() length = 0 for key in range(my_lattice.size()): length += my_lattice[key].length print "The length of the lattice is {}.".format(length)
<commit_before># Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') myLattice = ap.machines.getLattice() length = 0 for key in range(myLattice.size()): length += myLattice[key].length print "The length of the lattice is {}.".format(length) <commit_msg>Change variable name to more suitable ones<commit_after>
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') my_lattice = ap.machines.getLattice() length = 0 for key in range(my_lattice.size()): length += my_lattice[key].length print "The length of the lattice is {}.".format(length)
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') myLattice = ap.machines.getLattice() length = 0 for key in range(myLattice.size()): length += myLattice[key].length print "The length of the lattice is {}.".format(length) Change variable name to more suitable ones# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') my_lattice = ap.machines.getLattice() length = 0 for key in range(my_lattice.size()): length += my_lattice[key].length print "The length of the lattice is {}.".format(length)
<commit_before># Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') myLattice = ap.machines.getLattice() length = 0 for key in range(myLattice.size()): length += myLattice[key].length print "The length of the lattice is {}.".format(length) <commit_msg>Change variable name to more suitable ones<commit_after># Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Load the machine ap.machines.load('SRI21') my_lattice = ap.machines.getLattice() length = 0 for key in range(my_lattice.size()): length += my_lattice[key].length print "The length of the lattice is {}.".format(length)