partition stringclasses 3 values | func_name stringlengths 1 134 | docstring stringlengths 1 46.9k | path stringlengths 4 223 | original_string stringlengths 75 104k | code stringlengths 75 104k | docstring_tokens listlengths 1 1.97k | repo stringlengths 7 55 | language stringclasses 1 value | url stringlengths 87 315 | code_tokens listlengths 19 28.4k | sha stringlengths 40 40 |
|---|---|---|---|---|---|---|---|---|---|---|---|
valid | Instruction.name | The instruction name/mnemonic | pyevmasm/evmasm.py | def name(self):
""" The instruction name/mnemonic """
if self._name == 'PUSH':
return 'PUSH%d' % self.operand_size
elif self._name == 'DUP':
return 'DUP%d' % self.pops
elif self._name == 'SWAP':
return 'SWAP%d' % (self.pops - 1)
elif self._name == 'LOG':
return 'LOG%d' % (self.pops - 2)
return self._name | def name(self):
""" The instruction name/mnemonic """
if self._name == 'PUSH':
return 'PUSH%d' % self.operand_size
elif self._name == 'DUP':
return 'DUP%d' % self.pops
elif self._name == 'SWAP':
return 'SWAP%d' % (self.pops - 1)
elif self._name == 'LOG':
return 'LOG%d' % (self.pops - 2)
return self._name | [
"The",
"instruction",
"name",
"/",
"mnemonic"
] | crytic/pyevmasm | python | https://github.com/crytic/pyevmasm/blob/d27daf19a36d630a31499e783b716cf1165798d8/pyevmasm/evmasm.py#L139-L149 | [
"def",
"name",
"(",
"self",
")",
":",
"if",
"self",
".",
"_name",
"==",
"'PUSH'",
":",
"return",
"'PUSH%d'",
"%",
"self",
".",
"operand_size",
"elif",
"self",
".",
"_name",
"==",
"'DUP'",
":",
"return",
"'DUP%d'",
"%",
"self",
".",
"pops",
"elif",
"self",
".",
"_name",
"==",
"'SWAP'",
":",
"return",
"'SWAP%d'",
"%",
"(",
"self",
".",
"pops",
"-",
"1",
")",
"elif",
"self",
".",
"_name",
"==",
"'LOG'",
":",
"return",
"'LOG%d'",
"%",
"(",
"self",
".",
"pops",
"-",
"2",
")",
"return",
"self",
".",
"_name"
] | d27daf19a36d630a31499e783b716cf1165798d8 |
valid | Instruction.parse_operand | Parses an operand from buf
:param buf: a buffer
:type buf: iterator/generator/string | pyevmasm/evmasm.py | def parse_operand(self, buf):
""" Parses an operand from buf
:param buf: a buffer
:type buf: iterator/generator/string
"""
buf = iter(buf)
try:
operand = 0
for _ in range(self.operand_size):
operand <<= 8
operand |= next(buf)
self._operand = operand
except StopIteration:
raise ParseError("Not enough data for decoding") | def parse_operand(self, buf):
""" Parses an operand from buf
:param buf: a buffer
:type buf: iterator/generator/string
"""
buf = iter(buf)
try:
operand = 0
for _ in range(self.operand_size):
operand <<= 8
operand |= next(buf)
self._operand = operand
except StopIteration:
raise ParseError("Not enough data for decoding") | [
"Parses",
"an",
"operand",
"from",
"buf"
] | crytic/pyevmasm | python | https://github.com/crytic/pyevmasm/blob/d27daf19a36d630a31499e783b716cf1165798d8/pyevmasm/evmasm.py#L151-L165 | [
"def",
"parse_operand",
"(",
"self",
",",
"buf",
")",
":",
"buf",
"=",
"iter",
"(",
"buf",
")",
"try",
":",
"operand",
"=",
"0",
"for",
"_",
"in",
"range",
"(",
"self",
".",
"operand_size",
")",
":",
"operand",
"<<=",
"8",
"operand",
"|=",
"next",
"(",
"buf",
")",
"self",
".",
"_operand",
"=",
"operand",
"except",
"StopIteration",
":",
"raise",
"ParseError",
"(",
"\"Not enough data for decoding\"",
")"
] | d27daf19a36d630a31499e783b716cf1165798d8 |
valid | Instruction.bytes | Encoded instruction | pyevmasm/evmasm.py | def bytes(self):
""" Encoded instruction """
b = [bytes([self._opcode])]
for offset in reversed(range(self.operand_size)):
b.append(bytes([(self.operand >> offset * 8) & 0xff]))
return b''.join(b) | def bytes(self):
""" Encoded instruction """
b = [bytes([self._opcode])]
for offset in reversed(range(self.operand_size)):
b.append(bytes([(self.operand >> offset * 8) & 0xff]))
return b''.join(b) | [
"Encoded",
"instruction"
] | crytic/pyevmasm | python | https://github.com/crytic/pyevmasm/blob/d27daf19a36d630a31499e783b716cf1165798d8/pyevmasm/evmasm.py#L220-L225 | [
"def",
"bytes",
"(",
"self",
")",
":",
"b",
"=",
"[",
"bytes",
"(",
"[",
"self",
".",
"_opcode",
"]",
")",
"]",
"for",
"offset",
"in",
"reversed",
"(",
"range",
"(",
"self",
".",
"operand_size",
")",
")",
":",
"b",
".",
"append",
"(",
"bytes",
"(",
"[",
"(",
"self",
".",
"operand",
">>",
"offset",
"*",
"8",
")",
"&",
"0xff",
"]",
")",
")",
"return",
"b''",
".",
"join",
"(",
"b",
")"
] | d27daf19a36d630a31499e783b716cf1165798d8 |
valid | reverse | reverse the key, value in dictionary file
:param rev_inputs: the files to be reversed
:return: None | helper/reverse.py | def reverse(rev_inputs=REV_INPUTS):
"""
reverse the key, value in dictionary file
:param rev_inputs: the files to be reversed
:return: None
"""
dirname = os.path.dirname(__file__)
for in_file in rev_inputs:
reversed_dict = {}
input_file = in_file + '.txt'
output_file = in_file + 'Rev.txt'
input_file = os.path.join(dirname, DICT_DIRECTORY, input_file)
output_file = os.path.join(dirname, DICT_DIRECTORY, output_file)
with open(input_file, 'r', encoding='utf-8') as f:
for line in f:
line = line.strip()
split = line.split('\t')
if len(split) < 2:
continue
term1 = split[0]
term2 = split[1]
for char in term2.split(' '):
if char in reversed_dict:
reversed_dict[char].append(term1)
else:
reversed_dict[char] = [term1]
with open(output_file, 'w', encoding='utf-8') as f:
for key in reversed_dict:
line = key + '\t' + ' '.join(reversed_dict[key]) + '\n'
f.write(line) | def reverse(rev_inputs=REV_INPUTS):
"""
reverse the key, value in dictionary file
:param rev_inputs: the files to be reversed
:return: None
"""
dirname = os.path.dirname(__file__)
for in_file in rev_inputs:
reversed_dict = {}
input_file = in_file + '.txt'
output_file = in_file + 'Rev.txt'
input_file = os.path.join(dirname, DICT_DIRECTORY, input_file)
output_file = os.path.join(dirname, DICT_DIRECTORY, output_file)
with open(input_file, 'r', encoding='utf-8') as f:
for line in f:
line = line.strip()
split = line.split('\t')
if len(split) < 2:
continue
term1 = split[0]
term2 = split[1]
for char in term2.split(' '):
if char in reversed_dict:
reversed_dict[char].append(term1)
else:
reversed_dict[char] = [term1]
with open(output_file, 'w', encoding='utf-8') as f:
for key in reversed_dict:
line = key + '\t' + ' '.join(reversed_dict[key]) + '\n'
f.write(line) | [
"reverse",
"the",
"key",
"value",
"in",
"dictionary",
"file",
":",
"param",
"rev_inputs",
":",
"the",
"files",
"to",
"be",
"reversed",
":",
"return",
":",
"None"
] | yichen0831/opencc-python | python | https://github.com/yichen0831/opencc-python/blob/75a7e882a250076c7d239e0fef0791815aae545d/helper/reverse.py#L22-L53 | [
"def",
"reverse",
"(",
"rev_inputs",
"=",
"REV_INPUTS",
")",
":",
"dirname",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
"for",
"in_file",
"in",
"rev_inputs",
":",
"reversed_dict",
"=",
"{",
"}",
"input_file",
"=",
"in_file",
"+",
"'.txt'",
"output_file",
"=",
"in_file",
"+",
"'Rev.txt'",
"input_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"DICT_DIRECTORY",
",",
"input_file",
")",
"output_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"DICT_DIRECTORY",
",",
"output_file",
")",
"with",
"open",
"(",
"input_file",
",",
"'r'",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"f",
":",
"for",
"line",
"in",
"f",
":",
"line",
"=",
"line",
".",
"strip",
"(",
")",
"split",
"=",
"line",
".",
"split",
"(",
"'\\t'",
")",
"if",
"len",
"(",
"split",
")",
"<",
"2",
":",
"continue",
"term1",
"=",
"split",
"[",
"0",
"]",
"term2",
"=",
"split",
"[",
"1",
"]",
"for",
"char",
"in",
"term2",
".",
"split",
"(",
"' '",
")",
":",
"if",
"char",
"in",
"reversed_dict",
":",
"reversed_dict",
"[",
"char",
"]",
".",
"append",
"(",
"term1",
")",
"else",
":",
"reversed_dict",
"[",
"char",
"]",
"=",
"[",
"term1",
"]",
"with",
"open",
"(",
"output_file",
",",
"'w'",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"f",
":",
"for",
"key",
"in",
"reversed_dict",
":",
"line",
"=",
"key",
"+",
"'\\t'",
"+",
"' '",
".",
"join",
"(",
"reversed_dict",
"[",
"key",
"]",
")",
"+",
"'\\n'",
"f",
".",
"write",
"(",
"line",
")"
] | 75a7e882a250076c7d239e0fef0791815aae545d |
valid | merge | merge the phrase files into one file
:param mer_inputs: the phrase files
:param mer_output: the output file
:return: None | helper/merge.py | def merge(mer_inputs=MER_INPUTS, mer_output=MER_OUTPUT):
"""
merge the phrase files into one file
:param mer_inputs: the phrase files
:param mer_output: the output file
:return: None
"""
dirname = os.path.dirname(__file__)
output_file = os.path.join(dirname, DICT_DIRECTORY, mer_output)
lines = []
for in_file in MER_INPUTS:
input_file = os.path.join(dirname, DICT_DIRECTORY, in_file)
with open(input_file, encoding='utf-8') as f:
for line in f:
lines.append(line)
with open(output_file, 'w', encoding='utf-8') as f:
for line in lines:
f.write(line) | def merge(mer_inputs=MER_INPUTS, mer_output=MER_OUTPUT):
"""
merge the phrase files into one file
:param mer_inputs: the phrase files
:param mer_output: the output file
:return: None
"""
dirname = os.path.dirname(__file__)
output_file = os.path.join(dirname, DICT_DIRECTORY, mer_output)
lines = []
for in_file in MER_INPUTS:
input_file = os.path.join(dirname, DICT_DIRECTORY, in_file)
with open(input_file, encoding='utf-8') as f:
for line in f:
lines.append(line)
with open(output_file, 'w', encoding='utf-8') as f:
for line in lines:
f.write(line) | [
"merge",
"the",
"phrase",
"files",
"into",
"one",
"file",
":",
"param",
"mer_inputs",
":",
"the",
"phrase",
"files",
":",
"param",
"mer_output",
":",
"the",
"output",
"file",
":",
"return",
":",
"None"
] | yichen0831/opencc-python | python | https://github.com/yichen0831/opencc-python/blob/75a7e882a250076c7d239e0fef0791815aae545d/helper/merge.py#L24-L42 | [
"def",
"merge",
"(",
"mer_inputs",
"=",
"MER_INPUTS",
",",
"mer_output",
"=",
"MER_OUTPUT",
")",
":",
"dirname",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
"output_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"DICT_DIRECTORY",
",",
"mer_output",
")",
"lines",
"=",
"[",
"]",
"for",
"in_file",
"in",
"MER_INPUTS",
":",
"input_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"DICT_DIRECTORY",
",",
"in_file",
")",
"with",
"open",
"(",
"input_file",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"f",
":",
"for",
"line",
"in",
"f",
":",
"lines",
".",
"append",
"(",
"line",
")",
"with",
"open",
"(",
"output_file",
",",
"'w'",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"f",
":",
"for",
"line",
"in",
"lines",
":",
"f",
".",
"write",
"(",
"line",
")"
] | 75a7e882a250076c7d239e0fef0791815aae545d |
valid | PassiveThrottle._adjust_delay | Define delay adjustment policy | pystock_crawler/throttle.py | def _adjust_delay(self, slot, response):
"""Define delay adjustment policy"""
if response.status in self.retry_http_codes:
new_delay = max(slot.delay, 1) * 4
new_delay = max(new_delay, self.mindelay)
new_delay = min(new_delay, self.maxdelay)
slot.delay = new_delay
self.stats.inc_value('delay_count')
elif response.status == 200:
new_delay = max(slot.delay / 2, self.mindelay)
if new_delay < 0.01:
new_delay = 0
slot.delay = new_delay | def _adjust_delay(self, slot, response):
"""Define delay adjustment policy"""
if response.status in self.retry_http_codes:
new_delay = max(slot.delay, 1) * 4
new_delay = max(new_delay, self.mindelay)
new_delay = min(new_delay, self.maxdelay)
slot.delay = new_delay
self.stats.inc_value('delay_count')
elif response.status == 200:
new_delay = max(slot.delay / 2, self.mindelay)
if new_delay < 0.01:
new_delay = 0
slot.delay = new_delay | [
"Define",
"delay",
"adjustment",
"policy"
] | eliangcs/pystock-crawler | python | https://github.com/eliangcs/pystock-crawler/blob/8b803c8944f36af46daf04c6767a74132e37a101/pystock_crawler/throttle.py#L66-L78 | [
"def",
"_adjust_delay",
"(",
"self",
",",
"slot",
",",
"response",
")",
":",
"if",
"response",
".",
"status",
"in",
"self",
".",
"retry_http_codes",
":",
"new_delay",
"=",
"max",
"(",
"slot",
".",
"delay",
",",
"1",
")",
"*",
"4",
"new_delay",
"=",
"max",
"(",
"new_delay",
",",
"self",
".",
"mindelay",
")",
"new_delay",
"=",
"min",
"(",
"new_delay",
",",
"self",
".",
"maxdelay",
")",
"slot",
".",
"delay",
"=",
"new_delay",
"self",
".",
"stats",
".",
"inc_value",
"(",
"'delay_count'",
")",
"elif",
"response",
".",
"status",
"==",
"200",
":",
"new_delay",
"=",
"max",
"(",
"slot",
".",
"delay",
"/",
"2",
",",
"self",
".",
"mindelay",
")",
"if",
"new_delay",
"<",
"0.01",
":",
"new_delay",
"=",
"0",
"slot",
".",
"delay",
"=",
"new_delay"
] | 8b803c8944f36af46daf04c6767a74132e37a101 |
valid | memberness | The likelihood that the context is a "member". | pystock_crawler/loaders.py | def memberness(context):
'''The likelihood that the context is a "member".'''
if context:
texts = context.xpath('.//*[local-name()="explicitMember"]/text()').extract()
text = str(texts).lower()
if len(texts) > 1:
return 2
elif 'country' in text:
return 2
elif 'member' not in text:
return 0
elif 'successor' in text:
# 'SuccessorMember' is a rare case that shouldn't be treated as member
return 1
elif 'parent' in text:
return 2
return 3 | def memberness(context):
'''The likelihood that the context is a "member".'''
if context:
texts = context.xpath('.//*[local-name()="explicitMember"]/text()').extract()
text = str(texts).lower()
if len(texts) > 1:
return 2
elif 'country' in text:
return 2
elif 'member' not in text:
return 0
elif 'successor' in text:
# 'SuccessorMember' is a rare case that shouldn't be treated as member
return 1
elif 'parent' in text:
return 2
return 3 | [
"The",
"likelihood",
"that",
"the",
"context",
"is",
"a",
"member",
"."
] | eliangcs/pystock-crawler | python | https://github.com/eliangcs/pystock-crawler/blob/8b803c8944f36af46daf04c6767a74132e37a101/pystock_crawler/loaders.py#L300-L317 | [
"def",
"memberness",
"(",
"context",
")",
":",
"if",
"context",
":",
"texts",
"=",
"context",
".",
"xpath",
"(",
"'.//*[local-name()=\"explicitMember\"]/text()'",
")",
".",
"extract",
"(",
")",
"text",
"=",
"str",
"(",
"texts",
")",
".",
"lower",
"(",
")",
"if",
"len",
"(",
"texts",
")",
">",
"1",
":",
"return",
"2",
"elif",
"'country'",
"in",
"text",
":",
"return",
"2",
"elif",
"'member'",
"not",
"in",
"text",
":",
"return",
"0",
"elif",
"'successor'",
"in",
"text",
":",
"# 'SuccessorMember' is a rare case that shouldn't be treated as member",
"return",
"1",
"elif",
"'parent'",
"in",
"text",
":",
"return",
"2",
"return",
"3"
] | 8b803c8944f36af46daf04c6767a74132e37a101 |
valid | EdgarSpider.parse_10qk | Parse 10-Q or 10-K XML report. | pystock_crawler/spiders/edgar.py | def parse_10qk(self, response):
'''Parse 10-Q or 10-K XML report.'''
loader = ReportItemLoader(response=response)
item = loader.load_item()
if 'doc_type' in item:
doc_type = item['doc_type']
if doc_type in ('10-Q', '10-K'):
return item
return None | def parse_10qk(self, response):
'''Parse 10-Q or 10-K XML report.'''
loader = ReportItemLoader(response=response)
item = loader.load_item()
if 'doc_type' in item:
doc_type = item['doc_type']
if doc_type in ('10-Q', '10-K'):
return item
return None | [
"Parse",
"10",
"-",
"Q",
"or",
"10",
"-",
"K",
"XML",
"report",
"."
] | eliangcs/pystock-crawler | python | https://github.com/eliangcs/pystock-crawler/blob/8b803c8944f36af46daf04c6767a74132e37a101/pystock_crawler/spiders/edgar.py#L57-L67 | [
"def",
"parse_10qk",
"(",
"self",
",",
"response",
")",
":",
"loader",
"=",
"ReportItemLoader",
"(",
"response",
"=",
"response",
")",
"item",
"=",
"loader",
".",
"load_item",
"(",
")",
"if",
"'doc_type'",
"in",
"item",
":",
"doc_type",
"=",
"item",
"[",
"'doc_type'",
"]",
"if",
"doc_type",
"in",
"(",
"'10-Q'",
",",
"'10-K'",
")",
":",
"return",
"item",
"return",
"None"
] | 8b803c8944f36af46daf04c6767a74132e37a101 |
valid | camelcase | Convert string into camel case.
Args:
string: String to convert.
Returns:
string: Camel case string. | stringcase.py | def camelcase(string):
""" Convert string into camel case.
Args:
string: String to convert.
Returns:
string: Camel case string.
"""
string = re.sub(r"^[\-_\.]", '', str(string))
if not string:
return string
return lowercase(string[0]) + re.sub(r"[\-_\.\s]([a-z])", lambda matched: uppercase(matched.group(1)), string[1:]) | def camelcase(string):
""" Convert string into camel case.
Args:
string: String to convert.
Returns:
string: Camel case string.
"""
string = re.sub(r"^[\-_\.]", '', str(string))
if not string:
return string
return lowercase(string[0]) + re.sub(r"[\-_\.\s]([a-z])", lambda matched: uppercase(matched.group(1)), string[1:]) | [
"Convert",
"string",
"into",
"camel",
"case",
"."
] | okunishinishi/python-stringcase | python | https://github.com/okunishinishi/python-stringcase/blob/700ad111be16b384aadaddcf8199f9390575c7b6/stringcase.py#L8-L22 | [
"def",
"camelcase",
"(",
"string",
")",
":",
"string",
"=",
"re",
".",
"sub",
"(",
"r\"^[\\-_\\.]\"",
",",
"''",
",",
"str",
"(",
"string",
")",
")",
"if",
"not",
"string",
":",
"return",
"string",
"return",
"lowercase",
"(",
"string",
"[",
"0",
"]",
")",
"+",
"re",
".",
"sub",
"(",
"r\"[\\-_\\.\\s]([a-z])\"",
",",
"lambda",
"matched",
":",
"uppercase",
"(",
"matched",
".",
"group",
"(",
"1",
")",
")",
",",
"string",
"[",
"1",
":",
"]",
")"
] | 700ad111be16b384aadaddcf8199f9390575c7b6 |
valid | capitalcase | Convert string into capital case.
First letters will be uppercase.
Args:
string: String to convert.
Returns:
string: Capital case string. | stringcase.py | def capitalcase(string):
"""Convert string into capital case.
First letters will be uppercase.
Args:
string: String to convert.
Returns:
string: Capital case string.
"""
string = str(string)
if not string:
return string
return uppercase(string[0]) + string[1:] | def capitalcase(string):
"""Convert string into capital case.
First letters will be uppercase.
Args:
string: String to convert.
Returns:
string: Capital case string.
"""
string = str(string)
if not string:
return string
return uppercase(string[0]) + string[1:] | [
"Convert",
"string",
"into",
"capital",
"case",
".",
"First",
"letters",
"will",
"be",
"uppercase",
"."
] | okunishinishi/python-stringcase | python | https://github.com/okunishinishi/python-stringcase/blob/700ad111be16b384aadaddcf8199f9390575c7b6/stringcase.py#L25-L40 | [
"def",
"capitalcase",
"(",
"string",
")",
":",
"string",
"=",
"str",
"(",
"string",
")",
"if",
"not",
"string",
":",
"return",
"string",
"return",
"uppercase",
"(",
"string",
"[",
"0",
"]",
")",
"+",
"string",
"[",
"1",
":",
"]"
] | 700ad111be16b384aadaddcf8199f9390575c7b6 |
valid | pathcase | Convert string into path case.
Join punctuation with slash.
Args:
string: String to convert.
Returns:
string: Path cased string. | stringcase.py | def pathcase(string):
"""Convert string into path case.
Join punctuation with slash.
Args:
string: String to convert.
Returns:
string: Path cased string.
"""
string = snakecase(string)
if not string:
return string
return re.sub(r"_", "/", string) | def pathcase(string):
"""Convert string into path case.
Join punctuation with slash.
Args:
string: String to convert.
Returns:
string: Path cased string.
"""
string = snakecase(string)
if not string:
return string
return re.sub(r"_", "/", string) | [
"Convert",
"string",
"into",
"path",
"case",
".",
"Join",
"punctuation",
"with",
"slash",
"."
] | okunishinishi/python-stringcase | python | https://github.com/okunishinishi/python-stringcase/blob/700ad111be16b384aadaddcf8199f9390575c7b6/stringcase.py#L86-L100 | [
"def",
"pathcase",
"(",
"string",
")",
":",
"string",
"=",
"snakecase",
"(",
"string",
")",
"if",
"not",
"string",
":",
"return",
"string",
"return",
"re",
".",
"sub",
"(",
"r\"_\"",
",",
"\"/\"",
",",
"string",
")"
] | 700ad111be16b384aadaddcf8199f9390575c7b6 |
valid | backslashcase | Convert string into spinal case.
Join punctuation with backslash.
Args:
string: String to convert.
Returns:
string: Spinal cased string. | stringcase.py | def backslashcase(string):
"""Convert string into spinal case.
Join punctuation with backslash.
Args:
string: String to convert.
Returns:
string: Spinal cased string.
"""
str1 = re.sub(r"_", r"\\", snakecase(string))
return str1 | def backslashcase(string):
"""Convert string into spinal case.
Join punctuation with backslash.
Args:
string: String to convert.
Returns:
string: Spinal cased string.
"""
str1 = re.sub(r"_", r"\\", snakecase(string))
return str1 | [
"Convert",
"string",
"into",
"spinal",
"case",
".",
"Join",
"punctuation",
"with",
"backslash",
"."
] | okunishinishi/python-stringcase | python | https://github.com/okunishinishi/python-stringcase/blob/700ad111be16b384aadaddcf8199f9390575c7b6/stringcase.py#L103-L116 | [
"def",
"backslashcase",
"(",
"string",
")",
":",
"str1",
"=",
"re",
".",
"sub",
"(",
"r\"_\"",
",",
"r\"\\\\\"",
",",
"snakecase",
"(",
"string",
")",
")",
"return",
"str1"
] | 700ad111be16b384aadaddcf8199f9390575c7b6 |
valid | sentencecase | Convert string into sentence case.
First letter capped and each punctuations are joined with space.
Args:
string: String to convert.
Returns:
string: Sentence cased string. | stringcase.py | def sentencecase(string):
"""Convert string into sentence case.
First letter capped and each punctuations are joined with space.
Args:
string: String to convert.
Returns:
string: Sentence cased string.
"""
joiner = ' '
string = re.sub(r"[\-_\.\s]", joiner, str(string))
if not string:
return string
return capitalcase(trimcase(
re.sub(r"[A-Z]", lambda matched: joiner +
lowercase(matched.group(0)), string)
)) | def sentencecase(string):
"""Convert string into sentence case.
First letter capped and each punctuations are joined with space.
Args:
string: String to convert.
Returns:
string: Sentence cased string.
"""
joiner = ' '
string = re.sub(r"[\-_\.\s]", joiner, str(string))
if not string:
return string
return capitalcase(trimcase(
re.sub(r"[A-Z]", lambda matched: joiner +
lowercase(matched.group(0)), string)
)) | [
"Convert",
"string",
"into",
"sentence",
"case",
".",
"First",
"letter",
"capped",
"and",
"each",
"punctuations",
"are",
"joined",
"with",
"space",
"."
] | okunishinishi/python-stringcase | python | https://github.com/okunishinishi/python-stringcase/blob/700ad111be16b384aadaddcf8199f9390575c7b6/stringcase.py#L120-L138 | [
"def",
"sentencecase",
"(",
"string",
")",
":",
"joiner",
"=",
"' '",
"string",
"=",
"re",
".",
"sub",
"(",
"r\"[\\-_\\.\\s]\"",
",",
"joiner",
",",
"str",
"(",
"string",
")",
")",
"if",
"not",
"string",
":",
"return",
"string",
"return",
"capitalcase",
"(",
"trimcase",
"(",
"re",
".",
"sub",
"(",
"r\"[A-Z]\"",
",",
"lambda",
"matched",
":",
"joiner",
"+",
"lowercase",
"(",
"matched",
".",
"group",
"(",
"0",
")",
")",
",",
"string",
")",
")",
")"
] | 700ad111be16b384aadaddcf8199f9390575c7b6 |
valid | snakecase | Convert string into snake case.
Join punctuation with underscore
Args:
string: String to convert.
Returns:
string: Snake cased string. | stringcase.py | def snakecase(string):
"""Convert string into snake case.
Join punctuation with underscore
Args:
string: String to convert.
Returns:
string: Snake cased string.
"""
string = re.sub(r"[\-\.\s]", '_', str(string))
if not string:
return string
return lowercase(string[0]) + re.sub(r"[A-Z]", lambda matched: '_' + lowercase(matched.group(0)), string[1:]) | def snakecase(string):
"""Convert string into snake case.
Join punctuation with underscore
Args:
string: String to convert.
Returns:
string: Snake cased string.
"""
string = re.sub(r"[\-\.\s]", '_', str(string))
if not string:
return string
return lowercase(string[0]) + re.sub(r"[A-Z]", lambda matched: '_' + lowercase(matched.group(0)), string[1:]) | [
"Convert",
"string",
"into",
"snake",
"case",
".",
"Join",
"punctuation",
"with",
"underscore"
] | okunishinishi/python-stringcase | python | https://github.com/okunishinishi/python-stringcase/blob/700ad111be16b384aadaddcf8199f9390575c7b6/stringcase.py#L141-L156 | [
"def",
"snakecase",
"(",
"string",
")",
":",
"string",
"=",
"re",
".",
"sub",
"(",
"r\"[\\-\\.\\s]\"",
",",
"'_'",
",",
"str",
"(",
"string",
")",
")",
"if",
"not",
"string",
":",
"return",
"string",
"return",
"lowercase",
"(",
"string",
"[",
"0",
"]",
")",
"+",
"re",
".",
"sub",
"(",
"r\"[A-Z]\"",
",",
"lambda",
"matched",
":",
"'_'",
"+",
"lowercase",
"(",
"matched",
".",
"group",
"(",
"0",
")",
")",
",",
"string",
"[",
"1",
":",
"]",
")"
] | 700ad111be16b384aadaddcf8199f9390575c7b6 |
valid | STree._check_input | Checks the validity of the input.
In case of an invalid input throws ValueError. | suffix_trees/STree.py | def _check_input(self, input):
"""Checks the validity of the input.
In case of an invalid input throws ValueError.
"""
if isinstance(input, str):
return 'st'
elif isinstance(input, list):
if all(isinstance(item, str) for item in input):
return 'gst'
raise ValueError("String argument should be of type String or"
" a list of strings") | def _check_input(self, input):
"""Checks the validity of the input.
In case of an invalid input throws ValueError.
"""
if isinstance(input, str):
return 'st'
elif isinstance(input, list):
if all(isinstance(item, str) for item in input):
return 'gst'
raise ValueError("String argument should be of type String or"
" a list of strings") | [
"Checks",
"the",
"validity",
"of",
"the",
"input",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L16-L28 | [
"def",
"_check_input",
"(",
"self",
",",
"input",
")",
":",
"if",
"isinstance",
"(",
"input",
",",
"str",
")",
":",
"return",
"'st'",
"elif",
"isinstance",
"(",
"input",
",",
"list",
")",
":",
"if",
"all",
"(",
"isinstance",
"(",
"item",
",",
"str",
")",
"for",
"item",
"in",
"input",
")",
":",
"return",
"'gst'",
"raise",
"ValueError",
"(",
"\"String argument should be of type String or\"",
"\" a list of strings\"",
")"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree.build | Builds the Suffix tree on the given input.
If the input is of type List of Strings:
Generalized Suffix Tree is built.
:param x: String or List of Strings | suffix_trees/STree.py | def build(self, x):
"""Builds the Suffix tree on the given input.
If the input is of type List of Strings:
Generalized Suffix Tree is built.
:param x: String or List of Strings
"""
type = self._check_input(x)
if type == 'st':
x += next(self._terminalSymbolsGenerator())
self._build(x)
if type == 'gst':
self._build_generalized(x) | def build(self, x):
"""Builds the Suffix tree on the given input.
If the input is of type List of Strings:
Generalized Suffix Tree is built.
:param x: String or List of Strings
"""
type = self._check_input(x)
if type == 'st':
x += next(self._terminalSymbolsGenerator())
self._build(x)
if type == 'gst':
self._build_generalized(x) | [
"Builds",
"the",
"Suffix",
"tree",
"on",
"the",
"given",
"input",
".",
"If",
"the",
"input",
"is",
"of",
"type",
"List",
"of",
"Strings",
":",
"Generalized",
"Suffix",
"Tree",
"is",
"built",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L30-L43 | [
"def",
"build",
"(",
"self",
",",
"x",
")",
":",
"type",
"=",
"self",
".",
"_check_input",
"(",
"x",
")",
"if",
"type",
"==",
"'st'",
":",
"x",
"+=",
"next",
"(",
"self",
".",
"_terminalSymbolsGenerator",
"(",
")",
")",
"self",
".",
"_build",
"(",
"x",
")",
"if",
"type",
"==",
"'gst'",
":",
"self",
".",
"_build_generalized",
"(",
"x",
")"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._build_McCreight | Builds a Suffix tree using McCreight O(n) algorithm.
Algorithm based on:
McCreight, Edward M. "A space-economical suffix tree construction algorithm." - ACM, 1976.
Implementation based on:
UH CS - 58093 String Processing Algorithms Lecture Notes | suffix_trees/STree.py | def _build_McCreight(self, x):
"""Builds a Suffix tree using McCreight O(n) algorithm.
Algorithm based on:
McCreight, Edward M. "A space-economical suffix tree construction algorithm." - ACM, 1976.
Implementation based on:
UH CS - 58093 String Processing Algorithms Lecture Notes
"""
u = self.root
d = 0
for i in range(len(x)):
while u.depth == d and u._has_transition(x[d+i]):
u = u._get_transition_link(x[d+i])
d = d + 1
while d < u.depth and x[u.idx + d] == x[i + d]:
d = d + 1
if d < u.depth:
u = self._create_node(x, u, d)
self._create_leaf(x, i, u, d)
if not u._get_suffix_link():
self._compute_slink(x, u)
u = u._get_suffix_link()
d = d - 1
if d < 0:
d = 0 | def _build_McCreight(self, x):
"""Builds a Suffix tree using McCreight O(n) algorithm.
Algorithm based on:
McCreight, Edward M. "A space-economical suffix tree construction algorithm." - ACM, 1976.
Implementation based on:
UH CS - 58093 String Processing Algorithms Lecture Notes
"""
u = self.root
d = 0
for i in range(len(x)):
while u.depth == d and u._has_transition(x[d+i]):
u = u._get_transition_link(x[d+i])
d = d + 1
while d < u.depth and x[u.idx + d] == x[i + d]:
d = d + 1
if d < u.depth:
u = self._create_node(x, u, d)
self._create_leaf(x, i, u, d)
if not u._get_suffix_link():
self._compute_slink(x, u)
u = u._get_suffix_link()
d = d - 1
if d < 0:
d = 0 | [
"Builds",
"a",
"Suffix",
"tree",
"using",
"McCreight",
"O",
"(",
"n",
")",
"algorithm",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L50-L74 | [
"def",
"_build_McCreight",
"(",
"self",
",",
"x",
")",
":",
"u",
"=",
"self",
".",
"root",
"d",
"=",
"0",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"x",
")",
")",
":",
"while",
"u",
".",
"depth",
"==",
"d",
"and",
"u",
".",
"_has_transition",
"(",
"x",
"[",
"d",
"+",
"i",
"]",
")",
":",
"u",
"=",
"u",
".",
"_get_transition_link",
"(",
"x",
"[",
"d",
"+",
"i",
"]",
")",
"d",
"=",
"d",
"+",
"1",
"while",
"d",
"<",
"u",
".",
"depth",
"and",
"x",
"[",
"u",
".",
"idx",
"+",
"d",
"]",
"==",
"x",
"[",
"i",
"+",
"d",
"]",
":",
"d",
"=",
"d",
"+",
"1",
"if",
"d",
"<",
"u",
".",
"depth",
":",
"u",
"=",
"self",
".",
"_create_node",
"(",
"x",
",",
"u",
",",
"d",
")",
"self",
".",
"_create_leaf",
"(",
"x",
",",
"i",
",",
"u",
",",
"d",
")",
"if",
"not",
"u",
".",
"_get_suffix_link",
"(",
")",
":",
"self",
".",
"_compute_slink",
"(",
"x",
",",
"u",
")",
"u",
"=",
"u",
".",
"_get_suffix_link",
"(",
")",
"d",
"=",
"d",
"-",
"1",
"if",
"d",
"<",
"0",
":",
"d",
"=",
"0"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._build_generalized | Builds a Generalized Suffix Tree (GST) from the array of strings provided. | suffix_trees/STree.py | def _build_generalized(self, xs):
"""Builds a Generalized Suffix Tree (GST) from the array of strings provided.
"""
terminal_gen = self._terminalSymbolsGenerator()
_xs = ''.join([x + next(terminal_gen) for x in xs])
self.word = _xs
self._generalized_word_starts(xs)
self._build(_xs)
self.root._traverse(self._label_generalized) | def _build_generalized(self, xs):
"""Builds a Generalized Suffix Tree (GST) from the array of strings provided.
"""
terminal_gen = self._terminalSymbolsGenerator()
_xs = ''.join([x + next(terminal_gen) for x in xs])
self.word = _xs
self._generalized_word_starts(xs)
self._build(_xs)
self.root._traverse(self._label_generalized) | [
"Builds",
"a",
"Generalized",
"Suffix",
"Tree",
"(",
"GST",
")",
"from",
"the",
"array",
"of",
"strings",
"provided",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L113-L122 | [
"def",
"_build_generalized",
"(",
"self",
",",
"xs",
")",
":",
"terminal_gen",
"=",
"self",
".",
"_terminalSymbolsGenerator",
"(",
")",
"_xs",
"=",
"''",
".",
"join",
"(",
"[",
"x",
"+",
"next",
"(",
"terminal_gen",
")",
"for",
"x",
"in",
"xs",
"]",
")",
"self",
".",
"word",
"=",
"_xs",
"self",
".",
"_generalized_word_starts",
"(",
"xs",
")",
"self",
".",
"_build",
"(",
"_xs",
")",
"self",
".",
"root",
".",
"_traverse",
"(",
"self",
".",
"_label_generalized",
")"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._label_generalized | Helper method that labels the nodes of GST with indexes of strings
found in their descendants. | suffix_trees/STree.py | def _label_generalized(self, node):
"""Helper method that labels the nodes of GST with indexes of strings
found in their descendants.
"""
if node.is_leaf():
x = {self._get_word_start_index(node.idx)}
else:
x = {n for ns in node.transition_links for n in ns[0].generalized_idxs}
node.generalized_idxs = x | def _label_generalized(self, node):
"""Helper method that labels the nodes of GST with indexes of strings
found in their descendants.
"""
if node.is_leaf():
x = {self._get_word_start_index(node.idx)}
else:
x = {n for ns in node.transition_links for n in ns[0].generalized_idxs}
node.generalized_idxs = x | [
"Helper",
"method",
"that",
"labels",
"the",
"nodes",
"of",
"GST",
"with",
"indexes",
"of",
"strings",
"found",
"in",
"their",
"descendants",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L124-L132 | [
"def",
"_label_generalized",
"(",
"self",
",",
"node",
")",
":",
"if",
"node",
".",
"is_leaf",
"(",
")",
":",
"x",
"=",
"{",
"self",
".",
"_get_word_start_index",
"(",
"node",
".",
"idx",
")",
"}",
"else",
":",
"x",
"=",
"{",
"n",
"for",
"ns",
"in",
"node",
".",
"transition_links",
"for",
"n",
"in",
"ns",
"[",
"0",
"]",
".",
"generalized_idxs",
"}",
"node",
".",
"generalized_idxs",
"=",
"x"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._get_word_start_index | Helper method that returns the index of the string based on node's
starting index | suffix_trees/STree.py | def _get_word_start_index(self, idx):
"""Helper method that returns the index of the string based on node's
starting index"""
i = 0
for _idx in self.word_starts[1:]:
if idx < _idx:
return i
else:
i+=1
return i | def _get_word_start_index(self, idx):
"""Helper method that returns the index of the string based on node's
starting index"""
i = 0
for _idx in self.word_starts[1:]:
if idx < _idx:
return i
else:
i+=1
return i | [
"Helper",
"method",
"that",
"returns",
"the",
"index",
"of",
"the",
"string",
"based",
"on",
"node",
"s",
"starting",
"index"
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L134-L143 | [
"def",
"_get_word_start_index",
"(",
"self",
",",
"idx",
")",
":",
"i",
"=",
"0",
"for",
"_idx",
"in",
"self",
".",
"word_starts",
"[",
"1",
":",
"]",
":",
"if",
"idx",
"<",
"_idx",
":",
"return",
"i",
"else",
":",
"i",
"+=",
"1",
"return",
"i"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree.lcs | Returns the Largest Common Substring of Strings provided in stringIdxs.
If stringIdxs is not provided, the LCS of all strings is returned.
::param stringIdxs: Optional: List of indexes of strings. | suffix_trees/STree.py | def lcs(self, stringIdxs=-1):
"""Returns the Largest Common Substring of Strings provided in stringIdxs.
If stringIdxs is not provided, the LCS of all strings is returned.
::param stringIdxs: Optional: List of indexes of strings.
"""
if stringIdxs == -1 or not isinstance(stringIdxs, list):
stringIdxs = set(range(len(self.word_starts)))
else:
stringIdxs = set(stringIdxs)
deepestNode = self._find_lcs(self.root, stringIdxs)
start = deepestNode.idx
end = deepestNode.idx + deepestNode.depth
return self.word[start:end] | def lcs(self, stringIdxs=-1):
"""Returns the Largest Common Substring of Strings provided in stringIdxs.
If stringIdxs is not provided, the LCS of all strings is returned.
::param stringIdxs: Optional: List of indexes of strings.
"""
if stringIdxs == -1 or not isinstance(stringIdxs, list):
stringIdxs = set(range(len(self.word_starts)))
else:
stringIdxs = set(stringIdxs)
deepestNode = self._find_lcs(self.root, stringIdxs)
start = deepestNode.idx
end = deepestNode.idx + deepestNode.depth
return self.word[start:end] | [
"Returns",
"the",
"Largest",
"Common",
"Substring",
"of",
"Strings",
"provided",
"in",
"stringIdxs",
".",
"If",
"stringIdxs",
"is",
"not",
"provided",
"the",
"LCS",
"of",
"all",
"strings",
"is",
"returned",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L145-L159 | [
"def",
"lcs",
"(",
"self",
",",
"stringIdxs",
"=",
"-",
"1",
")",
":",
"if",
"stringIdxs",
"==",
"-",
"1",
"or",
"not",
"isinstance",
"(",
"stringIdxs",
",",
"list",
")",
":",
"stringIdxs",
"=",
"set",
"(",
"range",
"(",
"len",
"(",
"self",
".",
"word_starts",
")",
")",
")",
"else",
":",
"stringIdxs",
"=",
"set",
"(",
"stringIdxs",
")",
"deepestNode",
"=",
"self",
".",
"_find_lcs",
"(",
"self",
".",
"root",
",",
"stringIdxs",
")",
"start",
"=",
"deepestNode",
".",
"idx",
"end",
"=",
"deepestNode",
".",
"idx",
"+",
"deepestNode",
".",
"depth",
"return",
"self",
".",
"word",
"[",
"start",
":",
"end",
"]"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._find_lcs | Helper method that finds LCS by traversing the labeled GSD. | suffix_trees/STree.py | def _find_lcs(self, node, stringIdxs):
"""Helper method that finds LCS by traversing the labeled GSD."""
nodes = [self._find_lcs(n, stringIdxs)
for (n,_) in node.transition_links
if n.generalized_idxs.issuperset(stringIdxs)]
if nodes == []:
return node
deepestNode = max(nodes, key=lambda n: n.depth)
return deepestNode | def _find_lcs(self, node, stringIdxs):
"""Helper method that finds LCS by traversing the labeled GSD."""
nodes = [self._find_lcs(n, stringIdxs)
for (n,_) in node.transition_links
if n.generalized_idxs.issuperset(stringIdxs)]
if nodes == []:
return node
deepestNode = max(nodes, key=lambda n: n.depth)
return deepestNode | [
"Helper",
"method",
"that",
"finds",
"LCS",
"by",
"traversing",
"the",
"labeled",
"GSD",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L161-L171 | [
"def",
"_find_lcs",
"(",
"self",
",",
"node",
",",
"stringIdxs",
")",
":",
"nodes",
"=",
"[",
"self",
".",
"_find_lcs",
"(",
"n",
",",
"stringIdxs",
")",
"for",
"(",
"n",
",",
"_",
")",
"in",
"node",
".",
"transition_links",
"if",
"n",
".",
"generalized_idxs",
".",
"issuperset",
"(",
"stringIdxs",
")",
"]",
"if",
"nodes",
"==",
"[",
"]",
":",
"return",
"node",
"deepestNode",
"=",
"max",
"(",
"nodes",
",",
"key",
"=",
"lambda",
"n",
":",
"n",
".",
"depth",
")",
"return",
"deepestNode"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._generalized_word_starts | Helper method returns the starting indexes of strings in GST | suffix_trees/STree.py | def _generalized_word_starts(self, xs):
"""Helper method returns the starting indexes of strings in GST"""
self.word_starts = []
i = 0
for n in range(len(xs)):
self.word_starts.append(i)
i += len(xs[n]) + 1 | def _generalized_word_starts(self, xs):
"""Helper method returns the starting indexes of strings in GST"""
self.word_starts = []
i = 0
for n in range(len(xs)):
self.word_starts.append(i)
i += len(xs[n]) + 1 | [
"Helper",
"method",
"returns",
"the",
"starting",
"indexes",
"of",
"strings",
"in",
"GST"
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L173-L179 | [
"def",
"_generalized_word_starts",
"(",
"self",
",",
"xs",
")",
":",
"self",
".",
"word_starts",
"=",
"[",
"]",
"i",
"=",
"0",
"for",
"n",
"in",
"range",
"(",
"len",
"(",
"xs",
")",
")",
":",
"self",
".",
"word_starts",
".",
"append",
"(",
"i",
")",
"i",
"+=",
"len",
"(",
"xs",
"[",
"n",
"]",
")",
"+",
"1"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree.find | Returns starting position of the substring y in the string used for
building the Suffix tree.
:param y: String
:return: Index of the starting position of string y in the string used for building the Suffix tree
-1 if y is not a substring. | suffix_trees/STree.py | def find(self, y):
"""Returns starting position of the substring y in the string used for
building the Suffix tree.
:param y: String
:return: Index of the starting position of string y in the string used for building the Suffix tree
-1 if y is not a substring.
"""
node = self.root
while True:
edge = self._edgeLabel(node, node.parent)
if edge.startswith(y):
return node.idx
i = 0
while(i < len(edge) and edge[i] == y[0]):
y = y[1:]
i += 1
if i != 0:
if i == len(edge) and y != '':
pass
else:
return -1
node = node._get_transition_link(y[0])
if not node:
return -1 | def find(self, y):
"""Returns starting position of the substring y in the string used for
building the Suffix tree.
:param y: String
:return: Index of the starting position of string y in the string used for building the Suffix tree
-1 if y is not a substring.
"""
node = self.root
while True:
edge = self._edgeLabel(node, node.parent)
if edge.startswith(y):
return node.idx
i = 0
while(i < len(edge) and edge[i] == y[0]):
y = y[1:]
i += 1
if i != 0:
if i == len(edge) and y != '':
pass
else:
return -1
node = node._get_transition_link(y[0])
if not node:
return -1 | [
"Returns",
"starting",
"position",
"of",
"the",
"substring",
"y",
"in",
"the",
"string",
"used",
"for",
"building",
"the",
"Suffix",
"tree",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L181-L208 | [
"def",
"find",
"(",
"self",
",",
"y",
")",
":",
"node",
"=",
"self",
".",
"root",
"while",
"True",
":",
"edge",
"=",
"self",
".",
"_edgeLabel",
"(",
"node",
",",
"node",
".",
"parent",
")",
"if",
"edge",
".",
"startswith",
"(",
"y",
")",
":",
"return",
"node",
".",
"idx",
"i",
"=",
"0",
"while",
"(",
"i",
"<",
"len",
"(",
"edge",
")",
"and",
"edge",
"[",
"i",
"]",
"==",
"y",
"[",
"0",
"]",
")",
":",
"y",
"=",
"y",
"[",
"1",
":",
"]",
"i",
"+=",
"1",
"if",
"i",
"!=",
"0",
":",
"if",
"i",
"==",
"len",
"(",
"edge",
")",
"and",
"y",
"!=",
"''",
":",
"pass",
"else",
":",
"return",
"-",
"1",
"node",
"=",
"node",
".",
"_get_transition_link",
"(",
"y",
"[",
"0",
"]",
")",
"if",
"not",
"node",
":",
"return",
"-",
"1"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._edgeLabel | Helper method, returns the edge label between a node and it's parent | suffix_trees/STree.py | def _edgeLabel(self, node, parent):
"""Helper method, returns the edge label between a node and it's parent"""
return self.word[node.idx + parent.depth : node.idx + node.depth] | def _edgeLabel(self, node, parent):
"""Helper method, returns the edge label between a node and it's parent"""
return self.word[node.idx + parent.depth : node.idx + node.depth] | [
"Helper",
"method",
"returns",
"the",
"edge",
"label",
"between",
"a",
"node",
"and",
"it",
"s",
"parent"
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L236-L238 | [
"def",
"_edgeLabel",
"(",
"self",
",",
"node",
",",
"parent",
")",
":",
"return",
"self",
".",
"word",
"[",
"node",
".",
"idx",
"+",
"parent",
".",
"depth",
":",
"node",
".",
"idx",
"+",
"node",
".",
"depth",
"]"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | STree._terminalSymbolsGenerator | Generator of unique terminal symbols used for building the Generalized Suffix Tree.
Unicode Private Use Area U+E000..U+F8FF is used to ensure that terminal symbols
are not part of the input string. | suffix_trees/STree.py | def _terminalSymbolsGenerator(self):
"""Generator of unique terminal symbols used for building the Generalized Suffix Tree.
Unicode Private Use Area U+E000..U+F8FF is used to ensure that terminal symbols
are not part of the input string.
"""
py2 = sys.version[0] < '3'
UPPAs = list(list(range(0xE000,0xF8FF+1)) + list(range(0xF0000,0xFFFFD+1)) + list(range(0x100000, 0x10FFFD+1)))
for i in UPPAs:
if py2:
yield(unichr(i))
else:
yield(chr(i))
raise ValueError("To many input strings.") | def _terminalSymbolsGenerator(self):
"""Generator of unique terminal symbols used for building the Generalized Suffix Tree.
Unicode Private Use Area U+E000..U+F8FF is used to ensure that terminal symbols
are not part of the input string.
"""
py2 = sys.version[0] < '3'
UPPAs = list(list(range(0xE000,0xF8FF+1)) + list(range(0xF0000,0xFFFFD+1)) + list(range(0x100000, 0x10FFFD+1)))
for i in UPPAs:
if py2:
yield(unichr(i))
else:
yield(chr(i))
raise ValueError("To many input strings.") | [
"Generator",
"of",
"unique",
"terminal",
"symbols",
"used",
"for",
"building",
"the",
"Generalized",
"Suffix",
"Tree",
".",
"Unicode",
"Private",
"Use",
"Area",
"U",
"+",
"E000",
"..",
"U",
"+",
"F8FF",
"is",
"used",
"to",
"ensure",
"that",
"terminal",
"symbols",
"are",
"not",
"part",
"of",
"the",
"input",
"string",
"."
] | ptrus/suffix-trees | python | https://github.com/ptrus/suffix-trees/blob/e7439c93a7b895523fad36c8c65a781710320b57/suffix_trees/STree.py#L241-L253 | [
"def",
"_terminalSymbolsGenerator",
"(",
"self",
")",
":",
"py2",
"=",
"sys",
".",
"version",
"[",
"0",
"]",
"<",
"'3'",
"UPPAs",
"=",
"list",
"(",
"list",
"(",
"range",
"(",
"0xE000",
",",
"0xF8FF",
"+",
"1",
")",
")",
"+",
"list",
"(",
"range",
"(",
"0xF0000",
",",
"0xFFFFD",
"+",
"1",
")",
")",
"+",
"list",
"(",
"range",
"(",
"0x100000",
",",
"0x10FFFD",
"+",
"1",
")",
")",
")",
"for",
"i",
"in",
"UPPAs",
":",
"if",
"py2",
":",
"yield",
"(",
"unichr",
"(",
"i",
")",
")",
"else",
":",
"yield",
"(",
"chr",
"(",
"i",
")",
")",
"raise",
"ValueError",
"(",
"\"To many input strings.\"",
")"
] | e7439c93a7b895523fad36c8c65a781710320b57 |
valid | MPCKMeans._dist | (x - y)^T A (x - y) | active_semi_clustering/semi_supervised/pairwise_constraints/mpckmeans.py | def _dist(self, x, y, A):
"(x - y)^T A (x - y)"
return scipy.spatial.distance.mahalanobis(x, y, A) ** 2 | def _dist(self, x, y, A):
"(x - y)^T A (x - y)"
return scipy.spatial.distance.mahalanobis(x, y, A) ** 2 | [
"(",
"x",
"-",
"y",
")",
"^T",
"A",
"(",
"x",
"-",
"y",
")"
] | datamole-ai/active-semi-supervised-clustering | python | https://github.com/datamole-ai/active-semi-supervised-clustering/blob/0dcab86ea22cd66ed7ea64234efdff1c6aca7981/active_semi_clustering/semi_supervised/pairwise_constraints/mpckmeans.py#L96-L98 | [
"def",
"_dist",
"(",
"self",
",",
"x",
",",
"y",
",",
"A",
")",
":",
"return",
"scipy",
".",
"spatial",
".",
"distance",
".",
"mahalanobis",
"(",
"x",
",",
"y",
",",
"A",
")",
"**",
"2"
] | 0dcab86ea22cd66ed7ea64234efdff1c6aca7981 |
valid | ExampleOracle.query | Query the oracle to find out whether i and j should be must-linked | active_semi_clustering/active/pairwise_constraints/example_oracle.py | def query(self, i, j):
"Query the oracle to find out whether i and j should be must-linked"
if self.queries_cnt < self.max_queries_cnt:
self.queries_cnt += 1
return self.labels[i] == self.labels[j]
else:
raise MaximumQueriesExceeded | def query(self, i, j):
"Query the oracle to find out whether i and j should be must-linked"
if self.queries_cnt < self.max_queries_cnt:
self.queries_cnt += 1
return self.labels[i] == self.labels[j]
else:
raise MaximumQueriesExceeded | [
"Query",
"the",
"oracle",
"to",
"find",
"out",
"whether",
"i",
"and",
"j",
"should",
"be",
"must",
"-",
"linked"
] | datamole-ai/active-semi-supervised-clustering | python | https://github.com/datamole-ai/active-semi-supervised-clustering/blob/0dcab86ea22cd66ed7ea64234efdff1c6aca7981/active_semi_clustering/active/pairwise_constraints/example_oracle.py#L11-L17 | [
"def",
"query",
"(",
"self",
",",
"i",
",",
"j",
")",
":",
"if",
"self",
".",
"queries_cnt",
"<",
"self",
".",
"max_queries_cnt",
":",
"self",
".",
"queries_cnt",
"+=",
"1",
"return",
"self",
".",
"labels",
"[",
"i",
"]",
"==",
"self",
".",
"labels",
"[",
"j",
"]",
"else",
":",
"raise",
"MaximumQueriesExceeded"
] | 0dcab86ea22cd66ed7ea64234efdff1c6aca7981 |
valid | preprocess_constraints | Create a graph of constraints for both must- and cannot-links | active_semi_clustering/semi_supervised/pairwise_constraints/constraints.py | def preprocess_constraints(ml, cl, n):
"Create a graph of constraints for both must- and cannot-links"
# Represent the graphs using adjacency-lists
ml_graph, cl_graph = {}, {}
for i in range(n):
ml_graph[i] = set()
cl_graph[i] = set()
def add_both(d, i, j):
d[i].add(j)
d[j].add(i)
for (i, j) in ml:
ml_graph[i].add(j)
ml_graph[j].add(i)
for (i, j) in cl:
cl_graph[i].add(j)
cl_graph[j].add(i)
def dfs(i, graph, visited, component):
visited[i] = True
for j in graph[i]:
if not visited[j]:
dfs(j, graph, visited, component)
component.append(i)
# Run DFS from each node to get all the graph's components
# and add an edge for each pair of nodes in the component (create a complete graph)
# See http://www.techiedelight.com/transitive-closure-graph/ for more details
visited = [False] * n
neighborhoods = []
for i in range(n):
if not visited[i] and ml_graph[i]:
component = []
dfs(i, ml_graph, visited, component)
for x1 in component:
for x2 in component:
if x1 != x2:
ml_graph[x1].add(x2)
neighborhoods.append(component)
for (i, j) in cl:
for x in ml_graph[i]:
add_both(cl_graph, x, j)
for y in ml_graph[j]:
add_both(cl_graph, i, y)
for x in ml_graph[i]:
for y in ml_graph[j]:
add_both(cl_graph, x, y)
for i in ml_graph:
for j in ml_graph[i]:
if j != i and j in cl_graph[i]:
raise InconsistentConstraintsException('Inconsistent constraints between {} and {}'.format(i, j))
return ml_graph, cl_graph, neighborhoods | def preprocess_constraints(ml, cl, n):
"Create a graph of constraints for both must- and cannot-links"
# Represent the graphs using adjacency-lists
ml_graph, cl_graph = {}, {}
for i in range(n):
ml_graph[i] = set()
cl_graph[i] = set()
def add_both(d, i, j):
d[i].add(j)
d[j].add(i)
for (i, j) in ml:
ml_graph[i].add(j)
ml_graph[j].add(i)
for (i, j) in cl:
cl_graph[i].add(j)
cl_graph[j].add(i)
def dfs(i, graph, visited, component):
visited[i] = True
for j in graph[i]:
if not visited[j]:
dfs(j, graph, visited, component)
component.append(i)
# Run DFS from each node to get all the graph's components
# and add an edge for each pair of nodes in the component (create a complete graph)
# See http://www.techiedelight.com/transitive-closure-graph/ for more details
visited = [False] * n
neighborhoods = []
for i in range(n):
if not visited[i] and ml_graph[i]:
component = []
dfs(i, ml_graph, visited, component)
for x1 in component:
for x2 in component:
if x1 != x2:
ml_graph[x1].add(x2)
neighborhoods.append(component)
for (i, j) in cl:
for x in ml_graph[i]:
add_both(cl_graph, x, j)
for y in ml_graph[j]:
add_both(cl_graph, i, y)
for x in ml_graph[i]:
for y in ml_graph[j]:
add_both(cl_graph, x, y)
for i in ml_graph:
for j in ml_graph[i]:
if j != i and j in cl_graph[i]:
raise InconsistentConstraintsException('Inconsistent constraints between {} and {}'.format(i, j))
return ml_graph, cl_graph, neighborhoods | [
"Create",
"a",
"graph",
"of",
"constraints",
"for",
"both",
"must",
"-",
"and",
"cannot",
"-",
"links"
] | datamole-ai/active-semi-supervised-clustering | python | https://github.com/datamole-ai/active-semi-supervised-clustering/blob/0dcab86ea22cd66ed7ea64234efdff1c6aca7981/active_semi_clustering/semi_supervised/pairwise_constraints/constraints.py#L5-L64 | [
"def",
"preprocess_constraints",
"(",
"ml",
",",
"cl",
",",
"n",
")",
":",
"# Represent the graphs using adjacency-lists",
"ml_graph",
",",
"cl_graph",
"=",
"{",
"}",
",",
"{",
"}",
"for",
"i",
"in",
"range",
"(",
"n",
")",
":",
"ml_graph",
"[",
"i",
"]",
"=",
"set",
"(",
")",
"cl_graph",
"[",
"i",
"]",
"=",
"set",
"(",
")",
"def",
"add_both",
"(",
"d",
",",
"i",
",",
"j",
")",
":",
"d",
"[",
"i",
"]",
".",
"add",
"(",
"j",
")",
"d",
"[",
"j",
"]",
".",
"add",
"(",
"i",
")",
"for",
"(",
"i",
",",
"j",
")",
"in",
"ml",
":",
"ml_graph",
"[",
"i",
"]",
".",
"add",
"(",
"j",
")",
"ml_graph",
"[",
"j",
"]",
".",
"add",
"(",
"i",
")",
"for",
"(",
"i",
",",
"j",
")",
"in",
"cl",
":",
"cl_graph",
"[",
"i",
"]",
".",
"add",
"(",
"j",
")",
"cl_graph",
"[",
"j",
"]",
".",
"add",
"(",
"i",
")",
"def",
"dfs",
"(",
"i",
",",
"graph",
",",
"visited",
",",
"component",
")",
":",
"visited",
"[",
"i",
"]",
"=",
"True",
"for",
"j",
"in",
"graph",
"[",
"i",
"]",
":",
"if",
"not",
"visited",
"[",
"j",
"]",
":",
"dfs",
"(",
"j",
",",
"graph",
",",
"visited",
",",
"component",
")",
"component",
".",
"append",
"(",
"i",
")",
"# Run DFS from each node to get all the graph's components",
"# and add an edge for each pair of nodes in the component (create a complete graph)",
"# See http://www.techiedelight.com/transitive-closure-graph/ for more details",
"visited",
"=",
"[",
"False",
"]",
"*",
"n",
"neighborhoods",
"=",
"[",
"]",
"for",
"i",
"in",
"range",
"(",
"n",
")",
":",
"if",
"not",
"visited",
"[",
"i",
"]",
"and",
"ml_graph",
"[",
"i",
"]",
":",
"component",
"=",
"[",
"]",
"dfs",
"(",
"i",
",",
"ml_graph",
",",
"visited",
",",
"component",
")",
"for",
"x1",
"in",
"component",
":",
"for",
"x2",
"in",
"component",
":",
"if",
"x1",
"!=",
"x2",
":",
"ml_graph",
"[",
"x1",
"]",
".",
"add",
"(",
"x2",
")",
"neighborhoods",
".",
"append",
"(",
"component",
")",
"for",
"(",
"i",
",",
"j",
")",
"in",
"cl",
":",
"for",
"x",
"in",
"ml_graph",
"[",
"i",
"]",
":",
"add_both",
"(",
"cl_graph",
",",
"x",
",",
"j",
")",
"for",
"y",
"in",
"ml_graph",
"[",
"j",
"]",
":",
"add_both",
"(",
"cl_graph",
",",
"i",
",",
"y",
")",
"for",
"x",
"in",
"ml_graph",
"[",
"i",
"]",
":",
"for",
"y",
"in",
"ml_graph",
"[",
"j",
"]",
":",
"add_both",
"(",
"cl_graph",
",",
"x",
",",
"y",
")",
"for",
"i",
"in",
"ml_graph",
":",
"for",
"j",
"in",
"ml_graph",
"[",
"i",
"]",
":",
"if",
"j",
"!=",
"i",
"and",
"j",
"in",
"cl_graph",
"[",
"i",
"]",
":",
"raise",
"InconsistentConstraintsException",
"(",
"'Inconsistent constraints between {} and {}'",
".",
"format",
"(",
"i",
",",
"j",
")",
")",
"return",
"ml_graph",
",",
"cl_graph",
",",
"neighborhoods"
] | 0dcab86ea22cd66ed7ea64234efdff1c6aca7981 |
valid | make_pmml_pipeline | Translates a regular Scikit-Learn estimator or pipeline to a PMML pipeline.
Parameters:
----------
obj: BaseEstimator
The object.
active_fields: list of strings, optional
Feature names. If missing, "x1", "x2", .., "xn" are assumed.
target_fields: list of strings, optional
Label name(s). If missing, "y" is assumed. | sklearn2pmml/__init__.py | def make_pmml_pipeline(obj, active_fields = None, target_fields = None):
"""Translates a regular Scikit-Learn estimator or pipeline to a PMML pipeline.
Parameters:
----------
obj: BaseEstimator
The object.
active_fields: list of strings, optional
Feature names. If missing, "x1", "x2", .., "xn" are assumed.
target_fields: list of strings, optional
Label name(s). If missing, "y" is assumed.
"""
steps = _filter_steps(_get_steps(obj))
pipeline = PMMLPipeline(steps)
if active_fields is not None:
pipeline.active_fields = numpy.asarray(active_fields)
if target_fields is not None:
pipeline.target_fields = numpy.asarray(target_fields)
return pipeline | def make_pmml_pipeline(obj, active_fields = None, target_fields = None):
"""Translates a regular Scikit-Learn estimator or pipeline to a PMML pipeline.
Parameters:
----------
obj: BaseEstimator
The object.
active_fields: list of strings, optional
Feature names. If missing, "x1", "x2", .., "xn" are assumed.
target_fields: list of strings, optional
Label name(s). If missing, "y" is assumed.
"""
steps = _filter_steps(_get_steps(obj))
pipeline = PMMLPipeline(steps)
if active_fields is not None:
pipeline.active_fields = numpy.asarray(active_fields)
if target_fields is not None:
pipeline.target_fields = numpy.asarray(target_fields)
return pipeline | [
"Translates",
"a",
"regular",
"Scikit",
"-",
"Learn",
"estimator",
"or",
"pipeline",
"to",
"a",
"PMML",
"pipeline",
"."
] | jpmml/sklearn2pmml | python | https://github.com/jpmml/sklearn2pmml/blob/0a455a54323989473c16f9efc9a77cef3ff64c1e/sklearn2pmml/__init__.py#L107-L128 | [
"def",
"make_pmml_pipeline",
"(",
"obj",
",",
"active_fields",
"=",
"None",
",",
"target_fields",
"=",
"None",
")",
":",
"steps",
"=",
"_filter_steps",
"(",
"_get_steps",
"(",
"obj",
")",
")",
"pipeline",
"=",
"PMMLPipeline",
"(",
"steps",
")",
"if",
"active_fields",
"is",
"not",
"None",
":",
"pipeline",
".",
"active_fields",
"=",
"numpy",
".",
"asarray",
"(",
"active_fields",
")",
"if",
"target_fields",
"is",
"not",
"None",
":",
"pipeline",
".",
"target_fields",
"=",
"numpy",
".",
"asarray",
"(",
"target_fields",
")",
"return",
"pipeline"
] | 0a455a54323989473c16f9efc9a77cef3ff64c1e |
valid | sklearn2pmml | Converts a fitted Scikit-Learn pipeline to PMML.
Parameters:
----------
pipeline: PMMLPipeline
The pipeline.
pmml: string
The path to where the PMML document should be stored.
user_classpath: list of strings, optional
The paths to JAR files that provide custom Transformer, Selector and/or Estimator converter classes.
The JPMML-SkLearn classpath is constructed by appending user JAR files to package JAR files.
with_repr: boolean, optional
If true, insert the string representation of pipeline into the PMML document.
debug: boolean, optional
If true, print information about the conversion process.
java_encoding: string, optional
The character encoding to use for decoding Java output and error byte streams. | sklearn2pmml/__init__.py | def sklearn2pmml(pipeline, pmml, user_classpath = [], with_repr = False, debug = False, java_encoding = "UTF-8"):
"""Converts a fitted Scikit-Learn pipeline to PMML.
Parameters:
----------
pipeline: PMMLPipeline
The pipeline.
pmml: string
The path to where the PMML document should be stored.
user_classpath: list of strings, optional
The paths to JAR files that provide custom Transformer, Selector and/or Estimator converter classes.
The JPMML-SkLearn classpath is constructed by appending user JAR files to package JAR files.
with_repr: boolean, optional
If true, insert the string representation of pipeline into the PMML document.
debug: boolean, optional
If true, print information about the conversion process.
java_encoding: string, optional
The character encoding to use for decoding Java output and error byte streams.
"""
if debug:
java_version = _java_version(java_encoding)
if java_version is None:
java_version = ("java", "N/A")
print("python: {0}".format(platform.python_version()))
print("sklearn: {0}".format(sklearn.__version__))
print("sklearn.externals.joblib: {0}".format(joblib.__version__))
print("pandas: {0}".format(pandas.__version__))
print("sklearn_pandas: {0}".format(sklearn_pandas.__version__))
print("sklearn2pmml: {0}".format(__version__))
print("{0}: {1}".format(java_version[0], java_version[1]))
if not isinstance(pipeline, PMMLPipeline):
raise TypeError("The pipeline object is not an instance of " + PMMLPipeline.__name__ + ". Use the 'sklearn2pmml.make_pmml_pipeline(obj)' utility function to translate a regular Scikit-Learn estimator or pipeline to a PMML pipeline")
estimator = pipeline._final_estimator
cmd = ["java", "-cp", os.pathsep.join(_classpath(user_classpath)), "org.jpmml.sklearn.Main"]
dumps = []
try:
if with_repr:
pipeline.repr_ = repr(pipeline)
# if isinstance(estimator, H2OEstimator):
if hasattr(estimator, "download_mojo"):
estimator_mojo = estimator.download_mojo()
dumps.append(estimator_mojo)
estimator._mojo_path = estimator_mojo
pipeline_pkl = _dump(pipeline, "pipeline")
cmd.extend(["--pkl-pipeline-input", pipeline_pkl])
dumps.append(pipeline_pkl)
cmd.extend(["--pmml-output", pmml])
if debug:
print("Executing command:\n{0}".format(" ".join(cmd)))
try:
process = Popen(cmd, stdout = PIPE, stderr = PIPE, bufsize = 1)
except OSError:
raise RuntimeError("Java is not installed, or the Java executable is not on system path")
output, error = process.communicate()
retcode = process.poll()
if debug or retcode:
if(len(output) > 0):
print("Standard output:\n{0}".format(_decode(output, java_encoding)))
else:
print("Standard output is empty")
if(len(error) > 0):
print("Standard error:\n{0}".format(_decode(error, java_encoding)))
else:
print("Standard error is empty")
if retcode:
raise RuntimeError("The JPMML-SkLearn conversion application has failed. The Java executable should have printed more information about the failure into its standard output and/or standard error streams")
finally:
if debug:
print("Preserved joblib dump file(s): {0}".format(" ".join(dumps)))
else:
for dump in dumps:
os.remove(dump) | def sklearn2pmml(pipeline, pmml, user_classpath = [], with_repr = False, debug = False, java_encoding = "UTF-8"):
"""Converts a fitted Scikit-Learn pipeline to PMML.
Parameters:
----------
pipeline: PMMLPipeline
The pipeline.
pmml: string
The path to where the PMML document should be stored.
user_classpath: list of strings, optional
The paths to JAR files that provide custom Transformer, Selector and/or Estimator converter classes.
The JPMML-SkLearn classpath is constructed by appending user JAR files to package JAR files.
with_repr: boolean, optional
If true, insert the string representation of pipeline into the PMML document.
debug: boolean, optional
If true, print information about the conversion process.
java_encoding: string, optional
The character encoding to use for decoding Java output and error byte streams.
"""
if debug:
java_version = _java_version(java_encoding)
if java_version is None:
java_version = ("java", "N/A")
print("python: {0}".format(platform.python_version()))
print("sklearn: {0}".format(sklearn.__version__))
print("sklearn.externals.joblib: {0}".format(joblib.__version__))
print("pandas: {0}".format(pandas.__version__))
print("sklearn_pandas: {0}".format(sklearn_pandas.__version__))
print("sklearn2pmml: {0}".format(__version__))
print("{0}: {1}".format(java_version[0], java_version[1]))
if not isinstance(pipeline, PMMLPipeline):
raise TypeError("The pipeline object is not an instance of " + PMMLPipeline.__name__ + ". Use the 'sklearn2pmml.make_pmml_pipeline(obj)' utility function to translate a regular Scikit-Learn estimator or pipeline to a PMML pipeline")
estimator = pipeline._final_estimator
cmd = ["java", "-cp", os.pathsep.join(_classpath(user_classpath)), "org.jpmml.sklearn.Main"]
dumps = []
try:
if with_repr:
pipeline.repr_ = repr(pipeline)
# if isinstance(estimator, H2OEstimator):
if hasattr(estimator, "download_mojo"):
estimator_mojo = estimator.download_mojo()
dumps.append(estimator_mojo)
estimator._mojo_path = estimator_mojo
pipeline_pkl = _dump(pipeline, "pipeline")
cmd.extend(["--pkl-pipeline-input", pipeline_pkl])
dumps.append(pipeline_pkl)
cmd.extend(["--pmml-output", pmml])
if debug:
print("Executing command:\n{0}".format(" ".join(cmd)))
try:
process = Popen(cmd, stdout = PIPE, stderr = PIPE, bufsize = 1)
except OSError:
raise RuntimeError("Java is not installed, or the Java executable is not on system path")
output, error = process.communicate()
retcode = process.poll()
if debug or retcode:
if(len(output) > 0):
print("Standard output:\n{0}".format(_decode(output, java_encoding)))
else:
print("Standard output is empty")
if(len(error) > 0):
print("Standard error:\n{0}".format(_decode(error, java_encoding)))
else:
print("Standard error is empty")
if retcode:
raise RuntimeError("The JPMML-SkLearn conversion application has failed. The Java executable should have printed more information about the failure into its standard output and/or standard error streams")
finally:
if debug:
print("Preserved joblib dump file(s): {0}".format(" ".join(dumps)))
else:
for dump in dumps:
os.remove(dump) | [
"Converts",
"a",
"fitted",
"Scikit",
"-",
"Learn",
"pipeline",
"to",
"PMML",
"."
] | jpmml/sklearn2pmml | python | https://github.com/jpmml/sklearn2pmml/blob/0a455a54323989473c16f9efc9a77cef3ff64c1e/sklearn2pmml/__init__.py#L181-L258 | [
"def",
"sklearn2pmml",
"(",
"pipeline",
",",
"pmml",
",",
"user_classpath",
"=",
"[",
"]",
",",
"with_repr",
"=",
"False",
",",
"debug",
"=",
"False",
",",
"java_encoding",
"=",
"\"UTF-8\"",
")",
":",
"if",
"debug",
":",
"java_version",
"=",
"_java_version",
"(",
"java_encoding",
")",
"if",
"java_version",
"is",
"None",
":",
"java_version",
"=",
"(",
"\"java\"",
",",
"\"N/A\"",
")",
"print",
"(",
"\"python: {0}\"",
".",
"format",
"(",
"platform",
".",
"python_version",
"(",
")",
")",
")",
"print",
"(",
"\"sklearn: {0}\"",
".",
"format",
"(",
"sklearn",
".",
"__version__",
")",
")",
"print",
"(",
"\"sklearn.externals.joblib: {0}\"",
".",
"format",
"(",
"joblib",
".",
"__version__",
")",
")",
"print",
"(",
"\"pandas: {0}\"",
".",
"format",
"(",
"pandas",
".",
"__version__",
")",
")",
"print",
"(",
"\"sklearn_pandas: {0}\"",
".",
"format",
"(",
"sklearn_pandas",
".",
"__version__",
")",
")",
"print",
"(",
"\"sklearn2pmml: {0}\"",
".",
"format",
"(",
"__version__",
")",
")",
"print",
"(",
"\"{0}: {1}\"",
".",
"format",
"(",
"java_version",
"[",
"0",
"]",
",",
"java_version",
"[",
"1",
"]",
")",
")",
"if",
"not",
"isinstance",
"(",
"pipeline",
",",
"PMMLPipeline",
")",
":",
"raise",
"TypeError",
"(",
"\"The pipeline object is not an instance of \"",
"+",
"PMMLPipeline",
".",
"__name__",
"+",
"\". Use the 'sklearn2pmml.make_pmml_pipeline(obj)' utility function to translate a regular Scikit-Learn estimator or pipeline to a PMML pipeline\"",
")",
"estimator",
"=",
"pipeline",
".",
"_final_estimator",
"cmd",
"=",
"[",
"\"java\"",
",",
"\"-cp\"",
",",
"os",
".",
"pathsep",
".",
"join",
"(",
"_classpath",
"(",
"user_classpath",
")",
")",
",",
"\"org.jpmml.sklearn.Main\"",
"]",
"dumps",
"=",
"[",
"]",
"try",
":",
"if",
"with_repr",
":",
"pipeline",
".",
"repr_",
"=",
"repr",
"(",
"pipeline",
")",
"# if isinstance(estimator, H2OEstimator):",
"if",
"hasattr",
"(",
"estimator",
",",
"\"download_mojo\"",
")",
":",
"estimator_mojo",
"=",
"estimator",
".",
"download_mojo",
"(",
")",
"dumps",
".",
"append",
"(",
"estimator_mojo",
")",
"estimator",
".",
"_mojo_path",
"=",
"estimator_mojo",
"pipeline_pkl",
"=",
"_dump",
"(",
"pipeline",
",",
"\"pipeline\"",
")",
"cmd",
".",
"extend",
"(",
"[",
"\"--pkl-pipeline-input\"",
",",
"pipeline_pkl",
"]",
")",
"dumps",
".",
"append",
"(",
"pipeline_pkl",
")",
"cmd",
".",
"extend",
"(",
"[",
"\"--pmml-output\"",
",",
"pmml",
"]",
")",
"if",
"debug",
":",
"print",
"(",
"\"Executing command:\\n{0}\"",
".",
"format",
"(",
"\" \"",
".",
"join",
"(",
"cmd",
")",
")",
")",
"try",
":",
"process",
"=",
"Popen",
"(",
"cmd",
",",
"stdout",
"=",
"PIPE",
",",
"stderr",
"=",
"PIPE",
",",
"bufsize",
"=",
"1",
")",
"except",
"OSError",
":",
"raise",
"RuntimeError",
"(",
"\"Java is not installed, or the Java executable is not on system path\"",
")",
"output",
",",
"error",
"=",
"process",
".",
"communicate",
"(",
")",
"retcode",
"=",
"process",
".",
"poll",
"(",
")",
"if",
"debug",
"or",
"retcode",
":",
"if",
"(",
"len",
"(",
"output",
")",
">",
"0",
")",
":",
"print",
"(",
"\"Standard output:\\n{0}\"",
".",
"format",
"(",
"_decode",
"(",
"output",
",",
"java_encoding",
")",
")",
")",
"else",
":",
"print",
"(",
"\"Standard output is empty\"",
")",
"if",
"(",
"len",
"(",
"error",
")",
">",
"0",
")",
":",
"print",
"(",
"\"Standard error:\\n{0}\"",
".",
"format",
"(",
"_decode",
"(",
"error",
",",
"java_encoding",
")",
")",
")",
"else",
":",
"print",
"(",
"\"Standard error is empty\"",
")",
"if",
"retcode",
":",
"raise",
"RuntimeError",
"(",
"\"The JPMML-SkLearn conversion application has failed. The Java executable should have printed more information about the failure into its standard output and/or standard error streams\"",
")",
"finally",
":",
"if",
"debug",
":",
"print",
"(",
"\"Preserved joblib dump file(s): {0}\"",
".",
"format",
"(",
"\" \"",
".",
"join",
"(",
"dumps",
")",
")",
")",
"else",
":",
"for",
"dump",
"in",
"dumps",
":",
"os",
".",
"remove",
"(",
"dump",
")"
] | 0a455a54323989473c16f9efc9a77cef3ff64c1e |
valid | make_tpot_pmml_config | Translates a regular TPOT configuration to a PMML-compatible TPOT configuration.
Parameters:
----------
obj: config
The configuration dictionary.
user_classpath: list of strings, optional
The paths to JAR files that provide custom Transformer, Selector and/or Estimator converter classes.
The JPMML-SkLearn classpath is constructed by appending user JAR files to package JAR files. | sklearn2pmml/__init__.py | def make_tpot_pmml_config(config, user_classpath = []):
"""Translates a regular TPOT configuration to a PMML-compatible TPOT configuration.
Parameters:
----------
obj: config
The configuration dictionary.
user_classpath: list of strings, optional
The paths to JAR files that provide custom Transformer, Selector and/or Estimator converter classes.
The JPMML-SkLearn classpath is constructed by appending user JAR files to package JAR files.
"""
tpot_keys = set(config.keys())
classes = _supported_classes(user_classpath)
pmml_keys = (set(classes)).union(set([_strip_module(class_) for class_ in classes]))
return { key : config[key] for key in (tpot_keys).intersection(pmml_keys)} | def make_tpot_pmml_config(config, user_classpath = []):
"""Translates a regular TPOT configuration to a PMML-compatible TPOT configuration.
Parameters:
----------
obj: config
The configuration dictionary.
user_classpath: list of strings, optional
The paths to JAR files that provide custom Transformer, Selector and/or Estimator converter classes.
The JPMML-SkLearn classpath is constructed by appending user JAR files to package JAR files.
"""
tpot_keys = set(config.keys())
classes = _supported_classes(user_classpath)
pmml_keys = (set(classes)).union(set([_strip_module(class_) for class_ in classes]))
return { key : config[key] for key in (tpot_keys).intersection(pmml_keys)} | [
"Translates",
"a",
"regular",
"TPOT",
"configuration",
"to",
"a",
"PMML",
"-",
"compatible",
"TPOT",
"configuration",
"."
] | jpmml/sklearn2pmml | python | https://github.com/jpmml/sklearn2pmml/blob/0a455a54323989473c16f9efc9a77cef3ff64c1e/sklearn2pmml/__init__.py#L284-L300 | [
"def",
"make_tpot_pmml_config",
"(",
"config",
",",
"user_classpath",
"=",
"[",
"]",
")",
":",
"tpot_keys",
"=",
"set",
"(",
"config",
".",
"keys",
"(",
")",
")",
"classes",
"=",
"_supported_classes",
"(",
"user_classpath",
")",
"pmml_keys",
"=",
"(",
"set",
"(",
"classes",
")",
")",
".",
"union",
"(",
"set",
"(",
"[",
"_strip_module",
"(",
"class_",
")",
"for",
"class_",
"in",
"classes",
"]",
")",
")",
"return",
"{",
"key",
":",
"config",
"[",
"key",
"]",
"for",
"key",
"in",
"(",
"tpot_keys",
")",
".",
"intersection",
"(",
"pmml_keys",
")",
"}"
] | 0a455a54323989473c16f9efc9a77cef3ff64c1e |
valid | BaseFormSetFactory.construct_formset | Returns an instance of the formset | extra_views/formsets.py | def construct_formset(self):
"""
Returns an instance of the formset
"""
formset_class = self.get_formset()
if hasattr(self, 'get_extra_form_kwargs'):
klass = type(self).__name__
raise DeprecationWarning(
'Calling {0}.get_extra_form_kwargs is no longer supported. '
'Set `form_kwargs` in {0}.formset_kwargs or override '
'{0}.get_formset_kwargs() directly.'.format(klass),
)
return formset_class(**self.get_formset_kwargs()) | def construct_formset(self):
"""
Returns an instance of the formset
"""
formset_class = self.get_formset()
if hasattr(self, 'get_extra_form_kwargs'):
klass = type(self).__name__
raise DeprecationWarning(
'Calling {0}.get_extra_form_kwargs is no longer supported. '
'Set `form_kwargs` in {0}.formset_kwargs or override '
'{0}.get_formset_kwargs() directly.'.format(klass),
)
return formset_class(**self.get_formset_kwargs()) | [
"Returns",
"an",
"instance",
"of",
"the",
"formset"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L24-L36 | [
"def",
"construct_formset",
"(",
"self",
")",
":",
"formset_class",
"=",
"self",
".",
"get_formset",
"(",
")",
"if",
"hasattr",
"(",
"self",
",",
"'get_extra_form_kwargs'",
")",
":",
"klass",
"=",
"type",
"(",
"self",
")",
".",
"__name__",
"raise",
"DeprecationWarning",
"(",
"'Calling {0}.get_extra_form_kwargs is no longer supported. '",
"'Set `form_kwargs` in {0}.formset_kwargs or override '",
"'{0}.get_formset_kwargs() directly.'",
".",
"format",
"(",
"klass",
")",
",",
")",
"return",
"formset_class",
"(",
"*",
"*",
"self",
".",
"get_formset_kwargs",
"(",
")",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseFormSetFactory.get_formset_kwargs | Returns the keyword arguments for instantiating the formset. | extra_views/formsets.py | def get_formset_kwargs(self):
"""
Returns the keyword arguments for instantiating the formset.
"""
kwargs = self.formset_kwargs.copy()
kwargs.update({
'initial': self.get_initial(),
'prefix': self.get_prefix(),
})
if self.request.method in ('POST', 'PUT'):
kwargs.update({
'data': self.request.POST.copy(),
'files': self.request.FILES,
})
return kwargs | def get_formset_kwargs(self):
"""
Returns the keyword arguments for instantiating the formset.
"""
kwargs = self.formset_kwargs.copy()
kwargs.update({
'initial': self.get_initial(),
'prefix': self.get_prefix(),
})
if self.request.method in ('POST', 'PUT'):
kwargs.update({
'data': self.request.POST.copy(),
'files': self.request.FILES,
})
return kwargs | [
"Returns",
"the",
"keyword",
"arguments",
"for",
"instantiating",
"the",
"formset",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L68-L83 | [
"def",
"get_formset_kwargs",
"(",
"self",
")",
":",
"kwargs",
"=",
"self",
".",
"formset_kwargs",
".",
"copy",
"(",
")",
"kwargs",
".",
"update",
"(",
"{",
"'initial'",
":",
"self",
".",
"get_initial",
"(",
")",
",",
"'prefix'",
":",
"self",
".",
"get_prefix",
"(",
")",
",",
"}",
")",
"if",
"self",
".",
"request",
".",
"method",
"in",
"(",
"'POST'",
",",
"'PUT'",
")",
":",
"kwargs",
".",
"update",
"(",
"{",
"'data'",
":",
"self",
".",
"request",
".",
"POST",
".",
"copy",
"(",
")",
",",
"'files'",
":",
"self",
".",
"request",
".",
"FILES",
",",
"}",
")",
"return",
"kwargs"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseFormSetFactory.get_factory_kwargs | Returns the keyword arguments for calling the formset factory | extra_views/formsets.py | def get_factory_kwargs(self):
"""
Returns the keyword arguments for calling the formset factory
"""
# Perform deprecation check
for attr in ['extra', 'max_num', 'can_order', 'can_delete', 'ct_field',
'formfield_callback', 'fk_name', 'widgets', 'ct_fk_field']:
if hasattr(self, attr):
klass = type(self).__name__
raise DeprecationWarning(
'Setting `{0}.{1}` at the class level is now deprecated. '
'Set `{0}.factory_kwargs` instead.'.format(klass, attr)
)
kwargs = self.factory_kwargs.copy()
if self.get_formset_class():
kwargs['formset'] = self.get_formset_class()
return kwargs | def get_factory_kwargs(self):
"""
Returns the keyword arguments for calling the formset factory
"""
# Perform deprecation check
for attr in ['extra', 'max_num', 'can_order', 'can_delete', 'ct_field',
'formfield_callback', 'fk_name', 'widgets', 'ct_fk_field']:
if hasattr(self, attr):
klass = type(self).__name__
raise DeprecationWarning(
'Setting `{0}.{1}` at the class level is now deprecated. '
'Set `{0}.factory_kwargs` instead.'.format(klass, attr)
)
kwargs = self.factory_kwargs.copy()
if self.get_formset_class():
kwargs['formset'] = self.get_formset_class()
return kwargs | [
"Returns",
"the",
"keyword",
"arguments",
"for",
"calling",
"the",
"formset",
"factory"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L85-L102 | [
"def",
"get_factory_kwargs",
"(",
"self",
")",
":",
"# Perform deprecation check",
"for",
"attr",
"in",
"[",
"'extra'",
",",
"'max_num'",
",",
"'can_order'",
",",
"'can_delete'",
",",
"'ct_field'",
",",
"'formfield_callback'",
",",
"'fk_name'",
",",
"'widgets'",
",",
"'ct_fk_field'",
"]",
":",
"if",
"hasattr",
"(",
"self",
",",
"attr",
")",
":",
"klass",
"=",
"type",
"(",
"self",
")",
".",
"__name__",
"raise",
"DeprecationWarning",
"(",
"'Setting `{0}.{1}` at the class level is now deprecated. '",
"'Set `{0}.factory_kwargs` instead.'",
".",
"format",
"(",
"klass",
",",
"attr",
")",
")",
"kwargs",
"=",
"self",
".",
"factory_kwargs",
".",
"copy",
"(",
")",
"if",
"self",
".",
"get_formset_class",
"(",
")",
":",
"kwargs",
"[",
"'formset'",
"]",
"=",
"self",
".",
"get_formset_class",
"(",
")",
"return",
"kwargs"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | FormSetMixin.get_success_url | Returns the supplied URL. | extra_views/formsets.py | def get_success_url(self):
"""
Returns the supplied URL.
"""
if self.success_url:
url = self.success_url
else:
# Default to returning to the same page
url = self.request.get_full_path()
return url | def get_success_url(self):
"""
Returns the supplied URL.
"""
if self.success_url:
url = self.success_url
else:
# Default to returning to the same page
url = self.request.get_full_path()
return url | [
"Returns",
"the",
"supplied",
"URL",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L120-L129 | [
"def",
"get_success_url",
"(",
"self",
")",
":",
"if",
"self",
".",
"success_url",
":",
"url",
"=",
"self",
".",
"success_url",
"else",
":",
"# Default to returning to the same page",
"url",
"=",
"self",
".",
"request",
".",
"get_full_path",
"(",
")",
"return",
"url"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ModelFormSetMixin.get_formset_kwargs | Returns the keyword arguments for instantiating the formset. | extra_views/formsets.py | def get_formset_kwargs(self):
"""
Returns the keyword arguments for instantiating the formset.
"""
kwargs = super(ModelFormSetMixin, self).get_formset_kwargs()
kwargs['queryset'] = self.get_queryset()
return kwargs | def get_formset_kwargs(self):
"""
Returns the keyword arguments for instantiating the formset.
"""
kwargs = super(ModelFormSetMixin, self).get_formset_kwargs()
kwargs['queryset'] = self.get_queryset()
return kwargs | [
"Returns",
"the",
"keyword",
"arguments",
"for",
"instantiating",
"the",
"formset",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L155-L161 | [
"def",
"get_formset_kwargs",
"(",
"self",
")",
":",
"kwargs",
"=",
"super",
"(",
"ModelFormSetMixin",
",",
"self",
")",
".",
"get_formset_kwargs",
"(",
")",
"kwargs",
"[",
"'queryset'",
"]",
"=",
"self",
".",
"get_queryset",
"(",
")",
"return",
"kwargs"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ModelFormSetMixin.formset_valid | If the formset is valid, save the associated models. | extra_views/formsets.py | def formset_valid(self, formset):
"""
If the formset is valid, save the associated models.
"""
self.object_list = formset.save()
return super(ModelFormSetMixin, self).formset_valid(formset) | def formset_valid(self, formset):
"""
If the formset is valid, save the associated models.
"""
self.object_list = formset.save()
return super(ModelFormSetMixin, self).formset_valid(formset) | [
"If",
"the",
"formset",
"is",
"valid",
"save",
"the",
"associated",
"models",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L181-L186 | [
"def",
"formset_valid",
"(",
"self",
",",
"formset",
")",
":",
"self",
".",
"object_list",
"=",
"formset",
".",
"save",
"(",
")",
"return",
"super",
"(",
"ModelFormSetMixin",
",",
"self",
")",
".",
"formset_valid",
"(",
"formset",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseInlineFormSetFactory.get_formset_kwargs | Returns the keyword arguments for instantiating the formset. | extra_views/formsets.py | def get_formset_kwargs(self):
"""
Returns the keyword arguments for instantiating the formset.
"""
# Perform deprecation check
if hasattr(self, 'save_as_new'):
klass = type(self).__name__
raise DeprecationWarning(
'Setting `{0}.save_as_new` at the class level is now '
'deprecated. Set `{0}.formset_kwargs` instead.'.format(klass)
)
kwargs = super(BaseInlineFormSetFactory, self).get_formset_kwargs()
kwargs['instance'] = self.object
return kwargs | def get_formset_kwargs(self):
"""
Returns the keyword arguments for instantiating the formset.
"""
# Perform deprecation check
if hasattr(self, 'save_as_new'):
klass = type(self).__name__
raise DeprecationWarning(
'Setting `{0}.save_as_new` at the class level is now '
'deprecated. Set `{0}.formset_kwargs` instead.'.format(klass)
)
kwargs = super(BaseInlineFormSetFactory, self).get_formset_kwargs()
kwargs['instance'] = self.object
return kwargs | [
"Returns",
"the",
"keyword",
"arguments",
"for",
"instantiating",
"the",
"formset",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L206-L219 | [
"def",
"get_formset_kwargs",
"(",
"self",
")",
":",
"# Perform deprecation check",
"if",
"hasattr",
"(",
"self",
",",
"'save_as_new'",
")",
":",
"klass",
"=",
"type",
"(",
"self",
")",
".",
"__name__",
"raise",
"DeprecationWarning",
"(",
"'Setting `{0}.save_as_new` at the class level is now '",
"'deprecated. Set `{0}.formset_kwargs` instead.'",
".",
"format",
"(",
"klass",
")",
")",
"kwargs",
"=",
"super",
"(",
"BaseInlineFormSetFactory",
",",
"self",
")",
".",
"get_formset_kwargs",
"(",
")",
"kwargs",
"[",
"'instance'",
"]",
"=",
"self",
".",
"object",
"return",
"kwargs"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseInlineFormSetFactory.get_factory_kwargs | Returns the keyword arguments for calling the formset factory | extra_views/formsets.py | def get_factory_kwargs(self):
"""
Returns the keyword arguments for calling the formset factory
"""
kwargs = super(BaseInlineFormSetFactory, self).get_factory_kwargs()
kwargs.setdefault('fields', self.fields)
kwargs.setdefault('exclude', self.exclude)
if self.get_form_class():
kwargs['form'] = self.get_form_class()
return kwargs | def get_factory_kwargs(self):
"""
Returns the keyword arguments for calling the formset factory
"""
kwargs = super(BaseInlineFormSetFactory, self).get_factory_kwargs()
kwargs.setdefault('fields', self.fields)
kwargs.setdefault('exclude', self.exclude)
if self.get_form_class():
kwargs['form'] = self.get_form_class()
return kwargs | [
"Returns",
"the",
"keyword",
"arguments",
"for",
"calling",
"the",
"formset",
"factory"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L221-L231 | [
"def",
"get_factory_kwargs",
"(",
"self",
")",
":",
"kwargs",
"=",
"super",
"(",
"BaseInlineFormSetFactory",
",",
"self",
")",
".",
"get_factory_kwargs",
"(",
")",
"kwargs",
".",
"setdefault",
"(",
"'fields'",
",",
"self",
".",
"fields",
")",
"kwargs",
".",
"setdefault",
"(",
"'exclude'",
",",
"self",
".",
"exclude",
")",
"if",
"self",
".",
"get_form_class",
"(",
")",
":",
"kwargs",
"[",
"'form'",
"]",
"=",
"self",
".",
"get_form_class",
"(",
")",
"return",
"kwargs"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ProcessFormSetView.get | Handles GET requests and instantiates a blank version of the formset. | extra_views/formsets.py | def get(self, request, *args, **kwargs):
"""
Handles GET requests and instantiates a blank version of the formset.
"""
formset = self.construct_formset()
return self.render_to_response(self.get_context_data(formset=formset)) | def get(self, request, *args, **kwargs):
"""
Handles GET requests and instantiates a blank version of the formset.
"""
formset = self.construct_formset()
return self.render_to_response(self.get_context_data(formset=formset)) | [
"Handles",
"GET",
"requests",
"and",
"instantiates",
"a",
"blank",
"version",
"of",
"the",
"formset",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L264-L269 | [
"def",
"get",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"formset",
"=",
"self",
".",
"construct_formset",
"(",
")",
"return",
"self",
".",
"render_to_response",
"(",
"self",
".",
"get_context_data",
"(",
"formset",
"=",
"formset",
")",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ProcessFormSetView.post | Handles POST requests, instantiating a formset instance with the passed
POST variables and then checked for validity. | extra_views/formsets.py | def post(self, request, *args, **kwargs):
"""
Handles POST requests, instantiating a formset instance with the passed
POST variables and then checked for validity.
"""
formset = self.construct_formset()
if formset.is_valid():
return self.formset_valid(formset)
else:
return self.formset_invalid(formset) | def post(self, request, *args, **kwargs):
"""
Handles POST requests, instantiating a formset instance with the passed
POST variables and then checked for validity.
"""
formset = self.construct_formset()
if formset.is_valid():
return self.formset_valid(formset)
else:
return self.formset_invalid(formset) | [
"Handles",
"POST",
"requests",
"instantiating",
"a",
"formset",
"instance",
"with",
"the",
"passed",
"POST",
"variables",
"and",
"then",
"checked",
"for",
"validity",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/formsets.py#L271-L280 | [
"def",
"post",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"formset",
"=",
"self",
".",
"construct_formset",
"(",
")",
"if",
"formset",
".",
"is_valid",
"(",
")",
":",
"return",
"self",
".",
"formset_valid",
"(",
"formset",
")",
"else",
":",
"return",
"self",
".",
"formset_invalid",
"(",
"formset",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | InlineFormSetFactory.construct_formset | Overrides construct_formset to attach the model class as
an attribute of the returned formset instance. | extra_views/advanced.py | def construct_formset(self):
"""
Overrides construct_formset to attach the model class as
an attribute of the returned formset instance.
"""
formset = super(InlineFormSetFactory, self).construct_formset()
formset.model = self.inline_model
return formset | def construct_formset(self):
"""
Overrides construct_formset to attach the model class as
an attribute of the returned formset instance.
"""
formset = super(InlineFormSetFactory, self).construct_formset()
formset.model = self.inline_model
return formset | [
"Overrides",
"construct_formset",
"to",
"attach",
"the",
"model",
"class",
"as",
"an",
"attribute",
"of",
"the",
"returned",
"formset",
"instance",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L28-L35 | [
"def",
"construct_formset",
"(",
"self",
")",
":",
"formset",
"=",
"super",
"(",
"InlineFormSetFactory",
",",
"self",
")",
".",
"construct_formset",
"(",
")",
"formset",
".",
"model",
"=",
"self",
".",
"inline_model",
"return",
"formset"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ModelFormWithInlinesMixin.forms_valid | If the form and formsets are valid, save the associated models. | extra_views/advanced.py | def forms_valid(self, form, inlines):
"""
If the form and formsets are valid, save the associated models.
"""
response = self.form_valid(form)
for formset in inlines:
formset.save()
return response | def forms_valid(self, form, inlines):
"""
If the form and formsets are valid, save the associated models.
"""
response = self.form_valid(form)
for formset in inlines:
formset.save()
return response | [
"If",
"the",
"form",
"and",
"formsets",
"are",
"valid",
"save",
"the",
"associated",
"models",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L61-L68 | [
"def",
"forms_valid",
"(",
"self",
",",
"form",
",",
"inlines",
")",
":",
"response",
"=",
"self",
".",
"form_valid",
"(",
"form",
")",
"for",
"formset",
"in",
"inlines",
":",
"formset",
".",
"save",
"(",
")",
"return",
"response"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ModelFormWithInlinesMixin.forms_invalid | If the form or formsets are invalid, re-render the context data with the
data-filled form and formsets and errors. | extra_views/advanced.py | def forms_invalid(self, form, inlines):
"""
If the form or formsets are invalid, re-render the context data with the
data-filled form and formsets and errors.
"""
return self.render_to_response(self.get_context_data(form=form, inlines=inlines)) | def forms_invalid(self, form, inlines):
"""
If the form or formsets are invalid, re-render the context data with the
data-filled form and formsets and errors.
"""
return self.render_to_response(self.get_context_data(form=form, inlines=inlines)) | [
"If",
"the",
"form",
"or",
"formsets",
"are",
"invalid",
"re",
"-",
"render",
"the",
"context",
"data",
"with",
"the",
"data",
"-",
"filled",
"form",
"and",
"formsets",
"and",
"errors",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L70-L75 | [
"def",
"forms_invalid",
"(",
"self",
",",
"form",
",",
"inlines",
")",
":",
"return",
"self",
".",
"render_to_response",
"(",
"self",
".",
"get_context_data",
"(",
"form",
"=",
"form",
",",
"inlines",
"=",
"inlines",
")",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ModelFormWithInlinesMixin.construct_inlines | Returns the inline formset instances | extra_views/advanced.py | def construct_inlines(self):
"""
Returns the inline formset instances
"""
inline_formsets = []
for inline_class in self.get_inlines():
inline_instance = inline_class(self.model, self.request, self.object, self.kwargs, self)
inline_formset = inline_instance.construct_formset()
inline_formsets.append(inline_formset)
return inline_formsets | def construct_inlines(self):
"""
Returns the inline formset instances
"""
inline_formsets = []
for inline_class in self.get_inlines():
inline_instance = inline_class(self.model, self.request, self.object, self.kwargs, self)
inline_formset = inline_instance.construct_formset()
inline_formsets.append(inline_formset)
return inline_formsets | [
"Returns",
"the",
"inline",
"formset",
"instances"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L77-L86 | [
"def",
"construct_inlines",
"(",
"self",
")",
":",
"inline_formsets",
"=",
"[",
"]",
"for",
"inline_class",
"in",
"self",
".",
"get_inlines",
"(",
")",
":",
"inline_instance",
"=",
"inline_class",
"(",
"self",
".",
"model",
",",
"self",
".",
"request",
",",
"self",
".",
"object",
",",
"self",
".",
"kwargs",
",",
"self",
")",
"inline_formset",
"=",
"inline_instance",
".",
"construct_formset",
"(",
")",
"inline_formsets",
".",
"append",
"(",
"inline_formset",
")",
"return",
"inline_formsets"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ProcessFormWithInlinesView.get | Handles GET requests and instantiates a blank version of the form and formsets. | extra_views/advanced.py | def get(self, request, *args, **kwargs):
"""
Handles GET requests and instantiates a blank version of the form and formsets.
"""
form_class = self.get_form_class()
form = self.get_form(form_class)
inlines = self.construct_inlines()
return self.render_to_response(self.get_context_data(form=form, inlines=inlines, **kwargs)) | def get(self, request, *args, **kwargs):
"""
Handles GET requests and instantiates a blank version of the form and formsets.
"""
form_class = self.get_form_class()
form = self.get_form(form_class)
inlines = self.construct_inlines()
return self.render_to_response(self.get_context_data(form=form, inlines=inlines, **kwargs)) | [
"Handles",
"GET",
"requests",
"and",
"instantiates",
"a",
"blank",
"version",
"of",
"the",
"form",
"and",
"formsets",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L94-L101 | [
"def",
"get",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"form_class",
"=",
"self",
".",
"get_form_class",
"(",
")",
"form",
"=",
"self",
".",
"get_form",
"(",
"form_class",
")",
"inlines",
"=",
"self",
".",
"construct_inlines",
"(",
")",
"return",
"self",
".",
"render_to_response",
"(",
"self",
".",
"get_context_data",
"(",
"form",
"=",
"form",
",",
"inlines",
"=",
"inlines",
",",
"*",
"*",
"kwargs",
")",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | ProcessFormWithInlinesView.post | Handles POST requests, instantiating a form and formset instances with the passed
POST variables and then checked for validity. | extra_views/advanced.py | def post(self, request, *args, **kwargs):
"""
Handles POST requests, instantiating a form and formset instances with the passed
POST variables and then checked for validity.
"""
form_class = self.get_form_class()
form = self.get_form(form_class)
if form.is_valid():
self.object = form.save(commit=False)
form_validated = True
else:
form_validated = False
inlines = self.construct_inlines()
if all_valid(inlines) and form_validated:
return self.forms_valid(form, inlines)
return self.forms_invalid(form, inlines) | def post(self, request, *args, **kwargs):
"""
Handles POST requests, instantiating a form and formset instances with the passed
POST variables and then checked for validity.
"""
form_class = self.get_form_class()
form = self.get_form(form_class)
if form.is_valid():
self.object = form.save(commit=False)
form_validated = True
else:
form_validated = False
inlines = self.construct_inlines()
if all_valid(inlines) and form_validated:
return self.forms_valid(form, inlines)
return self.forms_invalid(form, inlines) | [
"Handles",
"POST",
"requests",
"instantiating",
"a",
"form",
"and",
"formset",
"instances",
"with",
"the",
"passed",
"POST",
"variables",
"and",
"then",
"checked",
"for",
"validity",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L103-L121 | [
"def",
"post",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"form_class",
"=",
"self",
".",
"get_form_class",
"(",
")",
"form",
"=",
"self",
".",
"get_form",
"(",
"form_class",
")",
"if",
"form",
".",
"is_valid",
"(",
")",
":",
"self",
".",
"object",
"=",
"form",
".",
"save",
"(",
"commit",
"=",
"False",
")",
"form_validated",
"=",
"True",
"else",
":",
"form_validated",
"=",
"False",
"inlines",
"=",
"self",
".",
"construct_inlines",
"(",
")",
"if",
"all_valid",
"(",
"inlines",
")",
"and",
"form_validated",
":",
"return",
"self",
".",
"forms_valid",
"(",
"form",
",",
"inlines",
")",
"return",
"self",
".",
"forms_invalid",
"(",
"form",
",",
"inlines",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | NamedFormsetsMixin.get_context_data | If `inlines_names` has been defined, add each formset to the context under
its corresponding entry in `inlines_names` | extra_views/advanced.py | def get_context_data(self, **kwargs):
"""
If `inlines_names` has been defined, add each formset to the context under
its corresponding entry in `inlines_names`
"""
context = {}
inlines_names = self.get_inlines_names()
if inlines_names:
# We have formset or inlines in context, but never both
context.update(zip(inlines_names, kwargs.get('inlines', [])))
if 'formset' in kwargs:
context[inlines_names[0]] = kwargs['formset']
context.update(kwargs)
return super(NamedFormsetsMixin, self).get_context_data(**context) | def get_context_data(self, **kwargs):
"""
If `inlines_names` has been defined, add each formset to the context under
its corresponding entry in `inlines_names`
"""
context = {}
inlines_names = self.get_inlines_names()
if inlines_names:
# We have formset or inlines in context, but never both
context.update(zip(inlines_names, kwargs.get('inlines', [])))
if 'formset' in kwargs:
context[inlines_names[0]] = kwargs['formset']
context.update(kwargs)
return super(NamedFormsetsMixin, self).get_context_data(**context) | [
"If",
"inlines_names",
"has",
"been",
"defined",
"add",
"each",
"formset",
"to",
"the",
"context",
"under",
"its",
"corresponding",
"entry",
"in",
"inlines_names"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/advanced.py#L190-L204 | [
"def",
"get_context_data",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"context",
"=",
"{",
"}",
"inlines_names",
"=",
"self",
".",
"get_inlines_names",
"(",
")",
"if",
"inlines_names",
":",
"# We have formset or inlines in context, but never both",
"context",
".",
"update",
"(",
"zip",
"(",
"inlines_names",
",",
"kwargs",
".",
"get",
"(",
"'inlines'",
",",
"[",
"]",
")",
")",
")",
"if",
"'formset'",
"in",
"kwargs",
":",
"context",
"[",
"inlines_names",
"[",
"0",
"]",
"]",
"=",
"kwargs",
"[",
"'formset'",
"]",
"context",
".",
"update",
"(",
"kwargs",
")",
"return",
"super",
"(",
"NamedFormsetsMixin",
",",
"self",
")",
".",
"get_context_data",
"(",
"*",
"*",
"context",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | SearchableListMixin.try_convert_to_date | Tries to convert word to date(datetime) using search_date_formats
Return None if word fits no one format | extra_views/contrib/mixins.py | def try_convert_to_date(self, word):
"""
Tries to convert word to date(datetime) using search_date_formats
Return None if word fits no one format
"""
for frm in self.search_date_formats:
try:
return datetime.datetime.strptime(word, frm).date()
except ValueError:
pass
return None | def try_convert_to_date(self, word):
"""
Tries to convert word to date(datetime) using search_date_formats
Return None if word fits no one format
"""
for frm in self.search_date_formats:
try:
return datetime.datetime.strptime(word, frm).date()
except ValueError:
pass
return None | [
"Tries",
"to",
"convert",
"word",
"to",
"date",
"(",
"datetime",
")",
"using",
"search_date_formats",
"Return",
"None",
"if",
"word",
"fits",
"no",
"one",
"format"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/contrib/mixins.py#L60-L70 | [
"def",
"try_convert_to_date",
"(",
"self",
",",
"word",
")",
":",
"for",
"frm",
"in",
"self",
".",
"search_date_formats",
":",
"try",
":",
"return",
"datetime",
".",
"datetime",
".",
"strptime",
"(",
"word",
",",
"frm",
")",
".",
"date",
"(",
")",
"except",
"ValueError",
":",
"pass",
"return",
"None"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | SortHelper.get_params_for_field | If sort_type is None - inverse current sort for field, if no sorted - use asc | extra_views/contrib/mixins.py | def get_params_for_field(self, field_name, sort_type=None):
"""
If sort_type is None - inverse current sort for field, if no sorted - use asc
"""
if not sort_type:
if self.initial_sort == field_name:
sort_type = 'desc' if self.initial_sort_type == 'asc' else 'asc'
else:
sort_type = 'asc'
self.initial_params[self.sort_param_name] = self.sort_fields[field_name]
self.initial_params[self.sort_type_param_name] = sort_type
return '?%s' % self.initial_params.urlencode() | def get_params_for_field(self, field_name, sort_type=None):
"""
If sort_type is None - inverse current sort for field, if no sorted - use asc
"""
if not sort_type:
if self.initial_sort == field_name:
sort_type = 'desc' if self.initial_sort_type == 'asc' else 'asc'
else:
sort_type = 'asc'
self.initial_params[self.sort_param_name] = self.sort_fields[field_name]
self.initial_params[self.sort_type_param_name] = sort_type
return '?%s' % self.initial_params.urlencode() | [
"If",
"sort_type",
"is",
"None",
"-",
"inverse",
"current",
"sort",
"for",
"field",
"if",
"no",
"sorted",
"-",
"use",
"asc"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/contrib/mixins.py#L120-L131 | [
"def",
"get_params_for_field",
"(",
"self",
",",
"field_name",
",",
"sort_type",
"=",
"None",
")",
":",
"if",
"not",
"sort_type",
":",
"if",
"self",
".",
"initial_sort",
"==",
"field_name",
":",
"sort_type",
"=",
"'desc'",
"if",
"self",
".",
"initial_sort_type",
"==",
"'asc'",
"else",
"'asc'",
"else",
":",
"sort_type",
"=",
"'asc'",
"self",
".",
"initial_params",
"[",
"self",
".",
"sort_param_name",
"]",
"=",
"self",
".",
"sort_fields",
"[",
"field_name",
"]",
"self",
".",
"initial_params",
"[",
"self",
".",
"sort_type_param_name",
"]",
"=",
"sort_type",
"return",
"'?%s'",
"%",
"self",
".",
"initial_params",
".",
"urlencode",
"(",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseCalendarMonthView.get_start_date | Returns the start date for a model instance | extra_views/dates.py | def get_start_date(self, obj):
"""
Returns the start date for a model instance
"""
obj_date = getattr(obj, self.get_date_field())
try:
obj_date = obj_date.date()
except AttributeError:
# It's a date rather than datetime, so we use it as is
pass
return obj_date | def get_start_date(self, obj):
"""
Returns the start date for a model instance
"""
obj_date = getattr(obj, self.get_date_field())
try:
obj_date = obj_date.date()
except AttributeError:
# It's a date rather than datetime, so we use it as is
pass
return obj_date | [
"Returns",
"the",
"start",
"date",
"for",
"a",
"model",
"instance"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/dates.py#L56-L66 | [
"def",
"get_start_date",
"(",
"self",
",",
"obj",
")",
":",
"obj_date",
"=",
"getattr",
"(",
"obj",
",",
"self",
".",
"get_date_field",
"(",
")",
")",
"try",
":",
"obj_date",
"=",
"obj_date",
".",
"date",
"(",
")",
"except",
"AttributeError",
":",
"# It's a date rather than datetime, so we use it as is",
"pass",
"return",
"obj_date"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseCalendarMonthView.get_end_date | Returns the end date for a model instance | extra_views/dates.py | def get_end_date(self, obj):
"""
Returns the end date for a model instance
"""
obj_date = getattr(obj, self.get_end_date_field())
try:
obj_date = obj_date.date()
except AttributeError:
# It's a date rather than datetime, so we use it as is
pass
return obj_date | def get_end_date(self, obj):
"""
Returns the end date for a model instance
"""
obj_date = getattr(obj, self.get_end_date_field())
try:
obj_date = obj_date.date()
except AttributeError:
# It's a date rather than datetime, so we use it as is
pass
return obj_date | [
"Returns",
"the",
"end",
"date",
"for",
"a",
"model",
"instance"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/dates.py#L68-L78 | [
"def",
"get_end_date",
"(",
"self",
",",
"obj",
")",
":",
"obj_date",
"=",
"getattr",
"(",
"obj",
",",
"self",
".",
"get_end_date_field",
"(",
")",
")",
"try",
":",
"obj_date",
"=",
"obj_date",
".",
"date",
"(",
")",
"except",
"AttributeError",
":",
"# It's a date rather than datetime, so we use it as is",
"pass",
"return",
"obj_date"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseCalendarMonthView.get_first_of_week | Returns an integer representing the first day of the week.
0 represents Monday, 6 represents Sunday. | extra_views/dates.py | def get_first_of_week(self):
"""
Returns an integer representing the first day of the week.
0 represents Monday, 6 represents Sunday.
"""
if self.first_of_week is None:
raise ImproperlyConfigured("%s.first_of_week is required." % self.__class__.__name__)
if self.first_of_week not in range(7):
raise ImproperlyConfigured("%s.first_of_week must be an integer between 0 and 6." % self.__class__.__name__)
return self.first_of_week | def get_first_of_week(self):
"""
Returns an integer representing the first day of the week.
0 represents Monday, 6 represents Sunday.
"""
if self.first_of_week is None:
raise ImproperlyConfigured("%s.first_of_week is required." % self.__class__.__name__)
if self.first_of_week not in range(7):
raise ImproperlyConfigured("%s.first_of_week must be an integer between 0 and 6." % self.__class__.__name__)
return self.first_of_week | [
"Returns",
"an",
"integer",
"representing",
"the",
"first",
"day",
"of",
"the",
"week",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/dates.py#L80-L90 | [
"def",
"get_first_of_week",
"(",
"self",
")",
":",
"if",
"self",
".",
"first_of_week",
"is",
"None",
":",
"raise",
"ImproperlyConfigured",
"(",
"\"%s.first_of_week is required.\"",
"%",
"self",
".",
"__class__",
".",
"__name__",
")",
"if",
"self",
".",
"first_of_week",
"not",
"in",
"range",
"(",
"7",
")",
":",
"raise",
"ImproperlyConfigured",
"(",
"\"%s.first_of_week must be an integer between 0 and 6.\"",
"%",
"self",
".",
"__class__",
".",
"__name__",
")",
"return",
"self",
".",
"first_of_week"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseCalendarMonthView.get_queryset | Returns a queryset of models for the month requested | extra_views/dates.py | def get_queryset(self):
"""
Returns a queryset of models for the month requested
"""
qs = super(BaseCalendarMonthView, self).get_queryset()
year = self.get_year()
month = self.get_month()
date_field = self.get_date_field()
end_date_field = self.get_end_date_field()
date = _date_from_string(year, self.get_year_format(),
month, self.get_month_format())
since = date
until = self.get_next_month(date)
# Adjust our start and end dates to allow for next and previous
# month edges
if since.weekday() != self.get_first_of_week():
diff = math.fabs(since.weekday() - self.get_first_of_week())
since = since - datetime.timedelta(days=diff)
if until.weekday() != ((self.get_first_of_week() + 6) % 7):
diff = math.fabs(((self.get_first_of_week() + 6) % 7) - until.weekday())
until = until + datetime.timedelta(days=diff)
if end_date_field:
# 5 possible conditions for showing an event:
# 1) Single day event, starts after 'since'
# 2) Multi-day event, starts after 'since' and ends before 'until'
# 3) Starts before 'since' and ends after 'since' and before 'until'
# 4) Starts after 'since' but before 'until' and ends after 'until'
# 5) Starts before 'since' and ends after 'until'
predicate1 = Q(**{
'%s__gte' % date_field: since,
end_date_field: None
})
predicate2 = Q(**{
'%s__gte' % date_field: since,
'%s__lt' % end_date_field: until
})
predicate3 = Q(**{
'%s__lt' % date_field: since,
'%s__gte' % end_date_field: since,
'%s__lt' % end_date_field: until
})
predicate4 = Q(**{
'%s__gte' % date_field: since,
'%s__lt' % date_field: until,
'%s__gte' % end_date_field: until
})
predicate5 = Q(**{
'%s__lt' % date_field: since,
'%s__gte' % end_date_field: until
})
return qs.filter(predicate1 | predicate2 | predicate3 | predicate4 | predicate5)
return qs.filter(**{
'%s__gte' % date_field: since
}) | def get_queryset(self):
"""
Returns a queryset of models for the month requested
"""
qs = super(BaseCalendarMonthView, self).get_queryset()
year = self.get_year()
month = self.get_month()
date_field = self.get_date_field()
end_date_field = self.get_end_date_field()
date = _date_from_string(year, self.get_year_format(),
month, self.get_month_format())
since = date
until = self.get_next_month(date)
# Adjust our start and end dates to allow for next and previous
# month edges
if since.weekday() != self.get_first_of_week():
diff = math.fabs(since.weekday() - self.get_first_of_week())
since = since - datetime.timedelta(days=diff)
if until.weekday() != ((self.get_first_of_week() + 6) % 7):
diff = math.fabs(((self.get_first_of_week() + 6) % 7) - until.weekday())
until = until + datetime.timedelta(days=diff)
if end_date_field:
# 5 possible conditions for showing an event:
# 1) Single day event, starts after 'since'
# 2) Multi-day event, starts after 'since' and ends before 'until'
# 3) Starts before 'since' and ends after 'since' and before 'until'
# 4) Starts after 'since' but before 'until' and ends after 'until'
# 5) Starts before 'since' and ends after 'until'
predicate1 = Q(**{
'%s__gte' % date_field: since,
end_date_field: None
})
predicate2 = Q(**{
'%s__gte' % date_field: since,
'%s__lt' % end_date_field: until
})
predicate3 = Q(**{
'%s__lt' % date_field: since,
'%s__gte' % end_date_field: since,
'%s__lt' % end_date_field: until
})
predicate4 = Q(**{
'%s__gte' % date_field: since,
'%s__lt' % date_field: until,
'%s__gte' % end_date_field: until
})
predicate5 = Q(**{
'%s__lt' % date_field: since,
'%s__gte' % end_date_field: until
})
return qs.filter(predicate1 | predicate2 | predicate3 | predicate4 | predicate5)
return qs.filter(**{
'%s__gte' % date_field: since
}) | [
"Returns",
"a",
"queryset",
"of",
"models",
"for",
"the",
"month",
"requested"
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/dates.py#L92-L153 | [
"def",
"get_queryset",
"(",
"self",
")",
":",
"qs",
"=",
"super",
"(",
"BaseCalendarMonthView",
",",
"self",
")",
".",
"get_queryset",
"(",
")",
"year",
"=",
"self",
".",
"get_year",
"(",
")",
"month",
"=",
"self",
".",
"get_month",
"(",
")",
"date_field",
"=",
"self",
".",
"get_date_field",
"(",
")",
"end_date_field",
"=",
"self",
".",
"get_end_date_field",
"(",
")",
"date",
"=",
"_date_from_string",
"(",
"year",
",",
"self",
".",
"get_year_format",
"(",
")",
",",
"month",
",",
"self",
".",
"get_month_format",
"(",
")",
")",
"since",
"=",
"date",
"until",
"=",
"self",
".",
"get_next_month",
"(",
"date",
")",
"# Adjust our start and end dates to allow for next and previous",
"# month edges",
"if",
"since",
".",
"weekday",
"(",
")",
"!=",
"self",
".",
"get_first_of_week",
"(",
")",
":",
"diff",
"=",
"math",
".",
"fabs",
"(",
"since",
".",
"weekday",
"(",
")",
"-",
"self",
".",
"get_first_of_week",
"(",
")",
")",
"since",
"=",
"since",
"-",
"datetime",
".",
"timedelta",
"(",
"days",
"=",
"diff",
")",
"if",
"until",
".",
"weekday",
"(",
")",
"!=",
"(",
"(",
"self",
".",
"get_first_of_week",
"(",
")",
"+",
"6",
")",
"%",
"7",
")",
":",
"diff",
"=",
"math",
".",
"fabs",
"(",
"(",
"(",
"self",
".",
"get_first_of_week",
"(",
")",
"+",
"6",
")",
"%",
"7",
")",
"-",
"until",
".",
"weekday",
"(",
")",
")",
"until",
"=",
"until",
"+",
"datetime",
".",
"timedelta",
"(",
"days",
"=",
"diff",
")",
"if",
"end_date_field",
":",
"# 5 possible conditions for showing an event:",
"# 1) Single day event, starts after 'since'",
"# 2) Multi-day event, starts after 'since' and ends before 'until'",
"# 3) Starts before 'since' and ends after 'since' and before 'until'",
"# 4) Starts after 'since' but before 'until' and ends after 'until'",
"# 5) Starts before 'since' and ends after 'until'",
"predicate1",
"=",
"Q",
"(",
"*",
"*",
"{",
"'%s__gte'",
"%",
"date_field",
":",
"since",
",",
"end_date_field",
":",
"None",
"}",
")",
"predicate2",
"=",
"Q",
"(",
"*",
"*",
"{",
"'%s__gte'",
"%",
"date_field",
":",
"since",
",",
"'%s__lt'",
"%",
"end_date_field",
":",
"until",
"}",
")",
"predicate3",
"=",
"Q",
"(",
"*",
"*",
"{",
"'%s__lt'",
"%",
"date_field",
":",
"since",
",",
"'%s__gte'",
"%",
"end_date_field",
":",
"since",
",",
"'%s__lt'",
"%",
"end_date_field",
":",
"until",
"}",
")",
"predicate4",
"=",
"Q",
"(",
"*",
"*",
"{",
"'%s__gte'",
"%",
"date_field",
":",
"since",
",",
"'%s__lt'",
"%",
"date_field",
":",
"until",
",",
"'%s__gte'",
"%",
"end_date_field",
":",
"until",
"}",
")",
"predicate5",
"=",
"Q",
"(",
"*",
"*",
"{",
"'%s__lt'",
"%",
"date_field",
":",
"since",
",",
"'%s__gte'",
"%",
"end_date_field",
":",
"until",
"}",
")",
"return",
"qs",
".",
"filter",
"(",
"predicate1",
"|",
"predicate2",
"|",
"predicate3",
"|",
"predicate4",
"|",
"predicate5",
")",
"return",
"qs",
".",
"filter",
"(",
"*",
"*",
"{",
"'%s__gte'",
"%",
"date_field",
":",
"since",
"}",
")"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | BaseCalendarMonthView.get_context_data | Injects variables necessary for rendering the calendar into the context.
Variables added are: `calendar`, `weekdays`, `month`, `next_month` and `previous_month`. | extra_views/dates.py | def get_context_data(self, **kwargs):
"""
Injects variables necessary for rendering the calendar into the context.
Variables added are: `calendar`, `weekdays`, `month`, `next_month` and `previous_month`.
"""
data = super(BaseCalendarMonthView, self).get_context_data(**kwargs)
year = self.get_year()
month = self.get_month()
date = _date_from_string(year, self.get_year_format(),
month, self.get_month_format())
cal = Calendar(self.get_first_of_week())
month_calendar = []
now = datetime.datetime.utcnow()
date_lists = defaultdict(list)
multidate_objs = []
for obj in data['object_list']:
obj_date = self.get_start_date(obj)
end_date_field = self.get_end_date_field()
if end_date_field:
end_date = self.get_end_date(obj)
if end_date and end_date != obj_date:
multidate_objs.append({
'obj': obj,
'range': [x for x in daterange(obj_date, end_date)]
})
continue # We don't put multi-day events in date_lists
date_lists[obj_date].append(obj)
for week in cal.monthdatescalendar(date.year, date.month):
week_range = set(daterange(week[0], week[6]))
week_events = []
for val in multidate_objs:
intersect_length = len(week_range.intersection(val['range']))
if intersect_length:
# Event happens during this week
slot = 1
width = intersect_length # How many days is the event during this week?
nowrap_previous = True # Does the event continue from the previous week?
nowrap_next = True # Does the event continue to the next week?
if val['range'][0] >= week[0]:
slot = 1 + (val['range'][0] - week[0]).days
else:
nowrap_previous = False
if val['range'][-1] > week[6]:
nowrap_next = False
week_events.append({
'event': val['obj'],
'slot': slot,
'width': width,
'nowrap_previous': nowrap_previous,
'nowrap_next': nowrap_next,
})
week_calendar = {
'events': week_events,
'date_list': [],
}
for day in week:
week_calendar['date_list'].append({
'day': day,
'events': date_lists[day],
'today': day == now.date(),
'is_current_month': day.month == date.month,
})
month_calendar.append(week_calendar)
data['calendar'] = month_calendar
data['weekdays'] = [DAYS[x] for x in cal.iterweekdays()]
data['month'] = date
data['next_month'] = self.get_next_month(date)
data['previous_month'] = self.get_previous_month(date)
return data | def get_context_data(self, **kwargs):
"""
Injects variables necessary for rendering the calendar into the context.
Variables added are: `calendar`, `weekdays`, `month`, `next_month` and `previous_month`.
"""
data = super(BaseCalendarMonthView, self).get_context_data(**kwargs)
year = self.get_year()
month = self.get_month()
date = _date_from_string(year, self.get_year_format(),
month, self.get_month_format())
cal = Calendar(self.get_first_of_week())
month_calendar = []
now = datetime.datetime.utcnow()
date_lists = defaultdict(list)
multidate_objs = []
for obj in data['object_list']:
obj_date = self.get_start_date(obj)
end_date_field = self.get_end_date_field()
if end_date_field:
end_date = self.get_end_date(obj)
if end_date and end_date != obj_date:
multidate_objs.append({
'obj': obj,
'range': [x for x in daterange(obj_date, end_date)]
})
continue # We don't put multi-day events in date_lists
date_lists[obj_date].append(obj)
for week in cal.monthdatescalendar(date.year, date.month):
week_range = set(daterange(week[0], week[6]))
week_events = []
for val in multidate_objs:
intersect_length = len(week_range.intersection(val['range']))
if intersect_length:
# Event happens during this week
slot = 1
width = intersect_length # How many days is the event during this week?
nowrap_previous = True # Does the event continue from the previous week?
nowrap_next = True # Does the event continue to the next week?
if val['range'][0] >= week[0]:
slot = 1 + (val['range'][0] - week[0]).days
else:
nowrap_previous = False
if val['range'][-1] > week[6]:
nowrap_next = False
week_events.append({
'event': val['obj'],
'slot': slot,
'width': width,
'nowrap_previous': nowrap_previous,
'nowrap_next': nowrap_next,
})
week_calendar = {
'events': week_events,
'date_list': [],
}
for day in week:
week_calendar['date_list'].append({
'day': day,
'events': date_lists[day],
'today': day == now.date(),
'is_current_month': day.month == date.month,
})
month_calendar.append(week_calendar)
data['calendar'] = month_calendar
data['weekdays'] = [DAYS[x] for x in cal.iterweekdays()]
data['month'] = date
data['next_month'] = self.get_next_month(date)
data['previous_month'] = self.get_previous_month(date)
return data | [
"Injects",
"variables",
"necessary",
"for",
"rendering",
"the",
"calendar",
"into",
"the",
"context",
"."
] | AndrewIngram/django-extra-views | python | https://github.com/AndrewIngram/django-extra-views/blob/188e1bf1f15a44d9a599028d020083af9fb43ea7/extra_views/dates.py#L155-L239 | [
"def",
"get_context_data",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"data",
"=",
"super",
"(",
"BaseCalendarMonthView",
",",
"self",
")",
".",
"get_context_data",
"(",
"*",
"*",
"kwargs",
")",
"year",
"=",
"self",
".",
"get_year",
"(",
")",
"month",
"=",
"self",
".",
"get_month",
"(",
")",
"date",
"=",
"_date_from_string",
"(",
"year",
",",
"self",
".",
"get_year_format",
"(",
")",
",",
"month",
",",
"self",
".",
"get_month_format",
"(",
")",
")",
"cal",
"=",
"Calendar",
"(",
"self",
".",
"get_first_of_week",
"(",
")",
")",
"month_calendar",
"=",
"[",
"]",
"now",
"=",
"datetime",
".",
"datetime",
".",
"utcnow",
"(",
")",
"date_lists",
"=",
"defaultdict",
"(",
"list",
")",
"multidate_objs",
"=",
"[",
"]",
"for",
"obj",
"in",
"data",
"[",
"'object_list'",
"]",
":",
"obj_date",
"=",
"self",
".",
"get_start_date",
"(",
"obj",
")",
"end_date_field",
"=",
"self",
".",
"get_end_date_field",
"(",
")",
"if",
"end_date_field",
":",
"end_date",
"=",
"self",
".",
"get_end_date",
"(",
"obj",
")",
"if",
"end_date",
"and",
"end_date",
"!=",
"obj_date",
":",
"multidate_objs",
".",
"append",
"(",
"{",
"'obj'",
":",
"obj",
",",
"'range'",
":",
"[",
"x",
"for",
"x",
"in",
"daterange",
"(",
"obj_date",
",",
"end_date",
")",
"]",
"}",
")",
"continue",
"# We don't put multi-day events in date_lists",
"date_lists",
"[",
"obj_date",
"]",
".",
"append",
"(",
"obj",
")",
"for",
"week",
"in",
"cal",
".",
"monthdatescalendar",
"(",
"date",
".",
"year",
",",
"date",
".",
"month",
")",
":",
"week_range",
"=",
"set",
"(",
"daterange",
"(",
"week",
"[",
"0",
"]",
",",
"week",
"[",
"6",
"]",
")",
")",
"week_events",
"=",
"[",
"]",
"for",
"val",
"in",
"multidate_objs",
":",
"intersect_length",
"=",
"len",
"(",
"week_range",
".",
"intersection",
"(",
"val",
"[",
"'range'",
"]",
")",
")",
"if",
"intersect_length",
":",
"# Event happens during this week",
"slot",
"=",
"1",
"width",
"=",
"intersect_length",
"# How many days is the event during this week?",
"nowrap_previous",
"=",
"True",
"# Does the event continue from the previous week?",
"nowrap_next",
"=",
"True",
"# Does the event continue to the next week?",
"if",
"val",
"[",
"'range'",
"]",
"[",
"0",
"]",
">=",
"week",
"[",
"0",
"]",
":",
"slot",
"=",
"1",
"+",
"(",
"val",
"[",
"'range'",
"]",
"[",
"0",
"]",
"-",
"week",
"[",
"0",
"]",
")",
".",
"days",
"else",
":",
"nowrap_previous",
"=",
"False",
"if",
"val",
"[",
"'range'",
"]",
"[",
"-",
"1",
"]",
">",
"week",
"[",
"6",
"]",
":",
"nowrap_next",
"=",
"False",
"week_events",
".",
"append",
"(",
"{",
"'event'",
":",
"val",
"[",
"'obj'",
"]",
",",
"'slot'",
":",
"slot",
",",
"'width'",
":",
"width",
",",
"'nowrap_previous'",
":",
"nowrap_previous",
",",
"'nowrap_next'",
":",
"nowrap_next",
",",
"}",
")",
"week_calendar",
"=",
"{",
"'events'",
":",
"week_events",
",",
"'date_list'",
":",
"[",
"]",
",",
"}",
"for",
"day",
"in",
"week",
":",
"week_calendar",
"[",
"'date_list'",
"]",
".",
"append",
"(",
"{",
"'day'",
":",
"day",
",",
"'events'",
":",
"date_lists",
"[",
"day",
"]",
",",
"'today'",
":",
"day",
"==",
"now",
".",
"date",
"(",
")",
",",
"'is_current_month'",
":",
"day",
".",
"month",
"==",
"date",
".",
"month",
",",
"}",
")",
"month_calendar",
".",
"append",
"(",
"week_calendar",
")",
"data",
"[",
"'calendar'",
"]",
"=",
"month_calendar",
"data",
"[",
"'weekdays'",
"]",
"=",
"[",
"DAYS",
"[",
"x",
"]",
"for",
"x",
"in",
"cal",
".",
"iterweekdays",
"(",
")",
"]",
"data",
"[",
"'month'",
"]",
"=",
"date",
"data",
"[",
"'next_month'",
"]",
"=",
"self",
".",
"get_next_month",
"(",
"date",
")",
"data",
"[",
"'previous_month'",
"]",
"=",
"self",
".",
"get_previous_month",
"(",
"date",
")",
"return",
"data"
] | 188e1bf1f15a44d9a599028d020083af9fb43ea7 |
valid | read_version | Read version from __init__.py without loading any files | setup.py | def read_version():
"""Read version from __init__.py without loading any files"""
finder = VersionFinder()
path = os.path.join(PROJECT_ROOT, 'colorful', '__init__.py')
with codecs.open(path, 'r', encoding='utf-8') as fp:
file_data = fp.read().encode('utf-8')
finder.visit(ast.parse(file_data))
return finder.version | def read_version():
"""Read version from __init__.py without loading any files"""
finder = VersionFinder()
path = os.path.join(PROJECT_ROOT, 'colorful', '__init__.py')
with codecs.open(path, 'r', encoding='utf-8') as fp:
file_data = fp.read().encode('utf-8')
finder.visit(ast.parse(file_data))
return finder.version | [
"Read",
"version",
"from",
"__init__",
".",
"py",
"without",
"loading",
"any",
"files"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/setup.py#L48-L56 | [
"def",
"read_version",
"(",
")",
":",
"finder",
"=",
"VersionFinder",
"(",
")",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"PROJECT_ROOT",
",",
"'colorful'",
",",
"'__init__.py'",
")",
"with",
"codecs",
".",
"open",
"(",
"path",
",",
"'r'",
",",
"encoding",
"=",
"'utf-8'",
")",
"as",
"fp",
":",
"file_data",
"=",
"fp",
".",
"read",
"(",
")",
".",
"encode",
"(",
"'utf-8'",
")",
"finder",
".",
"visit",
"(",
"ast",
".",
"parse",
"(",
"file_data",
")",
")",
"return",
"finder",
".",
"version"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | ColorfulModule.with_setup | Return a new Colorful object with the given color config. | colorful/__init__.py | def with_setup(self, colormode=None, colorpalette=None, extend_colors=False):
"""
Return a new Colorful object with the given color config.
"""
colorful = Colorful(
colormode=self.colorful.colormode,
colorpalette=copy.copy(self.colorful.colorpalette)
)
colorful.setup(
colormode=colormode, colorpalette=colorpalette, extend_colors=extend_colors
)
yield colorful | def with_setup(self, colormode=None, colorpalette=None, extend_colors=False):
"""
Return a new Colorful object with the given color config.
"""
colorful = Colorful(
colormode=self.colorful.colormode,
colorpalette=copy.copy(self.colorful.colorpalette)
)
colorful.setup(
colormode=colormode, colorpalette=colorpalette, extend_colors=extend_colors
)
yield colorful | [
"Return",
"a",
"new",
"Colorful",
"object",
"with",
"the",
"given",
"color",
"config",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/__init__.py#L42-L54 | [
"def",
"with_setup",
"(",
"self",
",",
"colormode",
"=",
"None",
",",
"colorpalette",
"=",
"None",
",",
"extend_colors",
"=",
"False",
")",
":",
"colorful",
"=",
"Colorful",
"(",
"colormode",
"=",
"self",
".",
"colorful",
".",
"colormode",
",",
"colorpalette",
"=",
"copy",
".",
"copy",
"(",
"self",
".",
"colorful",
".",
"colorpalette",
")",
")",
"colorful",
".",
"setup",
"(",
"colormode",
"=",
"colormode",
",",
"colorpalette",
"=",
"colorpalette",
",",
"extend_colors",
"=",
"extend_colors",
")",
"yield",
"colorful"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | parse_colors | Parse the given color files.
Supported are:
* .txt for X11 colors
* .json for colornames | colorful/colors.py | def parse_colors(path):
"""Parse the given color files.
Supported are:
* .txt for X11 colors
* .json for colornames
"""
if path.endswith(".txt"):
return parse_rgb_txt_file(path)
elif path.endswith(".json"):
return parse_json_color_file(path)
raise TypeError("colorful only supports .txt and .json files for colors") | def parse_colors(path):
"""Parse the given color files.
Supported are:
* .txt for X11 colors
* .json for colornames
"""
if path.endswith(".txt"):
return parse_rgb_txt_file(path)
elif path.endswith(".json"):
return parse_json_color_file(path)
raise TypeError("colorful only supports .txt and .json files for colors") | [
"Parse",
"the",
"given",
"color",
"files",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/colors.py#L18-L30 | [
"def",
"parse_colors",
"(",
"path",
")",
":",
"if",
"path",
".",
"endswith",
"(",
"\".txt\"",
")",
":",
"return",
"parse_rgb_txt_file",
"(",
"path",
")",
"elif",
"path",
".",
"endswith",
"(",
"\".json\"",
")",
":",
"return",
"parse_json_color_file",
"(",
"path",
")",
"raise",
"TypeError",
"(",
"\"colorful only supports .txt and .json files for colors\"",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | parse_rgb_txt_file | Parse the given rgb.txt file into a Python dict.
See https://en.wikipedia.org/wiki/X11_color_names for more information
:param str path: the path to the X11 rgb.txt file | colorful/colors.py | def parse_rgb_txt_file(path):
"""
Parse the given rgb.txt file into a Python dict.
See https://en.wikipedia.org/wiki/X11_color_names for more information
:param str path: the path to the X11 rgb.txt file
"""
#: Holds the generated color dict
color_dict = {}
with open(path, 'r') as rgb_txt:
for line in rgb_txt:
line = line.strip()
if not line or line.startswith('!'):
continue # skip comments
parts = line.split()
color_dict[" ".join(parts[3:])] = (int(parts[0]), int(parts[1]), int(parts[2]))
return color_dict | def parse_rgb_txt_file(path):
"""
Parse the given rgb.txt file into a Python dict.
See https://en.wikipedia.org/wiki/X11_color_names for more information
:param str path: the path to the X11 rgb.txt file
"""
#: Holds the generated color dict
color_dict = {}
with open(path, 'r') as rgb_txt:
for line in rgb_txt:
line = line.strip()
if not line or line.startswith('!'):
continue # skip comments
parts = line.split()
color_dict[" ".join(parts[3:])] = (int(parts[0]), int(parts[1]), int(parts[2]))
return color_dict | [
"Parse",
"the",
"given",
"rgb",
".",
"txt",
"file",
"into",
"a",
"Python",
"dict",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/colors.py#L33-L53 | [
"def",
"parse_rgb_txt_file",
"(",
"path",
")",
":",
"#: Holds the generated color dict",
"color_dict",
"=",
"{",
"}",
"with",
"open",
"(",
"path",
",",
"'r'",
")",
"as",
"rgb_txt",
":",
"for",
"line",
"in",
"rgb_txt",
":",
"line",
"=",
"line",
".",
"strip",
"(",
")",
"if",
"not",
"line",
"or",
"line",
".",
"startswith",
"(",
"'!'",
")",
":",
"continue",
"# skip comments",
"parts",
"=",
"line",
".",
"split",
"(",
")",
"color_dict",
"[",
"\" \"",
".",
"join",
"(",
"parts",
"[",
"3",
":",
"]",
")",
"]",
"=",
"(",
"int",
"(",
"parts",
"[",
"0",
"]",
")",
",",
"int",
"(",
"parts",
"[",
"1",
"]",
")",
",",
"int",
"(",
"parts",
"[",
"2",
"]",
")",
")",
"return",
"color_dict"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | parse_json_color_file | Parse a JSON color file.
The JSON has to be in the following format:
.. code:: json
[{"name": "COLOR_NAME", "hex": "#HEX"}, ...]
:param str path: the path to the JSON color file | colorful/colors.py | def parse_json_color_file(path):
"""Parse a JSON color file.
The JSON has to be in the following format:
.. code:: json
[{"name": "COLOR_NAME", "hex": "#HEX"}, ...]
:param str path: the path to the JSON color file
"""
with open(path, "r") as color_file:
color_list = json.load(color_file)
# transform raw color list into color dict
color_dict = {c["name"]: c["hex"] for c in color_list}
return color_dict | def parse_json_color_file(path):
"""Parse a JSON color file.
The JSON has to be in the following format:
.. code:: json
[{"name": "COLOR_NAME", "hex": "#HEX"}, ...]
:param str path: the path to the JSON color file
"""
with open(path, "r") as color_file:
color_list = json.load(color_file)
# transform raw color list into color dict
color_dict = {c["name"]: c["hex"] for c in color_list}
return color_dict | [
"Parse",
"a",
"JSON",
"color",
"file",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/colors.py#L56-L72 | [
"def",
"parse_json_color_file",
"(",
"path",
")",
":",
"with",
"open",
"(",
"path",
",",
"\"r\"",
")",
"as",
"color_file",
":",
"color_list",
"=",
"json",
".",
"load",
"(",
"color_file",
")",
"# transform raw color list into color dict",
"color_dict",
"=",
"{",
"c",
"[",
"\"name\"",
"]",
":",
"c",
"[",
"\"hex\"",
"]",
"for",
"c",
"in",
"color_list",
"}",
"return",
"color_dict"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | sanitize_color_palette | Sanitze the given color palette so it can
be safely used by Colorful.
It will convert colors specified in hex RGB to
a RGB channel triplet. | colorful/colors.py | def sanitize_color_palette(colorpalette):
"""
Sanitze the given color palette so it can
be safely used by Colorful.
It will convert colors specified in hex RGB to
a RGB channel triplet.
"""
new_palette = {}
def __make_valid_color_name(name):
"""
Convert the given name into a valid colorname
"""
if len(name) == 1:
name = name[0]
return name[:1].lower() + name[1:]
return name[0].lower() + ''.join(word.capitalize() for word in name[1:])
for key, value in colorpalette.items():
if isinstance(value, str):
# we assume it's a hex RGB value
value = utils.hex_to_rgb(value)
new_palette[__make_valid_color_name(key.split())] = value
return new_palette | def sanitize_color_palette(colorpalette):
"""
Sanitze the given color palette so it can
be safely used by Colorful.
It will convert colors specified in hex RGB to
a RGB channel triplet.
"""
new_palette = {}
def __make_valid_color_name(name):
"""
Convert the given name into a valid colorname
"""
if len(name) == 1:
name = name[0]
return name[:1].lower() + name[1:]
return name[0].lower() + ''.join(word.capitalize() for word in name[1:])
for key, value in colorpalette.items():
if isinstance(value, str):
# we assume it's a hex RGB value
value = utils.hex_to_rgb(value)
new_palette[__make_valid_color_name(key.split())] = value
return new_palette | [
"Sanitze",
"the",
"given",
"color",
"palette",
"so",
"it",
"can",
"be",
"safely",
"used",
"by",
"Colorful",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/colors.py#L75-L101 | [
"def",
"sanitize_color_palette",
"(",
"colorpalette",
")",
":",
"new_palette",
"=",
"{",
"}",
"def",
"__make_valid_color_name",
"(",
"name",
")",
":",
"\"\"\"\n Convert the given name into a valid colorname\n \"\"\"",
"if",
"len",
"(",
"name",
")",
"==",
"1",
":",
"name",
"=",
"name",
"[",
"0",
"]",
"return",
"name",
"[",
":",
"1",
"]",
".",
"lower",
"(",
")",
"+",
"name",
"[",
"1",
":",
"]",
"return",
"name",
"[",
"0",
"]",
".",
"lower",
"(",
")",
"+",
"''",
".",
"join",
"(",
"word",
".",
"capitalize",
"(",
")",
"for",
"word",
"in",
"name",
"[",
"1",
":",
"]",
")",
"for",
"key",
",",
"value",
"in",
"colorpalette",
".",
"items",
"(",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"str",
")",
":",
"# we assume it's a hex RGB value",
"value",
"=",
"utils",
".",
"hex_to_rgb",
"(",
"value",
")",
"new_palette",
"[",
"__make_valid_color_name",
"(",
"key",
".",
"split",
"(",
")",
")",
"]",
"=",
"value",
"return",
"new_palette"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | show | Show the modifiers and colors | examples/colors.py | def show():
"""
Show the modifiers and colors
"""
# modifiers
sys.stdout.write(colorful.bold('bold') + ' ')
sys.stdout.write(colorful.dimmed('dimmed') + ' ')
sys.stdout.write(colorful.italic('italic') + ' ')
sys.stdout.write(colorful.underlined('underlined') + ' ')
sys.stdout.write(colorful.inversed('inversed') + ' ')
sys.stdout.write(colorful.concealed('concealed') + ' ')
sys.stdout.write(colorful.struckthrough('struckthrough') + '\n')
# foreground colors
sys.stdout.write(colorful.red('red') + ' ')
sys.stdout.write(colorful.green('green') + ' ')
sys.stdout.write(colorful.yellow('yellow') + ' ')
sys.stdout.write(colorful.blue('blue') + ' ')
sys.stdout.write(colorful.magenta('magenta') + ' ')
sys.stdout.write(colorful.cyan('cyan') + ' ')
sys.stdout.write(colorful.white('white') + '\n')
# background colors
sys.stdout.write(colorful.on_red('red') + ' ')
sys.stdout.write(colorful.on_green('green') + ' ')
sys.stdout.write(colorful.on_yellow('yellow') + ' ')
sys.stdout.write(colorful.on_blue('blue') + ' ')
sys.stdout.write(colorful.on_magenta('magenta') + ' ')
sys.stdout.write(colorful.on_cyan('cyan') + ' ')
sys.stdout.write(colorful.on_white('white') + '\n') | def show():
"""
Show the modifiers and colors
"""
# modifiers
sys.stdout.write(colorful.bold('bold') + ' ')
sys.stdout.write(colorful.dimmed('dimmed') + ' ')
sys.stdout.write(colorful.italic('italic') + ' ')
sys.stdout.write(colorful.underlined('underlined') + ' ')
sys.stdout.write(colorful.inversed('inversed') + ' ')
sys.stdout.write(colorful.concealed('concealed') + ' ')
sys.stdout.write(colorful.struckthrough('struckthrough') + '\n')
# foreground colors
sys.stdout.write(colorful.red('red') + ' ')
sys.stdout.write(colorful.green('green') + ' ')
sys.stdout.write(colorful.yellow('yellow') + ' ')
sys.stdout.write(colorful.blue('blue') + ' ')
sys.stdout.write(colorful.magenta('magenta') + ' ')
sys.stdout.write(colorful.cyan('cyan') + ' ')
sys.stdout.write(colorful.white('white') + '\n')
# background colors
sys.stdout.write(colorful.on_red('red') + ' ')
sys.stdout.write(colorful.on_green('green') + ' ')
sys.stdout.write(colorful.on_yellow('yellow') + ' ')
sys.stdout.write(colorful.on_blue('blue') + ' ')
sys.stdout.write(colorful.on_magenta('magenta') + ' ')
sys.stdout.write(colorful.on_cyan('cyan') + ' ')
sys.stdout.write(colorful.on_white('white') + '\n') | [
"Show",
"the",
"modifiers",
"and",
"colors"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/examples/colors.py#L18-L47 | [
"def",
"show",
"(",
")",
":",
"# modifiers",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"bold",
"(",
"'bold'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"dimmed",
"(",
"'dimmed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"italic",
"(",
"'italic'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"underlined",
"(",
"'underlined'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"inversed",
"(",
"'inversed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"concealed",
"(",
"'concealed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"struckthrough",
"(",
"'struckthrough'",
")",
"+",
"'\\n'",
")",
"# foreground colors",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"red",
"(",
"'red'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"green",
"(",
"'green'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"yellow",
"(",
"'yellow'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"blue",
"(",
"'blue'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"magenta",
"(",
"'magenta'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"cyan",
"(",
"'cyan'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"white",
"(",
"'white'",
")",
"+",
"'\\n'",
")",
"# background colors",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_red",
"(",
"'red'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_green",
"(",
"'green'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_yellow",
"(",
"'yellow'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_blue",
"(",
"'blue'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_magenta",
"(",
"'magenta'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_cyan",
"(",
"'cyan'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"colorful",
".",
"on_white",
"(",
"'white'",
")",
"+",
"'\\n'",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | show | Show the modifiers and colors | examples/monokai.py | def show():
"""
Show the modifiers and colors
"""
with colorful.with_style('monokai') as c:
# modifiers
sys.stdout.write(c.bold('bold') + ' ')
sys.stdout.write(c.dimmed('dimmed') + ' ')
sys.stdout.write(c.italic('italic') + ' ')
sys.stdout.write(c.underlined('underlined') + ' ')
sys.stdout.write(c.inversed('inversed') + ' ')
sys.stdout.write(c.concealed('concealed') + ' ')
sys.stdout.write(c.struckthrough('struckthrough') + '\n')
# foreground colors
sys.stdout.write(c.orange('orange') + ' ')
sys.stdout.write(c.magenta('magenta') + ' ')
sys.stdout.write(c.purple('purple') + ' ')
sys.stdout.write(c.blue('blue') + ' ')
sys.stdout.write(c.seaGreen('sea green') + ' ')
sys.stdout.write(c.green('green') + ' ')
sys.stdout.write(c.yellow('yellow') + '\n')
# background colors
sys.stdout.write(c.on_orange('orange') + ' ')
sys.stdout.write(c.on_magenta('magenta') + ' ')
sys.stdout.write(c.on_purple('purple') + ' ')
sys.stdout.write(c.on_blue('blue') + ' ')
sys.stdout.write(c.on_seaGreen('sea green') + ' ')
sys.stdout.write(c.gray_on_green('green') + ' ')
sys.stdout.write(c.gray_on_yellow('yellow') + '\n') | def show():
"""
Show the modifiers and colors
"""
with colorful.with_style('monokai') as c:
# modifiers
sys.stdout.write(c.bold('bold') + ' ')
sys.stdout.write(c.dimmed('dimmed') + ' ')
sys.stdout.write(c.italic('italic') + ' ')
sys.stdout.write(c.underlined('underlined') + ' ')
sys.stdout.write(c.inversed('inversed') + ' ')
sys.stdout.write(c.concealed('concealed') + ' ')
sys.stdout.write(c.struckthrough('struckthrough') + '\n')
# foreground colors
sys.stdout.write(c.orange('orange') + ' ')
sys.stdout.write(c.magenta('magenta') + ' ')
sys.stdout.write(c.purple('purple') + ' ')
sys.stdout.write(c.blue('blue') + ' ')
sys.stdout.write(c.seaGreen('sea green') + ' ')
sys.stdout.write(c.green('green') + ' ')
sys.stdout.write(c.yellow('yellow') + '\n')
# background colors
sys.stdout.write(c.on_orange('orange') + ' ')
sys.stdout.write(c.on_magenta('magenta') + ' ')
sys.stdout.write(c.on_purple('purple') + ' ')
sys.stdout.write(c.on_blue('blue') + ' ')
sys.stdout.write(c.on_seaGreen('sea green') + ' ')
sys.stdout.write(c.gray_on_green('green') + ' ')
sys.stdout.write(c.gray_on_yellow('yellow') + '\n') | [
"Show",
"the",
"modifiers",
"and",
"colors"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/examples/monokai.py#L19-L49 | [
"def",
"show",
"(",
")",
":",
"with",
"colorful",
".",
"with_style",
"(",
"'monokai'",
")",
"as",
"c",
":",
"# modifiers",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"bold",
"(",
"'bold'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"dimmed",
"(",
"'dimmed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"italic",
"(",
"'italic'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"underlined",
"(",
"'underlined'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"inversed",
"(",
"'inversed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"concealed",
"(",
"'concealed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"struckthrough",
"(",
"'struckthrough'",
")",
"+",
"'\\n'",
")",
"# foreground colors",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"orange",
"(",
"'orange'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"magenta",
"(",
"'magenta'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"purple",
"(",
"'purple'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"blue",
"(",
"'blue'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"seaGreen",
"(",
"'sea green'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"green",
"(",
"'green'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"yellow",
"(",
"'yellow'",
")",
"+",
"'\\n'",
")",
"# background colors",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_orange",
"(",
"'orange'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_magenta",
"(",
"'magenta'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_purple",
"(",
"'purple'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_blue",
"(",
"'blue'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_seaGreen",
"(",
"'sea green'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"gray_on_green",
"(",
"'green'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"gray_on_yellow",
"(",
"'yellow'",
")",
"+",
"'\\n'",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | detect_color_support | Detect what color palettes are supported.
It'll return a valid color mode to use
with colorful.
:param dict env: the environment dict like returned by ``os.envion`` | colorful/terminal.py | def detect_color_support(env): # noqa
"""
Detect what color palettes are supported.
It'll return a valid color mode to use
with colorful.
:param dict env: the environment dict like returned by ``os.envion``
"""
if env.get('COLORFUL_DISABLE', '0') == '1':
return NO_COLORS
if env.get('COLORFUL_FORCE_8_COLORS', '0') == '1':
return ANSI_8_COLORS
if env.get('COLORFUL_FORCE_16_COLORS', '0') == '1':
return ANSI_16_COLORS
if env.get('COLORFUL_FORCE_256_COLORS', '0') == '1':
return ANSI_256_COLORS
if env.get('COLORFUL_FORCE_TRUE_COLORS', '0') == '1':
return TRUE_COLORS
# if we are not a tty
if not sys.stdout.isatty():
return NO_COLORS
colorterm_env = env.get('COLORTERM')
if colorterm_env:
if colorterm_env in {'truecolor', '24bit'}:
return TRUE_COLORS
if colorterm_env in {'8bit'}:
return ANSI_256_COLORS
termprog_env = env.get('TERM_PROGRAM')
if termprog_env:
if termprog_env in {'iTerm.app', 'Hyper'}:
return TRUE_COLORS
if termprog_env in {'Apple_Terminal'}:
return ANSI_256_COLORS
term_env = env.get('TERM')
if term_env:
if term_env in {'screen-256', 'screen-256color', 'xterm-256', 'xterm-256color'}:
return ANSI_256_COLORS
if term_env in {'screen', 'xterm', 'vt100', 'color', 'ansi', 'cygwin', 'linux'}:
return ANSI_16_COLORS
if colorterm_env:
# if there was no match with $TERM either but we
# had one with $COLORTERM, we use it!
return ANSI_16_COLORS
return ANSI_8_COLORS | def detect_color_support(env): # noqa
"""
Detect what color palettes are supported.
It'll return a valid color mode to use
with colorful.
:param dict env: the environment dict like returned by ``os.envion``
"""
if env.get('COLORFUL_DISABLE', '0') == '1':
return NO_COLORS
if env.get('COLORFUL_FORCE_8_COLORS', '0') == '1':
return ANSI_8_COLORS
if env.get('COLORFUL_FORCE_16_COLORS', '0') == '1':
return ANSI_16_COLORS
if env.get('COLORFUL_FORCE_256_COLORS', '0') == '1':
return ANSI_256_COLORS
if env.get('COLORFUL_FORCE_TRUE_COLORS', '0') == '1':
return TRUE_COLORS
# if we are not a tty
if not sys.stdout.isatty():
return NO_COLORS
colorterm_env = env.get('COLORTERM')
if colorterm_env:
if colorterm_env in {'truecolor', '24bit'}:
return TRUE_COLORS
if colorterm_env in {'8bit'}:
return ANSI_256_COLORS
termprog_env = env.get('TERM_PROGRAM')
if termprog_env:
if termprog_env in {'iTerm.app', 'Hyper'}:
return TRUE_COLORS
if termprog_env in {'Apple_Terminal'}:
return ANSI_256_COLORS
term_env = env.get('TERM')
if term_env:
if term_env in {'screen-256', 'screen-256color', 'xterm-256', 'xterm-256color'}:
return ANSI_256_COLORS
if term_env in {'screen', 'xterm', 'vt100', 'color', 'ansi', 'cygwin', 'linux'}:
return ANSI_16_COLORS
if colorterm_env:
# if there was no match with $TERM either but we
# had one with $COLORTERM, we use it!
return ANSI_16_COLORS
return ANSI_8_COLORS | [
"Detect",
"what",
"color",
"palettes",
"are",
"supported",
".",
"It",
"ll",
"return",
"a",
"valid",
"color",
"mode",
"to",
"use",
"with",
"colorful",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/terminal.py#L24-L80 | [
"def",
"detect_color_support",
"(",
"env",
")",
":",
"# noqa",
"if",
"env",
".",
"get",
"(",
"'COLORFUL_DISABLE'",
",",
"'0'",
")",
"==",
"'1'",
":",
"return",
"NO_COLORS",
"if",
"env",
".",
"get",
"(",
"'COLORFUL_FORCE_8_COLORS'",
",",
"'0'",
")",
"==",
"'1'",
":",
"return",
"ANSI_8_COLORS",
"if",
"env",
".",
"get",
"(",
"'COLORFUL_FORCE_16_COLORS'",
",",
"'0'",
")",
"==",
"'1'",
":",
"return",
"ANSI_16_COLORS",
"if",
"env",
".",
"get",
"(",
"'COLORFUL_FORCE_256_COLORS'",
",",
"'0'",
")",
"==",
"'1'",
":",
"return",
"ANSI_256_COLORS",
"if",
"env",
".",
"get",
"(",
"'COLORFUL_FORCE_TRUE_COLORS'",
",",
"'0'",
")",
"==",
"'1'",
":",
"return",
"TRUE_COLORS",
"# if we are not a tty",
"if",
"not",
"sys",
".",
"stdout",
".",
"isatty",
"(",
")",
":",
"return",
"NO_COLORS",
"colorterm_env",
"=",
"env",
".",
"get",
"(",
"'COLORTERM'",
")",
"if",
"colorterm_env",
":",
"if",
"colorterm_env",
"in",
"{",
"'truecolor'",
",",
"'24bit'",
"}",
":",
"return",
"TRUE_COLORS",
"if",
"colorterm_env",
"in",
"{",
"'8bit'",
"}",
":",
"return",
"ANSI_256_COLORS",
"termprog_env",
"=",
"env",
".",
"get",
"(",
"'TERM_PROGRAM'",
")",
"if",
"termprog_env",
":",
"if",
"termprog_env",
"in",
"{",
"'iTerm.app'",
",",
"'Hyper'",
"}",
":",
"return",
"TRUE_COLORS",
"if",
"termprog_env",
"in",
"{",
"'Apple_Terminal'",
"}",
":",
"return",
"ANSI_256_COLORS",
"term_env",
"=",
"env",
".",
"get",
"(",
"'TERM'",
")",
"if",
"term_env",
":",
"if",
"term_env",
"in",
"{",
"'screen-256'",
",",
"'screen-256color'",
",",
"'xterm-256'",
",",
"'xterm-256color'",
"}",
":",
"return",
"ANSI_256_COLORS",
"if",
"term_env",
"in",
"{",
"'screen'",
",",
"'xterm'",
",",
"'vt100'",
",",
"'color'",
",",
"'ansi'",
",",
"'cygwin'",
",",
"'linux'",
"}",
":",
"return",
"ANSI_16_COLORS",
"if",
"colorterm_env",
":",
"# if there was no match with $TERM either but we",
"# had one with $COLORTERM, we use it!",
"return",
"ANSI_16_COLORS",
"return",
"ANSI_8_COLORS"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | rgb_to_ansi256 | Convert RGB to ANSI 256 color | colorful/ansi.py | def rgb_to_ansi256(r, g, b):
"""
Convert RGB to ANSI 256 color
"""
if r == g and g == b:
if r < 8:
return 16
if r > 248:
return 231
return round(((r - 8) / 247.0) * 24) + 232
ansi_r = 36 * round(r / 255.0 * 5.0)
ansi_g = 6 * round(g / 255.0 * 5.0)
ansi_b = round(b / 255.0 * 5.0)
ansi = 16 + ansi_r + ansi_g + ansi_b
return ansi | def rgb_to_ansi256(r, g, b):
"""
Convert RGB to ANSI 256 color
"""
if r == g and g == b:
if r < 8:
return 16
if r > 248:
return 231
return round(((r - 8) / 247.0) * 24) + 232
ansi_r = 36 * round(r / 255.0 * 5.0)
ansi_g = 6 * round(g / 255.0 * 5.0)
ansi_b = round(b / 255.0 * 5.0)
ansi = 16 + ansi_r + ansi_g + ansi_b
return ansi | [
"Convert",
"RGB",
"to",
"ANSI",
"256",
"color"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/ansi.py#L61-L77 | [
"def",
"rgb_to_ansi256",
"(",
"r",
",",
"g",
",",
"b",
")",
":",
"if",
"r",
"==",
"g",
"and",
"g",
"==",
"b",
":",
"if",
"r",
"<",
"8",
":",
"return",
"16",
"if",
"r",
">",
"248",
":",
"return",
"231",
"return",
"round",
"(",
"(",
"(",
"r",
"-",
"8",
")",
"/",
"247.0",
")",
"*",
"24",
")",
"+",
"232",
"ansi_r",
"=",
"36",
"*",
"round",
"(",
"r",
"/",
"255.0",
"*",
"5.0",
")",
"ansi_g",
"=",
"6",
"*",
"round",
"(",
"g",
"/",
"255.0",
"*",
"5.0",
")",
"ansi_b",
"=",
"round",
"(",
"b",
"/",
"255.0",
"*",
"5.0",
")",
"ansi",
"=",
"16",
"+",
"ansi_r",
"+",
"ansi_g",
"+",
"ansi_b",
"return",
"ansi"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | rgb_to_ansi16 | Convert RGB to ANSI 16 color | colorful/ansi.py | def rgb_to_ansi16(r, g, b, use_bright=False):
"""
Convert RGB to ANSI 16 color
"""
ansi_b = round(b / 255.0) << 2
ansi_g = round(g / 255.0) << 1
ansi_r = round(r / 255.0)
ansi = (90 if use_bright else 30) + (ansi_b | ansi_g | ansi_r)
return ansi | def rgb_to_ansi16(r, g, b, use_bright=False):
"""
Convert RGB to ANSI 16 color
"""
ansi_b = round(b / 255.0) << 2
ansi_g = round(g / 255.0) << 1
ansi_r = round(r / 255.0)
ansi = (90 if use_bright else 30) + (ansi_b | ansi_g | ansi_r)
return ansi | [
"Convert",
"RGB",
"to",
"ANSI",
"16",
"color"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/ansi.py#L80-L89 | [
"def",
"rgb_to_ansi16",
"(",
"r",
",",
"g",
",",
"b",
",",
"use_bright",
"=",
"False",
")",
":",
"ansi_b",
"=",
"round",
"(",
"b",
"/",
"255.0",
")",
"<<",
"2",
"ansi_g",
"=",
"round",
"(",
"g",
"/",
"255.0",
")",
"<<",
"1",
"ansi_r",
"=",
"round",
"(",
"r",
"/",
"255.0",
")",
"ansi",
"=",
"(",
"90",
"if",
"use_bright",
"else",
"30",
")",
"+",
"(",
"ansi_b",
"|",
"ansi_g",
"|",
"ansi_r",
")",
"return",
"ansi"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | hex_to_rgb | Convert the given hex string to a
valid RGB channel triplet. | colorful/utils.py | def hex_to_rgb(value):
"""
Convert the given hex string to a
valid RGB channel triplet.
"""
value = value.lstrip('#')
check_hex(value)
length = len(value)
step = int(length / 3)
return tuple(int(value[i:i+step], 16) for i in range(0, length, step)) | def hex_to_rgb(value):
"""
Convert the given hex string to a
valid RGB channel triplet.
"""
value = value.lstrip('#')
check_hex(value)
length = len(value)
step = int(length / 3)
return tuple(int(value[i:i+step], 16) for i in range(0, length, step)) | [
"Convert",
"the",
"given",
"hex",
"string",
"to",
"a",
"valid",
"RGB",
"channel",
"triplet",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/utils.py#L30-L40 | [
"def",
"hex_to_rgb",
"(",
"value",
")",
":",
"value",
"=",
"value",
".",
"lstrip",
"(",
"'#'",
")",
"check_hex",
"(",
"value",
")",
"length",
"=",
"len",
"(",
"value",
")",
"step",
"=",
"int",
"(",
"length",
"/",
"3",
")",
"return",
"tuple",
"(",
"int",
"(",
"value",
"[",
"i",
":",
"i",
"+",
"step",
"]",
",",
"16",
")",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"length",
",",
"step",
")",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | check_hex | Check if the given hex value is a valid RGB color
It should match the format: [0-9a-fA-F]
and be of length 3 or 6. | colorful/utils.py | def check_hex(value):
"""
Check if the given hex value is a valid RGB color
It should match the format: [0-9a-fA-F]
and be of length 3 or 6.
"""
length = len(value)
if length not in (3, 6):
raise ValueError('Hex string #{} is too long'.format(value))
regex = r'[0-9a-f]{{{length}}}'.format(length=length)
if not re.search(regex, value, re.I):
raise ValueError('Invalid Hex String: #{}'.format(value)) | def check_hex(value):
"""
Check if the given hex value is a valid RGB color
It should match the format: [0-9a-fA-F]
and be of length 3 or 6.
"""
length = len(value)
if length not in (3, 6):
raise ValueError('Hex string #{} is too long'.format(value))
regex = r'[0-9a-f]{{{length}}}'.format(length=length)
if not re.search(regex, value, re.I):
raise ValueError('Invalid Hex String: #{}'.format(value)) | [
"Check",
"if",
"the",
"given",
"hex",
"value",
"is",
"a",
"valid",
"RGB",
"color"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/utils.py#L43-L56 | [
"def",
"check_hex",
"(",
"value",
")",
":",
"length",
"=",
"len",
"(",
"value",
")",
"if",
"length",
"not",
"in",
"(",
"3",
",",
"6",
")",
":",
"raise",
"ValueError",
"(",
"'Hex string #{} is too long'",
".",
"format",
"(",
"value",
")",
")",
"regex",
"=",
"r'[0-9a-f]{{{length}}}'",
".",
"format",
"(",
"length",
"=",
"length",
")",
"if",
"not",
"re",
".",
"search",
"(",
"regex",
",",
"value",
",",
"re",
".",
"I",
")",
":",
"raise",
"ValueError",
"(",
"'Invalid Hex String: #{}'",
".",
"format",
"(",
"value",
")",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | show | Show the modifiers and colors | examples/solarized.py | def show():
"""
Show the modifiers and colors
"""
with colorful.with_style('solarized') as c:
# modifiers
sys.stdout.write(c.bold('bold') + ' ')
sys.stdout.write(c.dimmed('dimmed') + ' ')
sys.stdout.write(c.italic('italic') + ' ')
sys.stdout.write(c.underlined('underlined') + ' ')
sys.stdout.write(c.inversed('inversed') + ' ')
sys.stdout.write(c.concealed('concealed') + ' ')
sys.stdout.write(c.struckthrough('struckthrough') + '\n')
# foreground colors
sys.stdout.write(c.yellow('yellow') + ' ')
sys.stdout.write(c.red('orange') + ' ')
sys.stdout.write(c.red('red') + ' ')
sys.stdout.write(c.magenta('magenta') + ' ')
sys.stdout.write(c.magenta('violet') + ' ')
sys.stdout.write(c.blue('blue') + ' ')
sys.stdout.write(c.cyan('cyan') + ' ')
sys.stdout.write(c.green('green') + '\n')
# background colors
sys.stdout.write(c.on_yellow('yellow') + ' ')
sys.stdout.write(c.on_red('orange') + ' ')
sys.stdout.write(c.on_red('red') + ' ')
sys.stdout.write(c.on_magenta('magenta') + ' ')
sys.stdout.write(c.on_magenta('violet') + ' ')
sys.stdout.write(c.on_blue('blue') + ' ')
sys.stdout.write(c.on_cyan('cyan') + ' ')
sys.stdout.write(c.on_green('green') + '\n') | def show():
"""
Show the modifiers and colors
"""
with colorful.with_style('solarized') as c:
# modifiers
sys.stdout.write(c.bold('bold') + ' ')
sys.stdout.write(c.dimmed('dimmed') + ' ')
sys.stdout.write(c.italic('italic') + ' ')
sys.stdout.write(c.underlined('underlined') + ' ')
sys.stdout.write(c.inversed('inversed') + ' ')
sys.stdout.write(c.concealed('concealed') + ' ')
sys.stdout.write(c.struckthrough('struckthrough') + '\n')
# foreground colors
sys.stdout.write(c.yellow('yellow') + ' ')
sys.stdout.write(c.red('orange') + ' ')
sys.stdout.write(c.red('red') + ' ')
sys.stdout.write(c.magenta('magenta') + ' ')
sys.stdout.write(c.magenta('violet') + ' ')
sys.stdout.write(c.blue('blue') + ' ')
sys.stdout.write(c.cyan('cyan') + ' ')
sys.stdout.write(c.green('green') + '\n')
# background colors
sys.stdout.write(c.on_yellow('yellow') + ' ')
sys.stdout.write(c.on_red('orange') + ' ')
sys.stdout.write(c.on_red('red') + ' ')
sys.stdout.write(c.on_magenta('magenta') + ' ')
sys.stdout.write(c.on_magenta('violet') + ' ')
sys.stdout.write(c.on_blue('blue') + ' ')
sys.stdout.write(c.on_cyan('cyan') + ' ')
sys.stdout.write(c.on_green('green') + '\n') | [
"Show",
"the",
"modifiers",
"and",
"colors"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/examples/solarized.py#L17-L49 | [
"def",
"show",
"(",
")",
":",
"with",
"colorful",
".",
"with_style",
"(",
"'solarized'",
")",
"as",
"c",
":",
"# modifiers",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"bold",
"(",
"'bold'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"dimmed",
"(",
"'dimmed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"italic",
"(",
"'italic'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"underlined",
"(",
"'underlined'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"inversed",
"(",
"'inversed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"concealed",
"(",
"'concealed'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"struckthrough",
"(",
"'struckthrough'",
")",
"+",
"'\\n'",
")",
"# foreground colors",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"yellow",
"(",
"'yellow'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"red",
"(",
"'orange'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"red",
"(",
"'red'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"magenta",
"(",
"'magenta'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"magenta",
"(",
"'violet'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"blue",
"(",
"'blue'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"cyan",
"(",
"'cyan'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"green",
"(",
"'green'",
")",
"+",
"'\\n'",
")",
"# background colors",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_yellow",
"(",
"'yellow'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_red",
"(",
"'orange'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_red",
"(",
"'red'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_magenta",
"(",
"'magenta'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_magenta",
"(",
"'violet'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_blue",
"(",
"'blue'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_cyan",
"(",
"'cyan'",
")",
"+",
"' '",
")",
"sys",
".",
"stdout",
".",
"write",
"(",
"c",
".",
"on_green",
"(",
"'green'",
")",
"+",
"'\\n'",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | translate_rgb_to_ansi_code | Translate the given RGB color into the appropriate ANSI escape code
for the given color mode.
The offset is used for the base color which is used.
The ``colormode`` has to be one of:
* 0: no colors / disabled
* 8: use ANSI 8 colors
* 16: use ANSI 16 colors (same as 8 but with brightness)
* 256: use ANSI 256 colors
* 0xFFFFFF / 16777215: use 16 Million true colors
:param int red: the red channel value
:param int green: the green channel value
:param int blue: the blue channel value
:param int offset: the offset to use for the base color
:param int colormode: the color mode to use. See explanation above | colorful/core.py | def translate_rgb_to_ansi_code(red, green, blue, offset, colormode):
"""
Translate the given RGB color into the appropriate ANSI escape code
for the given color mode.
The offset is used for the base color which is used.
The ``colormode`` has to be one of:
* 0: no colors / disabled
* 8: use ANSI 8 colors
* 16: use ANSI 16 colors (same as 8 but with brightness)
* 256: use ANSI 256 colors
* 0xFFFFFF / 16777215: use 16 Million true colors
:param int red: the red channel value
:param int green: the green channel value
:param int blue: the blue channel value
:param int offset: the offset to use for the base color
:param int colormode: the color mode to use. See explanation above
"""
if colormode == terminal.NO_COLORS: # colors are disabled, thus return empty string
return '', ''
if colormode == terminal.ANSI_8_COLORS or colormode == terminal.ANSI_16_COLORS:
color_code = ansi.rgb_to_ansi16(red, green, blue)
start_code = ansi.ANSI_ESCAPE_CODE.format(
code=color_code + offset - ansi.FOREGROUND_COLOR_OFFSET)
end_code = ansi.ANSI_ESCAPE_CODE.format(code=offset + ansi.COLOR_CLOSE_OFFSET)
return start_code, end_code
if colormode == terminal.ANSI_256_COLORS:
color_code = ansi.rgb_to_ansi256(red, green, blue)
start_code = ansi.ANSI_ESCAPE_CODE.format(code='{base};5;{code}'.format(
base=8 + offset, code=color_code))
end_code = ansi.ANSI_ESCAPE_CODE.format(code=offset + ansi.COLOR_CLOSE_OFFSET)
return start_code, end_code
if colormode == terminal.TRUE_COLORS:
start_code = ansi.ANSI_ESCAPE_CODE.format(code='{base};2;{red};{green};{blue}'.format(
base=8 + offset, red=red, green=green, blue=blue))
end_code = ansi.ANSI_ESCAPE_CODE.format(code=offset + ansi.COLOR_CLOSE_OFFSET)
return start_code, end_code
raise ColorfulError('invalid color mode "{0}"'.format(colormode)) | def translate_rgb_to_ansi_code(red, green, blue, offset, colormode):
"""
Translate the given RGB color into the appropriate ANSI escape code
for the given color mode.
The offset is used for the base color which is used.
The ``colormode`` has to be one of:
* 0: no colors / disabled
* 8: use ANSI 8 colors
* 16: use ANSI 16 colors (same as 8 but with brightness)
* 256: use ANSI 256 colors
* 0xFFFFFF / 16777215: use 16 Million true colors
:param int red: the red channel value
:param int green: the green channel value
:param int blue: the blue channel value
:param int offset: the offset to use for the base color
:param int colormode: the color mode to use. See explanation above
"""
if colormode == terminal.NO_COLORS: # colors are disabled, thus return empty string
return '', ''
if colormode == terminal.ANSI_8_COLORS or colormode == terminal.ANSI_16_COLORS:
color_code = ansi.rgb_to_ansi16(red, green, blue)
start_code = ansi.ANSI_ESCAPE_CODE.format(
code=color_code + offset - ansi.FOREGROUND_COLOR_OFFSET)
end_code = ansi.ANSI_ESCAPE_CODE.format(code=offset + ansi.COLOR_CLOSE_OFFSET)
return start_code, end_code
if colormode == terminal.ANSI_256_COLORS:
color_code = ansi.rgb_to_ansi256(red, green, blue)
start_code = ansi.ANSI_ESCAPE_CODE.format(code='{base};5;{code}'.format(
base=8 + offset, code=color_code))
end_code = ansi.ANSI_ESCAPE_CODE.format(code=offset + ansi.COLOR_CLOSE_OFFSET)
return start_code, end_code
if colormode == terminal.TRUE_COLORS:
start_code = ansi.ANSI_ESCAPE_CODE.format(code='{base};2;{red};{green};{blue}'.format(
base=8 + offset, red=red, green=green, blue=blue))
end_code = ansi.ANSI_ESCAPE_CODE.format(code=offset + ansi.COLOR_CLOSE_OFFSET)
return start_code, end_code
raise ColorfulError('invalid color mode "{0}"'.format(colormode)) | [
"Translate",
"the",
"given",
"RGB",
"color",
"into",
"the",
"appropriate",
"ANSI",
"escape",
"code",
"for",
"the",
"given",
"color",
"mode",
".",
"The",
"offset",
"is",
"used",
"for",
"the",
"base",
"color",
"which",
"is",
"used",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L43-L85 | [
"def",
"translate_rgb_to_ansi_code",
"(",
"red",
",",
"green",
",",
"blue",
",",
"offset",
",",
"colormode",
")",
":",
"if",
"colormode",
"==",
"terminal",
".",
"NO_COLORS",
":",
"# colors are disabled, thus return empty string",
"return",
"''",
",",
"''",
"if",
"colormode",
"==",
"terminal",
".",
"ANSI_8_COLORS",
"or",
"colormode",
"==",
"terminal",
".",
"ANSI_16_COLORS",
":",
"color_code",
"=",
"ansi",
".",
"rgb_to_ansi16",
"(",
"red",
",",
"green",
",",
"blue",
")",
"start_code",
"=",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"color_code",
"+",
"offset",
"-",
"ansi",
".",
"FOREGROUND_COLOR_OFFSET",
")",
"end_code",
"=",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"offset",
"+",
"ansi",
".",
"COLOR_CLOSE_OFFSET",
")",
"return",
"start_code",
",",
"end_code",
"if",
"colormode",
"==",
"terminal",
".",
"ANSI_256_COLORS",
":",
"color_code",
"=",
"ansi",
".",
"rgb_to_ansi256",
"(",
"red",
",",
"green",
",",
"blue",
")",
"start_code",
"=",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"'{base};5;{code}'",
".",
"format",
"(",
"base",
"=",
"8",
"+",
"offset",
",",
"code",
"=",
"color_code",
")",
")",
"end_code",
"=",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"offset",
"+",
"ansi",
".",
"COLOR_CLOSE_OFFSET",
")",
"return",
"start_code",
",",
"end_code",
"if",
"colormode",
"==",
"terminal",
".",
"TRUE_COLORS",
":",
"start_code",
"=",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"'{base};2;{red};{green};{blue}'",
".",
"format",
"(",
"base",
"=",
"8",
"+",
"offset",
",",
"red",
"=",
"red",
",",
"green",
"=",
"green",
",",
"blue",
"=",
"blue",
")",
")",
"end_code",
"=",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"offset",
"+",
"ansi",
".",
"COLOR_CLOSE_OFFSET",
")",
"return",
"start_code",
",",
"end_code",
"raise",
"ColorfulError",
"(",
"'invalid color mode \"{0}\"'",
".",
"format",
"(",
"colormode",
")",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | translate_colorname_to_ansi_code | Translate the given color name to a valid
ANSI escape code.
:parma str colorname: the name of the color to resolve
:parma str offset: the offset for the color code
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the color palette to use for the color name mapping
:returns str: the color as ANSI escape code
:raises ColorfulError: if the given color name is invalid | colorful/core.py | def translate_colorname_to_ansi_code(colorname, offset, colormode, colorpalette):
"""
Translate the given color name to a valid
ANSI escape code.
:parma str colorname: the name of the color to resolve
:parma str offset: the offset for the color code
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the color palette to use for the color name mapping
:returns str: the color as ANSI escape code
:raises ColorfulError: if the given color name is invalid
"""
try:
red, green, blue = colorpalette[colorname]
except KeyError:
raise ColorfulError('the color "{0}" is unknown. Use a color in your color palette (by default: X11 rgb.txt)'.format( # noqa
colorname))
else:
return translate_rgb_to_ansi_code(red, green, blue, offset, colormode) | def translate_colorname_to_ansi_code(colorname, offset, colormode, colorpalette):
"""
Translate the given color name to a valid
ANSI escape code.
:parma str colorname: the name of the color to resolve
:parma str offset: the offset for the color code
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the color palette to use for the color name mapping
:returns str: the color as ANSI escape code
:raises ColorfulError: if the given color name is invalid
"""
try:
red, green, blue = colorpalette[colorname]
except KeyError:
raise ColorfulError('the color "{0}" is unknown. Use a color in your color palette (by default: X11 rgb.txt)'.format( # noqa
colorname))
else:
return translate_rgb_to_ansi_code(red, green, blue, offset, colormode) | [
"Translate",
"the",
"given",
"color",
"name",
"to",
"a",
"valid",
"ANSI",
"escape",
"code",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L88-L108 | [
"def",
"translate_colorname_to_ansi_code",
"(",
"colorname",
",",
"offset",
",",
"colormode",
",",
"colorpalette",
")",
":",
"try",
":",
"red",
",",
"green",
",",
"blue",
"=",
"colorpalette",
"[",
"colorname",
"]",
"except",
"KeyError",
":",
"raise",
"ColorfulError",
"(",
"'the color \"{0}\" is unknown. Use a color in your color palette (by default: X11 rgb.txt)'",
".",
"format",
"(",
"# noqa",
"colorname",
")",
")",
"else",
":",
"return",
"translate_rgb_to_ansi_code",
"(",
"red",
",",
"green",
",",
"blue",
",",
"offset",
",",
"colormode",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | resolve_modifier_to_ansi_code | Resolve the given modifier name to a valid
ANSI escape code.
:param str modifiername: the name of the modifier to resolve
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:returns str: the ANSI escape code for the modifier
:raises ColorfulError: if the given modifier name is invalid | colorful/core.py | def resolve_modifier_to_ansi_code(modifiername, colormode):
"""
Resolve the given modifier name to a valid
ANSI escape code.
:param str modifiername: the name of the modifier to resolve
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:returns str: the ANSI escape code for the modifier
:raises ColorfulError: if the given modifier name is invalid
"""
if colormode == terminal.NO_COLORS: # return empty string if colors are disabled
return '', ''
try:
start_code, end_code = ansi.MODIFIERS[modifiername]
except KeyError:
raise ColorfulError('the modifier "{0}" is unknown. Use one of: {1}'.format(
modifiername, ansi.MODIFIERS.keys()))
else:
return ansi.ANSI_ESCAPE_CODE.format(
code=start_code), ansi.ANSI_ESCAPE_CODE.format(
code=end_code) | def resolve_modifier_to_ansi_code(modifiername, colormode):
"""
Resolve the given modifier name to a valid
ANSI escape code.
:param str modifiername: the name of the modifier to resolve
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:returns str: the ANSI escape code for the modifier
:raises ColorfulError: if the given modifier name is invalid
"""
if colormode == terminal.NO_COLORS: # return empty string if colors are disabled
return '', ''
try:
start_code, end_code = ansi.MODIFIERS[modifiername]
except KeyError:
raise ColorfulError('the modifier "{0}" is unknown. Use one of: {1}'.format(
modifiername, ansi.MODIFIERS.keys()))
else:
return ansi.ANSI_ESCAPE_CODE.format(
code=start_code), ansi.ANSI_ESCAPE_CODE.format(
code=end_code) | [
"Resolve",
"the",
"given",
"modifier",
"name",
"to",
"a",
"valid",
"ANSI",
"escape",
"code",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L111-L134 | [
"def",
"resolve_modifier_to_ansi_code",
"(",
"modifiername",
",",
"colormode",
")",
":",
"if",
"colormode",
"==",
"terminal",
".",
"NO_COLORS",
":",
"# return empty string if colors are disabled",
"return",
"''",
",",
"''",
"try",
":",
"start_code",
",",
"end_code",
"=",
"ansi",
".",
"MODIFIERS",
"[",
"modifiername",
"]",
"except",
"KeyError",
":",
"raise",
"ColorfulError",
"(",
"'the modifier \"{0}\" is unknown. Use one of: {1}'",
".",
"format",
"(",
"modifiername",
",",
"ansi",
".",
"MODIFIERS",
".",
"keys",
"(",
")",
")",
")",
"else",
":",
"return",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"start_code",
")",
",",
"ansi",
".",
"ANSI_ESCAPE_CODE",
".",
"format",
"(",
"code",
"=",
"end_code",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | translate_style | Translate the given style to an ANSI escape code
sequence.
``style`` examples are:
* green
* bold
* red_on_black
* bold_green
* italic_yellow_on_cyan
:param str style: the style to translate
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the color palette to use for the color name mapping | colorful/core.py | def translate_style(style, colormode, colorpalette):
"""
Translate the given style to an ANSI escape code
sequence.
``style`` examples are:
* green
* bold
* red_on_black
* bold_green
* italic_yellow_on_cyan
:param str style: the style to translate
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the color palette to use for the color name mapping
"""
style_parts = iter(style.split('_'))
ansi_start_sequence = []
ansi_end_sequence = []
try:
# consume all modifiers
part = None
for mod_part in style_parts:
part = mod_part
if part not in ansi.MODIFIERS:
break # all modifiers have been consumed
mod_start_code, mod_end_code = resolve_modifier_to_ansi_code(part, colormode)
ansi_start_sequence.append(mod_start_code)
ansi_end_sequence.append(mod_end_code)
else: # we've consumed all parts, thus we can exit
raise StopIteration()
# next part has to be a foreground color or the 'on' keyword
# which means we have to consume background colors
if part != 'on':
ansi_start_code, ansi_end_code = translate_colorname_to_ansi_code(
part, ansi.FOREGROUND_COLOR_OFFSET, colormode, colorpalette)
ansi_start_sequence.append(ansi_start_code)
ansi_end_sequence.append(ansi_end_code)
# consume the required 'on' keyword after the foreground color
next(style_parts)
# next part has to be the background color
part = next(style_parts)
ansi_start_code, ansi_end_code = translate_colorname_to_ansi_code(
part, ansi.BACKGROUND_COLOR_OFFSET, colormode, colorpalette)
ansi_start_sequence.append(ansi_start_code)
ansi_end_sequence.append(ansi_end_code)
except StopIteration: # we've consumed all parts of the styling string
pass
# construct and return ANSI escape code sequence
return ''.join(ansi_start_sequence), ''.join(ansi_end_sequence) | def translate_style(style, colormode, colorpalette):
"""
Translate the given style to an ANSI escape code
sequence.
``style`` examples are:
* green
* bold
* red_on_black
* bold_green
* italic_yellow_on_cyan
:param str style: the style to translate
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the color palette to use for the color name mapping
"""
style_parts = iter(style.split('_'))
ansi_start_sequence = []
ansi_end_sequence = []
try:
# consume all modifiers
part = None
for mod_part in style_parts:
part = mod_part
if part not in ansi.MODIFIERS:
break # all modifiers have been consumed
mod_start_code, mod_end_code = resolve_modifier_to_ansi_code(part, colormode)
ansi_start_sequence.append(mod_start_code)
ansi_end_sequence.append(mod_end_code)
else: # we've consumed all parts, thus we can exit
raise StopIteration()
# next part has to be a foreground color or the 'on' keyword
# which means we have to consume background colors
if part != 'on':
ansi_start_code, ansi_end_code = translate_colorname_to_ansi_code(
part, ansi.FOREGROUND_COLOR_OFFSET, colormode, colorpalette)
ansi_start_sequence.append(ansi_start_code)
ansi_end_sequence.append(ansi_end_code)
# consume the required 'on' keyword after the foreground color
next(style_parts)
# next part has to be the background color
part = next(style_parts)
ansi_start_code, ansi_end_code = translate_colorname_to_ansi_code(
part, ansi.BACKGROUND_COLOR_OFFSET, colormode, colorpalette)
ansi_start_sequence.append(ansi_start_code)
ansi_end_sequence.append(ansi_end_code)
except StopIteration: # we've consumed all parts of the styling string
pass
# construct and return ANSI escape code sequence
return ''.join(ansi_start_sequence), ''.join(ansi_end_sequence) | [
"Translate",
"the",
"given",
"style",
"to",
"an",
"ANSI",
"escape",
"code",
"sequence",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L137-L193 | [
"def",
"translate_style",
"(",
"style",
",",
"colormode",
",",
"colorpalette",
")",
":",
"style_parts",
"=",
"iter",
"(",
"style",
".",
"split",
"(",
"'_'",
")",
")",
"ansi_start_sequence",
"=",
"[",
"]",
"ansi_end_sequence",
"=",
"[",
"]",
"try",
":",
"# consume all modifiers",
"part",
"=",
"None",
"for",
"mod_part",
"in",
"style_parts",
":",
"part",
"=",
"mod_part",
"if",
"part",
"not",
"in",
"ansi",
".",
"MODIFIERS",
":",
"break",
"# all modifiers have been consumed",
"mod_start_code",
",",
"mod_end_code",
"=",
"resolve_modifier_to_ansi_code",
"(",
"part",
",",
"colormode",
")",
"ansi_start_sequence",
".",
"append",
"(",
"mod_start_code",
")",
"ansi_end_sequence",
".",
"append",
"(",
"mod_end_code",
")",
"else",
":",
"# we've consumed all parts, thus we can exit",
"raise",
"StopIteration",
"(",
")",
"# next part has to be a foreground color or the 'on' keyword",
"# which means we have to consume background colors",
"if",
"part",
"!=",
"'on'",
":",
"ansi_start_code",
",",
"ansi_end_code",
"=",
"translate_colorname_to_ansi_code",
"(",
"part",
",",
"ansi",
".",
"FOREGROUND_COLOR_OFFSET",
",",
"colormode",
",",
"colorpalette",
")",
"ansi_start_sequence",
".",
"append",
"(",
"ansi_start_code",
")",
"ansi_end_sequence",
".",
"append",
"(",
"ansi_end_code",
")",
"# consume the required 'on' keyword after the foreground color",
"next",
"(",
"style_parts",
")",
"# next part has to be the background color",
"part",
"=",
"next",
"(",
"style_parts",
")",
"ansi_start_code",
",",
"ansi_end_code",
"=",
"translate_colorname_to_ansi_code",
"(",
"part",
",",
"ansi",
".",
"BACKGROUND_COLOR_OFFSET",
",",
"colormode",
",",
"colorpalette",
")",
"ansi_start_sequence",
".",
"append",
"(",
"ansi_start_code",
")",
"ansi_end_sequence",
".",
"append",
"(",
"ansi_end_code",
")",
"except",
"StopIteration",
":",
"# we've consumed all parts of the styling string",
"pass",
"# construct and return ANSI escape code sequence",
"return",
"''",
".",
"join",
"(",
"ansi_start_sequence",
")",
",",
"''",
".",
"join",
"(",
"ansi_end_sequence",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | style_string | Style the given string according to the given
ANSI style string.
:param str string: the string to style
:param tuple ansi_style: the styling string returned by ``translate_style``
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:returns: a string containing proper ANSI sequence | colorful/core.py | def style_string(string, ansi_style, colormode, nested=False):
"""
Style the given string according to the given
ANSI style string.
:param str string: the string to style
:param tuple ansi_style: the styling string returned by ``translate_style``
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:returns: a string containing proper ANSI sequence
"""
ansi_start_code, ansi_end_code = ansi_style
# replace nest placeholders with the current begin style
if PY2:
if isinstance(string, str):
string = string.decode(DEFAULT_ENCODING)
string = UNICODE(string).replace(ansi.NEST_PLACEHOLDER, ansi_start_code)
return '{start_code}{string}{end_code}{nest_ph}'.format(
start_code=ansi_start_code,
string=string,
end_code=ansi_end_code,
nest_ph=ansi.NEST_PLACEHOLDER if nested else '') | def style_string(string, ansi_style, colormode, nested=False):
"""
Style the given string according to the given
ANSI style string.
:param str string: the string to style
:param tuple ansi_style: the styling string returned by ``translate_style``
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:returns: a string containing proper ANSI sequence
"""
ansi_start_code, ansi_end_code = ansi_style
# replace nest placeholders with the current begin style
if PY2:
if isinstance(string, str):
string = string.decode(DEFAULT_ENCODING)
string = UNICODE(string).replace(ansi.NEST_PLACEHOLDER, ansi_start_code)
return '{start_code}{string}{end_code}{nest_ph}'.format(
start_code=ansi_start_code,
string=string,
end_code=ansi_end_code,
nest_ph=ansi.NEST_PLACEHOLDER if nested else '') | [
"Style",
"the",
"given",
"string",
"according",
"to",
"the",
"given",
"ANSI",
"style",
"string",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L196-L219 | [
"def",
"style_string",
"(",
"string",
",",
"ansi_style",
",",
"colormode",
",",
"nested",
"=",
"False",
")",
":",
"ansi_start_code",
",",
"ansi_end_code",
"=",
"ansi_style",
"# replace nest placeholders with the current begin style",
"if",
"PY2",
":",
"if",
"isinstance",
"(",
"string",
",",
"str",
")",
":",
"string",
"=",
"string",
".",
"decode",
"(",
"DEFAULT_ENCODING",
")",
"string",
"=",
"UNICODE",
"(",
"string",
")",
".",
"replace",
"(",
"ansi",
".",
"NEST_PLACEHOLDER",
",",
"ansi_start_code",
")",
"return",
"'{start_code}{string}{end_code}{nest_ph}'",
".",
"format",
"(",
"start_code",
"=",
"ansi_start_code",
",",
"string",
"=",
"string",
",",
"end_code",
"=",
"ansi_end_code",
",",
"nest_ph",
"=",
"ansi",
".",
"NEST_PLACEHOLDER",
"if",
"nested",
"else",
"''",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | Colorful.colorpalette | Set the colorpalette which should be used | colorful/core.py | def colorpalette(self, colorpalette):
"""
Set the colorpalette which should be used
"""
if isinstance(colorpalette, str): # we assume it's a path to a color file
colorpalette = colors.parse_colors(colorpalette)
self._colorpalette = colors.sanitize_color_palette(colorpalette) | def colorpalette(self, colorpalette):
"""
Set the colorpalette which should be used
"""
if isinstance(colorpalette, str): # we assume it's a path to a color file
colorpalette = colors.parse_colors(colorpalette)
self._colorpalette = colors.sanitize_color_palette(colorpalette) | [
"Set",
"the",
"colorpalette",
"which",
"should",
"be",
"used"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L354-L361 | [
"def",
"colorpalette",
"(",
"self",
",",
"colorpalette",
")",
":",
"if",
"isinstance",
"(",
"colorpalette",
",",
"str",
")",
":",
"# we assume it's a path to a color file",
"colorpalette",
"=",
"colors",
".",
"parse_colors",
"(",
"colorpalette",
")",
"self",
".",
"_colorpalette",
"=",
"colors",
".",
"sanitize_color_palette",
"(",
"colorpalette",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | Colorful.setup | Setup this colorful object by setting a ``colormode`` and
the ``colorpalette`. The ``extend_colors`` flag is used
to extend the currently active color palette instead of
replacing it.
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the colorpalette to use. This ``dict`` should map
color names to it's corresponding RGB value
:param bool extend_colors: extend the active color palette instead of replacing it | colorful/core.py | def setup(self, colormode=None, colorpalette=None, extend_colors=False):
"""
Setup this colorful object by setting a ``colormode`` and
the ``colorpalette`. The ``extend_colors`` flag is used
to extend the currently active color palette instead of
replacing it.
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the colorpalette to use. This ``dict`` should map
color names to it's corresponding RGB value
:param bool extend_colors: extend the active color palette instead of replacing it
"""
if colormode:
self.colormode = colormode
if colorpalette:
if extend_colors:
self.update_palette(colorpalette)
else:
self.colorpalette = colorpalette | def setup(self, colormode=None, colorpalette=None, extend_colors=False):
"""
Setup this colorful object by setting a ``colormode`` and
the ``colorpalette`. The ``extend_colors`` flag is used
to extend the currently active color palette instead of
replacing it.
:param int colormode: the color mode to use. See ``translate_rgb_to_ansi_code``
:parma dict colorpalette: the colorpalette to use. This ``dict`` should map
color names to it's corresponding RGB value
:param bool extend_colors: extend the active color palette instead of replacing it
"""
if colormode:
self.colormode = colormode
if colorpalette:
if extend_colors:
self.update_palette(colorpalette)
else:
self.colorpalette = colorpalette | [
"Setup",
"this",
"colorful",
"object",
"by",
"setting",
"a",
"colormode",
"and",
"the",
"colorpalette",
".",
"The",
"extend_colors",
"flag",
"is",
"used",
"to",
"extend",
"the",
"currently",
"active",
"color",
"palette",
"instead",
"of",
"replacing",
"it",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L363-L382 | [
"def",
"setup",
"(",
"self",
",",
"colormode",
"=",
"None",
",",
"colorpalette",
"=",
"None",
",",
"extend_colors",
"=",
"False",
")",
":",
"if",
"colormode",
":",
"self",
".",
"colormode",
"=",
"colormode",
"if",
"colorpalette",
":",
"if",
"extend_colors",
":",
"self",
".",
"update_palette",
"(",
"colorpalette",
")",
"else",
":",
"self",
".",
"colorpalette",
"=",
"colorpalette"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | Colorful.use_style | Use a predefined style as color palette
:param str style_name: the name of the style | colorful/core.py | def use_style(self, style_name):
"""
Use a predefined style as color palette
:param str style_name: the name of the style
"""
try:
style = getattr(styles, style_name.upper())
except AttributeError:
raise ColorfulError('the style "{0}" is undefined'.format(
style_name))
else:
self.colorpalette = style | def use_style(self, style_name):
"""
Use a predefined style as color palette
:param str style_name: the name of the style
"""
try:
style = getattr(styles, style_name.upper())
except AttributeError:
raise ColorfulError('the style "{0}" is undefined'.format(
style_name))
else:
self.colorpalette = style | [
"Use",
"a",
"predefined",
"style",
"as",
"color",
"palette"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L427-L439 | [
"def",
"use_style",
"(",
"self",
",",
"style_name",
")",
":",
"try",
":",
"style",
"=",
"getattr",
"(",
"styles",
",",
"style_name",
".",
"upper",
"(",
")",
")",
"except",
"AttributeError",
":",
"raise",
"ColorfulError",
"(",
"'the style \"{0}\" is undefined'",
".",
"format",
"(",
"style_name",
")",
")",
"else",
":",
"self",
".",
"colorpalette",
"=",
"style"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | Colorful.format | Format the given string with the given ``args`` and ``kwargs``.
The string can contain references to ``c`` which is provided by
this colorful object.
:param str string: the string to format | colorful/core.py | def format(self, string, *args, **kwargs):
"""
Format the given string with the given ``args`` and ``kwargs``.
The string can contain references to ``c`` which is provided by
this colorful object.
:param str string: the string to format
"""
return string.format(c=self, *args, **kwargs) | def format(self, string, *args, **kwargs):
"""
Format the given string with the given ``args`` and ``kwargs``.
The string can contain references to ``c`` which is provided by
this colorful object.
:param str string: the string to format
"""
return string.format(c=self, *args, **kwargs) | [
"Format",
"the",
"given",
"string",
"with",
"the",
"given",
"args",
"and",
"kwargs",
".",
"The",
"string",
"can",
"contain",
"references",
"to",
"c",
"which",
"is",
"provided",
"by",
"this",
"colorful",
"object",
"."
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L441-L449 | [
"def",
"format",
"(",
"self",
",",
"string",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"string",
".",
"format",
"(",
"c",
"=",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | Colorful.print | Print the given objects to the given file stream.
See https://docs.python.org/3/library/functions.html#print
The only difference to the ``print()`` built-in is that
``Colorful.print()`` formats the string with ``c=self``.
With that stylings are possible
:param str sep: the seperater between the objects
:param str end: the ending delimiter after all objects
:param file: the file stream to write to
:param bool flush: if the stream should be flushed | colorful/core.py | def print(self, *objects, **options):
"""
Print the given objects to the given file stream.
See https://docs.python.org/3/library/functions.html#print
The only difference to the ``print()`` built-in is that
``Colorful.print()`` formats the string with ``c=self``.
With that stylings are possible
:param str sep: the seperater between the objects
:param str end: the ending delimiter after all objects
:param file: the file stream to write to
:param bool flush: if the stream should be flushed
"""
# NOTE: change signature to same as print() built-in function as
# soon as Python 2.7 is not supported anymore.
# If causes problems because of the keyword args after *args
allowed_options = {'sep', 'end', 'file', 'flush'}
given_options = set(options.keys())
if not given_options.issubset(allowed_options):
raise TypeError('Colorful.print() got unexpected keyword arguments: {0}'.format(
', '.join(given_options.difference(allowed_options))))
sep = options.get('sep', ' ')
end = options.get('end', '\n')
file = options.get('file', sys.stdout)
flush = options.get('flush', False)
styled_objects = [self.format(o) for o in objects]
print(*styled_objects, sep=sep, end=end, file=file)
# NOTE: if Python 2.7 support is dropped we can directly forward the
# flush keyword argument to the print() function.
if flush:
file.flush() | def print(self, *objects, **options):
"""
Print the given objects to the given file stream.
See https://docs.python.org/3/library/functions.html#print
The only difference to the ``print()`` built-in is that
``Colorful.print()`` formats the string with ``c=self``.
With that stylings are possible
:param str sep: the seperater between the objects
:param str end: the ending delimiter after all objects
:param file: the file stream to write to
:param bool flush: if the stream should be flushed
"""
# NOTE: change signature to same as print() built-in function as
# soon as Python 2.7 is not supported anymore.
# If causes problems because of the keyword args after *args
allowed_options = {'sep', 'end', 'file', 'flush'}
given_options = set(options.keys())
if not given_options.issubset(allowed_options):
raise TypeError('Colorful.print() got unexpected keyword arguments: {0}'.format(
', '.join(given_options.difference(allowed_options))))
sep = options.get('sep', ' ')
end = options.get('end', '\n')
file = options.get('file', sys.stdout)
flush = options.get('flush', False)
styled_objects = [self.format(o) for o in objects]
print(*styled_objects, sep=sep, end=end, file=file)
# NOTE: if Python 2.7 support is dropped we can directly forward the
# flush keyword argument to the print() function.
if flush:
file.flush() | [
"Print",
"the",
"given",
"objects",
"to",
"the",
"given",
"file",
"stream",
".",
"See",
"https",
":",
"//",
"docs",
".",
"python",
".",
"org",
"/",
"3",
"/",
"library",
"/",
"functions",
".",
"html#print"
] | timofurrer/colorful | python | https://github.com/timofurrer/colorful/blob/919fa6da17865cc5e01e6b16119193a97d180dc9/colorful/core.py#L469-L503 | [
"def",
"print",
"(",
"self",
",",
"*",
"objects",
",",
"*",
"*",
"options",
")",
":",
"# NOTE: change signature to same as print() built-in function as",
"# soon as Python 2.7 is not supported anymore.",
"# If causes problems because of the keyword args after *args",
"allowed_options",
"=",
"{",
"'sep'",
",",
"'end'",
",",
"'file'",
",",
"'flush'",
"}",
"given_options",
"=",
"set",
"(",
"options",
".",
"keys",
"(",
")",
")",
"if",
"not",
"given_options",
".",
"issubset",
"(",
"allowed_options",
")",
":",
"raise",
"TypeError",
"(",
"'Colorful.print() got unexpected keyword arguments: {0}'",
".",
"format",
"(",
"', '",
".",
"join",
"(",
"given_options",
".",
"difference",
"(",
"allowed_options",
")",
")",
")",
")",
"sep",
"=",
"options",
".",
"get",
"(",
"'sep'",
",",
"' '",
")",
"end",
"=",
"options",
".",
"get",
"(",
"'end'",
",",
"'\\n'",
")",
"file",
"=",
"options",
".",
"get",
"(",
"'file'",
",",
"sys",
".",
"stdout",
")",
"flush",
"=",
"options",
".",
"get",
"(",
"'flush'",
",",
"False",
")",
"styled_objects",
"=",
"[",
"self",
".",
"format",
"(",
"o",
")",
"for",
"o",
"in",
"objects",
"]",
"print",
"(",
"*",
"styled_objects",
",",
"sep",
"=",
"sep",
",",
"end",
"=",
"end",
",",
"file",
"=",
"file",
")",
"# NOTE: if Python 2.7 support is dropped we can directly forward the",
"# flush keyword argument to the print() function.",
"if",
"flush",
":",
"file",
".",
"flush",
"(",
")"
] | 919fa6da17865cc5e01e6b16119193a97d180dc9 |
valid | readattr | Read attribute from sysfs and return as string | temperusb/temper.py | def readattr(path, name):
"""
Read attribute from sysfs and return as string
"""
try:
f = open(USB_SYS_PREFIX + path + "/" + name)
return f.readline().rstrip("\n")
except IOError:
return None | def readattr(path, name):
"""
Read attribute from sysfs and return as string
"""
try:
f = open(USB_SYS_PREFIX + path + "/" + name)
return f.readline().rstrip("\n")
except IOError:
return None | [
"Read",
"attribute",
"from",
"sysfs",
"and",
"return",
"as",
"string"
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L38-L46 | [
"def",
"readattr",
"(",
"path",
",",
"name",
")",
":",
"try",
":",
"f",
"=",
"open",
"(",
"USB_SYS_PREFIX",
"+",
"path",
"+",
"\"/\"",
"+",
"name",
")",
"return",
"f",
".",
"readline",
"(",
")",
".",
"rstrip",
"(",
"\"\\n\"",
")",
"except",
"IOError",
":",
"return",
"None"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | find_ports | Find the port chain a device is plugged on.
This is done by searching sysfs for a device that matches the device
bus/address combination.
Useful when the underlying usb lib does not return device.port_number for
whatever reason. | temperusb/temper.py | def find_ports(device):
"""
Find the port chain a device is plugged on.
This is done by searching sysfs for a device that matches the device
bus/address combination.
Useful when the underlying usb lib does not return device.port_number for
whatever reason.
"""
bus_id = device.bus
dev_id = device.address
for dirent in os.listdir(USB_SYS_PREFIX):
matches = re.match(USB_PORTS_STR + '$', dirent)
if matches:
bus_str = readattr(dirent, 'busnum')
if bus_str:
busnum = float(bus_str)
else:
busnum = None
dev_str = readattr(dirent, 'devnum')
if dev_str:
devnum = float(dev_str)
else:
devnum = None
if busnum == bus_id and devnum == dev_id:
return str(matches.groups()[1]) | def find_ports(device):
"""
Find the port chain a device is plugged on.
This is done by searching sysfs for a device that matches the device
bus/address combination.
Useful when the underlying usb lib does not return device.port_number for
whatever reason.
"""
bus_id = device.bus
dev_id = device.address
for dirent in os.listdir(USB_SYS_PREFIX):
matches = re.match(USB_PORTS_STR + '$', dirent)
if matches:
bus_str = readattr(dirent, 'busnum')
if bus_str:
busnum = float(bus_str)
else:
busnum = None
dev_str = readattr(dirent, 'devnum')
if dev_str:
devnum = float(dev_str)
else:
devnum = None
if busnum == bus_id and devnum == dev_id:
return str(matches.groups()[1]) | [
"Find",
"the",
"port",
"chain",
"a",
"device",
"is",
"plugged",
"on",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L49-L75 | [
"def",
"find_ports",
"(",
"device",
")",
":",
"bus_id",
"=",
"device",
".",
"bus",
"dev_id",
"=",
"device",
".",
"address",
"for",
"dirent",
"in",
"os",
".",
"listdir",
"(",
"USB_SYS_PREFIX",
")",
":",
"matches",
"=",
"re",
".",
"match",
"(",
"USB_PORTS_STR",
"+",
"'$'",
",",
"dirent",
")",
"if",
"matches",
":",
"bus_str",
"=",
"readattr",
"(",
"dirent",
",",
"'busnum'",
")",
"if",
"bus_str",
":",
"busnum",
"=",
"float",
"(",
"bus_str",
")",
"else",
":",
"busnum",
"=",
"None",
"dev_str",
"=",
"readattr",
"(",
"dirent",
",",
"'devnum'",
")",
"if",
"dev_str",
":",
"devnum",
"=",
"float",
"(",
"dev_str",
")",
"else",
":",
"devnum",
"=",
"None",
"if",
"busnum",
"==",
"bus_id",
"and",
"devnum",
"==",
"dev_id",
":",
"return",
"str",
"(",
"matches",
".",
"groups",
"(",
")",
"[",
"1",
"]",
")"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice.set_calibration_data | Set device calibration data based on settings in /etc/temper.conf. | temperusb/temper.py | def set_calibration_data(self, scale=None, offset=None):
"""
Set device calibration data based on settings in /etc/temper.conf.
"""
if scale is not None and offset is not None:
self._scale = scale
self._offset = offset
elif scale is None and offset is None:
self._scale = 1.0
self._offset = 0.0
try:
f = open('/etc/temper.conf', 'r')
except IOError:
f = None
if f:
lines = f.read().split('\n')
f.close()
for line in lines:
matches = re.match(CALIB_LINE_STR, line)
if matches:
bus = int(matches.groups()[0])
ports = matches.groups()[1]
scale = float(matches.groups()[2])
offset = float(matches.groups()[3])
if (str(ports) == str(self._ports)) and (str(bus) == str(self._bus)):
self._scale = scale
self._offset = offset
else:
raise RuntimeError("Must set both scale and offset, or neither") | def set_calibration_data(self, scale=None, offset=None):
"""
Set device calibration data based on settings in /etc/temper.conf.
"""
if scale is not None and offset is not None:
self._scale = scale
self._offset = offset
elif scale is None and offset is None:
self._scale = 1.0
self._offset = 0.0
try:
f = open('/etc/temper.conf', 'r')
except IOError:
f = None
if f:
lines = f.read().split('\n')
f.close()
for line in lines:
matches = re.match(CALIB_LINE_STR, line)
if matches:
bus = int(matches.groups()[0])
ports = matches.groups()[1]
scale = float(matches.groups()[2])
offset = float(matches.groups()[3])
if (str(ports) == str(self._ports)) and (str(bus) == str(self._bus)):
self._scale = scale
self._offset = offset
else:
raise RuntimeError("Must set both scale and offset, or neither") | [
"Set",
"device",
"calibration",
"data",
"based",
"on",
"settings",
"in",
"/",
"etc",
"/",
"temper",
".",
"conf",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L107-L135 | [
"def",
"set_calibration_data",
"(",
"self",
",",
"scale",
"=",
"None",
",",
"offset",
"=",
"None",
")",
":",
"if",
"scale",
"is",
"not",
"None",
"and",
"offset",
"is",
"not",
"None",
":",
"self",
".",
"_scale",
"=",
"scale",
"self",
".",
"_offset",
"=",
"offset",
"elif",
"scale",
"is",
"None",
"and",
"offset",
"is",
"None",
":",
"self",
".",
"_scale",
"=",
"1.0",
"self",
".",
"_offset",
"=",
"0.0",
"try",
":",
"f",
"=",
"open",
"(",
"'/etc/temper.conf'",
",",
"'r'",
")",
"except",
"IOError",
":",
"f",
"=",
"None",
"if",
"f",
":",
"lines",
"=",
"f",
".",
"read",
"(",
")",
".",
"split",
"(",
"'\\n'",
")",
"f",
".",
"close",
"(",
")",
"for",
"line",
"in",
"lines",
":",
"matches",
"=",
"re",
".",
"match",
"(",
"CALIB_LINE_STR",
",",
"line",
")",
"if",
"matches",
":",
"bus",
"=",
"int",
"(",
"matches",
".",
"groups",
"(",
")",
"[",
"0",
"]",
")",
"ports",
"=",
"matches",
".",
"groups",
"(",
")",
"[",
"1",
"]",
"scale",
"=",
"float",
"(",
"matches",
".",
"groups",
"(",
")",
"[",
"2",
"]",
")",
"offset",
"=",
"float",
"(",
"matches",
".",
"groups",
"(",
")",
"[",
"3",
"]",
")",
"if",
"(",
"str",
"(",
"ports",
")",
"==",
"str",
"(",
"self",
".",
"_ports",
")",
")",
"and",
"(",
"str",
"(",
"bus",
")",
"==",
"str",
"(",
"self",
".",
"_bus",
")",
")",
":",
"self",
".",
"_scale",
"=",
"scale",
"self",
".",
"_offset",
"=",
"offset",
"else",
":",
"raise",
"RuntimeError",
"(",
"\"Must set both scale and offset, or neither\"",
")"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice.get_data | Get data from the USB device. | temperusb/temper.py | def get_data(self, reset_device=False):
"""
Get data from the USB device.
"""
try:
if reset_device:
self._device.reset()
# detach kernel driver from both interfaces if attached, so we can set_configuration()
for interface in [0,1]:
if self._device.is_kernel_driver_active(interface):
LOGGER.debug('Detaching kernel driver for interface %d '
'of %r on ports %r', interface, self._device, self._ports)
self._device.detach_kernel_driver(interface)
self._device.set_configuration()
# Prevent kernel message:
# "usbfs: process <PID> (python) did not claim interface x before use"
# This will become unnecessary once pull-request #124 for
# PyUSB has been accepted and we depend on a fixed release
# of PyUSB. Until then, and even with the fix applied, it
# does not hurt to explicitly claim the interface.
usb.util.claim_interface(self._device, INTERFACE)
# Turns out we don't actually need that ctrl_transfer.
# Disabling this reduces number of USBErrors from ~7/30 to 0!
#self._device.ctrl_transfer(bmRequestType=0x21, bRequest=0x09,
# wValue=0x0201, wIndex=0x00, data_or_wLength='\x01\x01',
# timeout=TIMEOUT)
# Magic: Our TEMPerV1.4 likes to be asked twice. When
# only asked once, it get's stuck on the next access and
# requires a reset.
self._control_transfer(COMMANDS['temp'])
self._interrupt_read()
# Turns out a whole lot of that magic seems unnecessary.
#self._control_transfer(COMMANDS['ini1'])
#self._interrupt_read()
#self._control_transfer(COMMANDS['ini2'])
#self._interrupt_read()
#self._interrupt_read()
# Get temperature
self._control_transfer(COMMANDS['temp'])
temp_data = self._interrupt_read()
# Get humidity
if self._device.product == 'TEMPer1F_H1_V1.4':
humidity_data = temp_data
else:
humidity_data = None
# Combine temperature and humidity data
data = {'temp_data': temp_data, 'humidity_data': humidity_data}
# Be a nice citizen and undo potential interface claiming.
# Also see: https://github.com/walac/pyusb/blob/master/docs/tutorial.rst#dont-be-selfish
usb.util.dispose_resources(self._device)
return data
except usb.USBError as err:
if not reset_device:
LOGGER.warning("Encountered %s, resetting %r and trying again.", err, self._device)
return self.get_data(True)
# Catch the permissions exception and add our message
if "not permitted" in str(err):
raise Exception(
"Permission problem accessing USB. "
"Maybe I need to run as root?")
else:
LOGGER.error(err)
raise | def get_data(self, reset_device=False):
"""
Get data from the USB device.
"""
try:
if reset_device:
self._device.reset()
# detach kernel driver from both interfaces if attached, so we can set_configuration()
for interface in [0,1]:
if self._device.is_kernel_driver_active(interface):
LOGGER.debug('Detaching kernel driver for interface %d '
'of %r on ports %r', interface, self._device, self._ports)
self._device.detach_kernel_driver(interface)
self._device.set_configuration()
# Prevent kernel message:
# "usbfs: process <PID> (python) did not claim interface x before use"
# This will become unnecessary once pull-request #124 for
# PyUSB has been accepted and we depend on a fixed release
# of PyUSB. Until then, and even with the fix applied, it
# does not hurt to explicitly claim the interface.
usb.util.claim_interface(self._device, INTERFACE)
# Turns out we don't actually need that ctrl_transfer.
# Disabling this reduces number of USBErrors from ~7/30 to 0!
#self._device.ctrl_transfer(bmRequestType=0x21, bRequest=0x09,
# wValue=0x0201, wIndex=0x00, data_or_wLength='\x01\x01',
# timeout=TIMEOUT)
# Magic: Our TEMPerV1.4 likes to be asked twice. When
# only asked once, it get's stuck on the next access and
# requires a reset.
self._control_transfer(COMMANDS['temp'])
self._interrupt_read()
# Turns out a whole lot of that magic seems unnecessary.
#self._control_transfer(COMMANDS['ini1'])
#self._interrupt_read()
#self._control_transfer(COMMANDS['ini2'])
#self._interrupt_read()
#self._interrupt_read()
# Get temperature
self._control_transfer(COMMANDS['temp'])
temp_data = self._interrupt_read()
# Get humidity
if self._device.product == 'TEMPer1F_H1_V1.4':
humidity_data = temp_data
else:
humidity_data = None
# Combine temperature and humidity data
data = {'temp_data': temp_data, 'humidity_data': humidity_data}
# Be a nice citizen and undo potential interface claiming.
# Also see: https://github.com/walac/pyusb/blob/master/docs/tutorial.rst#dont-be-selfish
usb.util.dispose_resources(self._device)
return data
except usb.USBError as err:
if not reset_device:
LOGGER.warning("Encountered %s, resetting %r and trying again.", err, self._device)
return self.get_data(True)
# Catch the permissions exception and add our message
if "not permitted" in str(err):
raise Exception(
"Permission problem accessing USB. "
"Maybe I need to run as root?")
else:
LOGGER.error(err)
raise | [
"Get",
"data",
"from",
"the",
"USB",
"device",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L207-L281 | [
"def",
"get_data",
"(",
"self",
",",
"reset_device",
"=",
"False",
")",
":",
"try",
":",
"if",
"reset_device",
":",
"self",
".",
"_device",
".",
"reset",
"(",
")",
"# detach kernel driver from both interfaces if attached, so we can set_configuration()",
"for",
"interface",
"in",
"[",
"0",
",",
"1",
"]",
":",
"if",
"self",
".",
"_device",
".",
"is_kernel_driver_active",
"(",
"interface",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'Detaching kernel driver for interface %d '",
"'of %r on ports %r'",
",",
"interface",
",",
"self",
".",
"_device",
",",
"self",
".",
"_ports",
")",
"self",
".",
"_device",
".",
"detach_kernel_driver",
"(",
"interface",
")",
"self",
".",
"_device",
".",
"set_configuration",
"(",
")",
"# Prevent kernel message:",
"# \"usbfs: process <PID> (python) did not claim interface x before use\"",
"# This will become unnecessary once pull-request #124 for",
"# PyUSB has been accepted and we depend on a fixed release",
"# of PyUSB. Until then, and even with the fix applied, it",
"# does not hurt to explicitly claim the interface.",
"usb",
".",
"util",
".",
"claim_interface",
"(",
"self",
".",
"_device",
",",
"INTERFACE",
")",
"# Turns out we don't actually need that ctrl_transfer.",
"# Disabling this reduces number of USBErrors from ~7/30 to 0!",
"#self._device.ctrl_transfer(bmRequestType=0x21, bRequest=0x09,",
"# wValue=0x0201, wIndex=0x00, data_or_wLength='\\x01\\x01',",
"# timeout=TIMEOUT)",
"# Magic: Our TEMPerV1.4 likes to be asked twice. When",
"# only asked once, it get's stuck on the next access and",
"# requires a reset.",
"self",
".",
"_control_transfer",
"(",
"COMMANDS",
"[",
"'temp'",
"]",
")",
"self",
".",
"_interrupt_read",
"(",
")",
"# Turns out a whole lot of that magic seems unnecessary.",
"#self._control_transfer(COMMANDS['ini1'])",
"#self._interrupt_read()",
"#self._control_transfer(COMMANDS['ini2'])",
"#self._interrupt_read()",
"#self._interrupt_read()",
"# Get temperature",
"self",
".",
"_control_transfer",
"(",
"COMMANDS",
"[",
"'temp'",
"]",
")",
"temp_data",
"=",
"self",
".",
"_interrupt_read",
"(",
")",
"# Get humidity",
"if",
"self",
".",
"_device",
".",
"product",
"==",
"'TEMPer1F_H1_V1.4'",
":",
"humidity_data",
"=",
"temp_data",
"else",
":",
"humidity_data",
"=",
"None",
"# Combine temperature and humidity data",
"data",
"=",
"{",
"'temp_data'",
":",
"temp_data",
",",
"'humidity_data'",
":",
"humidity_data",
"}",
"# Be a nice citizen and undo potential interface claiming.",
"# Also see: https://github.com/walac/pyusb/blob/master/docs/tutorial.rst#dont-be-selfish",
"usb",
".",
"util",
".",
"dispose_resources",
"(",
"self",
".",
"_device",
")",
"return",
"data",
"except",
"usb",
".",
"USBError",
"as",
"err",
":",
"if",
"not",
"reset_device",
":",
"LOGGER",
".",
"warning",
"(",
"\"Encountered %s, resetting %r and trying again.\"",
",",
"err",
",",
"self",
".",
"_device",
")",
"return",
"self",
".",
"get_data",
"(",
"True",
")",
"# Catch the permissions exception and add our message",
"if",
"\"not permitted\"",
"in",
"str",
"(",
"err",
")",
":",
"raise",
"Exception",
"(",
"\"Permission problem accessing USB. \"",
"\"Maybe I need to run as root?\"",
")",
"else",
":",
"LOGGER",
".",
"error",
"(",
"err",
")",
"raise"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice.get_temperature | Get device temperature reading. | temperusb/temper.py | def get_temperature(self, format='celsius', sensor=0):
"""
Get device temperature reading.
"""
results = self.get_temperatures(sensors=[sensor,])
if format == 'celsius':
return results[sensor]['temperature_c']
elif format == 'fahrenheit':
return results[sensor]['temperature_f']
elif format == 'millicelsius':
return results[sensor]['temperature_mc']
else:
raise ValueError("Unknown format") | def get_temperature(self, format='celsius', sensor=0):
"""
Get device temperature reading.
"""
results = self.get_temperatures(sensors=[sensor,])
if format == 'celsius':
return results[sensor]['temperature_c']
elif format == 'fahrenheit':
return results[sensor]['temperature_f']
elif format == 'millicelsius':
return results[sensor]['temperature_mc']
else:
raise ValueError("Unknown format") | [
"Get",
"device",
"temperature",
"reading",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L283-L296 | [
"def",
"get_temperature",
"(",
"self",
",",
"format",
"=",
"'celsius'",
",",
"sensor",
"=",
"0",
")",
":",
"results",
"=",
"self",
".",
"get_temperatures",
"(",
"sensors",
"=",
"[",
"sensor",
",",
"]",
")",
"if",
"format",
"==",
"'celsius'",
":",
"return",
"results",
"[",
"sensor",
"]",
"[",
"'temperature_c'",
"]",
"elif",
"format",
"==",
"'fahrenheit'",
":",
"return",
"results",
"[",
"sensor",
"]",
"[",
"'temperature_f'",
"]",
"elif",
"format",
"==",
"'millicelsius'",
":",
"return",
"results",
"[",
"sensor",
"]",
"[",
"'temperature_mc'",
"]",
"else",
":",
"raise",
"ValueError",
"(",
"\"Unknown format\"",
")"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice.get_temperatures | Get device temperature reading.
Params:
- sensors: optional list of sensors to get a reading for, examples:
[0,] - get reading for sensor 0
[0, 1,] - get reading for sensors 0 and 1
None - get readings for all sensors | temperusb/temper.py | def get_temperatures(self, sensors=None):
"""
Get device temperature reading.
Params:
- sensors: optional list of sensors to get a reading for, examples:
[0,] - get reading for sensor 0
[0, 1,] - get reading for sensors 0 and 1
None - get readings for all sensors
"""
_sensors = sensors
if _sensors is None:
_sensors = list(range(0, self._sensor_count))
if not set(_sensors).issubset(list(range(0, self._sensor_count))):
raise ValueError(
'Some or all of the sensors in the list %s are out of range '
'given a sensor_count of %d. Valid range: %s' % (
_sensors,
self._sensor_count,
list(range(0, self._sensor_count)),
)
)
data = self.get_data()
data = data['temp_data']
results = {}
# Interpret device response
for sensor in _sensors:
offset = self.lookup_offset(sensor)
celsius = struct.unpack_from('>h', data, offset)[0] / 256.0
# Apply scaling and offset (if any)
celsius = celsius * self._scale + self._offset
results[sensor] = {
'ports': self.get_ports(),
'bus': self.get_bus(),
'sensor': sensor,
'temperature_f': celsius * 1.8 + 32.0,
'temperature_c': celsius,
'temperature_mc': celsius * 1000,
'temperature_k': celsius + 273.15,
}
return results | def get_temperatures(self, sensors=None):
"""
Get device temperature reading.
Params:
- sensors: optional list of sensors to get a reading for, examples:
[0,] - get reading for sensor 0
[0, 1,] - get reading for sensors 0 and 1
None - get readings for all sensors
"""
_sensors = sensors
if _sensors is None:
_sensors = list(range(0, self._sensor_count))
if not set(_sensors).issubset(list(range(0, self._sensor_count))):
raise ValueError(
'Some or all of the sensors in the list %s are out of range '
'given a sensor_count of %d. Valid range: %s' % (
_sensors,
self._sensor_count,
list(range(0, self._sensor_count)),
)
)
data = self.get_data()
data = data['temp_data']
results = {}
# Interpret device response
for sensor in _sensors:
offset = self.lookup_offset(sensor)
celsius = struct.unpack_from('>h', data, offset)[0] / 256.0
# Apply scaling and offset (if any)
celsius = celsius * self._scale + self._offset
results[sensor] = {
'ports': self.get_ports(),
'bus': self.get_bus(),
'sensor': sensor,
'temperature_f': celsius * 1.8 + 32.0,
'temperature_c': celsius,
'temperature_mc': celsius * 1000,
'temperature_k': celsius + 273.15,
}
return results | [
"Get",
"device",
"temperature",
"reading",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L298-L343 | [
"def",
"get_temperatures",
"(",
"self",
",",
"sensors",
"=",
"None",
")",
":",
"_sensors",
"=",
"sensors",
"if",
"_sensors",
"is",
"None",
":",
"_sensors",
"=",
"list",
"(",
"range",
"(",
"0",
",",
"self",
".",
"_sensor_count",
")",
")",
"if",
"not",
"set",
"(",
"_sensors",
")",
".",
"issubset",
"(",
"list",
"(",
"range",
"(",
"0",
",",
"self",
".",
"_sensor_count",
")",
")",
")",
":",
"raise",
"ValueError",
"(",
"'Some or all of the sensors in the list %s are out of range '",
"'given a sensor_count of %d. Valid range: %s'",
"%",
"(",
"_sensors",
",",
"self",
".",
"_sensor_count",
",",
"list",
"(",
"range",
"(",
"0",
",",
"self",
".",
"_sensor_count",
")",
")",
",",
")",
")",
"data",
"=",
"self",
".",
"get_data",
"(",
")",
"data",
"=",
"data",
"[",
"'temp_data'",
"]",
"results",
"=",
"{",
"}",
"# Interpret device response",
"for",
"sensor",
"in",
"_sensors",
":",
"offset",
"=",
"self",
".",
"lookup_offset",
"(",
"sensor",
")",
"celsius",
"=",
"struct",
".",
"unpack_from",
"(",
"'>h'",
",",
"data",
",",
"offset",
")",
"[",
"0",
"]",
"/",
"256.0",
"# Apply scaling and offset (if any)",
"celsius",
"=",
"celsius",
"*",
"self",
".",
"_scale",
"+",
"self",
".",
"_offset",
"results",
"[",
"sensor",
"]",
"=",
"{",
"'ports'",
":",
"self",
".",
"get_ports",
"(",
")",
",",
"'bus'",
":",
"self",
".",
"get_bus",
"(",
")",
",",
"'sensor'",
":",
"sensor",
",",
"'temperature_f'",
":",
"celsius",
"*",
"1.8",
"+",
"32.0",
",",
"'temperature_c'",
":",
"celsius",
",",
"'temperature_mc'",
":",
"celsius",
"*",
"1000",
",",
"'temperature_k'",
":",
"celsius",
"+",
"273.15",
",",
"}",
"return",
"results"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice.get_humidity | Get device humidity reading.
Params:
- sensors: optional list of sensors to get a reading for, examples:
[0,] - get reading for sensor 0
[0, 1,] - get reading for sensors 0 and 1
None - get readings for all sensors | temperusb/temper.py | def get_humidity(self, sensors=None):
"""
Get device humidity reading.
Params:
- sensors: optional list of sensors to get a reading for, examples:
[0,] - get reading for sensor 0
[0, 1,] - get reading for sensors 0 and 1
None - get readings for all sensors
"""
_sensors = sensors
if _sensors is None:
_sensors = list(range(0, self._sensor_count))
if not set(_sensors).issubset(list(range(0, self._sensor_count))):
raise ValueError(
'Some or all of the sensors in the list %s are out of range '
'given a sensor_count of %d. Valid range: %s' % (
_sensors,
self._sensor_count,
list(range(0, self._sensor_count)),
)
)
data = self.get_data()
data = data['humidity_data']
results = {}
# Interpret device response
for sensor in _sensors:
offset = self.lookup_humidity_offset(sensor)
if offset is None:
continue
humidity = (struct.unpack_from('>H', data, offset)[0] * 32) / 1000.0
results[sensor] = {
'ports': self.get_ports(),
'bus': self.get_bus(),
'sensor': sensor,
'humidity_pc': humidity,
}
return results | def get_humidity(self, sensors=None):
"""
Get device humidity reading.
Params:
- sensors: optional list of sensors to get a reading for, examples:
[0,] - get reading for sensor 0
[0, 1,] - get reading for sensors 0 and 1
None - get readings for all sensors
"""
_sensors = sensors
if _sensors is None:
_sensors = list(range(0, self._sensor_count))
if not set(_sensors).issubset(list(range(0, self._sensor_count))):
raise ValueError(
'Some or all of the sensors in the list %s are out of range '
'given a sensor_count of %d. Valid range: %s' % (
_sensors,
self._sensor_count,
list(range(0, self._sensor_count)),
)
)
data = self.get_data()
data = data['humidity_data']
results = {}
# Interpret device response
for sensor in _sensors:
offset = self.lookup_humidity_offset(sensor)
if offset is None:
continue
humidity = (struct.unpack_from('>H', data, offset)[0] * 32) / 1000.0
results[sensor] = {
'ports': self.get_ports(),
'bus': self.get_bus(),
'sensor': sensor,
'humidity_pc': humidity,
}
return results | [
"Get",
"device",
"humidity",
"reading",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L345-L387 | [
"def",
"get_humidity",
"(",
"self",
",",
"sensors",
"=",
"None",
")",
":",
"_sensors",
"=",
"sensors",
"if",
"_sensors",
"is",
"None",
":",
"_sensors",
"=",
"list",
"(",
"range",
"(",
"0",
",",
"self",
".",
"_sensor_count",
")",
")",
"if",
"not",
"set",
"(",
"_sensors",
")",
".",
"issubset",
"(",
"list",
"(",
"range",
"(",
"0",
",",
"self",
".",
"_sensor_count",
")",
")",
")",
":",
"raise",
"ValueError",
"(",
"'Some or all of the sensors in the list %s are out of range '",
"'given a sensor_count of %d. Valid range: %s'",
"%",
"(",
"_sensors",
",",
"self",
".",
"_sensor_count",
",",
"list",
"(",
"range",
"(",
"0",
",",
"self",
".",
"_sensor_count",
")",
")",
",",
")",
")",
"data",
"=",
"self",
".",
"get_data",
"(",
")",
"data",
"=",
"data",
"[",
"'humidity_data'",
"]",
"results",
"=",
"{",
"}",
"# Interpret device response",
"for",
"sensor",
"in",
"_sensors",
":",
"offset",
"=",
"self",
".",
"lookup_humidity_offset",
"(",
"sensor",
")",
"if",
"offset",
"is",
"None",
":",
"continue",
"humidity",
"=",
"(",
"struct",
".",
"unpack_from",
"(",
"'>H'",
",",
"data",
",",
"offset",
")",
"[",
"0",
"]",
"*",
"32",
")",
"/",
"1000.0",
"results",
"[",
"sensor",
"]",
"=",
"{",
"'ports'",
":",
"self",
".",
"get_ports",
"(",
")",
",",
"'bus'",
":",
"self",
".",
"get_bus",
"(",
")",
",",
"'sensor'",
":",
"sensor",
",",
"'humidity_pc'",
":",
"humidity",
",",
"}",
"return",
"results"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice._control_transfer | Send device a control request with standard parameters and <data> as
payload. | temperusb/temper.py | def _control_transfer(self, data):
"""
Send device a control request with standard parameters and <data> as
payload.
"""
LOGGER.debug('Ctrl transfer: %r', data)
self._device.ctrl_transfer(bmRequestType=0x21, bRequest=0x09,
wValue=0x0200, wIndex=0x01, data_or_wLength=data, timeout=TIMEOUT) | def _control_transfer(self, data):
"""
Send device a control request with standard parameters and <data> as
payload.
"""
LOGGER.debug('Ctrl transfer: %r', data)
self._device.ctrl_transfer(bmRequestType=0x21, bRequest=0x09,
wValue=0x0200, wIndex=0x01, data_or_wLength=data, timeout=TIMEOUT) | [
"Send",
"device",
"a",
"control",
"request",
"with",
"standard",
"parameters",
"and",
"<data",
">",
"as",
"payload",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L389-L396 | [
"def",
"_control_transfer",
"(",
"self",
",",
"data",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'Ctrl transfer: %r'",
",",
"data",
")",
"self",
".",
"_device",
".",
"ctrl_transfer",
"(",
"bmRequestType",
"=",
"0x21",
",",
"bRequest",
"=",
"0x09",
",",
"wValue",
"=",
"0x0200",
",",
"wIndex",
"=",
"0x01",
",",
"data_or_wLength",
"=",
"data",
",",
"timeout",
"=",
"TIMEOUT",
")"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | TemperDevice._interrupt_read | Read data from device. | temperusb/temper.py | def _interrupt_read(self):
"""
Read data from device.
"""
data = self._device.read(ENDPOINT, REQ_INT_LEN, timeout=TIMEOUT)
LOGGER.debug('Read data: %r', data)
return data | def _interrupt_read(self):
"""
Read data from device.
"""
data = self._device.read(ENDPOINT, REQ_INT_LEN, timeout=TIMEOUT)
LOGGER.debug('Read data: %r', data)
return data | [
"Read",
"data",
"from",
"device",
"."
] | padelt/temper-python | python | https://github.com/padelt/temper-python/blob/cbdbace7e6755b1d91a2603ab63c9cb778078f79/temperusb/temper.py#L398-L404 | [
"def",
"_interrupt_read",
"(",
"self",
")",
":",
"data",
"=",
"self",
".",
"_device",
".",
"read",
"(",
"ENDPOINT",
",",
"REQ_INT_LEN",
",",
"timeout",
"=",
"TIMEOUT",
")",
"LOGGER",
".",
"debug",
"(",
"'Read data: %r'",
",",
"data",
")",
"return",
"data"
] | cbdbace7e6755b1d91a2603ab63c9cb778078f79 |
valid | Github.__check_looks_like_uri | Checks the URI looks like a RAW uri in github:
- 'https://raw.githubusercontent.com/github/hubot/master/README.md'
- 'https://github.com/github/hubot/raw/master/README.md'
:param uri: uri of the file | sirmordred/github.py | def __check_looks_like_uri(self, uri):
"""Checks the URI looks like a RAW uri in github:
- 'https://raw.githubusercontent.com/github/hubot/master/README.md'
- 'https://github.com/github/hubot/raw/master/README.md'
:param uri: uri of the file
"""
if uri.split('/')[2] == 'raw.githubusercontent.com':
return True
elif uri.split('/')[2] == 'github.com':
if uri.split('/')[5] == 'raw':
return True
else:
raise GithubFileNotFound('URI %s is not a valid link to a raw file in Github' % uri) | def __check_looks_like_uri(self, uri):
"""Checks the URI looks like a RAW uri in github:
- 'https://raw.githubusercontent.com/github/hubot/master/README.md'
- 'https://github.com/github/hubot/raw/master/README.md'
:param uri: uri of the file
"""
if uri.split('/')[2] == 'raw.githubusercontent.com':
return True
elif uri.split('/')[2] == 'github.com':
if uri.split('/')[5] == 'raw':
return True
else:
raise GithubFileNotFound('URI %s is not a valid link to a raw file in Github' % uri) | [
"Checks",
"the",
"URI",
"looks",
"like",
"a",
"RAW",
"uri",
"in",
"github",
":"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/github.py#L37-L51 | [
"def",
"__check_looks_like_uri",
"(",
"self",
",",
"uri",
")",
":",
"if",
"uri",
".",
"split",
"(",
"'/'",
")",
"[",
"2",
"]",
"==",
"'raw.githubusercontent.com'",
":",
"return",
"True",
"elif",
"uri",
".",
"split",
"(",
"'/'",
")",
"[",
"2",
"]",
"==",
"'github.com'",
":",
"if",
"uri",
".",
"split",
"(",
"'/'",
")",
"[",
"5",
"]",
"==",
"'raw'",
":",
"return",
"True",
"else",
":",
"raise",
"GithubFileNotFound",
"(",
"'URI %s is not a valid link to a raw file in Github'",
"%",
"uri",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | Github.read_file_from_uri | Reads the file from Github
:param uri: URI of the Github raw File
:returns: UTF-8 text with the content | sirmordred/github.py | def read_file_from_uri(self, uri):
"""Reads the file from Github
:param uri: URI of the Github raw File
:returns: UTF-8 text with the content
"""
logger.debug("Reading %s" % (uri))
self.__check_looks_like_uri(uri)
try:
req = urllib.request.Request(uri)
req.add_header('Authorization', 'token %s' % self.token)
r = urllib.request.urlopen(req)
except urllib.error.HTTPError as err:
if err.code == 404:
raise GithubFileNotFound('File %s is not available. Check the URL to ensure it really exists' % uri)
else:
raise
return r.read().decode("utf-8") | def read_file_from_uri(self, uri):
"""Reads the file from Github
:param uri: URI of the Github raw File
:returns: UTF-8 text with the content
"""
logger.debug("Reading %s" % (uri))
self.__check_looks_like_uri(uri)
try:
req = urllib.request.Request(uri)
req.add_header('Authorization', 'token %s' % self.token)
r = urllib.request.urlopen(req)
except urllib.error.HTTPError as err:
if err.code == 404:
raise GithubFileNotFound('File %s is not available. Check the URL to ensure it really exists' % uri)
else:
raise
return r.read().decode("utf-8") | [
"Reads",
"the",
"file",
"from",
"Github"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/github.py#L53-L74 | [
"def",
"read_file_from_uri",
"(",
"self",
",",
"uri",
")",
":",
"logger",
".",
"debug",
"(",
"\"Reading %s\"",
"%",
"(",
"uri",
")",
")",
"self",
".",
"__check_looks_like_uri",
"(",
"uri",
")",
"try",
":",
"req",
"=",
"urllib",
".",
"request",
".",
"Request",
"(",
"uri",
")",
"req",
".",
"add_header",
"(",
"'Authorization'",
",",
"'token %s'",
"%",
"self",
".",
"token",
")",
"r",
"=",
"urllib",
".",
"request",
".",
"urlopen",
"(",
"req",
")",
"except",
"urllib",
".",
"error",
".",
"HTTPError",
"as",
"err",
":",
"if",
"err",
".",
"code",
"==",
"404",
":",
"raise",
"GithubFileNotFound",
"(",
"'File %s is not available. Check the URL to ensure it really exists'",
"%",
"uri",
")",
"else",
":",
"raise",
"return",
"r",
".",
"read",
"(",
")",
".",
"decode",
"(",
"\"utf-8\"",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | TaskRawDataArthurCollection.measure_memory | Recursively finds size of objects | sirmordred/task_collection.py | def measure_memory(cls, obj, seen=None):
"""Recursively finds size of objects"""
size = sys.getsizeof(obj)
if seen is None:
seen = set()
obj_id = id(obj)
if obj_id in seen:
return 0
# Important mark as seen *before* entering recursion to gracefully handle
# self-referential objects
seen.add(obj_id)
if isinstance(obj, dict):
size += sum([cls.measure_memory(v, seen) for v in obj.values()])
size += sum([cls.measure_memory(k, seen) for k in obj.keys()])
elif hasattr(obj, '__dict__'):
size += cls.measure_memory(obj.__dict__, seen)
elif hasattr(obj, '__iter__') and not isinstance(obj, (str, bytes, bytearray)):
size += sum([cls.measure_memory(i, seen) for i in obj])
return size | def measure_memory(cls, obj, seen=None):
"""Recursively finds size of objects"""
size = sys.getsizeof(obj)
if seen is None:
seen = set()
obj_id = id(obj)
if obj_id in seen:
return 0
# Important mark as seen *before* entering recursion to gracefully handle
# self-referential objects
seen.add(obj_id)
if isinstance(obj, dict):
size += sum([cls.measure_memory(v, seen) for v in obj.values()])
size += sum([cls.measure_memory(k, seen) for k in obj.keys()])
elif hasattr(obj, '__dict__'):
size += cls.measure_memory(obj.__dict__, seen)
elif hasattr(obj, '__iter__') and not isinstance(obj, (str, bytes, bytearray)):
size += sum([cls.measure_memory(i, seen) for i in obj])
return size | [
"Recursively",
"finds",
"size",
"of",
"objects"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/task_collection.py#L173-L191 | [
"def",
"measure_memory",
"(",
"cls",
",",
"obj",
",",
"seen",
"=",
"None",
")",
":",
"size",
"=",
"sys",
".",
"getsizeof",
"(",
"obj",
")",
"if",
"seen",
"is",
"None",
":",
"seen",
"=",
"set",
"(",
")",
"obj_id",
"=",
"id",
"(",
"obj",
")",
"if",
"obj_id",
"in",
"seen",
":",
"return",
"0",
"# Important mark as seen *before* entering recursion to gracefully handle",
"# self-referential objects",
"seen",
".",
"add",
"(",
"obj_id",
")",
"if",
"isinstance",
"(",
"obj",
",",
"dict",
")",
":",
"size",
"+=",
"sum",
"(",
"[",
"cls",
".",
"measure_memory",
"(",
"v",
",",
"seen",
")",
"for",
"v",
"in",
"obj",
".",
"values",
"(",
")",
"]",
")",
"size",
"+=",
"sum",
"(",
"[",
"cls",
".",
"measure_memory",
"(",
"k",
",",
"seen",
")",
"for",
"k",
"in",
"obj",
".",
"keys",
"(",
")",
"]",
")",
"elif",
"hasattr",
"(",
"obj",
",",
"'__dict__'",
")",
":",
"size",
"+=",
"cls",
".",
"measure_memory",
"(",
"obj",
".",
"__dict__",
",",
"seen",
")",
"elif",
"hasattr",
"(",
"obj",
",",
"'__iter__'",
")",
"and",
"not",
"isinstance",
"(",
"obj",
",",
"(",
"str",
",",
"bytes",
",",
"bytearray",
")",
")",
":",
"size",
"+=",
"sum",
"(",
"[",
"cls",
".",
"measure_memory",
"(",
"i",
",",
"seen",
")",
"for",
"i",
"in",
"obj",
"]",
")",
"return",
"size"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | TaskRawDataArthurCollection.__feed_arthur | Feed Ocean with backend data collected from arthur redis queue | sirmordred/task_collection.py | def __feed_arthur(self):
""" Feed Ocean with backend data collected from arthur redis queue"""
with self.ARTHUR_FEED_LOCK:
# This is a expensive operation so don't do it always
if (time.time() - self.ARTHUR_LAST_MEMORY_CHECK) > 5 * self.ARTHUR_LAST_MEMORY_CHECK_TIME:
self.ARTHUR_LAST_MEMORY_CHECK = time.time()
logger.debug("Measuring the memory used by the raw items dict ...")
try:
memory_size = self.measure_memory(self.arthur_items) / (1024 * 1024)
except RuntimeError as ex:
# During memory usage measure, other thread could change the dict
logger.warning("Can't get the memory used by the raw items dict: %s", ex)
memory_size = self.ARTHUR_LAST_MEMORY_SIZE
self.ARTHUR_LAST_MEMORY_CHECK_TIME = time.time() - self.ARTHUR_LAST_MEMORY_CHECK
logger.debug("Arthur items memory size: %0.2f MB (%is to check)",
memory_size, self.ARTHUR_LAST_MEMORY_CHECK_TIME)
self.ARTHUR_LAST_MEMORY_SIZE = memory_size
# Don't feed items from redis if the current python dict is
# larger than ARTHUR_MAX_MEMORY_SIZE
if self.ARTHUR_LAST_MEMORY_SIZE > self.ARTHUR_MAX_MEMORY_SIZE:
logger.debug("Items queue full. Not collecting items from redis queue.")
return
logger.info("Collecting items from redis queue")
db_url = self.config.get_conf()['es_collection']['redis_url']
conn = redis.StrictRedis.from_url(db_url)
logger.debug("Redis connection stablished with %s.", db_url)
# Get and remove queued items in an atomic transaction
pipe = conn.pipeline()
# pipe.lrange(Q_STORAGE_ITEMS, 0, -1)
pipe.lrange(Q_STORAGE_ITEMS, 0, self.ARTHUR_REDIS_ITEMS - 1)
pipe.ltrim(Q_STORAGE_ITEMS, self.ARTHUR_REDIS_ITEMS, -1)
items = pipe.execute()[0]
for item in items:
arthur_item = pickle.loads(item)
if arthur_item['tag'] not in self.arthur_items:
self.arthur_items[arthur_item['tag']] = []
self.arthur_items[arthur_item['tag']].append(arthur_item)
for tag in self.arthur_items:
if self.arthur_items[tag]:
logger.debug("Arthur items for %s: %i", tag, len(self.arthur_items[tag])) | def __feed_arthur(self):
""" Feed Ocean with backend data collected from arthur redis queue"""
with self.ARTHUR_FEED_LOCK:
# This is a expensive operation so don't do it always
if (time.time() - self.ARTHUR_LAST_MEMORY_CHECK) > 5 * self.ARTHUR_LAST_MEMORY_CHECK_TIME:
self.ARTHUR_LAST_MEMORY_CHECK = time.time()
logger.debug("Measuring the memory used by the raw items dict ...")
try:
memory_size = self.measure_memory(self.arthur_items) / (1024 * 1024)
except RuntimeError as ex:
# During memory usage measure, other thread could change the dict
logger.warning("Can't get the memory used by the raw items dict: %s", ex)
memory_size = self.ARTHUR_LAST_MEMORY_SIZE
self.ARTHUR_LAST_MEMORY_CHECK_TIME = time.time() - self.ARTHUR_LAST_MEMORY_CHECK
logger.debug("Arthur items memory size: %0.2f MB (%is to check)",
memory_size, self.ARTHUR_LAST_MEMORY_CHECK_TIME)
self.ARTHUR_LAST_MEMORY_SIZE = memory_size
# Don't feed items from redis if the current python dict is
# larger than ARTHUR_MAX_MEMORY_SIZE
if self.ARTHUR_LAST_MEMORY_SIZE > self.ARTHUR_MAX_MEMORY_SIZE:
logger.debug("Items queue full. Not collecting items from redis queue.")
return
logger.info("Collecting items from redis queue")
db_url = self.config.get_conf()['es_collection']['redis_url']
conn = redis.StrictRedis.from_url(db_url)
logger.debug("Redis connection stablished with %s.", db_url)
# Get and remove queued items in an atomic transaction
pipe = conn.pipeline()
# pipe.lrange(Q_STORAGE_ITEMS, 0, -1)
pipe.lrange(Q_STORAGE_ITEMS, 0, self.ARTHUR_REDIS_ITEMS - 1)
pipe.ltrim(Q_STORAGE_ITEMS, self.ARTHUR_REDIS_ITEMS, -1)
items = pipe.execute()[0]
for item in items:
arthur_item = pickle.loads(item)
if arthur_item['tag'] not in self.arthur_items:
self.arthur_items[arthur_item['tag']] = []
self.arthur_items[arthur_item['tag']].append(arthur_item)
for tag in self.arthur_items:
if self.arthur_items[tag]:
logger.debug("Arthur items for %s: %i", tag, len(self.arthur_items[tag])) | [
"Feed",
"Ocean",
"with",
"backend",
"data",
"collected",
"from",
"arthur",
"redis",
"queue"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/task_collection.py#L193-L242 | [
"def",
"__feed_arthur",
"(",
"self",
")",
":",
"with",
"self",
".",
"ARTHUR_FEED_LOCK",
":",
"# This is a expensive operation so don't do it always",
"if",
"(",
"time",
".",
"time",
"(",
")",
"-",
"self",
".",
"ARTHUR_LAST_MEMORY_CHECK",
")",
">",
"5",
"*",
"self",
".",
"ARTHUR_LAST_MEMORY_CHECK_TIME",
":",
"self",
".",
"ARTHUR_LAST_MEMORY_CHECK",
"=",
"time",
".",
"time",
"(",
")",
"logger",
".",
"debug",
"(",
"\"Measuring the memory used by the raw items dict ...\"",
")",
"try",
":",
"memory_size",
"=",
"self",
".",
"measure_memory",
"(",
"self",
".",
"arthur_items",
")",
"/",
"(",
"1024",
"*",
"1024",
")",
"except",
"RuntimeError",
"as",
"ex",
":",
"# During memory usage measure, other thread could change the dict",
"logger",
".",
"warning",
"(",
"\"Can't get the memory used by the raw items dict: %s\"",
",",
"ex",
")",
"memory_size",
"=",
"self",
".",
"ARTHUR_LAST_MEMORY_SIZE",
"self",
".",
"ARTHUR_LAST_MEMORY_CHECK_TIME",
"=",
"time",
".",
"time",
"(",
")",
"-",
"self",
".",
"ARTHUR_LAST_MEMORY_CHECK",
"logger",
".",
"debug",
"(",
"\"Arthur items memory size: %0.2f MB (%is to check)\"",
",",
"memory_size",
",",
"self",
".",
"ARTHUR_LAST_MEMORY_CHECK_TIME",
")",
"self",
".",
"ARTHUR_LAST_MEMORY_SIZE",
"=",
"memory_size",
"# Don't feed items from redis if the current python dict is",
"# larger than ARTHUR_MAX_MEMORY_SIZE",
"if",
"self",
".",
"ARTHUR_LAST_MEMORY_SIZE",
">",
"self",
".",
"ARTHUR_MAX_MEMORY_SIZE",
":",
"logger",
".",
"debug",
"(",
"\"Items queue full. Not collecting items from redis queue.\"",
")",
"return",
"logger",
".",
"info",
"(",
"\"Collecting items from redis queue\"",
")",
"db_url",
"=",
"self",
".",
"config",
".",
"get_conf",
"(",
")",
"[",
"'es_collection'",
"]",
"[",
"'redis_url'",
"]",
"conn",
"=",
"redis",
".",
"StrictRedis",
".",
"from_url",
"(",
"db_url",
")",
"logger",
".",
"debug",
"(",
"\"Redis connection stablished with %s.\"",
",",
"db_url",
")",
"# Get and remove queued items in an atomic transaction",
"pipe",
"=",
"conn",
".",
"pipeline",
"(",
")",
"# pipe.lrange(Q_STORAGE_ITEMS, 0, -1)",
"pipe",
".",
"lrange",
"(",
"Q_STORAGE_ITEMS",
",",
"0",
",",
"self",
".",
"ARTHUR_REDIS_ITEMS",
"-",
"1",
")",
"pipe",
".",
"ltrim",
"(",
"Q_STORAGE_ITEMS",
",",
"self",
".",
"ARTHUR_REDIS_ITEMS",
",",
"-",
"1",
")",
"items",
"=",
"pipe",
".",
"execute",
"(",
")",
"[",
"0",
"]",
"for",
"item",
"in",
"items",
":",
"arthur_item",
"=",
"pickle",
".",
"loads",
"(",
"item",
")",
"if",
"arthur_item",
"[",
"'tag'",
"]",
"not",
"in",
"self",
".",
"arthur_items",
":",
"self",
".",
"arthur_items",
"[",
"arthur_item",
"[",
"'tag'",
"]",
"]",
"=",
"[",
"]",
"self",
".",
"arthur_items",
"[",
"arthur_item",
"[",
"'tag'",
"]",
"]",
".",
"append",
"(",
"arthur_item",
")",
"for",
"tag",
"in",
"self",
".",
"arthur_items",
":",
"if",
"self",
".",
"arthur_items",
"[",
"tag",
"]",
":",
"logger",
".",
"debug",
"(",
"\"Arthur items for %s: %i\"",
",",
"tag",
",",
"len",
"(",
"self",
".",
"arthur_items",
"[",
"tag",
"]",
")",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | TaskRawDataArthurCollection.__feed_backend_arthur | Feed Ocean with backend data collected from arthur redis queue | sirmordred/task_collection.py | def __feed_backend_arthur(self, repo):
""" Feed Ocean with backend data collected from arthur redis queue"""
# Always get pending items from arthur for all data sources
self.__feed_arthur()
tag = self.backend_tag(repo)
logger.debug("Arthur items available for %s", self.arthur_items.keys())
logger.debug("Getting arthur items for %s.", tag)
if tag in self.arthur_items:
logger.debug("Found items for %s.", tag)
while self.arthur_items[tag]:
yield self.arthur_items[tag].pop() | def __feed_backend_arthur(self, repo):
""" Feed Ocean with backend data collected from arthur redis queue"""
# Always get pending items from arthur for all data sources
self.__feed_arthur()
tag = self.backend_tag(repo)
logger.debug("Arthur items available for %s", self.arthur_items.keys())
logger.debug("Getting arthur items for %s.", tag)
if tag in self.arthur_items:
logger.debug("Found items for %s.", tag)
while self.arthur_items[tag]:
yield self.arthur_items[tag].pop() | [
"Feed",
"Ocean",
"with",
"backend",
"data",
"collected",
"from",
"arthur",
"redis",
"queue"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/task_collection.py#L257-L272 | [
"def",
"__feed_backend_arthur",
"(",
"self",
",",
"repo",
")",
":",
"# Always get pending items from arthur for all data sources",
"self",
".",
"__feed_arthur",
"(",
")",
"tag",
"=",
"self",
".",
"backend_tag",
"(",
"repo",
")",
"logger",
".",
"debug",
"(",
"\"Arthur items available for %s\"",
",",
"self",
".",
"arthur_items",
".",
"keys",
"(",
")",
")",
"logger",
".",
"debug",
"(",
"\"Getting arthur items for %s.\"",
",",
"tag",
")",
"if",
"tag",
"in",
"self",
".",
"arthur_items",
":",
"logger",
".",
"debug",
"(",
"\"Found items for %s.\"",
",",
"tag",
")",
"while",
"self",
".",
"arthur_items",
"[",
"tag",
"]",
":",
"yield",
"self",
".",
"arthur_items",
"[",
"tag",
"]",
".",
"pop",
"(",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | TaskRawDataArthurCollection.__create_arthur_json | Create the JSON for configuring arthur to collect data
https://github.com/grimoirelab/arthur#adding-tasks
Sample for git:
{
"tasks": [
{
"task_id": "arthur.git",
"backend": "git",
"backend_args": {
"gitpath": "/tmp/arthur_git/",
"uri": "https://github.com/grimoirelab/arthur.git"
},
"category": "commit",
"archive_args": {
"archive_path": '/tmp/test_archives',
"fetch_from_archive": false,
"archive_after": None
},
"scheduler_args": {
"delay": 10
}
}
]
} | sirmordred/task_collection.py | def __create_arthur_json(self, repo, backend_args):
""" Create the JSON for configuring arthur to collect data
https://github.com/grimoirelab/arthur#adding-tasks
Sample for git:
{
"tasks": [
{
"task_id": "arthur.git",
"backend": "git",
"backend_args": {
"gitpath": "/tmp/arthur_git/",
"uri": "https://github.com/grimoirelab/arthur.git"
},
"category": "commit",
"archive_args": {
"archive_path": '/tmp/test_archives',
"fetch_from_archive": false,
"archive_after": None
},
"scheduler_args": {
"delay": 10
}
}
]
}
"""
backend_args = self._compose_arthur_params(self.backend_section, repo)
if self.backend_section == 'git':
backend_args['gitpath'] = os.path.join(self.REPOSITORY_DIR, repo)
backend_args['tag'] = self.backend_tag(repo)
ajson = {"tasks": [{}]}
# This is the perceval tag
ajson["tasks"][0]['task_id'] = self.backend_tag(repo)
ajson["tasks"][0]['backend'] = self.backend_section.split(":")[0]
ajson["tasks"][0]['backend_args'] = backend_args
ajson["tasks"][0]['category'] = backend_args['category']
ajson["tasks"][0]['archive'] = {}
ajson["tasks"][0]['scheduler'] = {"delay": self.ARTHUR_TASK_DELAY}
# from-date or offset param must be added
es_col_url = self._get_collection_url()
es_index = self.conf[self.backend_section]['raw_index']
# Get the last activity for the data source
es = ElasticSearch(es_col_url, es_index)
connector = get_connector_from_name(self.backend_section)
klass = connector[0] # Backend for the connector
signature = inspect.signature(klass.fetch)
last_activity = None
filter_ = {"name": "tag", "value": backend_args['tag']}
if 'from_date' in signature.parameters:
last_activity = es.get_last_item_field('metadata__updated_on', [filter_])
if last_activity:
ajson["tasks"][0]['backend_args']['from_date'] = last_activity.isoformat()
elif 'offset' in signature.parameters:
last_activity = es.get_last_item_field('offset', [filter_])
if last_activity:
ajson["tasks"][0]['backend_args']['offset'] = last_activity
if last_activity:
logging.info("Getting raw item with arthur since %s", last_activity)
return(ajson) | def __create_arthur_json(self, repo, backend_args):
""" Create the JSON for configuring arthur to collect data
https://github.com/grimoirelab/arthur#adding-tasks
Sample for git:
{
"tasks": [
{
"task_id": "arthur.git",
"backend": "git",
"backend_args": {
"gitpath": "/tmp/arthur_git/",
"uri": "https://github.com/grimoirelab/arthur.git"
},
"category": "commit",
"archive_args": {
"archive_path": '/tmp/test_archives',
"fetch_from_archive": false,
"archive_after": None
},
"scheduler_args": {
"delay": 10
}
}
]
}
"""
backend_args = self._compose_arthur_params(self.backend_section, repo)
if self.backend_section == 'git':
backend_args['gitpath'] = os.path.join(self.REPOSITORY_DIR, repo)
backend_args['tag'] = self.backend_tag(repo)
ajson = {"tasks": [{}]}
# This is the perceval tag
ajson["tasks"][0]['task_id'] = self.backend_tag(repo)
ajson["tasks"][0]['backend'] = self.backend_section.split(":")[0]
ajson["tasks"][0]['backend_args'] = backend_args
ajson["tasks"][0]['category'] = backend_args['category']
ajson["tasks"][0]['archive'] = {}
ajson["tasks"][0]['scheduler'] = {"delay": self.ARTHUR_TASK_DELAY}
# from-date or offset param must be added
es_col_url = self._get_collection_url()
es_index = self.conf[self.backend_section]['raw_index']
# Get the last activity for the data source
es = ElasticSearch(es_col_url, es_index)
connector = get_connector_from_name(self.backend_section)
klass = connector[0] # Backend for the connector
signature = inspect.signature(klass.fetch)
last_activity = None
filter_ = {"name": "tag", "value": backend_args['tag']}
if 'from_date' in signature.parameters:
last_activity = es.get_last_item_field('metadata__updated_on', [filter_])
if last_activity:
ajson["tasks"][0]['backend_args']['from_date'] = last_activity.isoformat()
elif 'offset' in signature.parameters:
last_activity = es.get_last_item_field('offset', [filter_])
if last_activity:
ajson["tasks"][0]['backend_args']['offset'] = last_activity
if last_activity:
logging.info("Getting raw item with arthur since %s", last_activity)
return(ajson) | [
"Create",
"the",
"JSON",
"for",
"configuring",
"arthur",
"to",
"collect",
"data"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/task_collection.py#L274-L340 | [
"def",
"__create_arthur_json",
"(",
"self",
",",
"repo",
",",
"backend_args",
")",
":",
"backend_args",
"=",
"self",
".",
"_compose_arthur_params",
"(",
"self",
".",
"backend_section",
",",
"repo",
")",
"if",
"self",
".",
"backend_section",
"==",
"'git'",
":",
"backend_args",
"[",
"'gitpath'",
"]",
"=",
"os",
".",
"path",
".",
"join",
"(",
"self",
".",
"REPOSITORY_DIR",
",",
"repo",
")",
"backend_args",
"[",
"'tag'",
"]",
"=",
"self",
".",
"backend_tag",
"(",
"repo",
")",
"ajson",
"=",
"{",
"\"tasks\"",
":",
"[",
"{",
"}",
"]",
"}",
"# This is the perceval tag",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'task_id'",
"]",
"=",
"self",
".",
"backend_tag",
"(",
"repo",
")",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'backend'",
"]",
"=",
"self",
".",
"backend_section",
".",
"split",
"(",
"\":\"",
")",
"[",
"0",
"]",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'backend_args'",
"]",
"=",
"backend_args",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'category'",
"]",
"=",
"backend_args",
"[",
"'category'",
"]",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'archive'",
"]",
"=",
"{",
"}",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'scheduler'",
"]",
"=",
"{",
"\"delay\"",
":",
"self",
".",
"ARTHUR_TASK_DELAY",
"}",
"# from-date or offset param must be added",
"es_col_url",
"=",
"self",
".",
"_get_collection_url",
"(",
")",
"es_index",
"=",
"self",
".",
"conf",
"[",
"self",
".",
"backend_section",
"]",
"[",
"'raw_index'",
"]",
"# Get the last activity for the data source",
"es",
"=",
"ElasticSearch",
"(",
"es_col_url",
",",
"es_index",
")",
"connector",
"=",
"get_connector_from_name",
"(",
"self",
".",
"backend_section",
")",
"klass",
"=",
"connector",
"[",
"0",
"]",
"# Backend for the connector",
"signature",
"=",
"inspect",
".",
"signature",
"(",
"klass",
".",
"fetch",
")",
"last_activity",
"=",
"None",
"filter_",
"=",
"{",
"\"name\"",
":",
"\"tag\"",
",",
"\"value\"",
":",
"backend_args",
"[",
"'tag'",
"]",
"}",
"if",
"'from_date'",
"in",
"signature",
".",
"parameters",
":",
"last_activity",
"=",
"es",
".",
"get_last_item_field",
"(",
"'metadata__updated_on'",
",",
"[",
"filter_",
"]",
")",
"if",
"last_activity",
":",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'backend_args'",
"]",
"[",
"'from_date'",
"]",
"=",
"last_activity",
".",
"isoformat",
"(",
")",
"elif",
"'offset'",
"in",
"signature",
".",
"parameters",
":",
"last_activity",
"=",
"es",
".",
"get_last_item_field",
"(",
"'offset'",
",",
"[",
"filter_",
"]",
")",
"if",
"last_activity",
":",
"ajson",
"[",
"\"tasks\"",
"]",
"[",
"0",
"]",
"[",
"'backend_args'",
"]",
"[",
"'offset'",
"]",
"=",
"last_activity",
"if",
"last_activity",
":",
"logging",
".",
"info",
"(",
"\"Getting raw item with arthur since %s\"",
",",
"last_activity",
")",
"return",
"(",
"ajson",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | TaskIdentitiesExport.sha_github_file | Return the GitHub SHA for a file in the repository | sirmordred/task_identities.py | def sha_github_file(cls, config, repo_file, repository_api, repository_branch):
""" Return the GitHub SHA for a file in the repository """
repo_file_sha = None
cfg = config.get_conf()
github_token = cfg['sortinghat']['identities_api_token']
headers = {"Authorization": "token " + github_token}
url_dir = repository_api + "/git/trees/" + repository_branch
logger.debug("Gettting sha data from tree: %s", url_dir)
raw_repo_file_info = requests.get(url_dir, headers=headers)
raw_repo_file_info.raise_for_status()
for rfile in raw_repo_file_info.json()['tree']:
if rfile['path'] == repo_file:
logger.debug("SHA found: %s, ", rfile["sha"])
repo_file_sha = rfile["sha"]
break
return repo_file_sha | def sha_github_file(cls, config, repo_file, repository_api, repository_branch):
""" Return the GitHub SHA for a file in the repository """
repo_file_sha = None
cfg = config.get_conf()
github_token = cfg['sortinghat']['identities_api_token']
headers = {"Authorization": "token " + github_token}
url_dir = repository_api + "/git/trees/" + repository_branch
logger.debug("Gettting sha data from tree: %s", url_dir)
raw_repo_file_info = requests.get(url_dir, headers=headers)
raw_repo_file_info.raise_for_status()
for rfile in raw_repo_file_info.json()['tree']:
if rfile['path'] == repo_file:
logger.debug("SHA found: %s, ", rfile["sha"])
repo_file_sha = rfile["sha"]
break
return repo_file_sha | [
"Return",
"the",
"GitHub",
"SHA",
"for",
"a",
"file",
"in",
"the",
"repository"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/task_identities.py#L318-L337 | [
"def",
"sha_github_file",
"(",
"cls",
",",
"config",
",",
"repo_file",
",",
"repository_api",
",",
"repository_branch",
")",
":",
"repo_file_sha",
"=",
"None",
"cfg",
"=",
"config",
".",
"get_conf",
"(",
")",
"github_token",
"=",
"cfg",
"[",
"'sortinghat'",
"]",
"[",
"'identities_api_token'",
"]",
"headers",
"=",
"{",
"\"Authorization\"",
":",
"\"token \"",
"+",
"github_token",
"}",
"url_dir",
"=",
"repository_api",
"+",
"\"/git/trees/\"",
"+",
"repository_branch",
"logger",
".",
"debug",
"(",
"\"Gettting sha data from tree: %s\"",
",",
"url_dir",
")",
"raw_repo_file_info",
"=",
"requests",
".",
"get",
"(",
"url_dir",
",",
"headers",
"=",
"headers",
")",
"raw_repo_file_info",
".",
"raise_for_status",
"(",
")",
"for",
"rfile",
"in",
"raw_repo_file_info",
".",
"json",
"(",
")",
"[",
"'tree'",
"]",
":",
"if",
"rfile",
"[",
"'path'",
"]",
"==",
"repo_file",
":",
"logger",
".",
"debug",
"(",
"\"SHA found: %s, \"",
",",
"rfile",
"[",
"\"sha\"",
"]",
")",
"repo_file_sha",
"=",
"rfile",
"[",
"\"sha\"",
"]",
"break",
"return",
"repo_file_sha"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | TaskIdentitiesMerge.__get_uuids_from_profile_name | Get the uuid for a profile name | sirmordred/task_identities.py | def __get_uuids_from_profile_name(self, profile_name):
""" Get the uuid for a profile name """
uuids = []
with self.db.connect() as session:
query = session.query(Profile).\
filter(Profile.name == profile_name)
profiles = query.all()
if profiles:
for p in profiles:
uuids.append(p.uuid)
return uuids | def __get_uuids_from_profile_name(self, profile_name):
""" Get the uuid for a profile name """
uuids = []
with self.db.connect() as session:
query = session.query(Profile).\
filter(Profile.name == profile_name)
profiles = query.all()
if profiles:
for p in profiles:
uuids.append(p.uuid)
return uuids | [
"Get",
"the",
"uuid",
"for",
"a",
"profile",
"name"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/sirmordred/task_identities.py#L427-L438 | [
"def",
"__get_uuids_from_profile_name",
"(",
"self",
",",
"profile_name",
")",
":",
"uuids",
"=",
"[",
"]",
"with",
"self",
".",
"db",
".",
"connect",
"(",
")",
"as",
"session",
":",
"query",
"=",
"session",
".",
"query",
"(",
"Profile",
")",
".",
"filter",
"(",
"Profile",
".",
"name",
"==",
"profile_name",
")",
"profiles",
"=",
"query",
".",
"all",
"(",
")",
"if",
"profiles",
":",
"for",
"p",
"in",
"profiles",
":",
"uuids",
".",
"append",
"(",
"p",
".",
"uuid",
")",
"return",
"uuids"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | micro_mordred | Execute the raw and/or the enrich phases of a given backend section defined in a Mordred configuration file.
:param cfg_path: the path of a Mordred configuration file
:param backend_sections: the backend sections where the raw and/or enrich phases will be executed
:param raw: if true, it activates the collection of raw data
:param arthur: if true, it enables Arthur to collect the raw data
:param identities: if true, it activates the identities merge in SortingHat
:param enrich: if true, it activates the collection of enrich data
:param panels: if true, it activates the upload of panels | utils/micro.py | def micro_mordred(cfg_path, backend_sections, raw, arthur, identities, enrich, panels):
"""Execute the raw and/or the enrich phases of a given backend section defined in a Mordred configuration file.
:param cfg_path: the path of a Mordred configuration file
:param backend_sections: the backend sections where the raw and/or enrich phases will be executed
:param raw: if true, it activates the collection of raw data
:param arthur: if true, it enables Arthur to collect the raw data
:param identities: if true, it activates the identities merge in SortingHat
:param enrich: if true, it activates the collection of enrich data
:param panels: if true, it activates the upload of panels
"""
config = Config(cfg_path)
if raw:
for backend in backend_sections:
get_raw(config, backend, arthur)
if identities:
get_identities(config)
if enrich:
for backend in backend_sections:
get_enrich(config, backend)
if panels:
get_panels(config) | def micro_mordred(cfg_path, backend_sections, raw, arthur, identities, enrich, panels):
"""Execute the raw and/or the enrich phases of a given backend section defined in a Mordred configuration file.
:param cfg_path: the path of a Mordred configuration file
:param backend_sections: the backend sections where the raw and/or enrich phases will be executed
:param raw: if true, it activates the collection of raw data
:param arthur: if true, it enables Arthur to collect the raw data
:param identities: if true, it activates the identities merge in SortingHat
:param enrich: if true, it activates the collection of enrich data
:param panels: if true, it activates the upload of panels
"""
config = Config(cfg_path)
if raw:
for backend in backend_sections:
get_raw(config, backend, arthur)
if identities:
get_identities(config)
if enrich:
for backend in backend_sections:
get_enrich(config, backend)
if panels:
get_panels(config) | [
"Execute",
"the",
"raw",
"and",
"/",
"or",
"the",
"enrich",
"phases",
"of",
"a",
"given",
"backend",
"section",
"defined",
"in",
"a",
"Mordred",
"configuration",
"file",
"."
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/utils/micro.py#L41-L67 | [
"def",
"micro_mordred",
"(",
"cfg_path",
",",
"backend_sections",
",",
"raw",
",",
"arthur",
",",
"identities",
",",
"enrich",
",",
"panels",
")",
":",
"config",
"=",
"Config",
"(",
"cfg_path",
")",
"if",
"raw",
":",
"for",
"backend",
"in",
"backend_sections",
":",
"get_raw",
"(",
"config",
",",
"backend",
",",
"arthur",
")",
"if",
"identities",
":",
"get_identities",
"(",
"config",
")",
"if",
"enrich",
":",
"for",
"backend",
"in",
"backend_sections",
":",
"get_enrich",
"(",
"config",
",",
"backend",
")",
"if",
"panels",
":",
"get_panels",
"(",
"config",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | get_raw | Execute the raw phase for a given backend section, optionally using Arthur
:param config: a Mordred config object
:param backend_section: the backend section where the raw phase is executed
:param arthur: if true, it enables Arthur to collect the raw data | utils/micro.py | def get_raw(config, backend_section, arthur):
"""Execute the raw phase for a given backend section, optionally using Arthur
:param config: a Mordred config object
:param backend_section: the backend section where the raw phase is executed
:param arthur: if true, it enables Arthur to collect the raw data
"""
if arthur:
task = TaskRawDataArthurCollection(config, backend_section=backend_section)
else:
task = TaskRawDataCollection(config, backend_section=backend_section)
TaskProjects(config).execute()
try:
task.execute()
logging.info("Loading raw data finished!")
except Exception as e:
logging.error(str(e))
sys.exit(-1) | def get_raw(config, backend_section, arthur):
"""Execute the raw phase for a given backend section, optionally using Arthur
:param config: a Mordred config object
:param backend_section: the backend section where the raw phase is executed
:param arthur: if true, it enables Arthur to collect the raw data
"""
if arthur:
task = TaskRawDataArthurCollection(config, backend_section=backend_section)
else:
task = TaskRawDataCollection(config, backend_section=backend_section)
TaskProjects(config).execute()
try:
task.execute()
logging.info("Loading raw data finished!")
except Exception as e:
logging.error(str(e))
sys.exit(-1) | [
"Execute",
"the",
"raw",
"phase",
"for",
"a",
"given",
"backend",
"section",
"optionally",
"using",
"Arthur"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/utils/micro.py#L70-L89 | [
"def",
"get_raw",
"(",
"config",
",",
"backend_section",
",",
"arthur",
")",
":",
"if",
"arthur",
":",
"task",
"=",
"TaskRawDataArthurCollection",
"(",
"config",
",",
"backend_section",
"=",
"backend_section",
")",
"else",
":",
"task",
"=",
"TaskRawDataCollection",
"(",
"config",
",",
"backend_section",
"=",
"backend_section",
")",
"TaskProjects",
"(",
"config",
")",
".",
"execute",
"(",
")",
"try",
":",
"task",
".",
"execute",
"(",
")",
"logging",
".",
"info",
"(",
"\"Loading raw data finished!\"",
")",
"except",
"Exception",
"as",
"e",
":",
"logging",
".",
"error",
"(",
"str",
"(",
"e",
")",
")",
"sys",
".",
"exit",
"(",
"-",
"1",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
valid | get_identities | Execute the merge identities phase
:param config: a Mordred config object | utils/micro.py | def get_identities(config):
"""Execute the merge identities phase
:param config: a Mordred config object
"""
TaskProjects(config).execute()
task = TaskIdentitiesMerge(config)
task.execute()
logging.info("Merging identities finished!") | def get_identities(config):
"""Execute the merge identities phase
:param config: a Mordred config object
"""
TaskProjects(config).execute()
task = TaskIdentitiesMerge(config)
task.execute()
logging.info("Merging identities finished!") | [
"Execute",
"the",
"merge",
"identities",
"phase"
] | chaoss/grimoirelab-sirmordred | python | https://github.com/chaoss/grimoirelab-sirmordred/blob/d6ac94d28d707fae23170064d078f1edf937d13e/utils/micro.py#L92-L101 | [
"def",
"get_identities",
"(",
"config",
")",
":",
"TaskProjects",
"(",
"config",
")",
".",
"execute",
"(",
")",
"task",
"=",
"TaskIdentitiesMerge",
"(",
"config",
")",
"task",
".",
"execute",
"(",
")",
"logging",
".",
"info",
"(",
"\"Merging identities finished!\"",
")"
] | d6ac94d28d707fae23170064d078f1edf937d13e |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.