partition
stringclasses
3 values
func_name
stringlengths
1
134
docstring
stringlengths
1
46.9k
path
stringlengths
4
223
original_string
stringlengths
75
104k
code
stringlengths
75
104k
docstring_tokens
listlengths
1
1.97k
repo
stringlengths
7
55
language
stringclasses
1 value
url
stringlengths
87
315
code_tokens
listlengths
19
28.4k
sha
stringlengths
40
40
train
Docstring.method_returns_something
Check if the docstrings method can return something. Bare returns, returns valued None and returns from nested functions are disconsidered. Returns ------- bool Whether the docstrings method can return something.
scripts/validate_docstrings.py
def method_returns_something(self): ''' Check if the docstrings method can return something. Bare returns, returns valued None and returns from nested functions are disconsidered. Returns ------- bool Whether the docstrings method can return something. ''' def get_returns_not_on_nested_functions(node): returns = [node] if isinstance(node, ast.Return) else [] for child in ast.iter_child_nodes(node): # Ignore nested functions and its subtrees. if not isinstance(child, ast.FunctionDef): child_returns = get_returns_not_on_nested_functions(child) returns.extend(child_returns) return returns tree = ast.parse(self.method_source).body if tree: returns = get_returns_not_on_nested_functions(tree[0]) return_values = [r.value for r in returns] # Replace NameConstant nodes valued None for None. for i, v in enumerate(return_values): if isinstance(v, ast.NameConstant) and v.value is None: return_values[i] = None return any(return_values) else: return False
def method_returns_something(self): ''' Check if the docstrings method can return something. Bare returns, returns valued None and returns from nested functions are disconsidered. Returns ------- bool Whether the docstrings method can return something. ''' def get_returns_not_on_nested_functions(node): returns = [node] if isinstance(node, ast.Return) else [] for child in ast.iter_child_nodes(node): # Ignore nested functions and its subtrees. if not isinstance(child, ast.FunctionDef): child_returns = get_returns_not_on_nested_functions(child) returns.extend(child_returns) return returns tree = ast.parse(self.method_source).body if tree: returns = get_returns_not_on_nested_functions(tree[0]) return_values = [r.value for r in returns] # Replace NameConstant nodes valued None for None. for i, v in enumerate(return_values): if isinstance(v, ast.NameConstant) and v.value is None: return_values[i] = None return any(return_values) else: return False
[ "Check", "if", "the", "docstrings", "method", "can", "return", "something", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/scripts/validate_docstrings.py#L503-L535
[ "def", "method_returns_something", "(", "self", ")", ":", "def", "get_returns_not_on_nested_functions", "(", "node", ")", ":", "returns", "=", "[", "node", "]", "if", "isinstance", "(", "node", ",", "ast", ".", "Return", ")", "else", "[", "]", "for", "child", "in", "ast", ".", "iter_child_nodes", "(", "node", ")", ":", "# Ignore nested functions and its subtrees.", "if", "not", "isinstance", "(", "child", ",", "ast", ".", "FunctionDef", ")", ":", "child_returns", "=", "get_returns_not_on_nested_functions", "(", "child", ")", "returns", ".", "extend", "(", "child_returns", ")", "return", "returns", "tree", "=", "ast", ".", "parse", "(", "self", ".", "method_source", ")", ".", "body", "if", "tree", ":", "returns", "=", "get_returns_not_on_nested_functions", "(", "tree", "[", "0", "]", ")", "return_values", "=", "[", "r", ".", "value", "for", "r", "in", "returns", "]", "# Replace NameConstant nodes valued None for None.", "for", "i", ",", "v", "in", "enumerate", "(", "return_values", ")", ":", "if", "isinstance", "(", "v", ",", "ast", ".", "NameConstant", ")", "and", "v", ".", "value", "is", "None", ":", "return_values", "[", "i", "]", "=", "None", "return", "any", "(", "return_values", ")", "else", ":", "return", "False" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ExcelWriter._value_with_fmt
Convert numpy types to Python types for the Excel writers. Parameters ---------- val : object Value to be written into cells Returns ------- Tuple with the first element being the converted value and the second being an optional format
pandas/io/excel/_base.py
def _value_with_fmt(self, val): """Convert numpy types to Python types for the Excel writers. Parameters ---------- val : object Value to be written into cells Returns ------- Tuple with the first element being the converted value and the second being an optional format """ fmt = None if is_integer(val): val = int(val) elif is_float(val): val = float(val) elif is_bool(val): val = bool(val) elif isinstance(val, datetime): fmt = self.datetime_format elif isinstance(val, date): fmt = self.date_format elif isinstance(val, timedelta): val = val.total_seconds() / float(86400) fmt = '0' else: val = compat.to_str(val) return val, fmt
def _value_with_fmt(self, val): """Convert numpy types to Python types for the Excel writers. Parameters ---------- val : object Value to be written into cells Returns ------- Tuple with the first element being the converted value and the second being an optional format """ fmt = None if is_integer(val): val = int(val) elif is_float(val): val = float(val) elif is_bool(val): val = bool(val) elif isinstance(val, datetime): fmt = self.datetime_format elif isinstance(val, date): fmt = self.date_format elif isinstance(val, timedelta): val = val.total_seconds() / float(86400) fmt = '0' else: val = compat.to_str(val) return val, fmt
[ "Convert", "numpy", "types", "to", "Python", "types", "for", "the", "Excel", "writers", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/excel/_base.py#L675-L706
[ "def", "_value_with_fmt", "(", "self", ",", "val", ")", ":", "fmt", "=", "None", "if", "is_integer", "(", "val", ")", ":", "val", "=", "int", "(", "val", ")", "elif", "is_float", "(", "val", ")", ":", "val", "=", "float", "(", "val", ")", "elif", "is_bool", "(", "val", ")", ":", "val", "=", "bool", "(", "val", ")", "elif", "isinstance", "(", "val", ",", "datetime", ")", ":", "fmt", "=", "self", ".", "datetime_format", "elif", "isinstance", "(", "val", ",", "date", ")", ":", "fmt", "=", "self", ".", "date_format", "elif", "isinstance", "(", "val", ",", "timedelta", ")", ":", "val", "=", "val", ".", "total_seconds", "(", ")", "/", "float", "(", "86400", ")", "fmt", "=", "'0'", "else", ":", "val", "=", "compat", ".", "to_str", "(", "val", ")", "return", "val", ",", "fmt" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ExcelWriter.check_extension
checks that path's extension against the Writer's supported extensions. If it isn't supported, raises UnsupportedFiletypeError.
pandas/io/excel/_base.py
def check_extension(cls, ext): """checks that path's extension against the Writer's supported extensions. If it isn't supported, raises UnsupportedFiletypeError.""" if ext.startswith('.'): ext = ext[1:] if not any(ext in extension for extension in cls.supported_extensions): msg = ("Invalid extension for engine '{engine}': '{ext}'" .format(engine=pprint_thing(cls.engine), ext=pprint_thing(ext))) raise ValueError(msg) else: return True
def check_extension(cls, ext): """checks that path's extension against the Writer's supported extensions. If it isn't supported, raises UnsupportedFiletypeError.""" if ext.startswith('.'): ext = ext[1:] if not any(ext in extension for extension in cls.supported_extensions): msg = ("Invalid extension for engine '{engine}': '{ext}'" .format(engine=pprint_thing(cls.engine), ext=pprint_thing(ext))) raise ValueError(msg) else: return True
[ "checks", "that", "path", "s", "extension", "against", "the", "Writer", "s", "supported", "extensions", ".", "If", "it", "isn", "t", "supported", "raises", "UnsupportedFiletypeError", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/excel/_base.py#L709-L720
[ "def", "check_extension", "(", "cls", ",", "ext", ")", ":", "if", "ext", ".", "startswith", "(", "'.'", ")", ":", "ext", "=", "ext", "[", "1", ":", "]", "if", "not", "any", "(", "ext", "in", "extension", "for", "extension", "in", "cls", ".", "supported_extensions", ")", ":", "msg", "=", "(", "\"Invalid extension for engine '{engine}': '{ext}'\"", ".", "format", "(", "engine", "=", "pprint_thing", "(", "cls", ".", "engine", ")", ",", "ext", "=", "pprint_thing", "(", "ext", ")", ")", ")", "raise", "ValueError", "(", "msg", ")", "else", ":", "return", "True" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ExcelFile.parse
Parse specified sheet(s) into a DataFrame Equivalent to read_excel(ExcelFile, ...) See the read_excel docstring for more info on accepted parameters
pandas/io/excel/_base.py
def parse(self, sheet_name=0, header=0, names=None, index_col=None, usecols=None, squeeze=False, converters=None, true_values=None, false_values=None, skiprows=None, nrows=None, na_values=None, parse_dates=False, date_parser=None, thousands=None, comment=None, skipfooter=0, convert_float=True, mangle_dupe_cols=True, **kwds): """ Parse specified sheet(s) into a DataFrame Equivalent to read_excel(ExcelFile, ...) See the read_excel docstring for more info on accepted parameters """ # Can't use _deprecate_kwarg since sheetname=None has a special meaning if is_integer(sheet_name) and sheet_name == 0 and 'sheetname' in kwds: warnings.warn("The `sheetname` keyword is deprecated, use " "`sheet_name` instead", FutureWarning, stacklevel=2) sheet_name = kwds.pop("sheetname") elif 'sheetname' in kwds: raise TypeError("Cannot specify both `sheet_name` " "and `sheetname`. Use just `sheet_name`") if 'chunksize' in kwds: raise NotImplementedError("chunksize keyword of read_excel " "is not implemented") return self._reader.parse(sheet_name=sheet_name, header=header, names=names, index_col=index_col, usecols=usecols, squeeze=squeeze, converters=converters, true_values=true_values, false_values=false_values, skiprows=skiprows, nrows=nrows, na_values=na_values, parse_dates=parse_dates, date_parser=date_parser, thousands=thousands, comment=comment, skipfooter=skipfooter, convert_float=convert_float, mangle_dupe_cols=mangle_dupe_cols, **kwds)
def parse(self, sheet_name=0, header=0, names=None, index_col=None, usecols=None, squeeze=False, converters=None, true_values=None, false_values=None, skiprows=None, nrows=None, na_values=None, parse_dates=False, date_parser=None, thousands=None, comment=None, skipfooter=0, convert_float=True, mangle_dupe_cols=True, **kwds): """ Parse specified sheet(s) into a DataFrame Equivalent to read_excel(ExcelFile, ...) See the read_excel docstring for more info on accepted parameters """ # Can't use _deprecate_kwarg since sheetname=None has a special meaning if is_integer(sheet_name) and sheet_name == 0 and 'sheetname' in kwds: warnings.warn("The `sheetname` keyword is deprecated, use " "`sheet_name` instead", FutureWarning, stacklevel=2) sheet_name = kwds.pop("sheetname") elif 'sheetname' in kwds: raise TypeError("Cannot specify both `sheet_name` " "and `sheetname`. Use just `sheet_name`") if 'chunksize' in kwds: raise NotImplementedError("chunksize keyword of read_excel " "is not implemented") return self._reader.parse(sheet_name=sheet_name, header=header, names=names, index_col=index_col, usecols=usecols, squeeze=squeeze, converters=converters, true_values=true_values, false_values=false_values, skiprows=skiprows, nrows=nrows, na_values=na_values, parse_dates=parse_dates, date_parser=date_parser, thousands=thousands, comment=comment, skipfooter=skipfooter, convert_float=convert_float, mangle_dupe_cols=mangle_dupe_cols, **kwds)
[ "Parse", "specified", "sheet", "(", "s", ")", "into", "a", "DataFrame" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/excel/_base.py#L771-L831
[ "def", "parse", "(", "self", ",", "sheet_name", "=", "0", ",", "header", "=", "0", ",", "names", "=", "None", ",", "index_col", "=", "None", ",", "usecols", "=", "None", ",", "squeeze", "=", "False", ",", "converters", "=", "None", ",", "true_values", "=", "None", ",", "false_values", "=", "None", ",", "skiprows", "=", "None", ",", "nrows", "=", "None", ",", "na_values", "=", "None", ",", "parse_dates", "=", "False", ",", "date_parser", "=", "None", ",", "thousands", "=", "None", ",", "comment", "=", "None", ",", "skipfooter", "=", "0", ",", "convert_float", "=", "True", ",", "mangle_dupe_cols", "=", "True", ",", "*", "*", "kwds", ")", ":", "# Can't use _deprecate_kwarg since sheetname=None has a special meaning", "if", "is_integer", "(", "sheet_name", ")", "and", "sheet_name", "==", "0", "and", "'sheetname'", "in", "kwds", ":", "warnings", ".", "warn", "(", "\"The `sheetname` keyword is deprecated, use \"", "\"`sheet_name` instead\"", ",", "FutureWarning", ",", "stacklevel", "=", "2", ")", "sheet_name", "=", "kwds", ".", "pop", "(", "\"sheetname\"", ")", "elif", "'sheetname'", "in", "kwds", ":", "raise", "TypeError", "(", "\"Cannot specify both `sheet_name` \"", "\"and `sheetname`. Use just `sheet_name`\"", ")", "if", "'chunksize'", "in", "kwds", ":", "raise", "NotImplementedError", "(", "\"chunksize keyword of read_excel \"", "\"is not implemented\"", ")", "return", "self", ".", "_reader", ".", "parse", "(", "sheet_name", "=", "sheet_name", ",", "header", "=", "header", ",", "names", "=", "names", ",", "index_col", "=", "index_col", ",", "usecols", "=", "usecols", ",", "squeeze", "=", "squeeze", ",", "converters", "=", "converters", ",", "true_values", "=", "true_values", ",", "false_values", "=", "false_values", ",", "skiprows", "=", "skiprows", ",", "nrows", "=", "nrows", ",", "na_values", "=", "na_values", ",", "parse_dates", "=", "parse_dates", ",", "date_parser", "=", "date_parser", ",", "thousands", "=", "thousands", ",", "comment", "=", "comment", ",", "skipfooter", "=", "skipfooter", ",", "convert_float", "=", "convert_float", ",", "mangle_dupe_cols", "=", "mangle_dupe_cols", ",", "*", "*", "kwds", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_validate_where
Validate that the where statement is of the right type. The type may either be String, Expr, or list-like of Exprs. Parameters ---------- w : String term expression, Expr, or list-like of Exprs. Returns ------- where : The original where clause if the check was successful. Raises ------ TypeError : An invalid data type was passed in for w (e.g. dict).
pandas/core/computation/pytables.py
def _validate_where(w): """ Validate that the where statement is of the right type. The type may either be String, Expr, or list-like of Exprs. Parameters ---------- w : String term expression, Expr, or list-like of Exprs. Returns ------- where : The original where clause if the check was successful. Raises ------ TypeError : An invalid data type was passed in for w (e.g. dict). """ if not (isinstance(w, (Expr, str)) or is_list_like(w)): raise TypeError("where must be passed as a string, Expr, " "or list-like of Exprs") return w
def _validate_where(w): """ Validate that the where statement is of the right type. The type may either be String, Expr, or list-like of Exprs. Parameters ---------- w : String term expression, Expr, or list-like of Exprs. Returns ------- where : The original where clause if the check was successful. Raises ------ TypeError : An invalid data type was passed in for w (e.g. dict). """ if not (isinstance(w, (Expr, str)) or is_list_like(w)): raise TypeError("where must be passed as a string, Expr, " "or list-like of Exprs") return w
[ "Validate", "that", "the", "where", "statement", "is", "of", "the", "right", "type", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L460-L483
[ "def", "_validate_where", "(", "w", ")", ":", "if", "not", "(", "isinstance", "(", "w", ",", "(", "Expr", ",", "str", ")", ")", "or", "is_list_like", "(", "w", ")", ")", ":", "raise", "TypeError", "(", "\"where must be passed as a string, Expr, \"", "\"or list-like of Exprs\"", ")", "return", "w" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
maybe_expression
loose checking if s is a pytables-acceptable expression
pandas/core/computation/pytables.py
def maybe_expression(s): """ loose checking if s is a pytables-acceptable expression """ if not isinstance(s, str): return False ops = ExprVisitor.binary_ops + ExprVisitor.unary_ops + ('=',) # make sure we have an op at least return any(op in s for op in ops)
def maybe_expression(s): """ loose checking if s is a pytables-acceptable expression """ if not isinstance(s, str): return False ops = ExprVisitor.binary_ops + ExprVisitor.unary_ops + ('=',) # make sure we have an op at least return any(op in s for op in ops)
[ "loose", "checking", "if", "s", "is", "a", "pytables", "-", "acceptable", "expression" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L598-L605
[ "def", "maybe_expression", "(", "s", ")", ":", "if", "not", "isinstance", "(", "s", ",", "str", ")", ":", "return", "False", "ops", "=", "ExprVisitor", ".", "binary_ops", "+", "ExprVisitor", ".", "unary_ops", "+", "(", "'='", ",", ")", "# make sure we have an op at least", "return", "any", "(", "op", "in", "s", "for", "op", "in", "ops", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
BinOp.conform
inplace conform rhs
pandas/core/computation/pytables.py
def conform(self, rhs): """ inplace conform rhs """ if not is_list_like(rhs): rhs = [rhs] if isinstance(rhs, np.ndarray): rhs = rhs.ravel() return rhs
def conform(self, rhs): """ inplace conform rhs """ if not is_list_like(rhs): rhs = [rhs] if isinstance(rhs, np.ndarray): rhs = rhs.ravel() return rhs
[ "inplace", "conform", "rhs" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L132-L138
[ "def", "conform", "(", "self", ",", "rhs", ")", ":", "if", "not", "is_list_like", "(", "rhs", ")", ":", "rhs", "=", "[", "rhs", "]", "if", "isinstance", "(", "rhs", ",", "np", ".", "ndarray", ")", ":", "rhs", "=", "rhs", ".", "ravel", "(", ")", "return", "rhs" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
BinOp.generate
create and return the op string for this TermValue
pandas/core/computation/pytables.py
def generate(self, v): """ create and return the op string for this TermValue """ val = v.tostring(self.encoding) return "({lhs} {op} {val})".format(lhs=self.lhs, op=self.op, val=val)
def generate(self, v): """ create and return the op string for this TermValue """ val = v.tostring(self.encoding) return "({lhs} {op} {val})".format(lhs=self.lhs, op=self.op, val=val)
[ "create", "and", "return", "the", "op", "string", "for", "this", "TermValue" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L166-L169
[ "def", "generate", "(", "self", ",", "v", ")", ":", "val", "=", "v", ".", "tostring", "(", "self", ".", "encoding", ")", "return", "\"({lhs} {op} {val})\"", ".", "format", "(", "lhs", "=", "self", ".", "lhs", ",", "op", "=", "self", ".", "op", ",", "val", "=", "val", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
BinOp.convert_value
convert the expression that is in the term to something that is accepted by pytables
pandas/core/computation/pytables.py
def convert_value(self, v): """ convert the expression that is in the term to something that is accepted by pytables """ def stringify(value): if self.encoding is not None: encoder = partial(pprint_thing_encoded, encoding=self.encoding) else: encoder = pprint_thing return encoder(value) kind = _ensure_decoded(self.kind) meta = _ensure_decoded(self.meta) if kind == 'datetime64' or kind == 'datetime': if isinstance(v, (int, float)): v = stringify(v) v = _ensure_decoded(v) v = Timestamp(v) if v.tz is not None: v = v.tz_convert('UTC') return TermValue(v, v.value, kind) elif kind == 'timedelta64' or kind == 'timedelta': v = Timedelta(v, unit='s').value return TermValue(int(v), v, kind) elif meta == 'category': metadata = com.values_from_object(self.metadata) result = metadata.searchsorted(v, side='left') # result returns 0 if v is first element or if v is not in metadata # check that metadata contains v if not result and v not in metadata: result = -1 return TermValue(result, result, 'integer') elif kind == 'integer': v = int(float(v)) return TermValue(v, v, kind) elif kind == 'float': v = float(v) return TermValue(v, v, kind) elif kind == 'bool': if isinstance(v, str): v = not v.strip().lower() in ['false', 'f', 'no', 'n', 'none', '0', '[]', '{}', ''] else: v = bool(v) return TermValue(v, v, kind) elif isinstance(v, str): # string quoting return TermValue(v, stringify(v), 'string') else: raise TypeError("Cannot compare {v} of type {typ} to {kind} column" .format(v=v, typ=type(v), kind=kind))
def convert_value(self, v): """ convert the expression that is in the term to something that is accepted by pytables """ def stringify(value): if self.encoding is not None: encoder = partial(pprint_thing_encoded, encoding=self.encoding) else: encoder = pprint_thing return encoder(value) kind = _ensure_decoded(self.kind) meta = _ensure_decoded(self.meta) if kind == 'datetime64' or kind == 'datetime': if isinstance(v, (int, float)): v = stringify(v) v = _ensure_decoded(v) v = Timestamp(v) if v.tz is not None: v = v.tz_convert('UTC') return TermValue(v, v.value, kind) elif kind == 'timedelta64' or kind == 'timedelta': v = Timedelta(v, unit='s').value return TermValue(int(v), v, kind) elif meta == 'category': metadata = com.values_from_object(self.metadata) result = metadata.searchsorted(v, side='left') # result returns 0 if v is first element or if v is not in metadata # check that metadata contains v if not result and v not in metadata: result = -1 return TermValue(result, result, 'integer') elif kind == 'integer': v = int(float(v)) return TermValue(v, v, kind) elif kind == 'float': v = float(v) return TermValue(v, v, kind) elif kind == 'bool': if isinstance(v, str): v = not v.strip().lower() in ['false', 'f', 'no', 'n', 'none', '0', '[]', '{}', ''] else: v = bool(v) return TermValue(v, v, kind) elif isinstance(v, str): # string quoting return TermValue(v, stringify(v), 'string') else: raise TypeError("Cannot compare {v} of type {typ} to {kind} column" .format(v=v, typ=type(v), kind=kind))
[ "convert", "the", "expression", "that", "is", "in", "the", "term", "to", "something", "that", "is", "accepted", "by", "pytables" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L171-L224
[ "def", "convert_value", "(", "self", ",", "v", ")", ":", "def", "stringify", "(", "value", ")", ":", "if", "self", ".", "encoding", "is", "not", "None", ":", "encoder", "=", "partial", "(", "pprint_thing_encoded", ",", "encoding", "=", "self", ".", "encoding", ")", "else", ":", "encoder", "=", "pprint_thing", "return", "encoder", "(", "value", ")", "kind", "=", "_ensure_decoded", "(", "self", ".", "kind", ")", "meta", "=", "_ensure_decoded", "(", "self", ".", "meta", ")", "if", "kind", "==", "'datetime64'", "or", "kind", "==", "'datetime'", ":", "if", "isinstance", "(", "v", ",", "(", "int", ",", "float", ")", ")", ":", "v", "=", "stringify", "(", "v", ")", "v", "=", "_ensure_decoded", "(", "v", ")", "v", "=", "Timestamp", "(", "v", ")", "if", "v", ".", "tz", "is", "not", "None", ":", "v", "=", "v", ".", "tz_convert", "(", "'UTC'", ")", "return", "TermValue", "(", "v", ",", "v", ".", "value", ",", "kind", ")", "elif", "kind", "==", "'timedelta64'", "or", "kind", "==", "'timedelta'", ":", "v", "=", "Timedelta", "(", "v", ",", "unit", "=", "'s'", ")", ".", "value", "return", "TermValue", "(", "int", "(", "v", ")", ",", "v", ",", "kind", ")", "elif", "meta", "==", "'category'", ":", "metadata", "=", "com", ".", "values_from_object", "(", "self", ".", "metadata", ")", "result", "=", "metadata", ".", "searchsorted", "(", "v", ",", "side", "=", "'left'", ")", "# result returns 0 if v is first element or if v is not in metadata", "# check that metadata contains v", "if", "not", "result", "and", "v", "not", "in", "metadata", ":", "result", "=", "-", "1", "return", "TermValue", "(", "result", ",", "result", ",", "'integer'", ")", "elif", "kind", "==", "'integer'", ":", "v", "=", "int", "(", "float", "(", "v", ")", ")", "return", "TermValue", "(", "v", ",", "v", ",", "kind", ")", "elif", "kind", "==", "'float'", ":", "v", "=", "float", "(", "v", ")", "return", "TermValue", "(", "v", ",", "v", ",", "kind", ")", "elif", "kind", "==", "'bool'", ":", "if", "isinstance", "(", "v", ",", "str", ")", ":", "v", "=", "not", "v", ".", "strip", "(", ")", ".", "lower", "(", ")", "in", "[", "'false'", ",", "'f'", ",", "'no'", ",", "'n'", ",", "'none'", ",", "'0'", ",", "'[]'", ",", "'{}'", ",", "''", "]", "else", ":", "v", "=", "bool", "(", "v", ")", "return", "TermValue", "(", "v", ",", "v", ",", "kind", ")", "elif", "isinstance", "(", "v", ",", "str", ")", ":", "# string quoting", "return", "TermValue", "(", "v", ",", "stringify", "(", "v", ")", ",", "'string'", ")", "else", ":", "raise", "TypeError", "(", "\"Cannot compare {v} of type {typ} to {kind} column\"", ".", "format", "(", "v", "=", "v", ",", "typ", "=", "type", "(", "v", ")", ",", "kind", "=", "kind", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
FilterBinOp.invert
invert the filter
pandas/core/computation/pytables.py
def invert(self): """ invert the filter """ if self.filter is not None: f = list(self.filter) f[1] = self.generate_filter_op(invert=True) self.filter = tuple(f) return self
def invert(self): """ invert the filter """ if self.filter is not None: f = list(self.filter) f[1] = self.generate_filter_op(invert=True) self.filter = tuple(f) return self
[ "invert", "the", "filter" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L236-L242
[ "def", "invert", "(", "self", ")", ":", "if", "self", ".", "filter", "is", "not", "None", ":", "f", "=", "list", "(", "self", ".", "filter", ")", "f", "[", "1", "]", "=", "self", ".", "generate_filter_op", "(", "invert", "=", "True", ")", "self", ".", "filter", "=", "tuple", "(", "f", ")", "return", "self" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
Expr.evaluate
create and return the numexpr condition and filter
pandas/core/computation/pytables.py
def evaluate(self): """ create and return the numexpr condition and filter """ try: self.condition = self.terms.prune(ConditionBinOp) except AttributeError: raise ValueError("cannot process expression [{expr}], [{slf}] " "is not a valid condition".format(expr=self.expr, slf=self)) try: self.filter = self.terms.prune(FilterBinOp) except AttributeError: raise ValueError("cannot process expression [{expr}], [{slf}] " "is not a valid filter".format(expr=self.expr, slf=self)) return self.condition, self.filter
def evaluate(self): """ create and return the numexpr condition and filter """ try: self.condition = self.terms.prune(ConditionBinOp) except AttributeError: raise ValueError("cannot process expression [{expr}], [{slf}] " "is not a valid condition".format(expr=self.expr, slf=self)) try: self.filter = self.terms.prune(FilterBinOp) except AttributeError: raise ValueError("cannot process expression [{expr}], [{slf}] " "is not a valid filter".format(expr=self.expr, slf=self)) return self.condition, self.filter
[ "create", "and", "return", "the", "numexpr", "condition", "and", "filter" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L556-L572
[ "def", "evaluate", "(", "self", ")", ":", "try", ":", "self", ".", "condition", "=", "self", ".", "terms", ".", "prune", "(", "ConditionBinOp", ")", "except", "AttributeError", ":", "raise", "ValueError", "(", "\"cannot process expression [{expr}], [{slf}] \"", "\"is not a valid condition\"", ".", "format", "(", "expr", "=", "self", ".", "expr", ",", "slf", "=", "self", ")", ")", "try", ":", "self", ".", "filter", "=", "self", ".", "terms", ".", "prune", "(", "FilterBinOp", ")", "except", "AttributeError", ":", "raise", "ValueError", "(", "\"cannot process expression [{expr}], [{slf}] \"", "\"is not a valid filter\"", ".", "format", "(", "expr", "=", "self", ".", "expr", ",", "slf", "=", "self", ")", ")", "return", "self", ".", "condition", ",", "self", ".", "filter" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
TermValue.tostring
quote the string if not encoded else encode and return
pandas/core/computation/pytables.py
def tostring(self, encoding): """ quote the string if not encoded else encode and return """ if self.kind == 'string': if encoding is not None: return self.converted return '"{converted}"'.format(converted=self.converted) elif self.kind == 'float': # python 2 str(float) is not always # round-trippable so use repr() return repr(self.converted) return self.converted
def tostring(self, encoding): """ quote the string if not encoded else encode and return """ if self.kind == 'string': if encoding is not None: return self.converted return '"{converted}"'.format(converted=self.converted) elif self.kind == 'float': # python 2 str(float) is not always # round-trippable so use repr() return repr(self.converted) return self.converted
[ "quote", "the", "string", "if", "not", "encoded", "else", "encode", "and", "return" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/pytables.py#L584-L595
[ "def", "tostring", "(", "self", ",", "encoding", ")", ":", "if", "self", ".", "kind", "==", "'string'", ":", "if", "encoding", "is", "not", "None", ":", "return", "self", ".", "converted", "return", "'\"{converted}\"'", ".", "format", "(", "converted", "=", "self", ".", "converted", ")", "elif", "self", ".", "kind", "==", "'float'", ":", "# python 2 str(float) is not always", "# round-trippable so use repr()", "return", "repr", "(", "self", ".", "converted", ")", "return", "self", ".", "converted" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_ensure_decoded
if we have bytes, decode them to unicode
pandas/core/computation/common.py
def _ensure_decoded(s): """ if we have bytes, decode them to unicode """ if isinstance(s, (np.bytes_, bytes)): s = s.decode(pd.get_option('display.encoding')) return s
def _ensure_decoded(s): """ if we have bytes, decode them to unicode """ if isinstance(s, (np.bytes_, bytes)): s = s.decode(pd.get_option('display.encoding')) return s
[ "if", "we", "have", "bytes", "decode", "them", "to", "unicode" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/common.py#L11-L15
[ "def", "_ensure_decoded", "(", "s", ")", ":", "if", "isinstance", "(", "s", ",", "(", "np", ".", "bytes_", ",", "bytes", ")", ")", ":", "s", "=", "s", ".", "decode", "(", "pd", ".", "get_option", "(", "'display.encoding'", ")", ")", "return", "s" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_result_type_many
wrapper around numpy.result_type which overcomes the NPY_MAXARGS (32) argument limit
pandas/core/computation/common.py
def _result_type_many(*arrays_and_dtypes): """ wrapper around numpy.result_type which overcomes the NPY_MAXARGS (32) argument limit """ try: return np.result_type(*arrays_and_dtypes) except ValueError: # we have > NPY_MAXARGS terms in our expression return reduce(np.result_type, arrays_and_dtypes)
def _result_type_many(*arrays_and_dtypes): """ wrapper around numpy.result_type which overcomes the NPY_MAXARGS (32) argument limit """ try: return np.result_type(*arrays_and_dtypes) except ValueError: # we have > NPY_MAXARGS terms in our expression return reduce(np.result_type, arrays_and_dtypes)
[ "wrapper", "around", "numpy", ".", "result_type", "which", "overcomes", "the", "NPY_MAXARGS", "(", "32", ")", "argument", "limit" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/common.py#L18-L25
[ "def", "_result_type_many", "(", "*", "arrays_and_dtypes", ")", ":", "try", ":", "return", "np", ".", "result_type", "(", "*", "arrays_and_dtypes", ")", "except", "ValueError", ":", "# we have > NPY_MAXARGS terms in our expression", "return", "reduce", "(", "np", ".", "result_type", ",", "arrays_and_dtypes", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_argmin_with_skipna
If 'Series.argmin' is called via the 'numpy' library, the third parameter in its signature is 'out', which takes either an ndarray or 'None', so check if the 'skipna' parameter is either an instance of ndarray or is None, since 'skipna' itself should be a boolean
pandas/compat/numpy/function.py
def validate_argmin_with_skipna(skipna, args, kwargs): """ If 'Series.argmin' is called via the 'numpy' library, the third parameter in its signature is 'out', which takes either an ndarray or 'None', so check if the 'skipna' parameter is either an instance of ndarray or is None, since 'skipna' itself should be a boolean """ skipna, args = process_skipna(skipna, args) validate_argmin(args, kwargs) return skipna
def validate_argmin_with_skipna(skipna, args, kwargs): """ If 'Series.argmin' is called via the 'numpy' library, the third parameter in its signature is 'out', which takes either an ndarray or 'None', so check if the 'skipna' parameter is either an instance of ndarray or is None, since 'skipna' itself should be a boolean """ skipna, args = process_skipna(skipna, args) validate_argmin(args, kwargs) return skipna
[ "If", "Series", ".", "argmin", "is", "called", "via", "the", "numpy", "library", "the", "third", "parameter", "in", "its", "signature", "is", "out", "which", "takes", "either", "an", "ndarray", "or", "None", "so", "check", "if", "the", "skipna", "parameter", "is", "either", "an", "instance", "of", "ndarray", "or", "is", "None", "since", "skipna", "itself", "should", "be", "a", "boolean" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L77-L88
[ "def", "validate_argmin_with_skipna", "(", "skipna", ",", "args", ",", "kwargs", ")", ":", "skipna", ",", "args", "=", "process_skipna", "(", "skipna", ",", "args", ")", "validate_argmin", "(", "args", ",", "kwargs", ")", "return", "skipna" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_argmax_with_skipna
If 'Series.argmax' is called via the 'numpy' library, the third parameter in its signature is 'out', which takes either an ndarray or 'None', so check if the 'skipna' parameter is either an instance of ndarray or is None, since 'skipna' itself should be a boolean
pandas/compat/numpy/function.py
def validate_argmax_with_skipna(skipna, args, kwargs): """ If 'Series.argmax' is called via the 'numpy' library, the third parameter in its signature is 'out', which takes either an ndarray or 'None', so check if the 'skipna' parameter is either an instance of ndarray or is None, since 'skipna' itself should be a boolean """ skipna, args = process_skipna(skipna, args) validate_argmax(args, kwargs) return skipna
def validate_argmax_with_skipna(skipna, args, kwargs): """ If 'Series.argmax' is called via the 'numpy' library, the third parameter in its signature is 'out', which takes either an ndarray or 'None', so check if the 'skipna' parameter is either an instance of ndarray or is None, since 'skipna' itself should be a boolean """ skipna, args = process_skipna(skipna, args) validate_argmax(args, kwargs) return skipna
[ "If", "Series", ".", "argmax", "is", "called", "via", "the", "numpy", "library", "the", "third", "parameter", "in", "its", "signature", "is", "out", "which", "takes", "either", "an", "ndarray", "or", "None", "so", "check", "if", "the", "skipna", "parameter", "is", "either", "an", "instance", "of", "ndarray", "or", "is", "None", "since", "skipna", "itself", "should", "be", "a", "boolean" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L91-L102
[ "def", "validate_argmax_with_skipna", "(", "skipna", ",", "args", ",", "kwargs", ")", ":", "skipna", ",", "args", "=", "process_skipna", "(", "skipna", ",", "args", ")", "validate_argmax", "(", "args", ",", "kwargs", ")", "return", "skipna" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_argsort_with_ascending
If 'Categorical.argsort' is called via the 'numpy' library, the first parameter in its signature is 'axis', which takes either an integer or 'None', so check if the 'ascending' parameter has either integer type or is None, since 'ascending' itself should be a boolean
pandas/compat/numpy/function.py
def validate_argsort_with_ascending(ascending, args, kwargs): """ If 'Categorical.argsort' is called via the 'numpy' library, the first parameter in its signature is 'axis', which takes either an integer or 'None', so check if the 'ascending' parameter has either integer type or is None, since 'ascending' itself should be a boolean """ if is_integer(ascending) or ascending is None: args = (ascending,) + args ascending = True validate_argsort_kind(args, kwargs, max_fname_arg_count=3) return ascending
def validate_argsort_with_ascending(ascending, args, kwargs): """ If 'Categorical.argsort' is called via the 'numpy' library, the first parameter in its signature is 'axis', which takes either an integer or 'None', so check if the 'ascending' parameter has either integer type or is None, since 'ascending' itself should be a boolean """ if is_integer(ascending) or ascending is None: args = (ascending,) + args ascending = True validate_argsort_kind(args, kwargs, max_fname_arg_count=3) return ascending
[ "If", "Categorical", ".", "argsort", "is", "called", "via", "the", "numpy", "library", "the", "first", "parameter", "in", "its", "signature", "is", "axis", "which", "takes", "either", "an", "integer", "or", "None", "so", "check", "if", "the", "ascending", "parameter", "has", "either", "integer", "type", "or", "is", "None", "since", "ascending", "itself", "should", "be", "a", "boolean" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L123-L137
[ "def", "validate_argsort_with_ascending", "(", "ascending", ",", "args", ",", "kwargs", ")", ":", "if", "is_integer", "(", "ascending", ")", "or", "ascending", "is", "None", ":", "args", "=", "(", "ascending", ",", ")", "+", "args", "ascending", "=", "True", "validate_argsort_kind", "(", "args", ",", "kwargs", ",", "max_fname_arg_count", "=", "3", ")", "return", "ascending" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_clip_with_axis
If 'NDFrame.clip' is called via the numpy library, the third parameter in its signature is 'out', which can takes an ndarray, so check if the 'axis' parameter is an instance of ndarray, since 'axis' itself should either be an integer or None
pandas/compat/numpy/function.py
def validate_clip_with_axis(axis, args, kwargs): """ If 'NDFrame.clip' is called via the numpy library, the third parameter in its signature is 'out', which can takes an ndarray, so check if the 'axis' parameter is an instance of ndarray, since 'axis' itself should either be an integer or None """ if isinstance(axis, ndarray): args = (axis,) + args axis = None validate_clip(args, kwargs) return axis
def validate_clip_with_axis(axis, args, kwargs): """ If 'NDFrame.clip' is called via the numpy library, the third parameter in its signature is 'out', which can takes an ndarray, so check if the 'axis' parameter is an instance of ndarray, since 'axis' itself should either be an integer or None """ if isinstance(axis, ndarray): args = (axis,) + args axis = None validate_clip(args, kwargs) return axis
[ "If", "NDFrame", ".", "clip", "is", "called", "via", "the", "numpy", "library", "the", "third", "parameter", "in", "its", "signature", "is", "out", "which", "can", "takes", "an", "ndarray", "so", "check", "if", "the", "axis", "parameter", "is", "an", "instance", "of", "ndarray", "since", "axis", "itself", "should", "either", "be", "an", "integer", "or", "None" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L145-L158
[ "def", "validate_clip_with_axis", "(", "axis", ",", "args", ",", "kwargs", ")", ":", "if", "isinstance", "(", "axis", ",", "ndarray", ")", ":", "args", "=", "(", "axis", ",", ")", "+", "args", "axis", "=", "None", "validate_clip", "(", "args", ",", "kwargs", ")", "return", "axis" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_cum_func_with_skipna
If this function is called via the 'numpy' library, the third parameter in its signature is 'dtype', which takes either a 'numpy' dtype or 'None', so check if the 'skipna' parameter is a boolean or not
pandas/compat/numpy/function.py
def validate_cum_func_with_skipna(skipna, args, kwargs, name): """ If this function is called via the 'numpy' library, the third parameter in its signature is 'dtype', which takes either a 'numpy' dtype or 'None', so check if the 'skipna' parameter is a boolean or not """ if not is_bool(skipna): args = (skipna,) + args skipna = True validate_cum_func(args, kwargs, fname=name) return skipna
def validate_cum_func_with_skipna(skipna, args, kwargs, name): """ If this function is called via the 'numpy' library, the third parameter in its signature is 'dtype', which takes either a 'numpy' dtype or 'None', so check if the 'skipna' parameter is a boolean or not """ if not is_bool(skipna): args = (skipna,) + args skipna = True validate_cum_func(args, kwargs, fname=name) return skipna
[ "If", "this", "function", "is", "called", "via", "the", "numpy", "library", "the", "third", "parameter", "in", "its", "signature", "is", "dtype", "which", "takes", "either", "a", "numpy", "dtype", "or", "None", "so", "check", "if", "the", "skipna", "parameter", "is", "a", "boolean", "or", "not" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L176-L188
[ "def", "validate_cum_func_with_skipna", "(", "skipna", ",", "args", ",", "kwargs", ",", "name", ")", ":", "if", "not", "is_bool", "(", "skipna", ")", ":", "args", "=", "(", "skipna", ",", ")", "+", "args", "skipna", "=", "True", "validate_cum_func", "(", "args", ",", "kwargs", ",", "fname", "=", "name", ")", "return", "skipna" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_take_with_convert
If this function is called via the 'numpy' library, the third parameter in its signature is 'axis', which takes either an ndarray or 'None', so check if the 'convert' parameter is either an instance of ndarray or is None
pandas/compat/numpy/function.py
def validate_take_with_convert(convert, args, kwargs): """ If this function is called via the 'numpy' library, the third parameter in its signature is 'axis', which takes either an ndarray or 'None', so check if the 'convert' parameter is either an instance of ndarray or is None """ if isinstance(convert, ndarray) or convert is None: args = (convert,) + args convert = True validate_take(args, kwargs, max_fname_arg_count=3, method='both') return convert
def validate_take_with_convert(convert, args, kwargs): """ If this function is called via the 'numpy' library, the third parameter in its signature is 'axis', which takes either an ndarray or 'None', so check if the 'convert' parameter is either an instance of ndarray or is None """ if isinstance(convert, ndarray) or convert is None: args = (convert,) + args convert = True validate_take(args, kwargs, max_fname_arg_count=3, method='both') return convert
[ "If", "this", "function", "is", "called", "via", "the", "numpy", "library", "the", "third", "parameter", "in", "its", "signature", "is", "axis", "which", "takes", "either", "an", "ndarray", "or", "None", "so", "check", "if", "the", "convert", "parameter", "is", "either", "an", "instance", "of", "ndarray", "or", "is", "None" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L269-L282
[ "def", "validate_take_with_convert", "(", "convert", ",", "args", ",", "kwargs", ")", ":", "if", "isinstance", "(", "convert", ",", "ndarray", ")", "or", "convert", "is", "None", ":", "args", "=", "(", "convert", ",", ")", "+", "args", "convert", "=", "True", "validate_take", "(", "args", ",", "kwargs", ",", "max_fname_arg_count", "=", "3", ",", "method", "=", "'both'", ")", "return", "convert" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_groupby_func
'args' and 'kwargs' should be empty, except for allowed kwargs because all of their necessary parameters are explicitly listed in the function signature
pandas/compat/numpy/function.py
def validate_groupby_func(name, args, kwargs, allowed=None): """ 'args' and 'kwargs' should be empty, except for allowed kwargs because all of their necessary parameters are explicitly listed in the function signature """ if allowed is None: allowed = [] kwargs = set(kwargs) - set(allowed) if len(args) + len(kwargs) > 0: raise UnsupportedFunctionCall(( "numpy operations are not valid " "with groupby. Use .groupby(...)." "{func}() instead".format(func=name)))
def validate_groupby_func(name, args, kwargs, allowed=None): """ 'args' and 'kwargs' should be empty, except for allowed kwargs because all of their necessary parameters are explicitly listed in the function signature """ if allowed is None: allowed = [] kwargs = set(kwargs) - set(allowed) if len(args) + len(kwargs) > 0: raise UnsupportedFunctionCall(( "numpy operations are not valid " "with groupby. Use .groupby(...)." "{func}() instead".format(func=name)))
[ "args", "and", "kwargs", "should", "be", "empty", "except", "for", "allowed", "kwargs", "because", "all", "of", "their", "necessary", "parameters", "are", "explicitly", "listed", "in", "the", "function", "signature" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L349-L365
[ "def", "validate_groupby_func", "(", "name", ",", "args", ",", "kwargs", ",", "allowed", "=", "None", ")", ":", "if", "allowed", "is", "None", ":", "allowed", "=", "[", "]", "kwargs", "=", "set", "(", "kwargs", ")", "-", "set", "(", "allowed", ")", "if", "len", "(", "args", ")", "+", "len", "(", "kwargs", ")", ">", "0", ":", "raise", "UnsupportedFunctionCall", "(", "(", "\"numpy operations are not valid \"", "\"with groupby. Use .groupby(...).\"", "\"{func}() instead\"", ".", "format", "(", "func", "=", "name", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_resampler_func
'args' and 'kwargs' should be empty because all of their necessary parameters are explicitly listed in the function signature
pandas/compat/numpy/function.py
def validate_resampler_func(method, args, kwargs): """ 'args' and 'kwargs' should be empty because all of their necessary parameters are explicitly listed in the function signature """ if len(args) + len(kwargs) > 0: if method in RESAMPLER_NUMPY_OPS: raise UnsupportedFunctionCall(( "numpy operations are not valid " "with resample. Use .resample(...)." "{func}() instead".format(func=method))) else: raise TypeError("too many arguments passed in")
def validate_resampler_func(method, args, kwargs): """ 'args' and 'kwargs' should be empty because all of their necessary parameters are explicitly listed in the function signature """ if len(args) + len(kwargs) > 0: if method in RESAMPLER_NUMPY_OPS: raise UnsupportedFunctionCall(( "numpy operations are not valid " "with resample. Use .resample(...)." "{func}() instead".format(func=method))) else: raise TypeError("too many arguments passed in")
[ "args", "and", "kwargs", "should", "be", "empty", "because", "all", "of", "their", "necessary", "parameters", "are", "explicitly", "listed", "in", "the", "function", "signature" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L372-L385
[ "def", "validate_resampler_func", "(", "method", ",", "args", ",", "kwargs", ")", ":", "if", "len", "(", "args", ")", "+", "len", "(", "kwargs", ")", ">", "0", ":", "if", "method", "in", "RESAMPLER_NUMPY_OPS", ":", "raise", "UnsupportedFunctionCall", "(", "(", "\"numpy operations are not valid \"", "\"with resample. Use .resample(...).\"", "\"{func}() instead\"", ".", "format", "(", "func", "=", "method", ")", ")", ")", "else", ":", "raise", "TypeError", "(", "\"too many arguments passed in\"", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
validate_minmax_axis
Ensure that the axis argument passed to min, max, argmin, or argmax is zero or None, as otherwise it will be incorrectly ignored. Parameters ---------- axis : int or None Raises ------ ValueError
pandas/compat/numpy/function.py
def validate_minmax_axis(axis): """ Ensure that the axis argument passed to min, max, argmin, or argmax is zero or None, as otherwise it will be incorrectly ignored. Parameters ---------- axis : int or None Raises ------ ValueError """ ndim = 1 # hard-coded for Index if axis is None: return if axis >= ndim or (axis < 0 and ndim + axis < 0): raise ValueError("`axis` must be fewer than the number of " "dimensions ({ndim})".format(ndim=ndim))
def validate_minmax_axis(axis): """ Ensure that the axis argument passed to min, max, argmin, or argmax is zero or None, as otherwise it will be incorrectly ignored. Parameters ---------- axis : int or None Raises ------ ValueError """ ndim = 1 # hard-coded for Index if axis is None: return if axis >= ndim or (axis < 0 and ndim + axis < 0): raise ValueError("`axis` must be fewer than the number of " "dimensions ({ndim})".format(ndim=ndim))
[ "Ensure", "that", "the", "axis", "argument", "passed", "to", "min", "max", "argmin", "or", "argmax", "is", "zero", "or", "None", "as", "otherwise", "it", "will", "be", "incorrectly", "ignored", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/compat/numpy/function.py#L388-L406
[ "def", "validate_minmax_axis", "(", "axis", ")", ":", "ndim", "=", "1", "# hard-coded for Index", "if", "axis", "is", "None", ":", "return", "if", "axis", ">=", "ndim", "or", "(", "axis", "<", "0", "and", "ndim", "+", "axis", "<", "0", ")", ":", "raise", "ValueError", "(", "\"`axis` must be fewer than the number of \"", "\"dimensions ({ndim})\"", ".", "format", "(", "ndim", "=", "ndim", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
to_msgpack
msgpack (serialize) object to input file path THIS IS AN EXPERIMENTAL LIBRARY and the storage format may not be stable until a future release. Parameters ---------- path_or_buf : string File path, buffer-like, or None if None, return generated string args : an object or objects to serialize encoding : encoding for unicode objects append : boolean whether to append to an existing msgpack (default is False) compress : type of compressor (zlib or blosc), default to None (no compression)
pandas/io/packers.py
def to_msgpack(path_or_buf, *args, **kwargs): """ msgpack (serialize) object to input file path THIS IS AN EXPERIMENTAL LIBRARY and the storage format may not be stable until a future release. Parameters ---------- path_or_buf : string File path, buffer-like, or None if None, return generated string args : an object or objects to serialize encoding : encoding for unicode objects append : boolean whether to append to an existing msgpack (default is False) compress : type of compressor (zlib or blosc), default to None (no compression) """ global compressor compressor = kwargs.pop('compress', None) append = kwargs.pop('append', None) if append: mode = 'a+b' else: mode = 'wb' def writer(fh): for a in args: fh.write(pack(a, **kwargs)) path_or_buf = _stringify_path(path_or_buf) if isinstance(path_or_buf, str): with open(path_or_buf, mode) as fh: writer(fh) elif path_or_buf is None: buf = BytesIO() writer(buf) return buf.getvalue() else: writer(path_or_buf)
def to_msgpack(path_or_buf, *args, **kwargs): """ msgpack (serialize) object to input file path THIS IS AN EXPERIMENTAL LIBRARY and the storage format may not be stable until a future release. Parameters ---------- path_or_buf : string File path, buffer-like, or None if None, return generated string args : an object or objects to serialize encoding : encoding for unicode objects append : boolean whether to append to an existing msgpack (default is False) compress : type of compressor (zlib or blosc), default to None (no compression) """ global compressor compressor = kwargs.pop('compress', None) append = kwargs.pop('append', None) if append: mode = 'a+b' else: mode = 'wb' def writer(fh): for a in args: fh.write(pack(a, **kwargs)) path_or_buf = _stringify_path(path_or_buf) if isinstance(path_or_buf, str): with open(path_or_buf, mode) as fh: writer(fh) elif path_or_buf is None: buf = BytesIO() writer(buf) return buf.getvalue() else: writer(path_or_buf)
[ "msgpack", "(", "serialize", ")", "object", "to", "input", "file", "path" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L118-L157
[ "def", "to_msgpack", "(", "path_or_buf", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "global", "compressor", "compressor", "=", "kwargs", ".", "pop", "(", "'compress'", ",", "None", ")", "append", "=", "kwargs", ".", "pop", "(", "'append'", ",", "None", ")", "if", "append", ":", "mode", "=", "'a+b'", "else", ":", "mode", "=", "'wb'", "def", "writer", "(", "fh", ")", ":", "for", "a", "in", "args", ":", "fh", ".", "write", "(", "pack", "(", "a", ",", "*", "*", "kwargs", ")", ")", "path_or_buf", "=", "_stringify_path", "(", "path_or_buf", ")", "if", "isinstance", "(", "path_or_buf", ",", "str", ")", ":", "with", "open", "(", "path_or_buf", ",", "mode", ")", "as", "fh", ":", "writer", "(", "fh", ")", "elif", "path_or_buf", "is", "None", ":", "buf", "=", "BytesIO", "(", ")", "writer", "(", "buf", ")", "return", "buf", ".", "getvalue", "(", ")", "else", ":", "writer", "(", "path_or_buf", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
read_msgpack
Load msgpack pandas object from the specified file path THIS IS AN EXPERIMENTAL LIBRARY and the storage format may not be stable until a future release. Parameters ---------- path_or_buf : string File path, BytesIO like or string encoding : Encoding for decoding msgpack str type iterator : boolean, if True, return an iterator to the unpacker (default is False) Returns ------- obj : same type as object stored in file
pandas/io/packers.py
def read_msgpack(path_or_buf, encoding='utf-8', iterator=False, **kwargs): """ Load msgpack pandas object from the specified file path THIS IS AN EXPERIMENTAL LIBRARY and the storage format may not be stable until a future release. Parameters ---------- path_or_buf : string File path, BytesIO like or string encoding : Encoding for decoding msgpack str type iterator : boolean, if True, return an iterator to the unpacker (default is False) Returns ------- obj : same type as object stored in file """ path_or_buf, _, _, should_close = get_filepath_or_buffer(path_or_buf) if iterator: return Iterator(path_or_buf) def read(fh): unpacked_obj = list(unpack(fh, encoding=encoding, **kwargs)) if len(unpacked_obj) == 1: return unpacked_obj[0] if should_close: try: path_or_buf.close() except IOError: pass return unpacked_obj # see if we have an actual file if isinstance(path_or_buf, str): try: exists = os.path.exists(path_or_buf) except (TypeError, ValueError): exists = False if exists: with open(path_or_buf, 'rb') as fh: return read(fh) if isinstance(path_or_buf, bytes): # treat as a binary-like fh = None try: fh = BytesIO(path_or_buf) return read(fh) finally: if fh is not None: fh.close() elif hasattr(path_or_buf, 'read') and callable(path_or_buf.read): # treat as a buffer like return read(path_or_buf) raise ValueError('path_or_buf needs to be a string file path or file-like')
def read_msgpack(path_or_buf, encoding='utf-8', iterator=False, **kwargs): """ Load msgpack pandas object from the specified file path THIS IS AN EXPERIMENTAL LIBRARY and the storage format may not be stable until a future release. Parameters ---------- path_or_buf : string File path, BytesIO like or string encoding : Encoding for decoding msgpack str type iterator : boolean, if True, return an iterator to the unpacker (default is False) Returns ------- obj : same type as object stored in file """ path_or_buf, _, _, should_close = get_filepath_or_buffer(path_or_buf) if iterator: return Iterator(path_or_buf) def read(fh): unpacked_obj = list(unpack(fh, encoding=encoding, **kwargs)) if len(unpacked_obj) == 1: return unpacked_obj[0] if should_close: try: path_or_buf.close() except IOError: pass return unpacked_obj # see if we have an actual file if isinstance(path_or_buf, str): try: exists = os.path.exists(path_or_buf) except (TypeError, ValueError): exists = False if exists: with open(path_or_buf, 'rb') as fh: return read(fh) if isinstance(path_or_buf, bytes): # treat as a binary-like fh = None try: fh = BytesIO(path_or_buf) return read(fh) finally: if fh is not None: fh.close() elif hasattr(path_or_buf, 'read') and callable(path_or_buf.read): # treat as a buffer like return read(path_or_buf) raise ValueError('path_or_buf needs to be a string file path or file-like')
[ "Load", "msgpack", "pandas", "object", "from", "the", "specified", "file", "path" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L160-L219
[ "def", "read_msgpack", "(", "path_or_buf", ",", "encoding", "=", "'utf-8'", ",", "iterator", "=", "False", ",", "*", "*", "kwargs", ")", ":", "path_or_buf", ",", "_", ",", "_", ",", "should_close", "=", "get_filepath_or_buffer", "(", "path_or_buf", ")", "if", "iterator", ":", "return", "Iterator", "(", "path_or_buf", ")", "def", "read", "(", "fh", ")", ":", "unpacked_obj", "=", "list", "(", "unpack", "(", "fh", ",", "encoding", "=", "encoding", ",", "*", "*", "kwargs", ")", ")", "if", "len", "(", "unpacked_obj", ")", "==", "1", ":", "return", "unpacked_obj", "[", "0", "]", "if", "should_close", ":", "try", ":", "path_or_buf", ".", "close", "(", ")", "except", "IOError", ":", "pass", "return", "unpacked_obj", "# see if we have an actual file", "if", "isinstance", "(", "path_or_buf", ",", "str", ")", ":", "try", ":", "exists", "=", "os", ".", "path", ".", "exists", "(", "path_or_buf", ")", "except", "(", "TypeError", ",", "ValueError", ")", ":", "exists", "=", "False", "if", "exists", ":", "with", "open", "(", "path_or_buf", ",", "'rb'", ")", "as", "fh", ":", "return", "read", "(", "fh", ")", "if", "isinstance", "(", "path_or_buf", ",", "bytes", ")", ":", "# treat as a binary-like", "fh", "=", "None", "try", ":", "fh", "=", "BytesIO", "(", "path_or_buf", ")", "return", "read", "(", "fh", ")", "finally", ":", "if", "fh", "is", "not", "None", ":", "fh", ".", "close", "(", ")", "elif", "hasattr", "(", "path_or_buf", ",", "'read'", ")", "and", "callable", "(", "path_or_buf", ".", "read", ")", ":", "# treat as a buffer like", "return", "read", "(", "path_or_buf", ")", "raise", "ValueError", "(", "'path_or_buf needs to be a string file path or file-like'", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
dtype_for
return my dtype mapping, whether number or name
pandas/io/packers.py
def dtype_for(t): """ return my dtype mapping, whether number or name """ if t in dtype_dict: return dtype_dict[t] return np.typeDict.get(t, t)
def dtype_for(t): """ return my dtype mapping, whether number or name """ if t in dtype_dict: return dtype_dict[t] return np.typeDict.get(t, t)
[ "return", "my", "dtype", "mapping", "whether", "number", "or", "name" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L236-L240
[ "def", "dtype_for", "(", "t", ")", ":", "if", "t", "in", "dtype_dict", ":", "return", "dtype_dict", "[", "t", "]", "return", "np", ".", "typeDict", ".", "get", "(", "t", ",", "t", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
c2f
Convert strings to complex number instance with specified numpy type.
pandas/io/packers.py
def c2f(r, i, ctype_name): """ Convert strings to complex number instance with specified numpy type. """ ftype = c2f_dict[ctype_name] return np.typeDict[ctype_name](ftype(r) + 1j * ftype(i))
def c2f(r, i, ctype_name): """ Convert strings to complex number instance with specified numpy type. """ ftype = c2f_dict[ctype_name] return np.typeDict[ctype_name](ftype(r) + 1j * ftype(i))
[ "Convert", "strings", "to", "complex", "number", "instance", "with", "specified", "numpy", "type", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L252-L258
[ "def", "c2f", "(", "r", ",", "i", ",", "ctype_name", ")", ":", "ftype", "=", "c2f_dict", "[", "ctype_name", "]", "return", "np", ".", "typeDict", "[", "ctype_name", "]", "(", "ftype", "(", "r", ")", "+", "1j", "*", "ftype", "(", "i", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
convert
convert the numpy values to a list
pandas/io/packers.py
def convert(values): """ convert the numpy values to a list """ dtype = values.dtype if is_categorical_dtype(values): return values elif is_object_dtype(dtype): return values.ravel().tolist() if needs_i8_conversion(dtype): values = values.view('i8') v = values.ravel() if compressor == 'zlib': _check_zlib() # return string arrays like they are if dtype == np.object_: return v.tolist() # convert to a bytes array v = v.tostring() return ExtType(0, zlib.compress(v)) elif compressor == 'blosc': _check_blosc() # return string arrays like they are if dtype == np.object_: return v.tolist() # convert to a bytes array v = v.tostring() return ExtType(0, blosc.compress(v, typesize=dtype.itemsize)) # ndarray (on original dtype) return ExtType(0, v.tostring())
def convert(values): """ convert the numpy values to a list """ dtype = values.dtype if is_categorical_dtype(values): return values elif is_object_dtype(dtype): return values.ravel().tolist() if needs_i8_conversion(dtype): values = values.view('i8') v = values.ravel() if compressor == 'zlib': _check_zlib() # return string arrays like they are if dtype == np.object_: return v.tolist() # convert to a bytes array v = v.tostring() return ExtType(0, zlib.compress(v)) elif compressor == 'blosc': _check_blosc() # return string arrays like they are if dtype == np.object_: return v.tolist() # convert to a bytes array v = v.tostring() return ExtType(0, blosc.compress(v, typesize=dtype.itemsize)) # ndarray (on original dtype) return ExtType(0, v.tostring())
[ "convert", "the", "numpy", "values", "to", "a", "list" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L261-L299
[ "def", "convert", "(", "values", ")", ":", "dtype", "=", "values", ".", "dtype", "if", "is_categorical_dtype", "(", "values", ")", ":", "return", "values", "elif", "is_object_dtype", "(", "dtype", ")", ":", "return", "values", ".", "ravel", "(", ")", ".", "tolist", "(", ")", "if", "needs_i8_conversion", "(", "dtype", ")", ":", "values", "=", "values", ".", "view", "(", "'i8'", ")", "v", "=", "values", ".", "ravel", "(", ")", "if", "compressor", "==", "'zlib'", ":", "_check_zlib", "(", ")", "# return string arrays like they are", "if", "dtype", "==", "np", ".", "object_", ":", "return", "v", ".", "tolist", "(", ")", "# convert to a bytes array", "v", "=", "v", ".", "tostring", "(", ")", "return", "ExtType", "(", "0", ",", "zlib", ".", "compress", "(", "v", ")", ")", "elif", "compressor", "==", "'blosc'", ":", "_check_blosc", "(", ")", "# return string arrays like they are", "if", "dtype", "==", "np", ".", "object_", ":", "return", "v", ".", "tolist", "(", ")", "# convert to a bytes array", "v", "=", "v", ".", "tostring", "(", ")", "return", "ExtType", "(", "0", ",", "blosc", ".", "compress", "(", "v", ",", "typesize", "=", "dtype", ".", "itemsize", ")", ")", "# ndarray (on original dtype)", "return", "ExtType", "(", "0", ",", "v", ".", "tostring", "(", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
encode
Data encoder
pandas/io/packers.py
def encode(obj): """ Data encoder """ tobj = type(obj) if isinstance(obj, Index): if isinstance(obj, RangeIndex): return {'typ': 'range_index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'start': getattr(obj, '_start', None), 'stop': getattr(obj, '_stop', None), 'step': getattr(obj, '_step', None)} elif isinstance(obj, PeriodIndex): return {'typ': 'period_index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'freq': getattr(obj, 'freqstr', None), 'dtype': obj.dtype.name, 'data': convert(obj.asi8), 'compress': compressor} elif isinstance(obj, DatetimeIndex): tz = getattr(obj, 'tz', None) # store tz info and data as UTC if tz is not None: tz = tz.zone obj = obj.tz_convert('UTC') return {'typ': 'datetime_index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'dtype': obj.dtype.name, 'data': convert(obj.asi8), 'freq': getattr(obj, 'freqstr', None), 'tz': tz, 'compress': compressor} elif isinstance(obj, (IntervalIndex, IntervalArray)): if isinstance(obj, IntervalIndex): typ = 'interval_index' else: typ = 'interval_array' return {'typ': typ, 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'left': getattr(obj, 'left', None), 'right': getattr(obj, 'right', None), 'closed': getattr(obj, 'closed', None)} elif isinstance(obj, MultiIndex): return {'typ': 'multi_index', 'klass': obj.__class__.__name__, 'names': getattr(obj, 'names', None), 'dtype': obj.dtype.name, 'data': convert(obj.values), 'compress': compressor} else: return {'typ': 'index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'dtype': obj.dtype.name, 'data': convert(obj.values), 'compress': compressor} elif isinstance(obj, Categorical): return {'typ': 'category', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'codes': obj.codes, 'categories': obj.categories, 'ordered': obj.ordered, 'compress': compressor} elif isinstance(obj, Series): if isinstance(obj, SparseSeries): raise NotImplementedError( 'msgpack sparse series is not implemented' ) # d = {'typ': 'sparse_series', # 'klass': obj.__class__.__name__, # 'dtype': obj.dtype.name, # 'index': obj.index, # 'sp_index': obj.sp_index, # 'sp_values': convert(obj.sp_values), # 'compress': compressor} # for f in ['name', 'fill_value', 'kind']: # d[f] = getattr(obj, f, None) # return d else: return {'typ': 'series', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'index': obj.index, 'dtype': obj.dtype.name, 'data': convert(obj.values), 'compress': compressor} elif issubclass(tobj, NDFrame): if isinstance(obj, SparseDataFrame): raise NotImplementedError( 'msgpack sparse frame is not implemented' ) # d = {'typ': 'sparse_dataframe', # 'klass': obj.__class__.__name__, # 'columns': obj.columns} # for f in ['default_fill_value', 'default_kind']: # d[f] = getattr(obj, f, None) # d['data'] = dict([(name, ss) # for name, ss in obj.items()]) # return d else: data = obj._data if not data.is_consolidated(): data = data.consolidate() # the block manager return {'typ': 'block_manager', 'klass': obj.__class__.__name__, 'axes': data.axes, 'blocks': [{'locs': b.mgr_locs.as_array, 'values': convert(b.values), 'shape': b.values.shape, 'dtype': b.dtype.name, 'klass': b.__class__.__name__, 'compress': compressor} for b in data.blocks] } elif isinstance(obj, (datetime, date, np.datetime64, timedelta, np.timedelta64)) or obj is NaT: if isinstance(obj, Timestamp): tz = obj.tzinfo if tz is not None: tz = tz.zone freq = obj.freq if freq is not None: freq = freq.freqstr return {'typ': 'timestamp', 'value': obj.value, 'freq': freq, 'tz': tz} if obj is NaT: return {'typ': 'nat'} elif isinstance(obj, np.timedelta64): return {'typ': 'timedelta64', 'data': obj.view('i8')} elif isinstance(obj, timedelta): return {'typ': 'timedelta', 'data': (obj.days, obj.seconds, obj.microseconds)} elif isinstance(obj, np.datetime64): return {'typ': 'datetime64', 'data': str(obj)} elif isinstance(obj, datetime): return {'typ': 'datetime', 'data': obj.isoformat()} elif isinstance(obj, date): return {'typ': 'date', 'data': obj.isoformat()} raise Exception( "cannot encode this datetimelike object: {obj}".format(obj=obj)) elif isinstance(obj, Period): return {'typ': 'period', 'ordinal': obj.ordinal, 'freq': obj.freqstr} elif isinstance(obj, Interval): return {'typ': 'interval', 'left': obj.left, 'right': obj.right, 'closed': obj.closed} elif isinstance(obj, BlockIndex): return {'typ': 'block_index', 'klass': obj.__class__.__name__, 'blocs': obj.blocs, 'blengths': obj.blengths, 'length': obj.length} elif isinstance(obj, IntIndex): return {'typ': 'int_index', 'klass': obj.__class__.__name__, 'indices': obj.indices, 'length': obj.length} elif isinstance(obj, np.ndarray): return {'typ': 'ndarray', 'shape': obj.shape, 'ndim': obj.ndim, 'dtype': obj.dtype.name, 'data': convert(obj), 'compress': compressor} elif isinstance(obj, np.number): if np.iscomplexobj(obj): return {'typ': 'np_scalar', 'sub_typ': 'np_complex', 'dtype': obj.dtype.name, 'real': obj.real.__repr__(), 'imag': obj.imag.__repr__()} else: return {'typ': 'np_scalar', 'dtype': obj.dtype.name, 'data': obj.__repr__()} elif isinstance(obj, complex): return {'typ': 'np_complex', 'real': obj.real.__repr__(), 'imag': obj.imag.__repr__()} return obj
def encode(obj): """ Data encoder """ tobj = type(obj) if isinstance(obj, Index): if isinstance(obj, RangeIndex): return {'typ': 'range_index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'start': getattr(obj, '_start', None), 'stop': getattr(obj, '_stop', None), 'step': getattr(obj, '_step', None)} elif isinstance(obj, PeriodIndex): return {'typ': 'period_index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'freq': getattr(obj, 'freqstr', None), 'dtype': obj.dtype.name, 'data': convert(obj.asi8), 'compress': compressor} elif isinstance(obj, DatetimeIndex): tz = getattr(obj, 'tz', None) # store tz info and data as UTC if tz is not None: tz = tz.zone obj = obj.tz_convert('UTC') return {'typ': 'datetime_index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'dtype': obj.dtype.name, 'data': convert(obj.asi8), 'freq': getattr(obj, 'freqstr', None), 'tz': tz, 'compress': compressor} elif isinstance(obj, (IntervalIndex, IntervalArray)): if isinstance(obj, IntervalIndex): typ = 'interval_index' else: typ = 'interval_array' return {'typ': typ, 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'left': getattr(obj, 'left', None), 'right': getattr(obj, 'right', None), 'closed': getattr(obj, 'closed', None)} elif isinstance(obj, MultiIndex): return {'typ': 'multi_index', 'klass': obj.__class__.__name__, 'names': getattr(obj, 'names', None), 'dtype': obj.dtype.name, 'data': convert(obj.values), 'compress': compressor} else: return {'typ': 'index', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'dtype': obj.dtype.name, 'data': convert(obj.values), 'compress': compressor} elif isinstance(obj, Categorical): return {'typ': 'category', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'codes': obj.codes, 'categories': obj.categories, 'ordered': obj.ordered, 'compress': compressor} elif isinstance(obj, Series): if isinstance(obj, SparseSeries): raise NotImplementedError( 'msgpack sparse series is not implemented' ) # d = {'typ': 'sparse_series', # 'klass': obj.__class__.__name__, # 'dtype': obj.dtype.name, # 'index': obj.index, # 'sp_index': obj.sp_index, # 'sp_values': convert(obj.sp_values), # 'compress': compressor} # for f in ['name', 'fill_value', 'kind']: # d[f] = getattr(obj, f, None) # return d else: return {'typ': 'series', 'klass': obj.__class__.__name__, 'name': getattr(obj, 'name', None), 'index': obj.index, 'dtype': obj.dtype.name, 'data': convert(obj.values), 'compress': compressor} elif issubclass(tobj, NDFrame): if isinstance(obj, SparseDataFrame): raise NotImplementedError( 'msgpack sparse frame is not implemented' ) # d = {'typ': 'sparse_dataframe', # 'klass': obj.__class__.__name__, # 'columns': obj.columns} # for f in ['default_fill_value', 'default_kind']: # d[f] = getattr(obj, f, None) # d['data'] = dict([(name, ss) # for name, ss in obj.items()]) # return d else: data = obj._data if not data.is_consolidated(): data = data.consolidate() # the block manager return {'typ': 'block_manager', 'klass': obj.__class__.__name__, 'axes': data.axes, 'blocks': [{'locs': b.mgr_locs.as_array, 'values': convert(b.values), 'shape': b.values.shape, 'dtype': b.dtype.name, 'klass': b.__class__.__name__, 'compress': compressor} for b in data.blocks] } elif isinstance(obj, (datetime, date, np.datetime64, timedelta, np.timedelta64)) or obj is NaT: if isinstance(obj, Timestamp): tz = obj.tzinfo if tz is not None: tz = tz.zone freq = obj.freq if freq is not None: freq = freq.freqstr return {'typ': 'timestamp', 'value': obj.value, 'freq': freq, 'tz': tz} if obj is NaT: return {'typ': 'nat'} elif isinstance(obj, np.timedelta64): return {'typ': 'timedelta64', 'data': obj.view('i8')} elif isinstance(obj, timedelta): return {'typ': 'timedelta', 'data': (obj.days, obj.seconds, obj.microseconds)} elif isinstance(obj, np.datetime64): return {'typ': 'datetime64', 'data': str(obj)} elif isinstance(obj, datetime): return {'typ': 'datetime', 'data': obj.isoformat()} elif isinstance(obj, date): return {'typ': 'date', 'data': obj.isoformat()} raise Exception( "cannot encode this datetimelike object: {obj}".format(obj=obj)) elif isinstance(obj, Period): return {'typ': 'period', 'ordinal': obj.ordinal, 'freq': obj.freqstr} elif isinstance(obj, Interval): return {'typ': 'interval', 'left': obj.left, 'right': obj.right, 'closed': obj.closed} elif isinstance(obj, BlockIndex): return {'typ': 'block_index', 'klass': obj.__class__.__name__, 'blocs': obj.blocs, 'blengths': obj.blengths, 'length': obj.length} elif isinstance(obj, IntIndex): return {'typ': 'int_index', 'klass': obj.__class__.__name__, 'indices': obj.indices, 'length': obj.length} elif isinstance(obj, np.ndarray): return {'typ': 'ndarray', 'shape': obj.shape, 'ndim': obj.ndim, 'dtype': obj.dtype.name, 'data': convert(obj), 'compress': compressor} elif isinstance(obj, np.number): if np.iscomplexobj(obj): return {'typ': 'np_scalar', 'sub_typ': 'np_complex', 'dtype': obj.dtype.name, 'real': obj.real.__repr__(), 'imag': obj.imag.__repr__()} else: return {'typ': 'np_scalar', 'dtype': obj.dtype.name, 'data': obj.__repr__()} elif isinstance(obj, complex): return {'typ': 'np_complex', 'real': obj.real.__repr__(), 'imag': obj.imag.__repr__()} return obj
[ "Data", "encoder" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L360-L560
[ "def", "encode", "(", "obj", ")", ":", "tobj", "=", "type", "(", "obj", ")", "if", "isinstance", "(", "obj", ",", "Index", ")", ":", "if", "isinstance", "(", "obj", ",", "RangeIndex", ")", ":", "return", "{", "'typ'", ":", "'range_index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'start'", ":", "getattr", "(", "obj", ",", "'_start'", ",", "None", ")", ",", "'stop'", ":", "getattr", "(", "obj", ",", "'_stop'", ",", "None", ")", ",", "'step'", ":", "getattr", "(", "obj", ",", "'_step'", ",", "None", ")", "}", "elif", "isinstance", "(", "obj", ",", "PeriodIndex", ")", ":", "return", "{", "'typ'", ":", "'period_index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'freq'", ":", "getattr", "(", "obj", ",", "'freqstr'", ",", "None", ")", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "convert", "(", "obj", ".", "asi8", ")", ",", "'compress'", ":", "compressor", "}", "elif", "isinstance", "(", "obj", ",", "DatetimeIndex", ")", ":", "tz", "=", "getattr", "(", "obj", ",", "'tz'", ",", "None", ")", "# store tz info and data as UTC", "if", "tz", "is", "not", "None", ":", "tz", "=", "tz", ".", "zone", "obj", "=", "obj", ".", "tz_convert", "(", "'UTC'", ")", "return", "{", "'typ'", ":", "'datetime_index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "convert", "(", "obj", ".", "asi8", ")", ",", "'freq'", ":", "getattr", "(", "obj", ",", "'freqstr'", ",", "None", ")", ",", "'tz'", ":", "tz", ",", "'compress'", ":", "compressor", "}", "elif", "isinstance", "(", "obj", ",", "(", "IntervalIndex", ",", "IntervalArray", ")", ")", ":", "if", "isinstance", "(", "obj", ",", "IntervalIndex", ")", ":", "typ", "=", "'interval_index'", "else", ":", "typ", "=", "'interval_array'", "return", "{", "'typ'", ":", "typ", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'left'", ":", "getattr", "(", "obj", ",", "'left'", ",", "None", ")", ",", "'right'", ":", "getattr", "(", "obj", ",", "'right'", ",", "None", ")", ",", "'closed'", ":", "getattr", "(", "obj", ",", "'closed'", ",", "None", ")", "}", "elif", "isinstance", "(", "obj", ",", "MultiIndex", ")", ":", "return", "{", "'typ'", ":", "'multi_index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'names'", ":", "getattr", "(", "obj", ",", "'names'", ",", "None", ")", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "convert", "(", "obj", ".", "values", ")", ",", "'compress'", ":", "compressor", "}", "else", ":", "return", "{", "'typ'", ":", "'index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "convert", "(", "obj", ".", "values", ")", ",", "'compress'", ":", "compressor", "}", "elif", "isinstance", "(", "obj", ",", "Categorical", ")", ":", "return", "{", "'typ'", ":", "'category'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'codes'", ":", "obj", ".", "codes", ",", "'categories'", ":", "obj", ".", "categories", ",", "'ordered'", ":", "obj", ".", "ordered", ",", "'compress'", ":", "compressor", "}", "elif", "isinstance", "(", "obj", ",", "Series", ")", ":", "if", "isinstance", "(", "obj", ",", "SparseSeries", ")", ":", "raise", "NotImplementedError", "(", "'msgpack sparse series is not implemented'", ")", "# d = {'typ': 'sparse_series',", "# 'klass': obj.__class__.__name__,", "# 'dtype': obj.dtype.name,", "# 'index': obj.index,", "# 'sp_index': obj.sp_index,", "# 'sp_values': convert(obj.sp_values),", "# 'compress': compressor}", "# for f in ['name', 'fill_value', 'kind']:", "# d[f] = getattr(obj, f, None)", "# return d", "else", ":", "return", "{", "'typ'", ":", "'series'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'name'", ":", "getattr", "(", "obj", ",", "'name'", ",", "None", ")", ",", "'index'", ":", "obj", ".", "index", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "convert", "(", "obj", ".", "values", ")", ",", "'compress'", ":", "compressor", "}", "elif", "issubclass", "(", "tobj", ",", "NDFrame", ")", ":", "if", "isinstance", "(", "obj", ",", "SparseDataFrame", ")", ":", "raise", "NotImplementedError", "(", "'msgpack sparse frame is not implemented'", ")", "# d = {'typ': 'sparse_dataframe',", "# 'klass': obj.__class__.__name__,", "# 'columns': obj.columns}", "# for f in ['default_fill_value', 'default_kind']:", "# d[f] = getattr(obj, f, None)", "# d['data'] = dict([(name, ss)", "# for name, ss in obj.items()])", "# return d", "else", ":", "data", "=", "obj", ".", "_data", "if", "not", "data", ".", "is_consolidated", "(", ")", ":", "data", "=", "data", ".", "consolidate", "(", ")", "# the block manager", "return", "{", "'typ'", ":", "'block_manager'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'axes'", ":", "data", ".", "axes", ",", "'blocks'", ":", "[", "{", "'locs'", ":", "b", ".", "mgr_locs", ".", "as_array", ",", "'values'", ":", "convert", "(", "b", ".", "values", ")", ",", "'shape'", ":", "b", ".", "values", ".", "shape", ",", "'dtype'", ":", "b", ".", "dtype", ".", "name", ",", "'klass'", ":", "b", ".", "__class__", ".", "__name__", ",", "'compress'", ":", "compressor", "}", "for", "b", "in", "data", ".", "blocks", "]", "}", "elif", "isinstance", "(", "obj", ",", "(", "datetime", ",", "date", ",", "np", ".", "datetime64", ",", "timedelta", ",", "np", ".", "timedelta64", ")", ")", "or", "obj", "is", "NaT", ":", "if", "isinstance", "(", "obj", ",", "Timestamp", ")", ":", "tz", "=", "obj", ".", "tzinfo", "if", "tz", "is", "not", "None", ":", "tz", "=", "tz", ".", "zone", "freq", "=", "obj", ".", "freq", "if", "freq", "is", "not", "None", ":", "freq", "=", "freq", ".", "freqstr", "return", "{", "'typ'", ":", "'timestamp'", ",", "'value'", ":", "obj", ".", "value", ",", "'freq'", ":", "freq", ",", "'tz'", ":", "tz", "}", "if", "obj", "is", "NaT", ":", "return", "{", "'typ'", ":", "'nat'", "}", "elif", "isinstance", "(", "obj", ",", "np", ".", "timedelta64", ")", ":", "return", "{", "'typ'", ":", "'timedelta64'", ",", "'data'", ":", "obj", ".", "view", "(", "'i8'", ")", "}", "elif", "isinstance", "(", "obj", ",", "timedelta", ")", ":", "return", "{", "'typ'", ":", "'timedelta'", ",", "'data'", ":", "(", "obj", ".", "days", ",", "obj", ".", "seconds", ",", "obj", ".", "microseconds", ")", "}", "elif", "isinstance", "(", "obj", ",", "np", ".", "datetime64", ")", ":", "return", "{", "'typ'", ":", "'datetime64'", ",", "'data'", ":", "str", "(", "obj", ")", "}", "elif", "isinstance", "(", "obj", ",", "datetime", ")", ":", "return", "{", "'typ'", ":", "'datetime'", ",", "'data'", ":", "obj", ".", "isoformat", "(", ")", "}", "elif", "isinstance", "(", "obj", ",", "date", ")", ":", "return", "{", "'typ'", ":", "'date'", ",", "'data'", ":", "obj", ".", "isoformat", "(", ")", "}", "raise", "Exception", "(", "\"cannot encode this datetimelike object: {obj}\"", ".", "format", "(", "obj", "=", "obj", ")", ")", "elif", "isinstance", "(", "obj", ",", "Period", ")", ":", "return", "{", "'typ'", ":", "'period'", ",", "'ordinal'", ":", "obj", ".", "ordinal", ",", "'freq'", ":", "obj", ".", "freqstr", "}", "elif", "isinstance", "(", "obj", ",", "Interval", ")", ":", "return", "{", "'typ'", ":", "'interval'", ",", "'left'", ":", "obj", ".", "left", ",", "'right'", ":", "obj", ".", "right", ",", "'closed'", ":", "obj", ".", "closed", "}", "elif", "isinstance", "(", "obj", ",", "BlockIndex", ")", ":", "return", "{", "'typ'", ":", "'block_index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'blocs'", ":", "obj", ".", "blocs", ",", "'blengths'", ":", "obj", ".", "blengths", ",", "'length'", ":", "obj", ".", "length", "}", "elif", "isinstance", "(", "obj", ",", "IntIndex", ")", ":", "return", "{", "'typ'", ":", "'int_index'", ",", "'klass'", ":", "obj", ".", "__class__", ".", "__name__", ",", "'indices'", ":", "obj", ".", "indices", ",", "'length'", ":", "obj", ".", "length", "}", "elif", "isinstance", "(", "obj", ",", "np", ".", "ndarray", ")", ":", "return", "{", "'typ'", ":", "'ndarray'", ",", "'shape'", ":", "obj", ".", "shape", ",", "'ndim'", ":", "obj", ".", "ndim", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "convert", "(", "obj", ")", ",", "'compress'", ":", "compressor", "}", "elif", "isinstance", "(", "obj", ",", "np", ".", "number", ")", ":", "if", "np", ".", "iscomplexobj", "(", "obj", ")", ":", "return", "{", "'typ'", ":", "'np_scalar'", ",", "'sub_typ'", ":", "'np_complex'", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'real'", ":", "obj", ".", "real", ".", "__repr__", "(", ")", ",", "'imag'", ":", "obj", ".", "imag", ".", "__repr__", "(", ")", "}", "else", ":", "return", "{", "'typ'", ":", "'np_scalar'", ",", "'dtype'", ":", "obj", ".", "dtype", ".", "name", ",", "'data'", ":", "obj", ".", "__repr__", "(", ")", "}", "elif", "isinstance", "(", "obj", ",", "complex", ")", ":", "return", "{", "'typ'", ":", "'np_complex'", ",", "'real'", ":", "obj", ".", "real", ".", "__repr__", "(", ")", ",", "'imag'", ":", "obj", ".", "imag", ".", "__repr__", "(", ")", "}", "return", "obj" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
decode
Decoder for deserializing numpy data types.
pandas/io/packers.py
def decode(obj): """ Decoder for deserializing numpy data types. """ typ = obj.get('typ') if typ is None: return obj elif typ == 'timestamp': freq = obj['freq'] if 'freq' in obj else obj['offset'] return Timestamp(obj['value'], tz=obj['tz'], freq=freq) elif typ == 'nat': return NaT elif typ == 'period': return Period(ordinal=obj['ordinal'], freq=obj['freq']) elif typ == 'index': dtype = dtype_for(obj['dtype']) data = unconvert(obj['data'], dtype, obj.get('compress')) return Index(data, dtype=dtype, name=obj['name']) elif typ == 'range_index': return RangeIndex(obj['start'], obj['stop'], obj['step'], name=obj['name']) elif typ == 'multi_index': dtype = dtype_for(obj['dtype']) data = unconvert(obj['data'], dtype, obj.get('compress')) data = [tuple(x) for x in data] return MultiIndex.from_tuples(data, names=obj['names']) elif typ == 'period_index': data = unconvert(obj['data'], np.int64, obj.get('compress')) d = dict(name=obj['name'], freq=obj['freq']) freq = d.pop('freq', None) return PeriodIndex(PeriodArray(data, freq), **d) elif typ == 'datetime_index': data = unconvert(obj['data'], np.int64, obj.get('compress')) d = dict(name=obj['name'], freq=obj['freq']) result = DatetimeIndex(data, **d) tz = obj['tz'] # reverse tz conversion if tz is not None: result = result.tz_localize('UTC').tz_convert(tz) return result elif typ in ('interval_index', 'interval_array'): return globals()[obj['klass']].from_arrays(obj['left'], obj['right'], obj['closed'], name=obj['name']) elif typ == 'category': from_codes = globals()[obj['klass']].from_codes return from_codes(codes=obj['codes'], categories=obj['categories'], ordered=obj['ordered']) elif typ == 'interval': return Interval(obj['left'], obj['right'], obj['closed']) elif typ == 'series': dtype = dtype_for(obj['dtype']) pd_dtype = pandas_dtype(dtype) index = obj['index'] result = Series(unconvert(obj['data'], dtype, obj['compress']), index=index, dtype=pd_dtype, name=obj['name']) return result elif typ == 'block_manager': axes = obj['axes'] def create_block(b): values = _safe_reshape(unconvert( b['values'], dtype_for(b['dtype']), b['compress']), b['shape']) # locs handles duplicate column names, and should be used instead # of items; see GH 9618 if 'locs' in b: placement = b['locs'] else: placement = axes[0].get_indexer(b['items']) if is_datetime64tz_dtype(b['dtype']): assert isinstance(values, np.ndarray), type(values) assert values.dtype == 'M8[ns]', values.dtype values = DatetimeArray(values, dtype=b['dtype']) return make_block(values=values, klass=getattr(internals, b['klass']), placement=placement, dtype=b['dtype']) blocks = [create_block(b) for b in obj['blocks']] return globals()[obj['klass']](BlockManager(blocks, axes)) elif typ == 'datetime': return parse(obj['data']) elif typ == 'datetime64': return np.datetime64(parse(obj['data'])) elif typ == 'date': return parse(obj['data']).date() elif typ == 'timedelta': return timedelta(*obj['data']) elif typ == 'timedelta64': return np.timedelta64(int(obj['data'])) # elif typ == 'sparse_series': # dtype = dtype_for(obj['dtype']) # return SparseSeries( # unconvert(obj['sp_values'], dtype, obj['compress']), # sparse_index=obj['sp_index'], index=obj['index'], # fill_value=obj['fill_value'], kind=obj['kind'], name=obj['name']) # elif typ == 'sparse_dataframe': # return SparseDataFrame( # obj['data'], columns=obj['columns'], # default_fill_value=obj['default_fill_value'], # default_kind=obj['default_kind'] # ) # elif typ == 'sparse_panel': # return SparsePanel( # obj['data'], items=obj['items'], # default_fill_value=obj['default_fill_value'], # default_kind=obj['default_kind']) elif typ == 'block_index': return globals()[obj['klass']](obj['length'], obj['blocs'], obj['blengths']) elif typ == 'int_index': return globals()[obj['klass']](obj['length'], obj['indices']) elif typ == 'ndarray': return unconvert(obj['data'], np.typeDict[obj['dtype']], obj.get('compress')).reshape(obj['shape']) elif typ == 'np_scalar': if obj.get('sub_typ') == 'np_complex': return c2f(obj['real'], obj['imag'], obj['dtype']) else: dtype = dtype_for(obj['dtype']) try: return dtype(obj['data']) except (ValueError, TypeError): return dtype.type(obj['data']) elif typ == 'np_complex': return complex(obj['real'] + '+' + obj['imag'] + 'j') elif isinstance(obj, (dict, list, set)): return obj else: return obj
def decode(obj): """ Decoder for deserializing numpy data types. """ typ = obj.get('typ') if typ is None: return obj elif typ == 'timestamp': freq = obj['freq'] if 'freq' in obj else obj['offset'] return Timestamp(obj['value'], tz=obj['tz'], freq=freq) elif typ == 'nat': return NaT elif typ == 'period': return Period(ordinal=obj['ordinal'], freq=obj['freq']) elif typ == 'index': dtype = dtype_for(obj['dtype']) data = unconvert(obj['data'], dtype, obj.get('compress')) return Index(data, dtype=dtype, name=obj['name']) elif typ == 'range_index': return RangeIndex(obj['start'], obj['stop'], obj['step'], name=obj['name']) elif typ == 'multi_index': dtype = dtype_for(obj['dtype']) data = unconvert(obj['data'], dtype, obj.get('compress')) data = [tuple(x) for x in data] return MultiIndex.from_tuples(data, names=obj['names']) elif typ == 'period_index': data = unconvert(obj['data'], np.int64, obj.get('compress')) d = dict(name=obj['name'], freq=obj['freq']) freq = d.pop('freq', None) return PeriodIndex(PeriodArray(data, freq), **d) elif typ == 'datetime_index': data = unconvert(obj['data'], np.int64, obj.get('compress')) d = dict(name=obj['name'], freq=obj['freq']) result = DatetimeIndex(data, **d) tz = obj['tz'] # reverse tz conversion if tz is not None: result = result.tz_localize('UTC').tz_convert(tz) return result elif typ in ('interval_index', 'interval_array'): return globals()[obj['klass']].from_arrays(obj['left'], obj['right'], obj['closed'], name=obj['name']) elif typ == 'category': from_codes = globals()[obj['klass']].from_codes return from_codes(codes=obj['codes'], categories=obj['categories'], ordered=obj['ordered']) elif typ == 'interval': return Interval(obj['left'], obj['right'], obj['closed']) elif typ == 'series': dtype = dtype_for(obj['dtype']) pd_dtype = pandas_dtype(dtype) index = obj['index'] result = Series(unconvert(obj['data'], dtype, obj['compress']), index=index, dtype=pd_dtype, name=obj['name']) return result elif typ == 'block_manager': axes = obj['axes'] def create_block(b): values = _safe_reshape(unconvert( b['values'], dtype_for(b['dtype']), b['compress']), b['shape']) # locs handles duplicate column names, and should be used instead # of items; see GH 9618 if 'locs' in b: placement = b['locs'] else: placement = axes[0].get_indexer(b['items']) if is_datetime64tz_dtype(b['dtype']): assert isinstance(values, np.ndarray), type(values) assert values.dtype == 'M8[ns]', values.dtype values = DatetimeArray(values, dtype=b['dtype']) return make_block(values=values, klass=getattr(internals, b['klass']), placement=placement, dtype=b['dtype']) blocks = [create_block(b) for b in obj['blocks']] return globals()[obj['klass']](BlockManager(blocks, axes)) elif typ == 'datetime': return parse(obj['data']) elif typ == 'datetime64': return np.datetime64(parse(obj['data'])) elif typ == 'date': return parse(obj['data']).date() elif typ == 'timedelta': return timedelta(*obj['data']) elif typ == 'timedelta64': return np.timedelta64(int(obj['data'])) # elif typ == 'sparse_series': # dtype = dtype_for(obj['dtype']) # return SparseSeries( # unconvert(obj['sp_values'], dtype, obj['compress']), # sparse_index=obj['sp_index'], index=obj['index'], # fill_value=obj['fill_value'], kind=obj['kind'], name=obj['name']) # elif typ == 'sparse_dataframe': # return SparseDataFrame( # obj['data'], columns=obj['columns'], # default_fill_value=obj['default_fill_value'], # default_kind=obj['default_kind'] # ) # elif typ == 'sparse_panel': # return SparsePanel( # obj['data'], items=obj['items'], # default_fill_value=obj['default_fill_value'], # default_kind=obj['default_kind']) elif typ == 'block_index': return globals()[obj['klass']](obj['length'], obj['blocs'], obj['blengths']) elif typ == 'int_index': return globals()[obj['klass']](obj['length'], obj['indices']) elif typ == 'ndarray': return unconvert(obj['data'], np.typeDict[obj['dtype']], obj.get('compress')).reshape(obj['shape']) elif typ == 'np_scalar': if obj.get('sub_typ') == 'np_complex': return c2f(obj['real'], obj['imag'], obj['dtype']) else: dtype = dtype_for(obj['dtype']) try: return dtype(obj['data']) except (ValueError, TypeError): return dtype.type(obj['data']) elif typ == 'np_complex': return complex(obj['real'] + '+' + obj['imag'] + 'j') elif isinstance(obj, (dict, list, set)): return obj else: return obj
[ "Decoder", "for", "deserializing", "numpy", "data", "types", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L563-L711
[ "def", "decode", "(", "obj", ")", ":", "typ", "=", "obj", ".", "get", "(", "'typ'", ")", "if", "typ", "is", "None", ":", "return", "obj", "elif", "typ", "==", "'timestamp'", ":", "freq", "=", "obj", "[", "'freq'", "]", "if", "'freq'", "in", "obj", "else", "obj", "[", "'offset'", "]", "return", "Timestamp", "(", "obj", "[", "'value'", "]", ",", "tz", "=", "obj", "[", "'tz'", "]", ",", "freq", "=", "freq", ")", "elif", "typ", "==", "'nat'", ":", "return", "NaT", "elif", "typ", "==", "'period'", ":", "return", "Period", "(", "ordinal", "=", "obj", "[", "'ordinal'", "]", ",", "freq", "=", "obj", "[", "'freq'", "]", ")", "elif", "typ", "==", "'index'", ":", "dtype", "=", "dtype_for", "(", "obj", "[", "'dtype'", "]", ")", "data", "=", "unconvert", "(", "obj", "[", "'data'", "]", ",", "dtype", ",", "obj", ".", "get", "(", "'compress'", ")", ")", "return", "Index", "(", "data", ",", "dtype", "=", "dtype", ",", "name", "=", "obj", "[", "'name'", "]", ")", "elif", "typ", "==", "'range_index'", ":", "return", "RangeIndex", "(", "obj", "[", "'start'", "]", ",", "obj", "[", "'stop'", "]", ",", "obj", "[", "'step'", "]", ",", "name", "=", "obj", "[", "'name'", "]", ")", "elif", "typ", "==", "'multi_index'", ":", "dtype", "=", "dtype_for", "(", "obj", "[", "'dtype'", "]", ")", "data", "=", "unconvert", "(", "obj", "[", "'data'", "]", ",", "dtype", ",", "obj", ".", "get", "(", "'compress'", ")", ")", "data", "=", "[", "tuple", "(", "x", ")", "for", "x", "in", "data", "]", "return", "MultiIndex", ".", "from_tuples", "(", "data", ",", "names", "=", "obj", "[", "'names'", "]", ")", "elif", "typ", "==", "'period_index'", ":", "data", "=", "unconvert", "(", "obj", "[", "'data'", "]", ",", "np", ".", "int64", ",", "obj", ".", "get", "(", "'compress'", ")", ")", "d", "=", "dict", "(", "name", "=", "obj", "[", "'name'", "]", ",", "freq", "=", "obj", "[", "'freq'", "]", ")", "freq", "=", "d", ".", "pop", "(", "'freq'", ",", "None", ")", "return", "PeriodIndex", "(", "PeriodArray", "(", "data", ",", "freq", ")", ",", "*", "*", "d", ")", "elif", "typ", "==", "'datetime_index'", ":", "data", "=", "unconvert", "(", "obj", "[", "'data'", "]", ",", "np", ".", "int64", ",", "obj", ".", "get", "(", "'compress'", ")", ")", "d", "=", "dict", "(", "name", "=", "obj", "[", "'name'", "]", ",", "freq", "=", "obj", "[", "'freq'", "]", ")", "result", "=", "DatetimeIndex", "(", "data", ",", "*", "*", "d", ")", "tz", "=", "obj", "[", "'tz'", "]", "# reverse tz conversion", "if", "tz", "is", "not", "None", ":", "result", "=", "result", ".", "tz_localize", "(", "'UTC'", ")", ".", "tz_convert", "(", "tz", ")", "return", "result", "elif", "typ", "in", "(", "'interval_index'", ",", "'interval_array'", ")", ":", "return", "globals", "(", ")", "[", "obj", "[", "'klass'", "]", "]", ".", "from_arrays", "(", "obj", "[", "'left'", "]", ",", "obj", "[", "'right'", "]", ",", "obj", "[", "'closed'", "]", ",", "name", "=", "obj", "[", "'name'", "]", ")", "elif", "typ", "==", "'category'", ":", "from_codes", "=", "globals", "(", ")", "[", "obj", "[", "'klass'", "]", "]", ".", "from_codes", "return", "from_codes", "(", "codes", "=", "obj", "[", "'codes'", "]", ",", "categories", "=", "obj", "[", "'categories'", "]", ",", "ordered", "=", "obj", "[", "'ordered'", "]", ")", "elif", "typ", "==", "'interval'", ":", "return", "Interval", "(", "obj", "[", "'left'", "]", ",", "obj", "[", "'right'", "]", ",", "obj", "[", "'closed'", "]", ")", "elif", "typ", "==", "'series'", ":", "dtype", "=", "dtype_for", "(", "obj", "[", "'dtype'", "]", ")", "pd_dtype", "=", "pandas_dtype", "(", "dtype", ")", "index", "=", "obj", "[", "'index'", "]", "result", "=", "Series", "(", "unconvert", "(", "obj", "[", "'data'", "]", ",", "dtype", ",", "obj", "[", "'compress'", "]", ")", ",", "index", "=", "index", ",", "dtype", "=", "pd_dtype", ",", "name", "=", "obj", "[", "'name'", "]", ")", "return", "result", "elif", "typ", "==", "'block_manager'", ":", "axes", "=", "obj", "[", "'axes'", "]", "def", "create_block", "(", "b", ")", ":", "values", "=", "_safe_reshape", "(", "unconvert", "(", "b", "[", "'values'", "]", ",", "dtype_for", "(", "b", "[", "'dtype'", "]", ")", ",", "b", "[", "'compress'", "]", ")", ",", "b", "[", "'shape'", "]", ")", "# locs handles duplicate column names, and should be used instead", "# of items; see GH 9618", "if", "'locs'", "in", "b", ":", "placement", "=", "b", "[", "'locs'", "]", "else", ":", "placement", "=", "axes", "[", "0", "]", ".", "get_indexer", "(", "b", "[", "'items'", "]", ")", "if", "is_datetime64tz_dtype", "(", "b", "[", "'dtype'", "]", ")", ":", "assert", "isinstance", "(", "values", ",", "np", ".", "ndarray", ")", ",", "type", "(", "values", ")", "assert", "values", ".", "dtype", "==", "'M8[ns]'", ",", "values", ".", "dtype", "values", "=", "DatetimeArray", "(", "values", ",", "dtype", "=", "b", "[", "'dtype'", "]", ")", "return", "make_block", "(", "values", "=", "values", ",", "klass", "=", "getattr", "(", "internals", ",", "b", "[", "'klass'", "]", ")", ",", "placement", "=", "placement", ",", "dtype", "=", "b", "[", "'dtype'", "]", ")", "blocks", "=", "[", "create_block", "(", "b", ")", "for", "b", "in", "obj", "[", "'blocks'", "]", "]", "return", "globals", "(", ")", "[", "obj", "[", "'klass'", "]", "]", "(", "BlockManager", "(", "blocks", ",", "axes", ")", ")", "elif", "typ", "==", "'datetime'", ":", "return", "parse", "(", "obj", "[", "'data'", "]", ")", "elif", "typ", "==", "'datetime64'", ":", "return", "np", ".", "datetime64", "(", "parse", "(", "obj", "[", "'data'", "]", ")", ")", "elif", "typ", "==", "'date'", ":", "return", "parse", "(", "obj", "[", "'data'", "]", ")", ".", "date", "(", ")", "elif", "typ", "==", "'timedelta'", ":", "return", "timedelta", "(", "*", "obj", "[", "'data'", "]", ")", "elif", "typ", "==", "'timedelta64'", ":", "return", "np", ".", "timedelta64", "(", "int", "(", "obj", "[", "'data'", "]", ")", ")", "# elif typ == 'sparse_series':", "# dtype = dtype_for(obj['dtype'])", "# return SparseSeries(", "# unconvert(obj['sp_values'], dtype, obj['compress']),", "# sparse_index=obj['sp_index'], index=obj['index'],", "# fill_value=obj['fill_value'], kind=obj['kind'], name=obj['name'])", "# elif typ == 'sparse_dataframe':", "# return SparseDataFrame(", "# obj['data'], columns=obj['columns'],", "# default_fill_value=obj['default_fill_value'],", "# default_kind=obj['default_kind']", "# )", "# elif typ == 'sparse_panel':", "# return SparsePanel(", "# obj['data'], items=obj['items'],", "# default_fill_value=obj['default_fill_value'],", "# default_kind=obj['default_kind'])", "elif", "typ", "==", "'block_index'", ":", "return", "globals", "(", ")", "[", "obj", "[", "'klass'", "]", "]", "(", "obj", "[", "'length'", "]", ",", "obj", "[", "'blocs'", "]", ",", "obj", "[", "'blengths'", "]", ")", "elif", "typ", "==", "'int_index'", ":", "return", "globals", "(", ")", "[", "obj", "[", "'klass'", "]", "]", "(", "obj", "[", "'length'", "]", ",", "obj", "[", "'indices'", "]", ")", "elif", "typ", "==", "'ndarray'", ":", "return", "unconvert", "(", "obj", "[", "'data'", "]", ",", "np", ".", "typeDict", "[", "obj", "[", "'dtype'", "]", "]", ",", "obj", ".", "get", "(", "'compress'", ")", ")", ".", "reshape", "(", "obj", "[", "'shape'", "]", ")", "elif", "typ", "==", "'np_scalar'", ":", "if", "obj", ".", "get", "(", "'sub_typ'", ")", "==", "'np_complex'", ":", "return", "c2f", "(", "obj", "[", "'real'", "]", ",", "obj", "[", "'imag'", "]", ",", "obj", "[", "'dtype'", "]", ")", "else", ":", "dtype", "=", "dtype_for", "(", "obj", "[", "'dtype'", "]", ")", "try", ":", "return", "dtype", "(", "obj", "[", "'data'", "]", ")", "except", "(", "ValueError", ",", "TypeError", ")", ":", "return", "dtype", ".", "type", "(", "obj", "[", "'data'", "]", ")", "elif", "typ", "==", "'np_complex'", ":", "return", "complex", "(", "obj", "[", "'real'", "]", "+", "'+'", "+", "obj", "[", "'imag'", "]", "+", "'j'", ")", "elif", "isinstance", "(", "obj", ",", "(", "dict", ",", "list", ",", "set", ")", ")", ":", "return", "obj", "else", ":", "return", "obj" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
pack
Pack an object and return the packed bytes.
pandas/io/packers.py
def pack(o, default=encode, encoding='utf-8', unicode_errors='strict', use_single_float=False, autoreset=1, use_bin_type=1): """ Pack an object and return the packed bytes. """ return Packer(default=default, encoding=encoding, unicode_errors=unicode_errors, use_single_float=use_single_float, autoreset=autoreset, use_bin_type=use_bin_type).pack(o)
def pack(o, default=encode, encoding='utf-8', unicode_errors='strict', use_single_float=False, autoreset=1, use_bin_type=1): """ Pack an object and return the packed bytes. """ return Packer(default=default, encoding=encoding, unicode_errors=unicode_errors, use_single_float=use_single_float, autoreset=autoreset, use_bin_type=use_bin_type).pack(o)
[ "Pack", "an", "object", "and", "return", "the", "packed", "bytes", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L714-L725
[ "def", "pack", "(", "o", ",", "default", "=", "encode", ",", "encoding", "=", "'utf-8'", ",", "unicode_errors", "=", "'strict'", ",", "use_single_float", "=", "False", ",", "autoreset", "=", "1", ",", "use_bin_type", "=", "1", ")", ":", "return", "Packer", "(", "default", "=", "default", ",", "encoding", "=", "encoding", ",", "unicode_errors", "=", "unicode_errors", ",", "use_single_float", "=", "use_single_float", ",", "autoreset", "=", "autoreset", ",", "use_bin_type", "=", "use_bin_type", ")", ".", "pack", "(", "o", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
unpack
Unpack a packed object, return an iterator Note: packed lists will be returned as tuples
pandas/io/packers.py
def unpack(packed, object_hook=decode, list_hook=None, use_list=False, encoding='utf-8', unicode_errors='strict', object_pairs_hook=None, max_buffer_size=0, ext_hook=ExtType): """ Unpack a packed object, return an iterator Note: packed lists will be returned as tuples """ return Unpacker(packed, object_hook=object_hook, list_hook=list_hook, use_list=use_list, encoding=encoding, unicode_errors=unicode_errors, object_pairs_hook=object_pairs_hook, max_buffer_size=max_buffer_size, ext_hook=ext_hook)
def unpack(packed, object_hook=decode, list_hook=None, use_list=False, encoding='utf-8', unicode_errors='strict', object_pairs_hook=None, max_buffer_size=0, ext_hook=ExtType): """ Unpack a packed object, return an iterator Note: packed lists will be returned as tuples """ return Unpacker(packed, object_hook=object_hook, list_hook=list_hook, use_list=use_list, encoding=encoding, unicode_errors=unicode_errors, object_pairs_hook=object_pairs_hook, max_buffer_size=max_buffer_size, ext_hook=ext_hook)
[ "Unpack", "a", "packed", "object", "return", "an", "iterator", "Note", ":", "packed", "lists", "will", "be", "returned", "as", "tuples" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/packers.py#L728-L743
[ "def", "unpack", "(", "packed", ",", "object_hook", "=", "decode", ",", "list_hook", "=", "None", ",", "use_list", "=", "False", ",", "encoding", "=", "'utf-8'", ",", "unicode_errors", "=", "'strict'", ",", "object_pairs_hook", "=", "None", ",", "max_buffer_size", "=", "0", ",", "ext_hook", "=", "ExtType", ")", ":", "return", "Unpacker", "(", "packed", ",", "object_hook", "=", "object_hook", ",", "list_hook", "=", "list_hook", ",", "use_list", "=", "use_list", ",", "encoding", "=", "encoding", ",", "unicode_errors", "=", "unicode_errors", ",", "object_pairs_hook", "=", "object_pairs_hook", ",", "max_buffer_size", "=", "max_buffer_size", ",", "ext_hook", "=", "ext_hook", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
read_json
Convert a JSON string to pandas object. Parameters ---------- path_or_buf : a valid JSON string or file-like, default: None The string could be a URL. Valid URL schemes include http, ftp, s3, gcs, and file. For file URLs, a host is expected. For instance, a local file could be ``file://localhost/path/to/table.json`` orient : string, Indication of expected JSON string format. Compatible JSON strings can be produced by ``to_json()`` with a corresponding orient value. The set of possible orients is: - ``'split'`` : dict like ``{index -> [index], columns -> [columns], data -> [values]}`` - ``'records'`` : list like ``[{column -> value}, ... , {column -> value}]`` - ``'index'`` : dict like ``{index -> {column -> value}}`` - ``'columns'`` : dict like ``{column -> {index -> value}}`` - ``'values'`` : just the values array The allowed and default values depend on the value of the `typ` parameter. * when ``typ == 'series'``, - allowed orients are ``{'split','records','index'}`` - default is ``'index'`` - The Series index must be unique for orient ``'index'``. * when ``typ == 'frame'``, - allowed orients are ``{'split','records','index', 'columns','values', 'table'}`` - default is ``'columns'`` - The DataFrame index must be unique for orients ``'index'`` and ``'columns'``. - The DataFrame columns must be unique for orients ``'index'``, ``'columns'``, and ``'records'``. .. versionadded:: 0.23.0 'table' as an allowed value for the ``orient`` argument typ : type of object to recover (series or frame), default 'frame' dtype : boolean or dict, default None If True, infer dtypes; if a dict of column to dtype, then use those; if False, then don't infer dtypes at all, applies only to the data. For all ``orient`` values except ``'table'``, default is True. .. versionchanged:: 0.25.0 Not applicable for ``orient='table'``. convert_axes : boolean, default None Try to convert the axes to the proper dtypes. For all ``orient`` values except ``'table'``, default is True. .. versionchanged:: 0.25.0 Not applicable for ``orient='table'``. convert_dates : boolean, default True List of columns to parse for dates; If True, then try to parse datelike columns default is True; a column label is datelike if * it ends with ``'_at'``, * it ends with ``'_time'``, * it begins with ``'timestamp'``, * it is ``'modified'``, or * it is ``'date'`` keep_default_dates : boolean, default True If parsing dates, then parse the default datelike columns numpy : boolean, default False Direct decoding to numpy arrays. Supports numeric data only, but non-numeric column and index labels are supported. Note also that the JSON ordering MUST be the same for each term if numpy=True. precise_float : boolean, default False Set to enable usage of higher precision (strtod) function when decoding string to double values. Default (False) is to use fast but less precise builtin functionality date_unit : string, default None The timestamp unit to detect if converting dates. The default behaviour is to try and detect the correct precision, but if this is not desired then pass one of 's', 'ms', 'us' or 'ns' to force parsing only seconds, milliseconds, microseconds or nanoseconds respectively. encoding : str, default is 'utf-8' The encoding to use to decode py3 bytes. .. versionadded:: 0.19.0 lines : boolean, default False Read the file as a json object per line. .. versionadded:: 0.19.0 chunksize : integer, default None Return JsonReader object for iteration. See the `line-delimted json docs <http://pandas.pydata.org/pandas-docs/stable/io.html#io-jsonl>`_ for more information on ``chunksize``. This can only be passed if `lines=True`. If this is None, the file will be read into memory all at once. .. versionadded:: 0.21.0 compression : {'infer', 'gzip', 'bz2', 'zip', 'xz', None}, default 'infer' For on-the-fly decompression of on-disk data. If 'infer', then use gzip, bz2, zip or xz if path_or_buf is a string ending in '.gz', '.bz2', '.zip', or 'xz', respectively, and no decompression otherwise. If using 'zip', the ZIP file must contain only one data file to be read in. Set to None for no decompression. .. versionadded:: 0.21.0 Returns ------- result : Series or DataFrame, depending on the value of `typ`. See Also -------- DataFrame.to_json Notes ----- Specific to ``orient='table'``, if a :class:`DataFrame` with a literal :class:`Index` name of `index` gets written with :func:`to_json`, the subsequent read operation will incorrectly set the :class:`Index` name to ``None``. This is because `index` is also used by :func:`DataFrame.to_json` to denote a missing :class:`Index` name, and the subsequent :func:`read_json` operation cannot distinguish between the two. The same limitation is encountered with a :class:`MultiIndex` and any names beginning with ``'level_'``. Examples -------- >>> df = pd.DataFrame([['a', 'b'], ['c', 'd']], ... index=['row 1', 'row 2'], ... columns=['col 1', 'col 2']) Encoding/decoding a Dataframe using ``'split'`` formatted JSON: >>> df.to_json(orient='split') '{"columns":["col 1","col 2"], "index":["row 1","row 2"], "data":[["a","b"],["c","d"]]}' >>> pd.read_json(_, orient='split') col 1 col 2 row 1 a b row 2 c d Encoding/decoding a Dataframe using ``'index'`` formatted JSON: >>> df.to_json(orient='index') '{"row 1":{"col 1":"a","col 2":"b"},"row 2":{"col 1":"c","col 2":"d"}}' >>> pd.read_json(_, orient='index') col 1 col 2 row 1 a b row 2 c d Encoding/decoding a Dataframe using ``'records'`` formatted JSON. Note that index labels are not preserved with this encoding. >>> df.to_json(orient='records') '[{"col 1":"a","col 2":"b"},{"col 1":"c","col 2":"d"}]' >>> pd.read_json(_, orient='records') col 1 col 2 0 a b 1 c d Encoding with Table Schema >>> df.to_json(orient='table') '{"schema": {"fields": [{"name": "index", "type": "string"}, {"name": "col 1", "type": "string"}, {"name": "col 2", "type": "string"}], "primaryKey": "index", "pandas_version": "0.20.0"}, "data": [{"index": "row 1", "col 1": "a", "col 2": "b"}, {"index": "row 2", "col 1": "c", "col 2": "d"}]}'
pandas/io/json/json.py
def read_json(path_or_buf=None, orient=None, typ='frame', dtype=None, convert_axes=None, convert_dates=True, keep_default_dates=True, numpy=False, precise_float=False, date_unit=None, encoding=None, lines=False, chunksize=None, compression='infer'): """ Convert a JSON string to pandas object. Parameters ---------- path_or_buf : a valid JSON string or file-like, default: None The string could be a URL. Valid URL schemes include http, ftp, s3, gcs, and file. For file URLs, a host is expected. For instance, a local file could be ``file://localhost/path/to/table.json`` orient : string, Indication of expected JSON string format. Compatible JSON strings can be produced by ``to_json()`` with a corresponding orient value. The set of possible orients is: - ``'split'`` : dict like ``{index -> [index], columns -> [columns], data -> [values]}`` - ``'records'`` : list like ``[{column -> value}, ... , {column -> value}]`` - ``'index'`` : dict like ``{index -> {column -> value}}`` - ``'columns'`` : dict like ``{column -> {index -> value}}`` - ``'values'`` : just the values array The allowed and default values depend on the value of the `typ` parameter. * when ``typ == 'series'``, - allowed orients are ``{'split','records','index'}`` - default is ``'index'`` - The Series index must be unique for orient ``'index'``. * when ``typ == 'frame'``, - allowed orients are ``{'split','records','index', 'columns','values', 'table'}`` - default is ``'columns'`` - The DataFrame index must be unique for orients ``'index'`` and ``'columns'``. - The DataFrame columns must be unique for orients ``'index'``, ``'columns'``, and ``'records'``. .. versionadded:: 0.23.0 'table' as an allowed value for the ``orient`` argument typ : type of object to recover (series or frame), default 'frame' dtype : boolean or dict, default None If True, infer dtypes; if a dict of column to dtype, then use those; if False, then don't infer dtypes at all, applies only to the data. For all ``orient`` values except ``'table'``, default is True. .. versionchanged:: 0.25.0 Not applicable for ``orient='table'``. convert_axes : boolean, default None Try to convert the axes to the proper dtypes. For all ``orient`` values except ``'table'``, default is True. .. versionchanged:: 0.25.0 Not applicable for ``orient='table'``. convert_dates : boolean, default True List of columns to parse for dates; If True, then try to parse datelike columns default is True; a column label is datelike if * it ends with ``'_at'``, * it ends with ``'_time'``, * it begins with ``'timestamp'``, * it is ``'modified'``, or * it is ``'date'`` keep_default_dates : boolean, default True If parsing dates, then parse the default datelike columns numpy : boolean, default False Direct decoding to numpy arrays. Supports numeric data only, but non-numeric column and index labels are supported. Note also that the JSON ordering MUST be the same for each term if numpy=True. precise_float : boolean, default False Set to enable usage of higher precision (strtod) function when decoding string to double values. Default (False) is to use fast but less precise builtin functionality date_unit : string, default None The timestamp unit to detect if converting dates. The default behaviour is to try and detect the correct precision, but if this is not desired then pass one of 's', 'ms', 'us' or 'ns' to force parsing only seconds, milliseconds, microseconds or nanoseconds respectively. encoding : str, default is 'utf-8' The encoding to use to decode py3 bytes. .. versionadded:: 0.19.0 lines : boolean, default False Read the file as a json object per line. .. versionadded:: 0.19.0 chunksize : integer, default None Return JsonReader object for iteration. See the `line-delimted json docs <http://pandas.pydata.org/pandas-docs/stable/io.html#io-jsonl>`_ for more information on ``chunksize``. This can only be passed if `lines=True`. If this is None, the file will be read into memory all at once. .. versionadded:: 0.21.0 compression : {'infer', 'gzip', 'bz2', 'zip', 'xz', None}, default 'infer' For on-the-fly decompression of on-disk data. If 'infer', then use gzip, bz2, zip or xz if path_or_buf is a string ending in '.gz', '.bz2', '.zip', or 'xz', respectively, and no decompression otherwise. If using 'zip', the ZIP file must contain only one data file to be read in. Set to None for no decompression. .. versionadded:: 0.21.0 Returns ------- result : Series or DataFrame, depending on the value of `typ`. See Also -------- DataFrame.to_json Notes ----- Specific to ``orient='table'``, if a :class:`DataFrame` with a literal :class:`Index` name of `index` gets written with :func:`to_json`, the subsequent read operation will incorrectly set the :class:`Index` name to ``None``. This is because `index` is also used by :func:`DataFrame.to_json` to denote a missing :class:`Index` name, and the subsequent :func:`read_json` operation cannot distinguish between the two. The same limitation is encountered with a :class:`MultiIndex` and any names beginning with ``'level_'``. Examples -------- >>> df = pd.DataFrame([['a', 'b'], ['c', 'd']], ... index=['row 1', 'row 2'], ... columns=['col 1', 'col 2']) Encoding/decoding a Dataframe using ``'split'`` formatted JSON: >>> df.to_json(orient='split') '{"columns":["col 1","col 2"], "index":["row 1","row 2"], "data":[["a","b"],["c","d"]]}' >>> pd.read_json(_, orient='split') col 1 col 2 row 1 a b row 2 c d Encoding/decoding a Dataframe using ``'index'`` formatted JSON: >>> df.to_json(orient='index') '{"row 1":{"col 1":"a","col 2":"b"},"row 2":{"col 1":"c","col 2":"d"}}' >>> pd.read_json(_, orient='index') col 1 col 2 row 1 a b row 2 c d Encoding/decoding a Dataframe using ``'records'`` formatted JSON. Note that index labels are not preserved with this encoding. >>> df.to_json(orient='records') '[{"col 1":"a","col 2":"b"},{"col 1":"c","col 2":"d"}]' >>> pd.read_json(_, orient='records') col 1 col 2 0 a b 1 c d Encoding with Table Schema >>> df.to_json(orient='table') '{"schema": {"fields": [{"name": "index", "type": "string"}, {"name": "col 1", "type": "string"}, {"name": "col 2", "type": "string"}], "primaryKey": "index", "pandas_version": "0.20.0"}, "data": [{"index": "row 1", "col 1": "a", "col 2": "b"}, {"index": "row 2", "col 1": "c", "col 2": "d"}]}' """ if orient == 'table' and dtype: raise ValueError("cannot pass both dtype and orient='table'") if orient == 'table' and convert_axes: raise ValueError("cannot pass both convert_axes and orient='table'") if dtype is None and orient != 'table': dtype = True if convert_axes is None and orient != 'table': convert_axes = True compression = _infer_compression(path_or_buf, compression) filepath_or_buffer, _, compression, should_close = get_filepath_or_buffer( path_or_buf, encoding=encoding, compression=compression, ) json_reader = JsonReader( filepath_or_buffer, orient=orient, typ=typ, dtype=dtype, convert_axes=convert_axes, convert_dates=convert_dates, keep_default_dates=keep_default_dates, numpy=numpy, precise_float=precise_float, date_unit=date_unit, encoding=encoding, lines=lines, chunksize=chunksize, compression=compression, ) if chunksize: return json_reader result = json_reader.read() if should_close: try: filepath_or_buffer.close() except: # noqa: flake8 pass return result
def read_json(path_or_buf=None, orient=None, typ='frame', dtype=None, convert_axes=None, convert_dates=True, keep_default_dates=True, numpy=False, precise_float=False, date_unit=None, encoding=None, lines=False, chunksize=None, compression='infer'): """ Convert a JSON string to pandas object. Parameters ---------- path_or_buf : a valid JSON string or file-like, default: None The string could be a URL. Valid URL schemes include http, ftp, s3, gcs, and file. For file URLs, a host is expected. For instance, a local file could be ``file://localhost/path/to/table.json`` orient : string, Indication of expected JSON string format. Compatible JSON strings can be produced by ``to_json()`` with a corresponding orient value. The set of possible orients is: - ``'split'`` : dict like ``{index -> [index], columns -> [columns], data -> [values]}`` - ``'records'`` : list like ``[{column -> value}, ... , {column -> value}]`` - ``'index'`` : dict like ``{index -> {column -> value}}`` - ``'columns'`` : dict like ``{column -> {index -> value}}`` - ``'values'`` : just the values array The allowed and default values depend on the value of the `typ` parameter. * when ``typ == 'series'``, - allowed orients are ``{'split','records','index'}`` - default is ``'index'`` - The Series index must be unique for orient ``'index'``. * when ``typ == 'frame'``, - allowed orients are ``{'split','records','index', 'columns','values', 'table'}`` - default is ``'columns'`` - The DataFrame index must be unique for orients ``'index'`` and ``'columns'``. - The DataFrame columns must be unique for orients ``'index'``, ``'columns'``, and ``'records'``. .. versionadded:: 0.23.0 'table' as an allowed value for the ``orient`` argument typ : type of object to recover (series or frame), default 'frame' dtype : boolean or dict, default None If True, infer dtypes; if a dict of column to dtype, then use those; if False, then don't infer dtypes at all, applies only to the data. For all ``orient`` values except ``'table'``, default is True. .. versionchanged:: 0.25.0 Not applicable for ``orient='table'``. convert_axes : boolean, default None Try to convert the axes to the proper dtypes. For all ``orient`` values except ``'table'``, default is True. .. versionchanged:: 0.25.0 Not applicable for ``orient='table'``. convert_dates : boolean, default True List of columns to parse for dates; If True, then try to parse datelike columns default is True; a column label is datelike if * it ends with ``'_at'``, * it ends with ``'_time'``, * it begins with ``'timestamp'``, * it is ``'modified'``, or * it is ``'date'`` keep_default_dates : boolean, default True If parsing dates, then parse the default datelike columns numpy : boolean, default False Direct decoding to numpy arrays. Supports numeric data only, but non-numeric column and index labels are supported. Note also that the JSON ordering MUST be the same for each term if numpy=True. precise_float : boolean, default False Set to enable usage of higher precision (strtod) function when decoding string to double values. Default (False) is to use fast but less precise builtin functionality date_unit : string, default None The timestamp unit to detect if converting dates. The default behaviour is to try and detect the correct precision, but if this is not desired then pass one of 's', 'ms', 'us' or 'ns' to force parsing only seconds, milliseconds, microseconds or nanoseconds respectively. encoding : str, default is 'utf-8' The encoding to use to decode py3 bytes. .. versionadded:: 0.19.0 lines : boolean, default False Read the file as a json object per line. .. versionadded:: 0.19.0 chunksize : integer, default None Return JsonReader object for iteration. See the `line-delimted json docs <http://pandas.pydata.org/pandas-docs/stable/io.html#io-jsonl>`_ for more information on ``chunksize``. This can only be passed if `lines=True`. If this is None, the file will be read into memory all at once. .. versionadded:: 0.21.0 compression : {'infer', 'gzip', 'bz2', 'zip', 'xz', None}, default 'infer' For on-the-fly decompression of on-disk data. If 'infer', then use gzip, bz2, zip or xz if path_or_buf is a string ending in '.gz', '.bz2', '.zip', or 'xz', respectively, and no decompression otherwise. If using 'zip', the ZIP file must contain only one data file to be read in. Set to None for no decompression. .. versionadded:: 0.21.0 Returns ------- result : Series or DataFrame, depending on the value of `typ`. See Also -------- DataFrame.to_json Notes ----- Specific to ``orient='table'``, if a :class:`DataFrame` with a literal :class:`Index` name of `index` gets written with :func:`to_json`, the subsequent read operation will incorrectly set the :class:`Index` name to ``None``. This is because `index` is also used by :func:`DataFrame.to_json` to denote a missing :class:`Index` name, and the subsequent :func:`read_json` operation cannot distinguish between the two. The same limitation is encountered with a :class:`MultiIndex` and any names beginning with ``'level_'``. Examples -------- >>> df = pd.DataFrame([['a', 'b'], ['c', 'd']], ... index=['row 1', 'row 2'], ... columns=['col 1', 'col 2']) Encoding/decoding a Dataframe using ``'split'`` formatted JSON: >>> df.to_json(orient='split') '{"columns":["col 1","col 2"], "index":["row 1","row 2"], "data":[["a","b"],["c","d"]]}' >>> pd.read_json(_, orient='split') col 1 col 2 row 1 a b row 2 c d Encoding/decoding a Dataframe using ``'index'`` formatted JSON: >>> df.to_json(orient='index') '{"row 1":{"col 1":"a","col 2":"b"},"row 2":{"col 1":"c","col 2":"d"}}' >>> pd.read_json(_, orient='index') col 1 col 2 row 1 a b row 2 c d Encoding/decoding a Dataframe using ``'records'`` formatted JSON. Note that index labels are not preserved with this encoding. >>> df.to_json(orient='records') '[{"col 1":"a","col 2":"b"},{"col 1":"c","col 2":"d"}]' >>> pd.read_json(_, orient='records') col 1 col 2 0 a b 1 c d Encoding with Table Schema >>> df.to_json(orient='table') '{"schema": {"fields": [{"name": "index", "type": "string"}, {"name": "col 1", "type": "string"}, {"name": "col 2", "type": "string"}], "primaryKey": "index", "pandas_version": "0.20.0"}, "data": [{"index": "row 1", "col 1": "a", "col 2": "b"}, {"index": "row 2", "col 1": "c", "col 2": "d"}]}' """ if orient == 'table' and dtype: raise ValueError("cannot pass both dtype and orient='table'") if orient == 'table' and convert_axes: raise ValueError("cannot pass both convert_axes and orient='table'") if dtype is None and orient != 'table': dtype = True if convert_axes is None and orient != 'table': convert_axes = True compression = _infer_compression(path_or_buf, compression) filepath_or_buffer, _, compression, should_close = get_filepath_or_buffer( path_or_buf, encoding=encoding, compression=compression, ) json_reader = JsonReader( filepath_or_buffer, orient=orient, typ=typ, dtype=dtype, convert_axes=convert_axes, convert_dates=convert_dates, keep_default_dates=keep_default_dates, numpy=numpy, precise_float=precise_float, date_unit=date_unit, encoding=encoding, lines=lines, chunksize=chunksize, compression=compression, ) if chunksize: return json_reader result = json_reader.read() if should_close: try: filepath_or_buffer.close() except: # noqa: flake8 pass return result
[ "Convert", "a", "JSON", "string", "to", "pandas", "object", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L222-L450
[ "def", "read_json", "(", "path_or_buf", "=", "None", ",", "orient", "=", "None", ",", "typ", "=", "'frame'", ",", "dtype", "=", "None", ",", "convert_axes", "=", "None", ",", "convert_dates", "=", "True", ",", "keep_default_dates", "=", "True", ",", "numpy", "=", "False", ",", "precise_float", "=", "False", ",", "date_unit", "=", "None", ",", "encoding", "=", "None", ",", "lines", "=", "False", ",", "chunksize", "=", "None", ",", "compression", "=", "'infer'", ")", ":", "if", "orient", "==", "'table'", "and", "dtype", ":", "raise", "ValueError", "(", "\"cannot pass both dtype and orient='table'\"", ")", "if", "orient", "==", "'table'", "and", "convert_axes", ":", "raise", "ValueError", "(", "\"cannot pass both convert_axes and orient='table'\"", ")", "if", "dtype", "is", "None", "and", "orient", "!=", "'table'", ":", "dtype", "=", "True", "if", "convert_axes", "is", "None", "and", "orient", "!=", "'table'", ":", "convert_axes", "=", "True", "compression", "=", "_infer_compression", "(", "path_or_buf", ",", "compression", ")", "filepath_or_buffer", ",", "_", ",", "compression", ",", "should_close", "=", "get_filepath_or_buffer", "(", "path_or_buf", ",", "encoding", "=", "encoding", ",", "compression", "=", "compression", ",", ")", "json_reader", "=", "JsonReader", "(", "filepath_or_buffer", ",", "orient", "=", "orient", ",", "typ", "=", "typ", ",", "dtype", "=", "dtype", ",", "convert_axes", "=", "convert_axes", ",", "convert_dates", "=", "convert_dates", ",", "keep_default_dates", "=", "keep_default_dates", ",", "numpy", "=", "numpy", ",", "precise_float", "=", "precise_float", ",", "date_unit", "=", "date_unit", ",", "encoding", "=", "encoding", ",", "lines", "=", "lines", ",", "chunksize", "=", "chunksize", ",", "compression", "=", "compression", ",", ")", "if", "chunksize", ":", "return", "json_reader", "result", "=", "json_reader", ".", "read", "(", ")", "if", "should_close", ":", "try", ":", "filepath_or_buffer", ".", "close", "(", ")", "except", ":", "# noqa: flake8", "pass", "return", "result" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
FrameWriter._format_axes
Try to format axes if they are datelike.
pandas/io/json/json.py
def _format_axes(self): """ Try to format axes if they are datelike. """ if not self.obj.index.is_unique and self.orient in ( 'index', 'columns'): raise ValueError("DataFrame index must be unique for orient=" "'{orient}'.".format(orient=self.orient)) if not self.obj.columns.is_unique and self.orient in ( 'index', 'columns', 'records'): raise ValueError("DataFrame columns must be unique for orient=" "'{orient}'.".format(orient=self.orient))
def _format_axes(self): """ Try to format axes if they are datelike. """ if not self.obj.index.is_unique and self.orient in ( 'index', 'columns'): raise ValueError("DataFrame index must be unique for orient=" "'{orient}'.".format(orient=self.orient)) if not self.obj.columns.is_unique and self.orient in ( 'index', 'columns', 'records'): raise ValueError("DataFrame columns must be unique for orient=" "'{orient}'.".format(orient=self.orient))
[ "Try", "to", "format", "axes", "if", "they", "are", "datelike", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L138-L149
[ "def", "_format_axes", "(", "self", ")", ":", "if", "not", "self", ".", "obj", ".", "index", ".", "is_unique", "and", "self", ".", "orient", "in", "(", "'index'", ",", "'columns'", ")", ":", "raise", "ValueError", "(", "\"DataFrame index must be unique for orient=\"", "\"'{orient}'.\"", ".", "format", "(", "orient", "=", "self", ".", "orient", ")", ")", "if", "not", "self", ".", "obj", ".", "columns", ".", "is_unique", "and", "self", ".", "orient", "in", "(", "'index'", ",", "'columns'", ",", "'records'", ")", ":", "raise", "ValueError", "(", "\"DataFrame columns must be unique for orient=\"", "\"'{orient}'.\"", ".", "format", "(", "orient", "=", "self", ".", "orient", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
JsonReader._preprocess_data
At this point, the data either has a `read` attribute (e.g. a file object or a StringIO) or is a string that is a JSON document. If self.chunksize, we prepare the data for the `__next__` method. Otherwise, we read it into memory for the `read` method.
pandas/io/json/json.py
def _preprocess_data(self, data): """ At this point, the data either has a `read` attribute (e.g. a file object or a StringIO) or is a string that is a JSON document. If self.chunksize, we prepare the data for the `__next__` method. Otherwise, we read it into memory for the `read` method. """ if hasattr(data, 'read') and not self.chunksize: data = data.read() if not hasattr(data, 'read') and self.chunksize: data = StringIO(data) return data
def _preprocess_data(self, data): """ At this point, the data either has a `read` attribute (e.g. a file object or a StringIO) or is a string that is a JSON document. If self.chunksize, we prepare the data for the `__next__` method. Otherwise, we read it into memory for the `read` method. """ if hasattr(data, 'read') and not self.chunksize: data = data.read() if not hasattr(data, 'read') and self.chunksize: data = StringIO(data) return data
[ "At", "this", "point", "the", "data", "either", "has", "a", "read", "attribute", "(", "e", ".", "g", ".", "a", "file", "object", "or", "a", "StringIO", ")", "or", "is", "a", "string", "that", "is", "a", "JSON", "document", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L490-L503
[ "def", "_preprocess_data", "(", "self", ",", "data", ")", ":", "if", "hasattr", "(", "data", ",", "'read'", ")", "and", "not", "self", ".", "chunksize", ":", "data", "=", "data", ".", "read", "(", ")", "if", "not", "hasattr", "(", "data", ",", "'read'", ")", "and", "self", ".", "chunksize", ":", "data", "=", "StringIO", "(", "data", ")", "return", "data" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
JsonReader._get_data_from_filepath
The function read_json accepts three input types: 1. filepath (string-like) 2. file-like object (e.g. open file object, StringIO) 3. JSON string This method turns (1) into (2) to simplify the rest of the processing. It returns input types (2) and (3) unchanged.
pandas/io/json/json.py
def _get_data_from_filepath(self, filepath_or_buffer): """ The function read_json accepts three input types: 1. filepath (string-like) 2. file-like object (e.g. open file object, StringIO) 3. JSON string This method turns (1) into (2) to simplify the rest of the processing. It returns input types (2) and (3) unchanged. """ data = filepath_or_buffer exists = False if isinstance(data, str): try: exists = os.path.exists(filepath_or_buffer) # gh-5874: if the filepath is too long will raise here except (TypeError, ValueError): pass if exists or self.compression is not None: data, _ = _get_handle(filepath_or_buffer, 'r', encoding=self.encoding, compression=self.compression) self.should_close = True self.open_stream = data return data
def _get_data_from_filepath(self, filepath_or_buffer): """ The function read_json accepts three input types: 1. filepath (string-like) 2. file-like object (e.g. open file object, StringIO) 3. JSON string This method turns (1) into (2) to simplify the rest of the processing. It returns input types (2) and (3) unchanged. """ data = filepath_or_buffer exists = False if isinstance(data, str): try: exists = os.path.exists(filepath_or_buffer) # gh-5874: if the filepath is too long will raise here except (TypeError, ValueError): pass if exists or self.compression is not None: data, _ = _get_handle(filepath_or_buffer, 'r', encoding=self.encoding, compression=self.compression) self.should_close = True self.open_stream = data return data
[ "The", "function", "read_json", "accepts", "three", "input", "types", ":", "1", ".", "filepath", "(", "string", "-", "like", ")", "2", ".", "file", "-", "like", "object", "(", "e", ".", "g", ".", "open", "file", "object", "StringIO", ")", "3", ".", "JSON", "string" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L505-L532
[ "def", "_get_data_from_filepath", "(", "self", ",", "filepath_or_buffer", ")", ":", "data", "=", "filepath_or_buffer", "exists", "=", "False", "if", "isinstance", "(", "data", ",", "str", ")", ":", "try", ":", "exists", "=", "os", ".", "path", ".", "exists", "(", "filepath_or_buffer", ")", "# gh-5874: if the filepath is too long will raise here", "except", "(", "TypeError", ",", "ValueError", ")", ":", "pass", "if", "exists", "or", "self", ".", "compression", "is", "not", "None", ":", "data", ",", "_", "=", "_get_handle", "(", "filepath_or_buffer", ",", "'r'", ",", "encoding", "=", "self", ".", "encoding", ",", "compression", "=", "self", ".", "compression", ")", "self", ".", "should_close", "=", "True", "self", ".", "open_stream", "=", "data", "return", "data" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
JsonReader._combine_lines
Combines a list of JSON objects into one JSON object.
pandas/io/json/json.py
def _combine_lines(self, lines): """ Combines a list of JSON objects into one JSON object. """ lines = filter(None, map(lambda x: x.strip(), lines)) return '[' + ','.join(lines) + ']'
def _combine_lines(self, lines): """ Combines a list of JSON objects into one JSON object. """ lines = filter(None, map(lambda x: x.strip(), lines)) return '[' + ','.join(lines) + ']'
[ "Combines", "a", "list", "of", "JSON", "objects", "into", "one", "JSON", "object", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L534-L539
[ "def", "_combine_lines", "(", "self", ",", "lines", ")", ":", "lines", "=", "filter", "(", "None", ",", "map", "(", "lambda", "x", ":", "x", ".", "strip", "(", ")", ",", "lines", ")", ")", "return", "'['", "+", "','", ".", "join", "(", "lines", ")", "+", "']'" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
JsonReader.read
Read the whole JSON input into a pandas object.
pandas/io/json/json.py
def read(self): """ Read the whole JSON input into a pandas object. """ if self.lines and self.chunksize: obj = concat(self) elif self.lines: data = to_str(self.data) obj = self._get_object_parser( self._combine_lines(data.split('\n')) ) else: obj = self._get_object_parser(self.data) self.close() return obj
def read(self): """ Read the whole JSON input into a pandas object. """ if self.lines and self.chunksize: obj = concat(self) elif self.lines: data = to_str(self.data) obj = self._get_object_parser( self._combine_lines(data.split('\n')) ) else: obj = self._get_object_parser(self.data) self.close() return obj
[ "Read", "the", "whole", "JSON", "input", "into", "a", "pandas", "object", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L541-L556
[ "def", "read", "(", "self", ")", ":", "if", "self", ".", "lines", "and", "self", ".", "chunksize", ":", "obj", "=", "concat", "(", "self", ")", "elif", "self", ".", "lines", ":", "data", "=", "to_str", "(", "self", ".", "data", ")", "obj", "=", "self", ".", "_get_object_parser", "(", "self", ".", "_combine_lines", "(", "data", ".", "split", "(", "'\\n'", ")", ")", ")", "else", ":", "obj", "=", "self", ".", "_get_object_parser", "(", "self", ".", "data", ")", "self", ".", "close", "(", ")", "return", "obj" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
JsonReader._get_object_parser
Parses a json document into a pandas object.
pandas/io/json/json.py
def _get_object_parser(self, json): """ Parses a json document into a pandas object. """ typ = self.typ dtype = self.dtype kwargs = { "orient": self.orient, "dtype": self.dtype, "convert_axes": self.convert_axes, "convert_dates": self.convert_dates, "keep_default_dates": self.keep_default_dates, "numpy": self.numpy, "precise_float": self.precise_float, "date_unit": self.date_unit } obj = None if typ == 'frame': obj = FrameParser(json, **kwargs).parse() if typ == 'series' or obj is None: if not isinstance(dtype, bool): kwargs['dtype'] = dtype obj = SeriesParser(json, **kwargs).parse() return obj
def _get_object_parser(self, json): """ Parses a json document into a pandas object. """ typ = self.typ dtype = self.dtype kwargs = { "orient": self.orient, "dtype": self.dtype, "convert_axes": self.convert_axes, "convert_dates": self.convert_dates, "keep_default_dates": self.keep_default_dates, "numpy": self.numpy, "precise_float": self.precise_float, "date_unit": self.date_unit } obj = None if typ == 'frame': obj = FrameParser(json, **kwargs).parse() if typ == 'series' or obj is None: if not isinstance(dtype, bool): kwargs['dtype'] = dtype obj = SeriesParser(json, **kwargs).parse() return obj
[ "Parses", "a", "json", "document", "into", "a", "pandas", "object", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L558-L580
[ "def", "_get_object_parser", "(", "self", ",", "json", ")", ":", "typ", "=", "self", ".", "typ", "dtype", "=", "self", ".", "dtype", "kwargs", "=", "{", "\"orient\"", ":", "self", ".", "orient", ",", "\"dtype\"", ":", "self", ".", "dtype", ",", "\"convert_axes\"", ":", "self", ".", "convert_axes", ",", "\"convert_dates\"", ":", "self", ".", "convert_dates", ",", "\"keep_default_dates\"", ":", "self", ".", "keep_default_dates", ",", "\"numpy\"", ":", "self", ".", "numpy", ",", "\"precise_float\"", ":", "self", ".", "precise_float", ",", "\"date_unit\"", ":", "self", ".", "date_unit", "}", "obj", "=", "None", "if", "typ", "==", "'frame'", ":", "obj", "=", "FrameParser", "(", "json", ",", "*", "*", "kwargs", ")", ".", "parse", "(", ")", "if", "typ", "==", "'series'", "or", "obj", "is", "None", ":", "if", "not", "isinstance", "(", "dtype", ",", "bool", ")", ":", "kwargs", "[", "'dtype'", "]", "=", "dtype", "obj", "=", "SeriesParser", "(", "json", ",", "*", "*", "kwargs", ")", ".", "parse", "(", ")", "return", "obj" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
Parser.check_keys_split
Checks that dict has only the appropriate keys for orient='split'.
pandas/io/json/json.py
def check_keys_split(self, decoded): """ Checks that dict has only the appropriate keys for orient='split'. """ bad_keys = set(decoded.keys()).difference(set(self._split_keys)) if bad_keys: bad_keys = ", ".join(bad_keys) raise ValueError("JSON data had unexpected key(s): {bad_keys}" .format(bad_keys=pprint_thing(bad_keys)))
def check_keys_split(self, decoded): """ Checks that dict has only the appropriate keys for orient='split'. """ bad_keys = set(decoded.keys()).difference(set(self._split_keys)) if bad_keys: bad_keys = ", ".join(bad_keys) raise ValueError("JSON data had unexpected key(s): {bad_keys}" .format(bad_keys=pprint_thing(bad_keys)))
[ "Checks", "that", "dict", "has", "only", "the", "appropriate", "keys", "for", "orient", "=", "split", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L651-L659
[ "def", "check_keys_split", "(", "self", ",", "decoded", ")", ":", "bad_keys", "=", "set", "(", "decoded", ".", "keys", "(", ")", ")", ".", "difference", "(", "set", "(", "self", ".", "_split_keys", ")", ")", "if", "bad_keys", ":", "bad_keys", "=", "\", \"", ".", "join", "(", "bad_keys", ")", "raise", "ValueError", "(", "\"JSON data had unexpected key(s): {bad_keys}\"", ".", "format", "(", "bad_keys", "=", "pprint_thing", "(", "bad_keys", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
Parser._convert_axes
Try to convert axes.
pandas/io/json/json.py
def _convert_axes(self): """ Try to convert axes. """ for axis in self.obj._AXIS_NUMBERS.keys(): new_axis, result = self._try_convert_data( axis, self.obj._get_axis(axis), use_dtypes=False, convert_dates=True) if result: setattr(self.obj, axis, new_axis)
def _convert_axes(self): """ Try to convert axes. """ for axis in self.obj._AXIS_NUMBERS.keys(): new_axis, result = self._try_convert_data( axis, self.obj._get_axis(axis), use_dtypes=False, convert_dates=True) if result: setattr(self.obj, axis, new_axis)
[ "Try", "to", "convert", "axes", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L678-L687
[ "def", "_convert_axes", "(", "self", ")", ":", "for", "axis", "in", "self", ".", "obj", ".", "_AXIS_NUMBERS", ".", "keys", "(", ")", ":", "new_axis", ",", "result", "=", "self", ".", "_try_convert_data", "(", "axis", ",", "self", ".", "obj", ".", "_get_axis", "(", "axis", ")", ",", "use_dtypes", "=", "False", ",", "convert_dates", "=", "True", ")", "if", "result", ":", "setattr", "(", "self", ".", "obj", ",", "axis", ",", "new_axis", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
FrameParser._process_converter
Take a conversion function and possibly recreate the frame.
pandas/io/json/json.py
def _process_converter(self, f, filt=None): """ Take a conversion function and possibly recreate the frame. """ if filt is None: filt = lambda col, c: True needs_new_obj = False new_obj = dict() for i, (col, c) in enumerate(self.obj.iteritems()): if filt(col, c): new_data, result = f(col, c) if result: c = new_data needs_new_obj = True new_obj[i] = c if needs_new_obj: # possibly handle dup columns new_obj = DataFrame(new_obj, index=self.obj.index) new_obj.columns = self.obj.columns self.obj = new_obj
def _process_converter(self, f, filt=None): """ Take a conversion function and possibly recreate the frame. """ if filt is None: filt = lambda col, c: True needs_new_obj = False new_obj = dict() for i, (col, c) in enumerate(self.obj.iteritems()): if filt(col, c): new_data, result = f(col, c) if result: c = new_data needs_new_obj = True new_obj[i] = c if needs_new_obj: # possibly handle dup columns new_obj = DataFrame(new_obj, index=self.obj.index) new_obj.columns = self.obj.columns self.obj = new_obj
[ "Take", "a", "conversion", "function", "and", "possibly", "recreate", "the", "frame", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/json/json.py#L904-L927
[ "def", "_process_converter", "(", "self", ",", "f", ",", "filt", "=", "None", ")", ":", "if", "filt", "is", "None", ":", "filt", "=", "lambda", "col", ",", "c", ":", "True", "needs_new_obj", "=", "False", "new_obj", "=", "dict", "(", ")", "for", "i", ",", "(", "col", ",", "c", ")", "in", "enumerate", "(", "self", ".", "obj", ".", "iteritems", "(", ")", ")", ":", "if", "filt", "(", "col", ",", "c", ")", ":", "new_data", ",", "result", "=", "f", "(", "col", ",", "c", ")", "if", "result", ":", "c", "=", "new_data", "needs_new_obj", "=", "True", "new_obj", "[", "i", "]", "=", "c", "if", "needs_new_obj", ":", "# possibly handle dup columns", "new_obj", "=", "DataFrame", "(", "new_obj", ",", "index", "=", "self", ".", "obj", ".", "index", ")", "new_obj", ".", "columns", "=", "self", ".", "obj", ".", "columns", "self", ".", "obj", "=", "new_obj" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
format_array
Format an array for printing. Parameters ---------- values formatter float_format na_rep digits space justify decimal leading_space : bool, optional Whether the array should be formatted with a leading space. When an array as a column of a Series or DataFrame, we do want the leading space to pad between columns. When formatting an Index subclass (e.g. IntervalIndex._format_native_types), we don't want the leading space since it should be left-aligned. Returns ------- List[str]
pandas/io/formats/format.py
def format_array(values, formatter, float_format=None, na_rep='NaN', digits=None, space=None, justify='right', decimal='.', leading_space=None): """ Format an array for printing. Parameters ---------- values formatter float_format na_rep digits space justify decimal leading_space : bool, optional Whether the array should be formatted with a leading space. When an array as a column of a Series or DataFrame, we do want the leading space to pad between columns. When formatting an Index subclass (e.g. IntervalIndex._format_native_types), we don't want the leading space since it should be left-aligned. Returns ------- List[str] """ if is_datetime64_dtype(values.dtype): fmt_klass = Datetime64Formatter elif is_datetime64tz_dtype(values): fmt_klass = Datetime64TZFormatter elif is_timedelta64_dtype(values.dtype): fmt_klass = Timedelta64Formatter elif is_extension_array_dtype(values.dtype): fmt_klass = ExtensionArrayFormatter elif is_float_dtype(values.dtype) or is_complex_dtype(values.dtype): fmt_klass = FloatArrayFormatter elif is_integer_dtype(values.dtype): fmt_klass = IntArrayFormatter else: fmt_klass = GenericArrayFormatter if space is None: space = get_option("display.column_space") if float_format is None: float_format = get_option("display.float_format") if digits is None: digits = get_option("display.precision") fmt_obj = fmt_klass(values, digits=digits, na_rep=na_rep, float_format=float_format, formatter=formatter, space=space, justify=justify, decimal=decimal, leading_space=leading_space) return fmt_obj.get_result()
def format_array(values, formatter, float_format=None, na_rep='NaN', digits=None, space=None, justify='right', decimal='.', leading_space=None): """ Format an array for printing. Parameters ---------- values formatter float_format na_rep digits space justify decimal leading_space : bool, optional Whether the array should be formatted with a leading space. When an array as a column of a Series or DataFrame, we do want the leading space to pad between columns. When formatting an Index subclass (e.g. IntervalIndex._format_native_types), we don't want the leading space since it should be left-aligned. Returns ------- List[str] """ if is_datetime64_dtype(values.dtype): fmt_klass = Datetime64Formatter elif is_datetime64tz_dtype(values): fmt_klass = Datetime64TZFormatter elif is_timedelta64_dtype(values.dtype): fmt_klass = Timedelta64Formatter elif is_extension_array_dtype(values.dtype): fmt_klass = ExtensionArrayFormatter elif is_float_dtype(values.dtype) or is_complex_dtype(values.dtype): fmt_klass = FloatArrayFormatter elif is_integer_dtype(values.dtype): fmt_klass = IntArrayFormatter else: fmt_klass = GenericArrayFormatter if space is None: space = get_option("display.column_space") if float_format is None: float_format = get_option("display.float_format") if digits is None: digits = get_option("display.precision") fmt_obj = fmt_klass(values, digits=digits, na_rep=na_rep, float_format=float_format, formatter=formatter, space=space, justify=justify, decimal=decimal, leading_space=leading_space) return fmt_obj.get_result()
[ "Format", "an", "array", "for", "printing", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L853-L912
[ "def", "format_array", "(", "values", ",", "formatter", ",", "float_format", "=", "None", ",", "na_rep", "=", "'NaN'", ",", "digits", "=", "None", ",", "space", "=", "None", ",", "justify", "=", "'right'", ",", "decimal", "=", "'.'", ",", "leading_space", "=", "None", ")", ":", "if", "is_datetime64_dtype", "(", "values", ".", "dtype", ")", ":", "fmt_klass", "=", "Datetime64Formatter", "elif", "is_datetime64tz_dtype", "(", "values", ")", ":", "fmt_klass", "=", "Datetime64TZFormatter", "elif", "is_timedelta64_dtype", "(", "values", ".", "dtype", ")", ":", "fmt_klass", "=", "Timedelta64Formatter", "elif", "is_extension_array_dtype", "(", "values", ".", "dtype", ")", ":", "fmt_klass", "=", "ExtensionArrayFormatter", "elif", "is_float_dtype", "(", "values", ".", "dtype", ")", "or", "is_complex_dtype", "(", "values", ".", "dtype", ")", ":", "fmt_klass", "=", "FloatArrayFormatter", "elif", "is_integer_dtype", "(", "values", ".", "dtype", ")", ":", "fmt_klass", "=", "IntArrayFormatter", "else", ":", "fmt_klass", "=", "GenericArrayFormatter", "if", "space", "is", "None", ":", "space", "=", "get_option", "(", "\"display.column_space\"", ")", "if", "float_format", "is", "None", ":", "float_format", "=", "get_option", "(", "\"display.float_format\"", ")", "if", "digits", "is", "None", ":", "digits", "=", "get_option", "(", "\"display.precision\"", ")", "fmt_obj", "=", "fmt_klass", "(", "values", ",", "digits", "=", "digits", ",", "na_rep", "=", "na_rep", ",", "float_format", "=", "float_format", ",", "formatter", "=", "formatter", ",", "space", "=", "space", ",", "justify", "=", "justify", ",", "decimal", "=", "decimal", ",", "leading_space", "=", "leading_space", ")", "return", "fmt_obj", ".", "get_result", "(", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
format_percentiles
Outputs rounded and formatted percentiles. Parameters ---------- percentiles : list-like, containing floats from interval [0,1] Returns ------- formatted : list of strings Notes ----- Rounding precision is chosen so that: (1) if any two elements of ``percentiles`` differ, they remain different after rounding (2) no entry is *rounded* to 0% or 100%. Any non-integer is always rounded to at least 1 decimal place. Examples -------- Keeps all entries different after rounding: >>> format_percentiles([0.01999, 0.02001, 0.5, 0.666666, 0.9999]) ['1.999%', '2.001%', '50%', '66.667%', '99.99%'] No element is rounded to 0% or 100% (unless already equal to it). Duplicates are allowed: >>> format_percentiles([0, 0.5, 0.02001, 0.5, 0.666666, 0.9999]) ['0%', '50%', '2.0%', '50%', '66.67%', '99.99%']
pandas/io/formats/format.py
def format_percentiles(percentiles): """ Outputs rounded and formatted percentiles. Parameters ---------- percentiles : list-like, containing floats from interval [0,1] Returns ------- formatted : list of strings Notes ----- Rounding precision is chosen so that: (1) if any two elements of ``percentiles`` differ, they remain different after rounding (2) no entry is *rounded* to 0% or 100%. Any non-integer is always rounded to at least 1 decimal place. Examples -------- Keeps all entries different after rounding: >>> format_percentiles([0.01999, 0.02001, 0.5, 0.666666, 0.9999]) ['1.999%', '2.001%', '50%', '66.667%', '99.99%'] No element is rounded to 0% or 100% (unless already equal to it). Duplicates are allowed: >>> format_percentiles([0, 0.5, 0.02001, 0.5, 0.666666, 0.9999]) ['0%', '50%', '2.0%', '50%', '66.67%', '99.99%'] """ percentiles = np.asarray(percentiles) # It checks for np.NaN as well with np.errstate(invalid='ignore'): if not is_numeric_dtype(percentiles) or not np.all(percentiles >= 0) \ or not np.all(percentiles <= 1): raise ValueError("percentiles should all be in the interval [0,1]") percentiles = 100 * percentiles int_idx = (percentiles.astype(int) == percentiles) if np.all(int_idx): out = percentiles.astype(int).astype(str) return [i + '%' for i in out] unique_pcts = np.unique(percentiles) to_begin = unique_pcts[0] if unique_pcts[0] > 0 else None to_end = 100 - unique_pcts[-1] if unique_pcts[-1] < 100 else None # Least precision that keeps percentiles unique after rounding prec = -np.floor(np.log10(np.min( np.ediff1d(unique_pcts, to_begin=to_begin, to_end=to_end) ))).astype(int) prec = max(1, prec) out = np.empty_like(percentiles, dtype=object) out[int_idx] = percentiles[int_idx].astype(int).astype(str) out[~int_idx] = percentiles[~int_idx].round(prec).astype(str) return [i + '%' for i in out]
def format_percentiles(percentiles): """ Outputs rounded and formatted percentiles. Parameters ---------- percentiles : list-like, containing floats from interval [0,1] Returns ------- formatted : list of strings Notes ----- Rounding precision is chosen so that: (1) if any two elements of ``percentiles`` differ, they remain different after rounding (2) no entry is *rounded* to 0% or 100%. Any non-integer is always rounded to at least 1 decimal place. Examples -------- Keeps all entries different after rounding: >>> format_percentiles([0.01999, 0.02001, 0.5, 0.666666, 0.9999]) ['1.999%', '2.001%', '50%', '66.667%', '99.99%'] No element is rounded to 0% or 100% (unless already equal to it). Duplicates are allowed: >>> format_percentiles([0, 0.5, 0.02001, 0.5, 0.666666, 0.9999]) ['0%', '50%', '2.0%', '50%', '66.67%', '99.99%'] """ percentiles = np.asarray(percentiles) # It checks for np.NaN as well with np.errstate(invalid='ignore'): if not is_numeric_dtype(percentiles) or not np.all(percentiles >= 0) \ or not np.all(percentiles <= 1): raise ValueError("percentiles should all be in the interval [0,1]") percentiles = 100 * percentiles int_idx = (percentiles.astype(int) == percentiles) if np.all(int_idx): out = percentiles.astype(int).astype(str) return [i + '%' for i in out] unique_pcts = np.unique(percentiles) to_begin = unique_pcts[0] if unique_pcts[0] > 0 else None to_end = 100 - unique_pcts[-1] if unique_pcts[-1] < 100 else None # Least precision that keeps percentiles unique after rounding prec = -np.floor(np.log10(np.min( np.ediff1d(unique_pcts, to_begin=to_begin, to_end=to_end) ))).astype(int) prec = max(1, prec) out = np.empty_like(percentiles, dtype=object) out[int_idx] = percentiles[int_idx].astype(int).astype(str) out[~int_idx] = percentiles[~int_idx].round(prec).astype(str) return [i + '%' for i in out]
[ "Outputs", "rounded", "and", "formatted", "percentiles", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1208-L1268
[ "def", "format_percentiles", "(", "percentiles", ")", ":", "percentiles", "=", "np", ".", "asarray", "(", "percentiles", ")", "# It checks for np.NaN as well", "with", "np", ".", "errstate", "(", "invalid", "=", "'ignore'", ")", ":", "if", "not", "is_numeric_dtype", "(", "percentiles", ")", "or", "not", "np", ".", "all", "(", "percentiles", ">=", "0", ")", "or", "not", "np", ".", "all", "(", "percentiles", "<=", "1", ")", ":", "raise", "ValueError", "(", "\"percentiles should all be in the interval [0,1]\"", ")", "percentiles", "=", "100", "*", "percentiles", "int_idx", "=", "(", "percentiles", ".", "astype", "(", "int", ")", "==", "percentiles", ")", "if", "np", ".", "all", "(", "int_idx", ")", ":", "out", "=", "percentiles", ".", "astype", "(", "int", ")", ".", "astype", "(", "str", ")", "return", "[", "i", "+", "'%'", "for", "i", "in", "out", "]", "unique_pcts", "=", "np", ".", "unique", "(", "percentiles", ")", "to_begin", "=", "unique_pcts", "[", "0", "]", "if", "unique_pcts", "[", "0", "]", ">", "0", "else", "None", "to_end", "=", "100", "-", "unique_pcts", "[", "-", "1", "]", "if", "unique_pcts", "[", "-", "1", "]", "<", "100", "else", "None", "# Least precision that keeps percentiles unique after rounding", "prec", "=", "-", "np", ".", "floor", "(", "np", ".", "log10", "(", "np", ".", "min", "(", "np", ".", "ediff1d", "(", "unique_pcts", ",", "to_begin", "=", "to_begin", ",", "to_end", "=", "to_end", ")", ")", ")", ")", ".", "astype", "(", "int", ")", "prec", "=", "max", "(", "1", ",", "prec", ")", "out", "=", "np", ".", "empty_like", "(", "percentiles", ",", "dtype", "=", "object", ")", "out", "[", "int_idx", "]", "=", "percentiles", "[", "int_idx", "]", ".", "astype", "(", "int", ")", ".", "astype", "(", "str", ")", "out", "[", "~", "int_idx", "]", "=", "percentiles", "[", "~", "int_idx", "]", ".", "round", "(", "prec", ")", ".", "astype", "(", "str", ")", "return", "[", "i", "+", "'%'", "for", "i", "in", "out", "]" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_get_format_timedelta64
Return a formatter function for a range of timedeltas. These will all have the same format argument If box, then show the return in quotes
pandas/io/formats/format.py
def _get_format_timedelta64(values, nat_rep='NaT', box=False): """ Return a formatter function for a range of timedeltas. These will all have the same format argument If box, then show the return in quotes """ values_int = values.astype(np.int64) consider_values = values_int != iNaT one_day_nanos = (86400 * 1e9) even_days = np.logical_and(consider_values, values_int % one_day_nanos != 0).sum() == 0 all_sub_day = np.logical_and( consider_values, np.abs(values_int) >= one_day_nanos).sum() == 0 if even_days: format = None elif all_sub_day: format = 'sub_day' else: format = 'long' def _formatter(x): if x is None or (is_scalar(x) and isna(x)): return nat_rep if not isinstance(x, Timedelta): x = Timedelta(x) result = x._repr_base(format=format) if box: result = "'{res}'".format(res=result) return result return _formatter
def _get_format_timedelta64(values, nat_rep='NaT', box=False): """ Return a formatter function for a range of timedeltas. These will all have the same format argument If box, then show the return in quotes """ values_int = values.astype(np.int64) consider_values = values_int != iNaT one_day_nanos = (86400 * 1e9) even_days = np.logical_and(consider_values, values_int % one_day_nanos != 0).sum() == 0 all_sub_day = np.logical_and( consider_values, np.abs(values_int) >= one_day_nanos).sum() == 0 if even_days: format = None elif all_sub_day: format = 'sub_day' else: format = 'long' def _formatter(x): if x is None or (is_scalar(x) and isna(x)): return nat_rep if not isinstance(x, Timedelta): x = Timedelta(x) result = x._repr_base(format=format) if box: result = "'{res}'".format(res=result) return result return _formatter
[ "Return", "a", "formatter", "function", "for", "a", "range", "of", "timedeltas", ".", "These", "will", "all", "have", "the", "same", "format", "argument" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1360-L1396
[ "def", "_get_format_timedelta64", "(", "values", ",", "nat_rep", "=", "'NaT'", ",", "box", "=", "False", ")", ":", "values_int", "=", "values", ".", "astype", "(", "np", ".", "int64", ")", "consider_values", "=", "values_int", "!=", "iNaT", "one_day_nanos", "=", "(", "86400", "*", "1e9", ")", "even_days", "=", "np", ".", "logical_and", "(", "consider_values", ",", "values_int", "%", "one_day_nanos", "!=", "0", ")", ".", "sum", "(", ")", "==", "0", "all_sub_day", "=", "np", ".", "logical_and", "(", "consider_values", ",", "np", ".", "abs", "(", "values_int", ")", ">=", "one_day_nanos", ")", ".", "sum", "(", ")", "==", "0", "if", "even_days", ":", "format", "=", "None", "elif", "all_sub_day", ":", "format", "=", "'sub_day'", "else", ":", "format", "=", "'long'", "def", "_formatter", "(", "x", ")", ":", "if", "x", "is", "None", "or", "(", "is_scalar", "(", "x", ")", "and", "isna", "(", "x", ")", ")", ":", "return", "nat_rep", "if", "not", "isinstance", "(", "x", ",", "Timedelta", ")", ":", "x", "=", "Timedelta", "(", "x", ")", "result", "=", "x", ".", "_repr_base", "(", "format", "=", "format", ")", "if", "box", ":", "result", "=", "\"'{res}'\"", ".", "format", "(", "res", "=", "result", ")", "return", "result", "return", "_formatter" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_trim_zeros_complex
Separates the real and imaginary parts from the complex number, and executes the _trim_zeros_float method on each of those.
pandas/io/formats/format.py
def _trim_zeros_complex(str_complexes, na_rep='NaN'): """ Separates the real and imaginary parts from the complex number, and executes the _trim_zeros_float method on each of those. """ def separate_and_trim(str_complex, na_rep): num_arr = str_complex.split('+') return (_trim_zeros_float([num_arr[0]], na_rep) + ['+'] + _trim_zeros_float([num_arr[1][:-1]], na_rep) + ['j']) return [''.join(separate_and_trim(x, na_rep)) for x in str_complexes]
def _trim_zeros_complex(str_complexes, na_rep='NaN'): """ Separates the real and imaginary parts from the complex number, and executes the _trim_zeros_float method on each of those. """ def separate_and_trim(str_complex, na_rep): num_arr = str_complex.split('+') return (_trim_zeros_float([num_arr[0]], na_rep) + ['+'] + _trim_zeros_float([num_arr[1][:-1]], na_rep) + ['j']) return [''.join(separate_and_trim(x, na_rep)) for x in str_complexes]
[ "Separates", "the", "real", "and", "imaginary", "parts", "from", "the", "complex", "number", "and", "executes", "the", "_trim_zeros_float", "method", "on", "each", "of", "those", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1427-L1439
[ "def", "_trim_zeros_complex", "(", "str_complexes", ",", "na_rep", "=", "'NaN'", ")", ":", "def", "separate_and_trim", "(", "str_complex", ",", "na_rep", ")", ":", "num_arr", "=", "str_complex", ".", "split", "(", "'+'", ")", "return", "(", "_trim_zeros_float", "(", "[", "num_arr", "[", "0", "]", "]", ",", "na_rep", ")", "+", "[", "'+'", "]", "+", "_trim_zeros_float", "(", "[", "num_arr", "[", "1", "]", "[", ":", "-", "1", "]", "]", ",", "na_rep", ")", "+", "[", "'j'", "]", ")", "return", "[", "''", ".", "join", "(", "separate_and_trim", "(", "x", ",", "na_rep", ")", ")", "for", "x", "in", "str_complexes", "]" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_trim_zeros_float
Trims zeros, leaving just one before the decimal points if need be.
pandas/io/formats/format.py
def _trim_zeros_float(str_floats, na_rep='NaN'): """ Trims zeros, leaving just one before the decimal points if need be. """ trimmed = str_floats def _is_number(x): return (x != na_rep and not x.endswith('inf')) def _cond(values): finite = [x for x in values if _is_number(x)] return (len(finite) > 0 and all(x.endswith('0') for x in finite) and not (any(('e' in x) or ('E' in x) for x in finite))) while _cond(trimmed): trimmed = [x[:-1] if _is_number(x) else x for x in trimmed] # leave one 0 after the decimal points if need be. return [x + "0" if x.endswith('.') and _is_number(x) else x for x in trimmed]
def _trim_zeros_float(str_floats, na_rep='NaN'): """ Trims zeros, leaving just one before the decimal points if need be. """ trimmed = str_floats def _is_number(x): return (x != na_rep and not x.endswith('inf')) def _cond(values): finite = [x for x in values if _is_number(x)] return (len(finite) > 0 and all(x.endswith('0') for x in finite) and not (any(('e' in x) or ('E' in x) for x in finite))) while _cond(trimmed): trimmed = [x[:-1] if _is_number(x) else x for x in trimmed] # leave one 0 after the decimal points if need be. return [x + "0" if x.endswith('.') and _is_number(x) else x for x in trimmed]
[ "Trims", "zeros", "leaving", "just", "one", "before", "the", "decimal", "points", "if", "need", "be", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1442-L1461
[ "def", "_trim_zeros_float", "(", "str_floats", ",", "na_rep", "=", "'NaN'", ")", ":", "trimmed", "=", "str_floats", "def", "_is_number", "(", "x", ")", ":", "return", "(", "x", "!=", "na_rep", "and", "not", "x", ".", "endswith", "(", "'inf'", ")", ")", "def", "_cond", "(", "values", ")", ":", "finite", "=", "[", "x", "for", "x", "in", "values", "if", "_is_number", "(", "x", ")", "]", "return", "(", "len", "(", "finite", ")", ">", "0", "and", "all", "(", "x", ".", "endswith", "(", "'0'", ")", "for", "x", "in", "finite", ")", "and", "not", "(", "any", "(", "(", "'e'", "in", "x", ")", "or", "(", "'E'", "in", "x", ")", "for", "x", "in", "finite", ")", ")", ")", "while", "_cond", "(", "trimmed", ")", ":", "trimmed", "=", "[", "x", "[", ":", "-", "1", "]", "if", "_is_number", "(", "x", ")", "else", "x", "for", "x", "in", "trimmed", "]", "# leave one 0 after the decimal points if need be.", "return", "[", "x", "+", "\"0\"", "if", "x", ".", "endswith", "(", "'.'", ")", "and", "_is_number", "(", "x", ")", "else", "x", "for", "x", "in", "trimmed", "]" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
set_eng_float_format
Alter default behavior on how float is formatted in DataFrame. Format float in engineering format. By accuracy, we mean the number of decimal digits after the floating point. See also EngFormatter.
pandas/io/formats/format.py
def set_eng_float_format(accuracy=3, use_eng_prefix=False): """ Alter default behavior on how float is formatted in DataFrame. Format float in engineering format. By accuracy, we mean the number of decimal digits after the floating point. See also EngFormatter. """ set_option("display.float_format", EngFormatter(accuracy, use_eng_prefix)) set_option("display.column_space", max(12, accuracy + 9))
def set_eng_float_format(accuracy=3, use_eng_prefix=False): """ Alter default behavior on how float is formatted in DataFrame. Format float in engineering format. By accuracy, we mean the number of decimal digits after the floating point. See also EngFormatter. """ set_option("display.float_format", EngFormatter(accuracy, use_eng_prefix)) set_option("display.column_space", max(12, accuracy + 9))
[ "Alter", "default", "behavior", "on", "how", "float", "is", "formatted", "in", "DataFrame", ".", "Format", "float", "in", "engineering", "format", ".", "By", "accuracy", "we", "mean", "the", "number", "of", "decimal", "digits", "after", "the", "floating", "point", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1570-L1580
[ "def", "set_eng_float_format", "(", "accuracy", "=", "3", ",", "use_eng_prefix", "=", "False", ")", ":", "set_option", "(", "\"display.float_format\"", ",", "EngFormatter", "(", "accuracy", ",", "use_eng_prefix", ")", ")", "set_option", "(", "\"display.column_space\"", ",", "max", "(", "12", ",", "accuracy", "+", "9", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
get_level_lengths
For each index in each level the function returns lengths of indexes. Parameters ---------- levels : list of lists List of values on for level. sentinel : string, optional Value which states that no new index starts on there. Returns ---------- Returns list of maps. For each level returns map of indexes (key is index in row and value is length of index).
pandas/io/formats/format.py
def get_level_lengths(levels, sentinel=''): """For each index in each level the function returns lengths of indexes. Parameters ---------- levels : list of lists List of values on for level. sentinel : string, optional Value which states that no new index starts on there. Returns ---------- Returns list of maps. For each level returns map of indexes (key is index in row and value is length of index). """ if len(levels) == 0: return [] control = [True] * len(levels[0]) result = [] for level in levels: last_index = 0 lengths = {} for i, key in enumerate(level): if control[i] and key == sentinel: pass else: control[i] = False lengths[last_index] = i - last_index last_index = i lengths[last_index] = len(level) - last_index result.append(lengths) return result
def get_level_lengths(levels, sentinel=''): """For each index in each level the function returns lengths of indexes. Parameters ---------- levels : list of lists List of values on for level. sentinel : string, optional Value which states that no new index starts on there. Returns ---------- Returns list of maps. For each level returns map of indexes (key is index in row and value is length of index). """ if len(levels) == 0: return [] control = [True] * len(levels[0]) result = [] for level in levels: last_index = 0 lengths = {} for i, key in enumerate(level): if control[i] and key == sentinel: pass else: control[i] = False lengths[last_index] = i - last_index last_index = i lengths[last_index] = len(level) - last_index result.append(lengths) return result
[ "For", "each", "index", "in", "each", "level", "the", "function", "returns", "lengths", "of", "indexes", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1603-L1640
[ "def", "get_level_lengths", "(", "levels", ",", "sentinel", "=", "''", ")", ":", "if", "len", "(", "levels", ")", "==", "0", ":", "return", "[", "]", "control", "=", "[", "True", "]", "*", "len", "(", "levels", "[", "0", "]", ")", "result", "=", "[", "]", "for", "level", "in", "levels", ":", "last_index", "=", "0", "lengths", "=", "{", "}", "for", "i", ",", "key", "in", "enumerate", "(", "level", ")", ":", "if", "control", "[", "i", "]", "and", "key", "==", "sentinel", ":", "pass", "else", ":", "control", "[", "i", "]", "=", "False", "lengths", "[", "last_index", "]", "=", "i", "-", "last_index", "last_index", "=", "i", "lengths", "[", "last_index", "]", "=", "len", "(", "level", ")", "-", "last_index", "result", ".", "append", "(", "lengths", ")", "return", "result" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
buffer_put_lines
Appends lines to a buffer. Parameters ---------- buf The buffer to write to lines The lines to append.
pandas/io/formats/format.py
def buffer_put_lines(buf, lines): """ Appends lines to a buffer. Parameters ---------- buf The buffer to write to lines The lines to append. """ if any(isinstance(x, str) for x in lines): lines = [str(x) for x in lines] buf.write('\n'.join(lines))
def buffer_put_lines(buf, lines): """ Appends lines to a buffer. Parameters ---------- buf The buffer to write to lines The lines to append. """ if any(isinstance(x, str) for x in lines): lines = [str(x) for x in lines] buf.write('\n'.join(lines))
[ "Appends", "lines", "to", "a", "buffer", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1643-L1656
[ "def", "buffer_put_lines", "(", "buf", ",", "lines", ")", ":", "if", "any", "(", "isinstance", "(", "x", ",", "str", ")", "for", "x", "in", "lines", ")", ":", "lines", "=", "[", "str", "(", "x", ")", "for", "x", "in", "lines", "]", "buf", ".", "write", "(", "'\\n'", ".", "join", "(", "lines", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
EastAsianTextAdjustment.len
Calculate display width considering unicode East Asian Width
pandas/io/formats/format.py
def len(self, text): """ Calculate display width considering unicode East Asian Width """ if not isinstance(text, str): return len(text) return sum(self._EAW_MAP.get(east_asian_width(c), self.ambiguous_width) for c in text)
def len(self, text): """ Calculate display width considering unicode East Asian Width """ if not isinstance(text, str): return len(text) return sum(self._EAW_MAP.get(east_asian_width(c), self.ambiguous_width) for c in text)
[ "Calculate", "display", "width", "considering", "unicode", "East", "Asian", "Width" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L322-L330
[ "def", "len", "(", "self", ",", "text", ")", ":", "if", "not", "isinstance", "(", "text", ",", "str", ")", ":", "return", "len", "(", "text", ")", "return", "sum", "(", "self", ".", "_EAW_MAP", ".", "get", "(", "east_asian_width", "(", "c", ")", ",", "self", ".", "ambiguous_width", ")", "for", "c", "in", "text", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
DataFrameFormatter._to_str_columns
Render a DataFrame to a list of columns (as lists of strings).
pandas/io/formats/format.py
def _to_str_columns(self): """ Render a DataFrame to a list of columns (as lists of strings). """ frame = self.tr_frame # may include levels names also str_index = self._get_formatted_index(frame) if not is_list_like(self.header) and not self.header: stringified = [] for i, c in enumerate(frame): fmt_values = self._format_col(i) fmt_values = _make_fixed_width(fmt_values, self.justify, minimum=(self.col_space or 0), adj=self.adj) stringified.append(fmt_values) else: if is_list_like(self.header): if len(self.header) != len(self.columns): raise ValueError(('Writing {ncols} cols but got {nalias} ' 'aliases' .format(ncols=len(self.columns), nalias=len(self.header)))) str_columns = [[label] for label in self.header] else: str_columns = self._get_formatted_column_labels(frame) if self.show_row_idx_names: for x in str_columns: x.append('') stringified = [] for i, c in enumerate(frame): cheader = str_columns[i] header_colwidth = max(self.col_space or 0, *(self.adj.len(x) for x in cheader)) fmt_values = self._format_col(i) fmt_values = _make_fixed_width(fmt_values, self.justify, minimum=header_colwidth, adj=self.adj) max_len = max(max(self.adj.len(x) for x in fmt_values), header_colwidth) cheader = self.adj.justify(cheader, max_len, mode=self.justify) stringified.append(cheader + fmt_values) strcols = stringified if self.index: strcols.insert(0, str_index) # Add ... to signal truncated truncate_h = self.truncate_h truncate_v = self.truncate_v if truncate_h: col_num = self.tr_col_num strcols.insert(self.tr_col_num + 1, [' ...'] * (len(str_index))) if truncate_v: n_header_rows = len(str_index) - len(frame) row_num = self.tr_row_num for ix, col in enumerate(strcols): # infer from above row cwidth = self.adj.len(strcols[ix][row_num]) is_dot_col = False if truncate_h: is_dot_col = ix == col_num + 1 if cwidth > 3 or is_dot_col: my_str = '...' else: my_str = '..' if ix == 0: dot_mode = 'left' elif is_dot_col: cwidth = 4 dot_mode = 'right' else: dot_mode = 'right' dot_str = self.adj.justify([my_str], cwidth, mode=dot_mode)[0] strcols[ix].insert(row_num + n_header_rows, dot_str) return strcols
def _to_str_columns(self): """ Render a DataFrame to a list of columns (as lists of strings). """ frame = self.tr_frame # may include levels names also str_index = self._get_formatted_index(frame) if not is_list_like(self.header) and not self.header: stringified = [] for i, c in enumerate(frame): fmt_values = self._format_col(i) fmt_values = _make_fixed_width(fmt_values, self.justify, minimum=(self.col_space or 0), adj=self.adj) stringified.append(fmt_values) else: if is_list_like(self.header): if len(self.header) != len(self.columns): raise ValueError(('Writing {ncols} cols but got {nalias} ' 'aliases' .format(ncols=len(self.columns), nalias=len(self.header)))) str_columns = [[label] for label in self.header] else: str_columns = self._get_formatted_column_labels(frame) if self.show_row_idx_names: for x in str_columns: x.append('') stringified = [] for i, c in enumerate(frame): cheader = str_columns[i] header_colwidth = max(self.col_space or 0, *(self.adj.len(x) for x in cheader)) fmt_values = self._format_col(i) fmt_values = _make_fixed_width(fmt_values, self.justify, minimum=header_colwidth, adj=self.adj) max_len = max(max(self.adj.len(x) for x in fmt_values), header_colwidth) cheader = self.adj.justify(cheader, max_len, mode=self.justify) stringified.append(cheader + fmt_values) strcols = stringified if self.index: strcols.insert(0, str_index) # Add ... to signal truncated truncate_h = self.truncate_h truncate_v = self.truncate_v if truncate_h: col_num = self.tr_col_num strcols.insert(self.tr_col_num + 1, [' ...'] * (len(str_index))) if truncate_v: n_header_rows = len(str_index) - len(frame) row_num = self.tr_row_num for ix, col in enumerate(strcols): # infer from above row cwidth = self.adj.len(strcols[ix][row_num]) is_dot_col = False if truncate_h: is_dot_col = ix == col_num + 1 if cwidth > 3 or is_dot_col: my_str = '...' else: my_str = '..' if ix == 0: dot_mode = 'left' elif is_dot_col: cwidth = 4 dot_mode = 'right' else: dot_mode = 'right' dot_str = self.adj.justify([my_str], cwidth, mode=dot_mode)[0] strcols[ix].insert(row_num + n_header_rows, dot_str) return strcols
[ "Render", "a", "DataFrame", "to", "a", "list", "of", "columns", "(", "as", "lists", "of", "strings", ")", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L509-L590
[ "def", "_to_str_columns", "(", "self", ")", ":", "frame", "=", "self", ".", "tr_frame", "# may include levels names also", "str_index", "=", "self", ".", "_get_formatted_index", "(", "frame", ")", "if", "not", "is_list_like", "(", "self", ".", "header", ")", "and", "not", "self", ".", "header", ":", "stringified", "=", "[", "]", "for", "i", ",", "c", "in", "enumerate", "(", "frame", ")", ":", "fmt_values", "=", "self", ".", "_format_col", "(", "i", ")", "fmt_values", "=", "_make_fixed_width", "(", "fmt_values", ",", "self", ".", "justify", ",", "minimum", "=", "(", "self", ".", "col_space", "or", "0", ")", ",", "adj", "=", "self", ".", "adj", ")", "stringified", ".", "append", "(", "fmt_values", ")", "else", ":", "if", "is_list_like", "(", "self", ".", "header", ")", ":", "if", "len", "(", "self", ".", "header", ")", "!=", "len", "(", "self", ".", "columns", ")", ":", "raise", "ValueError", "(", "(", "'Writing {ncols} cols but got {nalias} '", "'aliases'", ".", "format", "(", "ncols", "=", "len", "(", "self", ".", "columns", ")", ",", "nalias", "=", "len", "(", "self", ".", "header", ")", ")", ")", ")", "str_columns", "=", "[", "[", "label", "]", "for", "label", "in", "self", ".", "header", "]", "else", ":", "str_columns", "=", "self", ".", "_get_formatted_column_labels", "(", "frame", ")", "if", "self", ".", "show_row_idx_names", ":", "for", "x", "in", "str_columns", ":", "x", ".", "append", "(", "''", ")", "stringified", "=", "[", "]", "for", "i", ",", "c", "in", "enumerate", "(", "frame", ")", ":", "cheader", "=", "str_columns", "[", "i", "]", "header_colwidth", "=", "max", "(", "self", ".", "col_space", "or", "0", ",", "*", "(", "self", ".", "adj", ".", "len", "(", "x", ")", "for", "x", "in", "cheader", ")", ")", "fmt_values", "=", "self", ".", "_format_col", "(", "i", ")", "fmt_values", "=", "_make_fixed_width", "(", "fmt_values", ",", "self", ".", "justify", ",", "minimum", "=", "header_colwidth", ",", "adj", "=", "self", ".", "adj", ")", "max_len", "=", "max", "(", "max", "(", "self", ".", "adj", ".", "len", "(", "x", ")", "for", "x", "in", "fmt_values", ")", ",", "header_colwidth", ")", "cheader", "=", "self", ".", "adj", ".", "justify", "(", "cheader", ",", "max_len", ",", "mode", "=", "self", ".", "justify", ")", "stringified", ".", "append", "(", "cheader", "+", "fmt_values", ")", "strcols", "=", "stringified", "if", "self", ".", "index", ":", "strcols", ".", "insert", "(", "0", ",", "str_index", ")", "# Add ... to signal truncated", "truncate_h", "=", "self", ".", "truncate_h", "truncate_v", "=", "self", ".", "truncate_v", "if", "truncate_h", ":", "col_num", "=", "self", ".", "tr_col_num", "strcols", ".", "insert", "(", "self", ".", "tr_col_num", "+", "1", ",", "[", "' ...'", "]", "*", "(", "len", "(", "str_index", ")", ")", ")", "if", "truncate_v", ":", "n_header_rows", "=", "len", "(", "str_index", ")", "-", "len", "(", "frame", ")", "row_num", "=", "self", ".", "tr_row_num", "for", "ix", ",", "col", "in", "enumerate", "(", "strcols", ")", ":", "# infer from above row", "cwidth", "=", "self", ".", "adj", ".", "len", "(", "strcols", "[", "ix", "]", "[", "row_num", "]", ")", "is_dot_col", "=", "False", "if", "truncate_h", ":", "is_dot_col", "=", "ix", "==", "col_num", "+", "1", "if", "cwidth", ">", "3", "or", "is_dot_col", ":", "my_str", "=", "'...'", "else", ":", "my_str", "=", "'..'", "if", "ix", "==", "0", ":", "dot_mode", "=", "'left'", "elif", "is_dot_col", ":", "cwidth", "=", "4", "dot_mode", "=", "'right'", "else", ":", "dot_mode", "=", "'right'", "dot_str", "=", "self", ".", "adj", ".", "justify", "(", "[", "my_str", "]", ",", "cwidth", ",", "mode", "=", "dot_mode", ")", "[", "0", "]", "strcols", "[", "ix", "]", ".", "insert", "(", "row_num", "+", "n_header_rows", ",", "dot_str", ")", "return", "strcols" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
DataFrameFormatter.to_string
Render a DataFrame to a console-friendly tabular output.
pandas/io/formats/format.py
def to_string(self): """ Render a DataFrame to a console-friendly tabular output. """ from pandas import Series frame = self.frame if len(frame.columns) == 0 or len(frame.index) == 0: info_line = ('Empty {name}\nColumns: {col}\nIndex: {idx}' .format(name=type(self.frame).__name__, col=pprint_thing(frame.columns), idx=pprint_thing(frame.index))) text = info_line else: strcols = self._to_str_columns() if self.line_width is None: # no need to wrap around just print # the whole frame text = self.adj.adjoin(1, *strcols) elif (not isinstance(self.max_cols, int) or self.max_cols > 0): # need to wrap around text = self._join_multiline(*strcols) else: # max_cols == 0. Try to fit frame to terminal text = self.adj.adjoin(1, *strcols).split('\n') max_len = Series(text).str.len().max() # plus truncate dot col dif = max_len - self.w # '+ 1' to avoid too wide repr (GH PR #17023) adj_dif = dif + 1 col_lens = Series([Series(ele).apply(len).max() for ele in strcols]) n_cols = len(col_lens) counter = 0 while adj_dif > 0 and n_cols > 1: counter += 1 mid = int(round(n_cols / 2.)) mid_ix = col_lens.index[mid] col_len = col_lens[mid_ix] # adjoin adds one adj_dif -= (col_len + 1) col_lens = col_lens.drop(mid_ix) n_cols = len(col_lens) # subtract index column max_cols_adj = n_cols - self.index # GH-21180. Ensure that we print at least two. max_cols_adj = max(max_cols_adj, 2) self.max_cols_adj = max_cols_adj # Call again _chk_truncate to cut frame appropriately # and then generate string representation self._chk_truncate() strcols = self._to_str_columns() text = self.adj.adjoin(1, *strcols) self.buf.writelines(text) if self.should_show_dimensions: self.buf.write("\n\n[{nrows} rows x {ncols} columns]" .format(nrows=len(frame), ncols=len(frame.columns)))
def to_string(self): """ Render a DataFrame to a console-friendly tabular output. """ from pandas import Series frame = self.frame if len(frame.columns) == 0 or len(frame.index) == 0: info_line = ('Empty {name}\nColumns: {col}\nIndex: {idx}' .format(name=type(self.frame).__name__, col=pprint_thing(frame.columns), idx=pprint_thing(frame.index))) text = info_line else: strcols = self._to_str_columns() if self.line_width is None: # no need to wrap around just print # the whole frame text = self.adj.adjoin(1, *strcols) elif (not isinstance(self.max_cols, int) or self.max_cols > 0): # need to wrap around text = self._join_multiline(*strcols) else: # max_cols == 0. Try to fit frame to terminal text = self.adj.adjoin(1, *strcols).split('\n') max_len = Series(text).str.len().max() # plus truncate dot col dif = max_len - self.w # '+ 1' to avoid too wide repr (GH PR #17023) adj_dif = dif + 1 col_lens = Series([Series(ele).apply(len).max() for ele in strcols]) n_cols = len(col_lens) counter = 0 while adj_dif > 0 and n_cols > 1: counter += 1 mid = int(round(n_cols / 2.)) mid_ix = col_lens.index[mid] col_len = col_lens[mid_ix] # adjoin adds one adj_dif -= (col_len + 1) col_lens = col_lens.drop(mid_ix) n_cols = len(col_lens) # subtract index column max_cols_adj = n_cols - self.index # GH-21180. Ensure that we print at least two. max_cols_adj = max(max_cols_adj, 2) self.max_cols_adj = max_cols_adj # Call again _chk_truncate to cut frame appropriately # and then generate string representation self._chk_truncate() strcols = self._to_str_columns() text = self.adj.adjoin(1, *strcols) self.buf.writelines(text) if self.should_show_dimensions: self.buf.write("\n\n[{nrows} rows x {ncols} columns]" .format(nrows=len(frame), ncols=len(frame.columns)))
[ "Render", "a", "DataFrame", "to", "a", "console", "-", "friendly", "tabular", "output", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L592-L650
[ "def", "to_string", "(", "self", ")", ":", "from", "pandas", "import", "Series", "frame", "=", "self", ".", "frame", "if", "len", "(", "frame", ".", "columns", ")", "==", "0", "or", "len", "(", "frame", ".", "index", ")", "==", "0", ":", "info_line", "=", "(", "'Empty {name}\\nColumns: {col}\\nIndex: {idx}'", ".", "format", "(", "name", "=", "type", "(", "self", ".", "frame", ")", ".", "__name__", ",", "col", "=", "pprint_thing", "(", "frame", ".", "columns", ")", ",", "idx", "=", "pprint_thing", "(", "frame", ".", "index", ")", ")", ")", "text", "=", "info_line", "else", ":", "strcols", "=", "self", ".", "_to_str_columns", "(", ")", "if", "self", ".", "line_width", "is", "None", ":", "# no need to wrap around just print", "# the whole frame", "text", "=", "self", ".", "adj", ".", "adjoin", "(", "1", ",", "*", "strcols", ")", "elif", "(", "not", "isinstance", "(", "self", ".", "max_cols", ",", "int", ")", "or", "self", ".", "max_cols", ">", "0", ")", ":", "# need to wrap around", "text", "=", "self", ".", "_join_multiline", "(", "*", "strcols", ")", "else", ":", "# max_cols == 0. Try to fit frame to terminal", "text", "=", "self", ".", "adj", ".", "adjoin", "(", "1", ",", "*", "strcols", ")", ".", "split", "(", "'\\n'", ")", "max_len", "=", "Series", "(", "text", ")", ".", "str", ".", "len", "(", ")", ".", "max", "(", ")", "# plus truncate dot col", "dif", "=", "max_len", "-", "self", ".", "w", "# '+ 1' to avoid too wide repr (GH PR #17023)", "adj_dif", "=", "dif", "+", "1", "col_lens", "=", "Series", "(", "[", "Series", "(", "ele", ")", ".", "apply", "(", "len", ")", ".", "max", "(", ")", "for", "ele", "in", "strcols", "]", ")", "n_cols", "=", "len", "(", "col_lens", ")", "counter", "=", "0", "while", "adj_dif", ">", "0", "and", "n_cols", ">", "1", ":", "counter", "+=", "1", "mid", "=", "int", "(", "round", "(", "n_cols", "/", "2.", ")", ")", "mid_ix", "=", "col_lens", ".", "index", "[", "mid", "]", "col_len", "=", "col_lens", "[", "mid_ix", "]", "# adjoin adds one", "adj_dif", "-=", "(", "col_len", "+", "1", ")", "col_lens", "=", "col_lens", ".", "drop", "(", "mid_ix", ")", "n_cols", "=", "len", "(", "col_lens", ")", "# subtract index column", "max_cols_adj", "=", "n_cols", "-", "self", ".", "index", "# GH-21180. Ensure that we print at least two.", "max_cols_adj", "=", "max", "(", "max_cols_adj", ",", "2", ")", "self", ".", "max_cols_adj", "=", "max_cols_adj", "# Call again _chk_truncate to cut frame appropriately", "# and then generate string representation", "self", ".", "_chk_truncate", "(", ")", "strcols", "=", "self", ".", "_to_str_columns", "(", ")", "text", "=", "self", ".", "adj", ".", "adjoin", "(", "1", ",", "*", "strcols", ")", "self", ".", "buf", ".", "writelines", "(", "text", ")", "if", "self", ".", "should_show_dimensions", ":", "self", ".", "buf", ".", "write", "(", "\"\\n\\n[{nrows} rows x {ncols} columns]\"", ".", "format", "(", "nrows", "=", "len", "(", "frame", ")", ",", "ncols", "=", "len", "(", "frame", ".", "columns", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
DataFrameFormatter.to_latex
Render a DataFrame to a LaTeX tabular/longtable environment output.
pandas/io/formats/format.py
def to_latex(self, column_format=None, longtable=False, encoding=None, multicolumn=False, multicolumn_format=None, multirow=False): """ Render a DataFrame to a LaTeX tabular/longtable environment output. """ from pandas.io.formats.latex import LatexFormatter latex_renderer = LatexFormatter(self, column_format=column_format, longtable=longtable, multicolumn=multicolumn, multicolumn_format=multicolumn_format, multirow=multirow) if encoding is None: encoding = 'utf-8' if hasattr(self.buf, 'write'): latex_renderer.write_result(self.buf) elif isinstance(self.buf, str): import codecs with codecs.open(self.buf, 'w', encoding=encoding) as f: latex_renderer.write_result(f) else: raise TypeError('buf is not a file name and it has no write ' 'method')
def to_latex(self, column_format=None, longtable=False, encoding=None, multicolumn=False, multicolumn_format=None, multirow=False): """ Render a DataFrame to a LaTeX tabular/longtable environment output. """ from pandas.io.formats.latex import LatexFormatter latex_renderer = LatexFormatter(self, column_format=column_format, longtable=longtable, multicolumn=multicolumn, multicolumn_format=multicolumn_format, multirow=multirow) if encoding is None: encoding = 'utf-8' if hasattr(self.buf, 'write'): latex_renderer.write_result(self.buf) elif isinstance(self.buf, str): import codecs with codecs.open(self.buf, 'w', encoding=encoding) as f: latex_renderer.write_result(f) else: raise TypeError('buf is not a file name and it has no write ' 'method')
[ "Render", "a", "DataFrame", "to", "a", "LaTeX", "tabular", "/", "longtable", "environment", "output", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L686-L710
[ "def", "to_latex", "(", "self", ",", "column_format", "=", "None", ",", "longtable", "=", "False", ",", "encoding", "=", "None", ",", "multicolumn", "=", "False", ",", "multicolumn_format", "=", "None", ",", "multirow", "=", "False", ")", ":", "from", "pandas", ".", "io", ".", "formats", ".", "latex", "import", "LatexFormatter", "latex_renderer", "=", "LatexFormatter", "(", "self", ",", "column_format", "=", "column_format", ",", "longtable", "=", "longtable", ",", "multicolumn", "=", "multicolumn", ",", "multicolumn_format", "=", "multicolumn_format", ",", "multirow", "=", "multirow", ")", "if", "encoding", "is", "None", ":", "encoding", "=", "'utf-8'", "if", "hasattr", "(", "self", ".", "buf", ",", "'write'", ")", ":", "latex_renderer", ".", "write_result", "(", "self", ".", "buf", ")", "elif", "isinstance", "(", "self", ".", "buf", ",", "str", ")", ":", "import", "codecs", "with", "codecs", ".", "open", "(", "self", ".", "buf", ",", "'w'", ",", "encoding", "=", "encoding", ")", "as", "f", ":", "latex_renderer", ".", "write_result", "(", "f", ")", "else", ":", "raise", "TypeError", "(", "'buf is not a file name and it has no write '", "'method'", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
DataFrameFormatter.to_html
Render a DataFrame to a html table. Parameters ---------- classes : str or list-like classes to include in the `class` attribute of the opening ``<table>`` tag, in addition to the default "dataframe". notebook : {True, False}, optional, default False Whether the generated HTML is for IPython Notebook. border : int A ``border=border`` attribute is included in the opening ``<table>`` tag. Default ``pd.options.html.border``. .. versionadded:: 0.19.0
pandas/io/formats/format.py
def to_html(self, classes=None, notebook=False, border=None): """ Render a DataFrame to a html table. Parameters ---------- classes : str or list-like classes to include in the `class` attribute of the opening ``<table>`` tag, in addition to the default "dataframe". notebook : {True, False}, optional, default False Whether the generated HTML is for IPython Notebook. border : int A ``border=border`` attribute is included in the opening ``<table>`` tag. Default ``pd.options.html.border``. .. versionadded:: 0.19.0 """ from pandas.io.formats.html import HTMLFormatter, NotebookFormatter Klass = NotebookFormatter if notebook else HTMLFormatter html = Klass(self, classes=classes, border=border).render() if hasattr(self.buf, 'write'): buffer_put_lines(self.buf, html) elif isinstance(self.buf, str): with open(self.buf, 'w') as f: buffer_put_lines(f, html) else: raise TypeError('buf is not a file name and it has no write ' ' method')
def to_html(self, classes=None, notebook=False, border=None): """ Render a DataFrame to a html table. Parameters ---------- classes : str or list-like classes to include in the `class` attribute of the opening ``<table>`` tag, in addition to the default "dataframe". notebook : {True, False}, optional, default False Whether the generated HTML is for IPython Notebook. border : int A ``border=border`` attribute is included in the opening ``<table>`` tag. Default ``pd.options.html.border``. .. versionadded:: 0.19.0 """ from pandas.io.formats.html import HTMLFormatter, NotebookFormatter Klass = NotebookFormatter if notebook else HTMLFormatter html = Klass(self, classes=classes, border=border).render() if hasattr(self.buf, 'write'): buffer_put_lines(self.buf, html) elif isinstance(self.buf, str): with open(self.buf, 'w') as f: buffer_put_lines(f, html) else: raise TypeError('buf is not a file name and it has no write ' ' method')
[ "Render", "a", "DataFrame", "to", "a", "html", "table", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L720-L747
[ "def", "to_html", "(", "self", ",", "classes", "=", "None", ",", "notebook", "=", "False", ",", "border", "=", "None", ")", ":", "from", "pandas", ".", "io", ".", "formats", ".", "html", "import", "HTMLFormatter", ",", "NotebookFormatter", "Klass", "=", "NotebookFormatter", "if", "notebook", "else", "HTMLFormatter", "html", "=", "Klass", "(", "self", ",", "classes", "=", "classes", ",", "border", "=", "border", ")", ".", "render", "(", ")", "if", "hasattr", "(", "self", ".", "buf", ",", "'write'", ")", ":", "buffer_put_lines", "(", "self", ".", "buf", ",", "html", ")", "elif", "isinstance", "(", "self", ".", "buf", ",", "str", ")", ":", "with", "open", "(", "self", ".", "buf", ",", "'w'", ")", "as", "f", ":", "buffer_put_lines", "(", "f", ",", "html", ")", "else", ":", "raise", "TypeError", "(", "'buf is not a file name and it has no write '", "' method'", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
FloatArrayFormatter._value_formatter
Returns a function to be applied on each value to format it
pandas/io/formats/format.py
def _value_formatter(self, float_format=None, threshold=None): """Returns a function to be applied on each value to format it """ # the float_format parameter supersedes self.float_format if float_format is None: float_format = self.float_format # we are going to compose different functions, to first convert to # a string, then replace the decimal symbol, and finally chop according # to the threshold # when there is no float_format, we use str instead of '%g' # because str(0.0) = '0.0' while '%g' % 0.0 = '0' if float_format: def base_formatter(v): return float_format(value=v) if notna(v) else self.na_rep else: def base_formatter(v): return str(v) if notna(v) else self.na_rep if self.decimal != '.': def decimal_formatter(v): return base_formatter(v).replace('.', self.decimal, 1) else: decimal_formatter = base_formatter if threshold is None: return decimal_formatter def formatter(value): if notna(value): if abs(value) > threshold: return decimal_formatter(value) else: return decimal_formatter(0.0) else: return self.na_rep return formatter
def _value_formatter(self, float_format=None, threshold=None): """Returns a function to be applied on each value to format it """ # the float_format parameter supersedes self.float_format if float_format is None: float_format = self.float_format # we are going to compose different functions, to first convert to # a string, then replace the decimal symbol, and finally chop according # to the threshold # when there is no float_format, we use str instead of '%g' # because str(0.0) = '0.0' while '%g' % 0.0 = '0' if float_format: def base_formatter(v): return float_format(value=v) if notna(v) else self.na_rep else: def base_formatter(v): return str(v) if notna(v) else self.na_rep if self.decimal != '.': def decimal_formatter(v): return base_formatter(v).replace('.', self.decimal, 1) else: decimal_formatter = base_formatter if threshold is None: return decimal_formatter def formatter(value): if notna(value): if abs(value) > threshold: return decimal_formatter(value) else: return decimal_formatter(0.0) else: return self.na_rep return formatter
[ "Returns", "a", "function", "to", "be", "applied", "on", "each", "value", "to", "format", "it" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1015-L1054
[ "def", "_value_formatter", "(", "self", ",", "float_format", "=", "None", ",", "threshold", "=", "None", ")", ":", "# the float_format parameter supersedes self.float_format", "if", "float_format", "is", "None", ":", "float_format", "=", "self", ".", "float_format", "# we are going to compose different functions, to first convert to", "# a string, then replace the decimal symbol, and finally chop according", "# to the threshold", "# when there is no float_format, we use str instead of '%g'", "# because str(0.0) = '0.0' while '%g' % 0.0 = '0'", "if", "float_format", ":", "def", "base_formatter", "(", "v", ")", ":", "return", "float_format", "(", "value", "=", "v", ")", "if", "notna", "(", "v", ")", "else", "self", ".", "na_rep", "else", ":", "def", "base_formatter", "(", "v", ")", ":", "return", "str", "(", "v", ")", "if", "notna", "(", "v", ")", "else", "self", ".", "na_rep", "if", "self", ".", "decimal", "!=", "'.'", ":", "def", "decimal_formatter", "(", "v", ")", ":", "return", "base_formatter", "(", "v", ")", ".", "replace", "(", "'.'", ",", "self", ".", "decimal", ",", "1", ")", "else", ":", "decimal_formatter", "=", "base_formatter", "if", "threshold", "is", "None", ":", "return", "decimal_formatter", "def", "formatter", "(", "value", ")", ":", "if", "notna", "(", "value", ")", ":", "if", "abs", "(", "value", ")", ">", "threshold", ":", "return", "decimal_formatter", "(", "value", ")", "else", ":", "return", "decimal_formatter", "(", "0.0", ")", "else", ":", "return", "self", ".", "na_rep", "return", "formatter" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
FloatArrayFormatter.get_result_as_array
Returns the float values converted into strings using the parameters given at initialisation, as a numpy array
pandas/io/formats/format.py
def get_result_as_array(self): """ Returns the float values converted into strings using the parameters given at initialisation, as a numpy array """ if self.formatter is not None: return np.array([self.formatter(x) for x in self.values]) if self.fixed_width: threshold = get_option("display.chop_threshold") else: threshold = None # if we have a fixed_width, we'll need to try different float_format def format_values_with(float_format): formatter = self._value_formatter(float_format, threshold) # default formatter leaves a space to the left when formatting # floats, must be consistent for left-justifying NaNs (GH #25061) if self.justify == 'left': na_rep = ' ' + self.na_rep else: na_rep = self.na_rep # separate the wheat from the chaff values = self.values is_complex = is_complex_dtype(values) mask = isna(values) if hasattr(values, 'to_dense'): # sparse numpy ndarray values = values.to_dense() values = np.array(values, dtype='object') values[mask] = na_rep imask = (~mask).ravel() values.flat[imask] = np.array([formatter(val) for val in values.ravel()[imask]]) if self.fixed_width: if is_complex: return _trim_zeros_complex(values, na_rep) else: return _trim_zeros_float(values, na_rep) return values # There is a special default string when we are fixed-width # The default is otherwise to use str instead of a formatting string if self.float_format is None: if self.fixed_width: float_format = partial('{value: .{digits:d}f}'.format, digits=self.digits) else: float_format = self.float_format else: float_format = lambda value: self.float_format % value formatted_values = format_values_with(float_format) if not self.fixed_width: return formatted_values # we need do convert to engineering format if some values are too small # and would appear as 0, or if some values are too big and take too # much space if len(formatted_values) > 0: maxlen = max(len(x) for x in formatted_values) too_long = maxlen > self.digits + 6 else: too_long = False with np.errstate(invalid='ignore'): abs_vals = np.abs(self.values) # this is pretty arbitrary for now # large values: more that 8 characters including decimal symbol # and first digit, hence > 1e6 has_large_values = (abs_vals > 1e6).any() has_small_values = ((abs_vals < 10**(-self.digits)) & (abs_vals > 0)).any() if has_small_values or (too_long and has_large_values): float_format = partial('{value: .{digits:d}e}'.format, digits=self.digits) formatted_values = format_values_with(float_format) return formatted_values
def get_result_as_array(self): """ Returns the float values converted into strings using the parameters given at initialisation, as a numpy array """ if self.formatter is not None: return np.array([self.formatter(x) for x in self.values]) if self.fixed_width: threshold = get_option("display.chop_threshold") else: threshold = None # if we have a fixed_width, we'll need to try different float_format def format_values_with(float_format): formatter = self._value_formatter(float_format, threshold) # default formatter leaves a space to the left when formatting # floats, must be consistent for left-justifying NaNs (GH #25061) if self.justify == 'left': na_rep = ' ' + self.na_rep else: na_rep = self.na_rep # separate the wheat from the chaff values = self.values is_complex = is_complex_dtype(values) mask = isna(values) if hasattr(values, 'to_dense'): # sparse numpy ndarray values = values.to_dense() values = np.array(values, dtype='object') values[mask] = na_rep imask = (~mask).ravel() values.flat[imask] = np.array([formatter(val) for val in values.ravel()[imask]]) if self.fixed_width: if is_complex: return _trim_zeros_complex(values, na_rep) else: return _trim_zeros_float(values, na_rep) return values # There is a special default string when we are fixed-width # The default is otherwise to use str instead of a formatting string if self.float_format is None: if self.fixed_width: float_format = partial('{value: .{digits:d}f}'.format, digits=self.digits) else: float_format = self.float_format else: float_format = lambda value: self.float_format % value formatted_values = format_values_with(float_format) if not self.fixed_width: return formatted_values # we need do convert to engineering format if some values are too small # and would appear as 0, or if some values are too big and take too # much space if len(formatted_values) > 0: maxlen = max(len(x) for x in formatted_values) too_long = maxlen > self.digits + 6 else: too_long = False with np.errstate(invalid='ignore'): abs_vals = np.abs(self.values) # this is pretty arbitrary for now # large values: more that 8 characters including decimal symbol # and first digit, hence > 1e6 has_large_values = (abs_vals > 1e6).any() has_small_values = ((abs_vals < 10**(-self.digits)) & (abs_vals > 0)).any() if has_small_values or (too_long and has_large_values): float_format = partial('{value: .{digits:d}e}'.format, digits=self.digits) formatted_values = format_values_with(float_format) return formatted_values
[ "Returns", "the", "float", "values", "converted", "into", "strings", "using", "the", "parameters", "given", "at", "initialisation", "as", "a", "numpy", "array" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1056-L1141
[ "def", "get_result_as_array", "(", "self", ")", ":", "if", "self", ".", "formatter", "is", "not", "None", ":", "return", "np", ".", "array", "(", "[", "self", ".", "formatter", "(", "x", ")", "for", "x", "in", "self", ".", "values", "]", ")", "if", "self", ".", "fixed_width", ":", "threshold", "=", "get_option", "(", "\"display.chop_threshold\"", ")", "else", ":", "threshold", "=", "None", "# if we have a fixed_width, we'll need to try different float_format", "def", "format_values_with", "(", "float_format", ")", ":", "formatter", "=", "self", ".", "_value_formatter", "(", "float_format", ",", "threshold", ")", "# default formatter leaves a space to the left when formatting", "# floats, must be consistent for left-justifying NaNs (GH #25061)", "if", "self", ".", "justify", "==", "'left'", ":", "na_rep", "=", "' '", "+", "self", ".", "na_rep", "else", ":", "na_rep", "=", "self", ".", "na_rep", "# separate the wheat from the chaff", "values", "=", "self", ".", "values", "is_complex", "=", "is_complex_dtype", "(", "values", ")", "mask", "=", "isna", "(", "values", ")", "if", "hasattr", "(", "values", ",", "'to_dense'", ")", ":", "# sparse numpy ndarray", "values", "=", "values", ".", "to_dense", "(", ")", "values", "=", "np", ".", "array", "(", "values", ",", "dtype", "=", "'object'", ")", "values", "[", "mask", "]", "=", "na_rep", "imask", "=", "(", "~", "mask", ")", ".", "ravel", "(", ")", "values", ".", "flat", "[", "imask", "]", "=", "np", ".", "array", "(", "[", "formatter", "(", "val", ")", "for", "val", "in", "values", ".", "ravel", "(", ")", "[", "imask", "]", "]", ")", "if", "self", ".", "fixed_width", ":", "if", "is_complex", ":", "return", "_trim_zeros_complex", "(", "values", ",", "na_rep", ")", "else", ":", "return", "_trim_zeros_float", "(", "values", ",", "na_rep", ")", "return", "values", "# There is a special default string when we are fixed-width", "# The default is otherwise to use str instead of a formatting string", "if", "self", ".", "float_format", "is", "None", ":", "if", "self", ".", "fixed_width", ":", "float_format", "=", "partial", "(", "'{value: .{digits:d}f}'", ".", "format", ",", "digits", "=", "self", ".", "digits", ")", "else", ":", "float_format", "=", "self", ".", "float_format", "else", ":", "float_format", "=", "lambda", "value", ":", "self", ".", "float_format", "%", "value", "formatted_values", "=", "format_values_with", "(", "float_format", ")", "if", "not", "self", ".", "fixed_width", ":", "return", "formatted_values", "# we need do convert to engineering format if some values are too small", "# and would appear as 0, or if some values are too big and take too", "# much space", "if", "len", "(", "formatted_values", ")", ">", "0", ":", "maxlen", "=", "max", "(", "len", "(", "x", ")", "for", "x", "in", "formatted_values", ")", "too_long", "=", "maxlen", ">", "self", ".", "digits", "+", "6", "else", ":", "too_long", "=", "False", "with", "np", ".", "errstate", "(", "invalid", "=", "'ignore'", ")", ":", "abs_vals", "=", "np", ".", "abs", "(", "self", ".", "values", ")", "# this is pretty arbitrary for now", "# large values: more that 8 characters including decimal symbol", "# and first digit, hence > 1e6", "has_large_values", "=", "(", "abs_vals", ">", "1e6", ")", ".", "any", "(", ")", "has_small_values", "=", "(", "(", "abs_vals", "<", "10", "**", "(", "-", "self", ".", "digits", ")", ")", "&", "(", "abs_vals", ">", "0", ")", ")", ".", "any", "(", ")", "if", "has_small_values", "or", "(", "too_long", "and", "has_large_values", ")", ":", "float_format", "=", "partial", "(", "'{value: .{digits:d}e}'", ".", "format", ",", "digits", "=", "self", ".", "digits", ")", "formatted_values", "=", "format_values_with", "(", "float_format", ")", "return", "formatted_values" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
Datetime64Formatter._format_strings
we by definition have DO NOT have a TZ
pandas/io/formats/format.py
def _format_strings(self): """ we by definition have DO NOT have a TZ """ values = self.values if not isinstance(values, DatetimeIndex): values = DatetimeIndex(values) if self.formatter is not None and callable(self.formatter): return [self.formatter(x) for x in values] fmt_values = format_array_from_datetime( values.asi8.ravel(), format=_get_format_datetime64_from_values(values, self.date_format), na_rep=self.nat_rep).reshape(values.shape) return fmt_values.tolist()
def _format_strings(self): """ we by definition have DO NOT have a TZ """ values = self.values if not isinstance(values, DatetimeIndex): values = DatetimeIndex(values) if self.formatter is not None and callable(self.formatter): return [self.formatter(x) for x in values] fmt_values = format_array_from_datetime( values.asi8.ravel(), format=_get_format_datetime64_from_values(values, self.date_format), na_rep=self.nat_rep).reshape(values.shape) return fmt_values.tolist()
[ "we", "by", "definition", "have", "DO", "NOT", "have", "a", "TZ" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1166-L1182
[ "def", "_format_strings", "(", "self", ")", ":", "values", "=", "self", ".", "values", "if", "not", "isinstance", "(", "values", ",", "DatetimeIndex", ")", ":", "values", "=", "DatetimeIndex", "(", "values", ")", "if", "self", ".", "formatter", "is", "not", "None", "and", "callable", "(", "self", ".", "formatter", ")", ":", "return", "[", "self", ".", "formatter", "(", "x", ")", "for", "x", "in", "values", "]", "fmt_values", "=", "format_array_from_datetime", "(", "values", ".", "asi8", ".", "ravel", "(", ")", ",", "format", "=", "_get_format_datetime64_from_values", "(", "values", ",", "self", ".", "date_format", ")", ",", "na_rep", "=", "self", ".", "nat_rep", ")", ".", "reshape", "(", "values", ".", "shape", ")", "return", "fmt_values", ".", "tolist", "(", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
Datetime64TZFormatter._format_strings
we by definition have a TZ
pandas/io/formats/format.py
def _format_strings(self): """ we by definition have a TZ """ values = self.values.astype(object) is_dates_only = _is_dates_only(values) formatter = (self.formatter or _get_format_datetime64(is_dates_only, date_format=self.date_format)) fmt_values = [formatter(x) for x in values] return fmt_values
def _format_strings(self): """ we by definition have a TZ """ values = self.values.astype(object) is_dates_only = _is_dates_only(values) formatter = (self.formatter or _get_format_datetime64(is_dates_only, date_format=self.date_format)) fmt_values = [formatter(x) for x in values] return fmt_values
[ "we", "by", "definition", "have", "a", "TZ" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/format.py#L1332-L1342
[ "def", "_format_strings", "(", "self", ")", ":", "values", "=", "self", ".", "values", ".", "astype", "(", "object", ")", "is_dates_only", "=", "_is_dates_only", "(", "values", ")", "formatter", "=", "(", "self", ".", "formatter", "or", "_get_format_datetime64", "(", "is_dates_only", ",", "date_format", "=", "self", ".", "date_format", ")", ")", "fmt_values", "=", "[", "formatter", "(", "x", ")", "for", "x", "in", "values", "]", "return", "fmt_values" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_get_interval_closed_bounds
Given an Interval or IntervalIndex, return the corresponding interval with closed bounds.
pandas/core/indexes/interval.py
def _get_interval_closed_bounds(interval): """ Given an Interval or IntervalIndex, return the corresponding interval with closed bounds. """ left, right = interval.left, interval.right if interval.open_left: left = _get_next_label(left) if interval.open_right: right = _get_prev_label(right) return left, right
def _get_interval_closed_bounds(interval): """ Given an Interval or IntervalIndex, return the corresponding interval with closed bounds. """ left, right = interval.left, interval.right if interval.open_left: left = _get_next_label(left) if interval.open_right: right = _get_prev_label(right) return left, right
[ "Given", "an", "Interval", "or", "IntervalIndex", "return", "the", "corresponding", "interval", "with", "closed", "bounds", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/indexes/interval.py#L79-L89
[ "def", "_get_interval_closed_bounds", "(", "interval", ")", ":", "left", ",", "right", "=", "interval", ".", "left", ",", "interval", ".", "right", "if", "interval", ".", "open_left", ":", "left", "=", "_get_next_label", "(", "left", ")", "if", "interval", ".", "open_right", ":", "right", "=", "_get_prev_label", "(", "right", ")", "return", "left", ",", "right" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_is_valid_endpoint
helper for interval_range to check if start/end are valid types
pandas/core/indexes/interval.py
def _is_valid_endpoint(endpoint): """helper for interval_range to check if start/end are valid types""" return any([is_number(endpoint), isinstance(endpoint, Timestamp), isinstance(endpoint, Timedelta), endpoint is None])
def _is_valid_endpoint(endpoint): """helper for interval_range to check if start/end are valid types""" return any([is_number(endpoint), isinstance(endpoint, Timestamp), isinstance(endpoint, Timedelta), endpoint is None])
[ "helper", "for", "interval_range", "to", "check", "if", "start", "/", "end", "are", "valid", "types" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/indexes/interval.py#L1138-L1143
[ "def", "_is_valid_endpoint", "(", "endpoint", ")", ":", "return", "any", "(", "[", "is_number", "(", "endpoint", ")", ",", "isinstance", "(", "endpoint", ",", "Timestamp", ")", ",", "isinstance", "(", "endpoint", ",", "Timedelta", ")", ",", "endpoint", "is", "None", "]", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_is_type_compatible
helper for interval_range to check type compat of start/end/freq
pandas/core/indexes/interval.py
def _is_type_compatible(a, b): """helper for interval_range to check type compat of start/end/freq""" is_ts_compat = lambda x: isinstance(x, (Timestamp, DateOffset)) is_td_compat = lambda x: isinstance(x, (Timedelta, DateOffset)) return ((is_number(a) and is_number(b)) or (is_ts_compat(a) and is_ts_compat(b)) or (is_td_compat(a) and is_td_compat(b)) or com._any_none(a, b))
def _is_type_compatible(a, b): """helper for interval_range to check type compat of start/end/freq""" is_ts_compat = lambda x: isinstance(x, (Timestamp, DateOffset)) is_td_compat = lambda x: isinstance(x, (Timedelta, DateOffset)) return ((is_number(a) and is_number(b)) or (is_ts_compat(a) and is_ts_compat(b)) or (is_td_compat(a) and is_td_compat(b)) or com._any_none(a, b))
[ "helper", "for", "interval_range", "to", "check", "type", "compat", "of", "start", "/", "end", "/", "freq" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/indexes/interval.py#L1146-L1153
[ "def", "_is_type_compatible", "(", "a", ",", "b", ")", ":", "is_ts_compat", "=", "lambda", "x", ":", "isinstance", "(", "x", ",", "(", "Timestamp", ",", "DateOffset", ")", ")", "is_td_compat", "=", "lambda", "x", ":", "isinstance", "(", "x", ",", "(", "Timedelta", ",", "DateOffset", ")", ")", "return", "(", "(", "is_number", "(", "a", ")", "and", "is_number", "(", "b", ")", ")", "or", "(", "is_ts_compat", "(", "a", ")", "and", "is_ts_compat", "(", "b", ")", ")", "or", "(", "is_td_compat", "(", "a", ")", "and", "is_td_compat", "(", "b", ")", ")", "or", "com", ".", "_any_none", "(", "a", ",", "b", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
interval_range
Return a fixed frequency IntervalIndex Parameters ---------- start : numeric or datetime-like, default None Left bound for generating intervals end : numeric or datetime-like, default None Right bound for generating intervals periods : integer, default None Number of periods to generate freq : numeric, string, or DateOffset, default None The length of each interval. Must be consistent with the type of start and end, e.g. 2 for numeric, or '5H' for datetime-like. Default is 1 for numeric and 'D' for datetime-like. name : string, default None Name of the resulting IntervalIndex closed : {'left', 'right', 'both', 'neither'}, default 'right' Whether the intervals are closed on the left-side, right-side, both or neither. Returns ------- rng : IntervalIndex See Also -------- IntervalIndex : An Index of intervals that are all closed on the same side. Notes ----- Of the four parameters ``start``, ``end``, ``periods``, and ``freq``, exactly three must be specified. If ``freq`` is omitted, the resulting ``IntervalIndex`` will have ``periods`` linearly spaced elements between ``start`` and ``end``, inclusively. To learn more about datetime-like frequency strings, please see `this link <http://pandas.pydata.org/pandas-docs/stable/timeseries.html#offset-aliases>`__. Examples -------- Numeric ``start`` and ``end`` is supported. >>> pd.interval_range(start=0, end=5) IntervalIndex([(0, 1], (1, 2], (2, 3], (3, 4], (4, 5]], closed='right', dtype='interval[int64]') Additionally, datetime-like input is also supported. >>> pd.interval_range(start=pd.Timestamp('2017-01-01'), ... end=pd.Timestamp('2017-01-04')) IntervalIndex([(2017-01-01, 2017-01-02], (2017-01-02, 2017-01-03], (2017-01-03, 2017-01-04]], closed='right', dtype='interval[datetime64[ns]]') The ``freq`` parameter specifies the frequency between the left and right. endpoints of the individual intervals within the ``IntervalIndex``. For numeric ``start`` and ``end``, the frequency must also be numeric. >>> pd.interval_range(start=0, periods=4, freq=1.5) IntervalIndex([(0.0, 1.5], (1.5, 3.0], (3.0, 4.5], (4.5, 6.0]], closed='right', dtype='interval[float64]') Similarly, for datetime-like ``start`` and ``end``, the frequency must be convertible to a DateOffset. >>> pd.interval_range(start=pd.Timestamp('2017-01-01'), ... periods=3, freq='MS') IntervalIndex([(2017-01-01, 2017-02-01], (2017-02-01, 2017-03-01], (2017-03-01, 2017-04-01]], closed='right', dtype='interval[datetime64[ns]]') Specify ``start``, ``end``, and ``periods``; the frequency is generated automatically (linearly spaced). >>> pd.interval_range(start=0, end=6, periods=4) IntervalIndex([(0.0, 1.5], (1.5, 3.0], (3.0, 4.5], (4.5, 6.0]], closed='right', dtype='interval[float64]') The ``closed`` parameter specifies which endpoints of the individual intervals within the ``IntervalIndex`` are closed. >>> pd.interval_range(end=5, periods=4, closed='both') IntervalIndex([[1, 2], [2, 3], [3, 4], [4, 5]], closed='both', dtype='interval[int64]')
pandas/core/indexes/interval.py
def interval_range(start=None, end=None, periods=None, freq=None, name=None, closed='right'): """ Return a fixed frequency IntervalIndex Parameters ---------- start : numeric or datetime-like, default None Left bound for generating intervals end : numeric or datetime-like, default None Right bound for generating intervals periods : integer, default None Number of periods to generate freq : numeric, string, or DateOffset, default None The length of each interval. Must be consistent with the type of start and end, e.g. 2 for numeric, or '5H' for datetime-like. Default is 1 for numeric and 'D' for datetime-like. name : string, default None Name of the resulting IntervalIndex closed : {'left', 'right', 'both', 'neither'}, default 'right' Whether the intervals are closed on the left-side, right-side, both or neither. Returns ------- rng : IntervalIndex See Also -------- IntervalIndex : An Index of intervals that are all closed on the same side. Notes ----- Of the four parameters ``start``, ``end``, ``periods``, and ``freq``, exactly three must be specified. If ``freq`` is omitted, the resulting ``IntervalIndex`` will have ``periods`` linearly spaced elements between ``start`` and ``end``, inclusively. To learn more about datetime-like frequency strings, please see `this link <http://pandas.pydata.org/pandas-docs/stable/timeseries.html#offset-aliases>`__. Examples -------- Numeric ``start`` and ``end`` is supported. >>> pd.interval_range(start=0, end=5) IntervalIndex([(0, 1], (1, 2], (2, 3], (3, 4], (4, 5]], closed='right', dtype='interval[int64]') Additionally, datetime-like input is also supported. >>> pd.interval_range(start=pd.Timestamp('2017-01-01'), ... end=pd.Timestamp('2017-01-04')) IntervalIndex([(2017-01-01, 2017-01-02], (2017-01-02, 2017-01-03], (2017-01-03, 2017-01-04]], closed='right', dtype='interval[datetime64[ns]]') The ``freq`` parameter specifies the frequency between the left and right. endpoints of the individual intervals within the ``IntervalIndex``. For numeric ``start`` and ``end``, the frequency must also be numeric. >>> pd.interval_range(start=0, periods=4, freq=1.5) IntervalIndex([(0.0, 1.5], (1.5, 3.0], (3.0, 4.5], (4.5, 6.0]], closed='right', dtype='interval[float64]') Similarly, for datetime-like ``start`` and ``end``, the frequency must be convertible to a DateOffset. >>> pd.interval_range(start=pd.Timestamp('2017-01-01'), ... periods=3, freq='MS') IntervalIndex([(2017-01-01, 2017-02-01], (2017-02-01, 2017-03-01], (2017-03-01, 2017-04-01]], closed='right', dtype='interval[datetime64[ns]]') Specify ``start``, ``end``, and ``periods``; the frequency is generated automatically (linearly spaced). >>> pd.interval_range(start=0, end=6, periods=4) IntervalIndex([(0.0, 1.5], (1.5, 3.0], (3.0, 4.5], (4.5, 6.0]], closed='right', dtype='interval[float64]') The ``closed`` parameter specifies which endpoints of the individual intervals within the ``IntervalIndex`` are closed. >>> pd.interval_range(end=5, periods=4, closed='both') IntervalIndex([[1, 2], [2, 3], [3, 4], [4, 5]], closed='both', dtype='interval[int64]') """ start = com.maybe_box_datetimelike(start) end = com.maybe_box_datetimelike(end) endpoint = start if start is not None else end if freq is None and com._any_none(periods, start, end): freq = 1 if is_number(endpoint) else 'D' if com.count_not_none(start, end, periods, freq) != 3: raise ValueError('Of the four parameters: start, end, periods, and ' 'freq, exactly three must be specified') if not _is_valid_endpoint(start): msg = 'start must be numeric or datetime-like, got {start}' raise ValueError(msg.format(start=start)) elif not _is_valid_endpoint(end): msg = 'end must be numeric or datetime-like, got {end}' raise ValueError(msg.format(end=end)) if is_float(periods): periods = int(periods) elif not is_integer(periods) and periods is not None: msg = 'periods must be a number, got {periods}' raise TypeError(msg.format(periods=periods)) if freq is not None and not is_number(freq): try: freq = to_offset(freq) except ValueError: raise ValueError('freq must be numeric or convertible to ' 'DateOffset, got {freq}'.format(freq=freq)) # verify type compatibility if not all([_is_type_compatible(start, end), _is_type_compatible(start, freq), _is_type_compatible(end, freq)]): raise TypeError("start, end, freq need to be type compatible") # +1 to convert interval count to breaks count (n breaks = n-1 intervals) if periods is not None: periods += 1 if is_number(endpoint): # force consistency between start/end/freq (lower end if freq skips it) if com._all_not_none(start, end, freq): end -= (end - start) % freq # compute the period/start/end if unspecified (at most one) if periods is None: periods = int((end - start) // freq) + 1 elif start is None: start = end - (periods - 1) * freq elif end is None: end = start + (periods - 1) * freq breaks = np.linspace(start, end, periods) if all(is_integer(x) for x in com._not_none(start, end, freq)): # np.linspace always produces float output breaks = maybe_downcast_to_dtype(breaks, 'int64') else: # delegate to the appropriate range function if isinstance(endpoint, Timestamp): range_func = date_range else: range_func = timedelta_range breaks = range_func(start=start, end=end, periods=periods, freq=freq) return IntervalIndex.from_breaks(breaks, name=name, closed=closed)
def interval_range(start=None, end=None, periods=None, freq=None, name=None, closed='right'): """ Return a fixed frequency IntervalIndex Parameters ---------- start : numeric or datetime-like, default None Left bound for generating intervals end : numeric or datetime-like, default None Right bound for generating intervals periods : integer, default None Number of periods to generate freq : numeric, string, or DateOffset, default None The length of each interval. Must be consistent with the type of start and end, e.g. 2 for numeric, or '5H' for datetime-like. Default is 1 for numeric and 'D' for datetime-like. name : string, default None Name of the resulting IntervalIndex closed : {'left', 'right', 'both', 'neither'}, default 'right' Whether the intervals are closed on the left-side, right-side, both or neither. Returns ------- rng : IntervalIndex See Also -------- IntervalIndex : An Index of intervals that are all closed on the same side. Notes ----- Of the four parameters ``start``, ``end``, ``periods``, and ``freq``, exactly three must be specified. If ``freq`` is omitted, the resulting ``IntervalIndex`` will have ``periods`` linearly spaced elements between ``start`` and ``end``, inclusively. To learn more about datetime-like frequency strings, please see `this link <http://pandas.pydata.org/pandas-docs/stable/timeseries.html#offset-aliases>`__. Examples -------- Numeric ``start`` and ``end`` is supported. >>> pd.interval_range(start=0, end=5) IntervalIndex([(0, 1], (1, 2], (2, 3], (3, 4], (4, 5]], closed='right', dtype='interval[int64]') Additionally, datetime-like input is also supported. >>> pd.interval_range(start=pd.Timestamp('2017-01-01'), ... end=pd.Timestamp('2017-01-04')) IntervalIndex([(2017-01-01, 2017-01-02], (2017-01-02, 2017-01-03], (2017-01-03, 2017-01-04]], closed='right', dtype='interval[datetime64[ns]]') The ``freq`` parameter specifies the frequency between the left and right. endpoints of the individual intervals within the ``IntervalIndex``. For numeric ``start`` and ``end``, the frequency must also be numeric. >>> pd.interval_range(start=0, periods=4, freq=1.5) IntervalIndex([(0.0, 1.5], (1.5, 3.0], (3.0, 4.5], (4.5, 6.0]], closed='right', dtype='interval[float64]') Similarly, for datetime-like ``start`` and ``end``, the frequency must be convertible to a DateOffset. >>> pd.interval_range(start=pd.Timestamp('2017-01-01'), ... periods=3, freq='MS') IntervalIndex([(2017-01-01, 2017-02-01], (2017-02-01, 2017-03-01], (2017-03-01, 2017-04-01]], closed='right', dtype='interval[datetime64[ns]]') Specify ``start``, ``end``, and ``periods``; the frequency is generated automatically (linearly spaced). >>> pd.interval_range(start=0, end=6, periods=4) IntervalIndex([(0.0, 1.5], (1.5, 3.0], (3.0, 4.5], (4.5, 6.0]], closed='right', dtype='interval[float64]') The ``closed`` parameter specifies which endpoints of the individual intervals within the ``IntervalIndex`` are closed. >>> pd.interval_range(end=5, periods=4, closed='both') IntervalIndex([[1, 2], [2, 3], [3, 4], [4, 5]], closed='both', dtype='interval[int64]') """ start = com.maybe_box_datetimelike(start) end = com.maybe_box_datetimelike(end) endpoint = start if start is not None else end if freq is None and com._any_none(periods, start, end): freq = 1 if is_number(endpoint) else 'D' if com.count_not_none(start, end, periods, freq) != 3: raise ValueError('Of the four parameters: start, end, periods, and ' 'freq, exactly three must be specified') if not _is_valid_endpoint(start): msg = 'start must be numeric or datetime-like, got {start}' raise ValueError(msg.format(start=start)) elif not _is_valid_endpoint(end): msg = 'end must be numeric or datetime-like, got {end}' raise ValueError(msg.format(end=end)) if is_float(periods): periods = int(periods) elif not is_integer(periods) and periods is not None: msg = 'periods must be a number, got {periods}' raise TypeError(msg.format(periods=periods)) if freq is not None and not is_number(freq): try: freq = to_offset(freq) except ValueError: raise ValueError('freq must be numeric or convertible to ' 'DateOffset, got {freq}'.format(freq=freq)) # verify type compatibility if not all([_is_type_compatible(start, end), _is_type_compatible(start, freq), _is_type_compatible(end, freq)]): raise TypeError("start, end, freq need to be type compatible") # +1 to convert interval count to breaks count (n breaks = n-1 intervals) if periods is not None: periods += 1 if is_number(endpoint): # force consistency between start/end/freq (lower end if freq skips it) if com._all_not_none(start, end, freq): end -= (end - start) % freq # compute the period/start/end if unspecified (at most one) if periods is None: periods = int((end - start) // freq) + 1 elif start is None: start = end - (periods - 1) * freq elif end is None: end = start + (periods - 1) * freq breaks = np.linspace(start, end, periods) if all(is_integer(x) for x in com._not_none(start, end, freq)): # np.linspace always produces float output breaks = maybe_downcast_to_dtype(breaks, 'int64') else: # delegate to the appropriate range function if isinstance(endpoint, Timestamp): range_func = date_range else: range_func = timedelta_range breaks = range_func(start=start, end=end, periods=periods, freq=freq) return IntervalIndex.from_breaks(breaks, name=name, closed=closed)
[ "Return", "a", "fixed", "frequency", "IntervalIndex" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/indexes/interval.py#L1156-L1312
[ "def", "interval_range", "(", "start", "=", "None", ",", "end", "=", "None", ",", "periods", "=", "None", ",", "freq", "=", "None", ",", "name", "=", "None", ",", "closed", "=", "'right'", ")", ":", "start", "=", "com", ".", "maybe_box_datetimelike", "(", "start", ")", "end", "=", "com", ".", "maybe_box_datetimelike", "(", "end", ")", "endpoint", "=", "start", "if", "start", "is", "not", "None", "else", "end", "if", "freq", "is", "None", "and", "com", ".", "_any_none", "(", "periods", ",", "start", ",", "end", ")", ":", "freq", "=", "1", "if", "is_number", "(", "endpoint", ")", "else", "'D'", "if", "com", ".", "count_not_none", "(", "start", ",", "end", ",", "periods", ",", "freq", ")", "!=", "3", ":", "raise", "ValueError", "(", "'Of the four parameters: start, end, periods, and '", "'freq, exactly three must be specified'", ")", "if", "not", "_is_valid_endpoint", "(", "start", ")", ":", "msg", "=", "'start must be numeric or datetime-like, got {start}'", "raise", "ValueError", "(", "msg", ".", "format", "(", "start", "=", "start", ")", ")", "elif", "not", "_is_valid_endpoint", "(", "end", ")", ":", "msg", "=", "'end must be numeric or datetime-like, got {end}'", "raise", "ValueError", "(", "msg", ".", "format", "(", "end", "=", "end", ")", ")", "if", "is_float", "(", "periods", ")", ":", "periods", "=", "int", "(", "periods", ")", "elif", "not", "is_integer", "(", "periods", ")", "and", "periods", "is", "not", "None", ":", "msg", "=", "'periods must be a number, got {periods}'", "raise", "TypeError", "(", "msg", ".", "format", "(", "periods", "=", "periods", ")", ")", "if", "freq", "is", "not", "None", "and", "not", "is_number", "(", "freq", ")", ":", "try", ":", "freq", "=", "to_offset", "(", "freq", ")", "except", "ValueError", ":", "raise", "ValueError", "(", "'freq must be numeric or convertible to '", "'DateOffset, got {freq}'", ".", "format", "(", "freq", "=", "freq", ")", ")", "# verify type compatibility", "if", "not", "all", "(", "[", "_is_type_compatible", "(", "start", ",", "end", ")", ",", "_is_type_compatible", "(", "start", ",", "freq", ")", ",", "_is_type_compatible", "(", "end", ",", "freq", ")", "]", ")", ":", "raise", "TypeError", "(", "\"start, end, freq need to be type compatible\"", ")", "# +1 to convert interval count to breaks count (n breaks = n-1 intervals)", "if", "periods", "is", "not", "None", ":", "periods", "+=", "1", "if", "is_number", "(", "endpoint", ")", ":", "# force consistency between start/end/freq (lower end if freq skips it)", "if", "com", ".", "_all_not_none", "(", "start", ",", "end", ",", "freq", ")", ":", "end", "-=", "(", "end", "-", "start", ")", "%", "freq", "# compute the period/start/end if unspecified (at most one)", "if", "periods", "is", "None", ":", "periods", "=", "int", "(", "(", "end", "-", "start", ")", "//", "freq", ")", "+", "1", "elif", "start", "is", "None", ":", "start", "=", "end", "-", "(", "periods", "-", "1", ")", "*", "freq", "elif", "end", "is", "None", ":", "end", "=", "start", "+", "(", "periods", "-", "1", ")", "*", "freq", "breaks", "=", "np", ".", "linspace", "(", "start", ",", "end", ",", "periods", ")", "if", "all", "(", "is_integer", "(", "x", ")", "for", "x", "in", "com", ".", "_not_none", "(", "start", ",", "end", ",", "freq", ")", ")", ":", "# np.linspace always produces float output", "breaks", "=", "maybe_downcast_to_dtype", "(", "breaks", ",", "'int64'", ")", "else", ":", "# delegate to the appropriate range function", "if", "isinstance", "(", "endpoint", ",", "Timestamp", ")", ":", "range_func", "=", "date_range", "else", ":", "range_func", "=", "timedelta_range", "breaks", "=", "range_func", "(", "start", "=", "start", ",", "end", "=", "end", ",", "periods", "=", "periods", ",", "freq", "=", "freq", ")", "return", "IntervalIndex", ".", "from_breaks", "(", "breaks", ",", "name", "=", "name", ",", "closed", "=", "closed", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
CSVFormatter.save
Create the writer & save
pandas/io/formats/csvs.py
def save(self): """ Create the writer & save """ # GH21227 internal compression is not used when file-like passed. if self.compression and hasattr(self.path_or_buf, 'write'): msg = ("compression has no effect when passing file-like " "object as input.") warnings.warn(msg, RuntimeWarning, stacklevel=2) # when zip compression is called. is_zip = isinstance(self.path_or_buf, ZipFile) or ( not hasattr(self.path_or_buf, 'write') and self.compression == 'zip') if is_zip: # zipfile doesn't support writing string to archive. uses string # buffer to receive csv writing and dump into zip compression # file handle. GH21241, GH21118 f = StringIO() close = False elif hasattr(self.path_or_buf, 'write'): f = self.path_or_buf close = False else: f, handles = _get_handle(self.path_or_buf, self.mode, encoding=self.encoding, compression=self.compression) close = True try: writer_kwargs = dict(lineterminator=self.line_terminator, delimiter=self.sep, quoting=self.quoting, doublequote=self.doublequote, escapechar=self.escapechar, quotechar=self.quotechar) if self.encoding == 'ascii': self.writer = csvlib.writer(f, **writer_kwargs) else: writer_kwargs['encoding'] = self.encoding self.writer = UnicodeWriter(f, **writer_kwargs) self._save() finally: if is_zip: # GH17778 handles zip compression separately. buf = f.getvalue() if hasattr(self.path_or_buf, 'write'): self.path_or_buf.write(buf) else: f, handles = _get_handle(self.path_or_buf, self.mode, encoding=self.encoding, compression=self.compression) f.write(buf) close = True if close: f.close() for _fh in handles: _fh.close()
def save(self): """ Create the writer & save """ # GH21227 internal compression is not used when file-like passed. if self.compression and hasattr(self.path_or_buf, 'write'): msg = ("compression has no effect when passing file-like " "object as input.") warnings.warn(msg, RuntimeWarning, stacklevel=2) # when zip compression is called. is_zip = isinstance(self.path_or_buf, ZipFile) or ( not hasattr(self.path_or_buf, 'write') and self.compression == 'zip') if is_zip: # zipfile doesn't support writing string to archive. uses string # buffer to receive csv writing and dump into zip compression # file handle. GH21241, GH21118 f = StringIO() close = False elif hasattr(self.path_or_buf, 'write'): f = self.path_or_buf close = False else: f, handles = _get_handle(self.path_or_buf, self.mode, encoding=self.encoding, compression=self.compression) close = True try: writer_kwargs = dict(lineterminator=self.line_terminator, delimiter=self.sep, quoting=self.quoting, doublequote=self.doublequote, escapechar=self.escapechar, quotechar=self.quotechar) if self.encoding == 'ascii': self.writer = csvlib.writer(f, **writer_kwargs) else: writer_kwargs['encoding'] = self.encoding self.writer = UnicodeWriter(f, **writer_kwargs) self._save() finally: if is_zip: # GH17778 handles zip compression separately. buf = f.getvalue() if hasattr(self.path_or_buf, 'write'): self.path_or_buf.write(buf) else: f, handles = _get_handle(self.path_or_buf, self.mode, encoding=self.encoding, compression=self.compression) f.write(buf) close = True if close: f.close() for _fh in handles: _fh.close()
[ "Create", "the", "writer", "&", "save" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/csvs.py#L125-L184
[ "def", "save", "(", "self", ")", ":", "# GH21227 internal compression is not used when file-like passed.", "if", "self", ".", "compression", "and", "hasattr", "(", "self", ".", "path_or_buf", ",", "'write'", ")", ":", "msg", "=", "(", "\"compression has no effect when passing file-like \"", "\"object as input.\"", ")", "warnings", ".", "warn", "(", "msg", ",", "RuntimeWarning", ",", "stacklevel", "=", "2", ")", "# when zip compression is called.", "is_zip", "=", "isinstance", "(", "self", ".", "path_or_buf", ",", "ZipFile", ")", "or", "(", "not", "hasattr", "(", "self", ".", "path_or_buf", ",", "'write'", ")", "and", "self", ".", "compression", "==", "'zip'", ")", "if", "is_zip", ":", "# zipfile doesn't support writing string to archive. uses string", "# buffer to receive csv writing and dump into zip compression", "# file handle. GH21241, GH21118", "f", "=", "StringIO", "(", ")", "close", "=", "False", "elif", "hasattr", "(", "self", ".", "path_or_buf", ",", "'write'", ")", ":", "f", "=", "self", ".", "path_or_buf", "close", "=", "False", "else", ":", "f", ",", "handles", "=", "_get_handle", "(", "self", ".", "path_or_buf", ",", "self", ".", "mode", ",", "encoding", "=", "self", ".", "encoding", ",", "compression", "=", "self", ".", "compression", ")", "close", "=", "True", "try", ":", "writer_kwargs", "=", "dict", "(", "lineterminator", "=", "self", ".", "line_terminator", ",", "delimiter", "=", "self", ".", "sep", ",", "quoting", "=", "self", ".", "quoting", ",", "doublequote", "=", "self", ".", "doublequote", ",", "escapechar", "=", "self", ".", "escapechar", ",", "quotechar", "=", "self", ".", "quotechar", ")", "if", "self", ".", "encoding", "==", "'ascii'", ":", "self", ".", "writer", "=", "csvlib", ".", "writer", "(", "f", ",", "*", "*", "writer_kwargs", ")", "else", ":", "writer_kwargs", "[", "'encoding'", "]", "=", "self", ".", "encoding", "self", ".", "writer", "=", "UnicodeWriter", "(", "f", ",", "*", "*", "writer_kwargs", ")", "self", ".", "_save", "(", ")", "finally", ":", "if", "is_zip", ":", "# GH17778 handles zip compression separately.", "buf", "=", "f", ".", "getvalue", "(", ")", "if", "hasattr", "(", "self", ".", "path_or_buf", ",", "'write'", ")", ":", "self", ".", "path_or_buf", ".", "write", "(", "buf", ")", "else", ":", "f", ",", "handles", "=", "_get_handle", "(", "self", ".", "path_or_buf", ",", "self", ".", "mode", ",", "encoding", "=", "self", ".", "encoding", ",", "compression", "=", "self", ".", "compression", ")", "f", ".", "write", "(", "buf", ")", "close", "=", "True", "if", "close", ":", "f", ".", "close", "(", ")", "for", "_fh", "in", "handles", ":", "_fh", ".", "close", "(", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
delegate_names
Add delegated names to a class using a class decorator. This provides an alternative usage to directly calling `_add_delegate_accessors` below a class definition. Parameters ---------- delegate : object the class to get methods/properties & doc-strings accessors : Sequence[str] List of accessor to add typ : {'property', 'method'} overwrite : boolean, default False overwrite the method/property in the target class if it exists Returns ------- callable A class decorator. Examples -------- @delegate_names(Categorical, ["categories", "ordered"], "property") class CategoricalAccessor(PandasDelegate): [...]
pandas/core/accessor.py
def delegate_names(delegate, accessors, typ, overwrite=False): """ Add delegated names to a class using a class decorator. This provides an alternative usage to directly calling `_add_delegate_accessors` below a class definition. Parameters ---------- delegate : object the class to get methods/properties & doc-strings accessors : Sequence[str] List of accessor to add typ : {'property', 'method'} overwrite : boolean, default False overwrite the method/property in the target class if it exists Returns ------- callable A class decorator. Examples -------- @delegate_names(Categorical, ["categories", "ordered"], "property") class CategoricalAccessor(PandasDelegate): [...] """ def add_delegate_accessors(cls): cls._add_delegate_accessors(delegate, accessors, typ, overwrite=overwrite) return cls return add_delegate_accessors
def delegate_names(delegate, accessors, typ, overwrite=False): """ Add delegated names to a class using a class decorator. This provides an alternative usage to directly calling `_add_delegate_accessors` below a class definition. Parameters ---------- delegate : object the class to get methods/properties & doc-strings accessors : Sequence[str] List of accessor to add typ : {'property', 'method'} overwrite : boolean, default False overwrite the method/property in the target class if it exists Returns ------- callable A class decorator. Examples -------- @delegate_names(Categorical, ["categories", "ordered"], "property") class CategoricalAccessor(PandasDelegate): [...] """ def add_delegate_accessors(cls): cls._add_delegate_accessors(delegate, accessors, typ, overwrite=overwrite) return cls return add_delegate_accessors
[ "Add", "delegated", "names", "to", "a", "class", "using", "a", "class", "decorator", ".", "This", "provides", "an", "alternative", "usage", "to", "directly", "calling", "_add_delegate_accessors", "below", "a", "class", "definition", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/accessor.py#L114-L146
[ "def", "delegate_names", "(", "delegate", ",", "accessors", ",", "typ", ",", "overwrite", "=", "False", ")", ":", "def", "add_delegate_accessors", "(", "cls", ")", ":", "cls", ".", "_add_delegate_accessors", "(", "delegate", ",", "accessors", ",", "typ", ",", "overwrite", "=", "overwrite", ")", "return", "cls", "return", "add_delegate_accessors" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
DirNamesMixin._dir_additions
Add additional __dir__ for this object.
pandas/core/accessor.py
def _dir_additions(self): """ Add additional __dir__ for this object. """ rv = set() for accessor in self._accessors: try: getattr(self, accessor) rv.add(accessor) except AttributeError: pass return rv
def _dir_additions(self): """ Add additional __dir__ for this object. """ rv = set() for accessor in self._accessors: try: getattr(self, accessor) rv.add(accessor) except AttributeError: pass return rv
[ "Add", "additional", "__dir__", "for", "this", "object", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/accessor.py#L24-L35
[ "def", "_dir_additions", "(", "self", ")", ":", "rv", "=", "set", "(", ")", "for", "accessor", "in", "self", ".", "_accessors", ":", "try", ":", "getattr", "(", "self", ",", "accessor", ")", "rv", ".", "add", "(", "accessor", ")", "except", "AttributeError", ":", "pass", "return", "rv" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
PandasDelegate._add_delegate_accessors
Add accessors to cls from the delegate class. Parameters ---------- cls : the class to add the methods/properties to delegate : the class to get methods/properties & doc-strings accessors : string list of accessors to add typ : 'property' or 'method' overwrite : boolean, default False overwrite the method/property in the target class if it exists.
pandas/core/accessor.py
def _add_delegate_accessors(cls, delegate, accessors, typ, overwrite=False): """ Add accessors to cls from the delegate class. Parameters ---------- cls : the class to add the methods/properties to delegate : the class to get methods/properties & doc-strings accessors : string list of accessors to add typ : 'property' or 'method' overwrite : boolean, default False overwrite the method/property in the target class if it exists. """ def _create_delegator_property(name): def _getter(self): return self._delegate_property_get(name) def _setter(self, new_values): return self._delegate_property_set(name, new_values) _getter.__name__ = name _setter.__name__ = name return property(fget=_getter, fset=_setter, doc=getattr(delegate, name).__doc__) def _create_delegator_method(name): def f(self, *args, **kwargs): return self._delegate_method(name, *args, **kwargs) f.__name__ = name f.__doc__ = getattr(delegate, name).__doc__ return f for name in accessors: if typ == 'property': f = _create_delegator_property(name) else: f = _create_delegator_method(name) # don't overwrite existing methods/properties if overwrite or not hasattr(cls, name): setattr(cls, name, f)
def _add_delegate_accessors(cls, delegate, accessors, typ, overwrite=False): """ Add accessors to cls from the delegate class. Parameters ---------- cls : the class to add the methods/properties to delegate : the class to get methods/properties & doc-strings accessors : string list of accessors to add typ : 'property' or 'method' overwrite : boolean, default False overwrite the method/property in the target class if it exists. """ def _create_delegator_property(name): def _getter(self): return self._delegate_property_get(name) def _setter(self, new_values): return self._delegate_property_set(name, new_values) _getter.__name__ = name _setter.__name__ = name return property(fget=_getter, fset=_setter, doc=getattr(delegate, name).__doc__) def _create_delegator_method(name): def f(self, *args, **kwargs): return self._delegate_method(name, *args, **kwargs) f.__name__ = name f.__doc__ = getattr(delegate, name).__doc__ return f for name in accessors: if typ == 'property': f = _create_delegator_property(name) else: f = _create_delegator_method(name) # don't overwrite existing methods/properties if overwrite or not hasattr(cls, name): setattr(cls, name, f)
[ "Add", "accessors", "to", "cls", "from", "the", "delegate", "class", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/accessor.py#L63-L111
[ "def", "_add_delegate_accessors", "(", "cls", ",", "delegate", ",", "accessors", ",", "typ", ",", "overwrite", "=", "False", ")", ":", "def", "_create_delegator_property", "(", "name", ")", ":", "def", "_getter", "(", "self", ")", ":", "return", "self", ".", "_delegate_property_get", "(", "name", ")", "def", "_setter", "(", "self", ",", "new_values", ")", ":", "return", "self", ".", "_delegate_property_set", "(", "name", ",", "new_values", ")", "_getter", ".", "__name__", "=", "name", "_setter", ".", "__name__", "=", "name", "return", "property", "(", "fget", "=", "_getter", ",", "fset", "=", "_setter", ",", "doc", "=", "getattr", "(", "delegate", ",", "name", ")", ".", "__doc__", ")", "def", "_create_delegator_method", "(", "name", ")", ":", "def", "f", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "self", ".", "_delegate_method", "(", "name", ",", "*", "args", ",", "*", "*", "kwargs", ")", "f", ".", "__name__", "=", "name", "f", ".", "__doc__", "=", "getattr", "(", "delegate", ",", "name", ")", ".", "__doc__", "return", "f", "for", "name", "in", "accessors", ":", "if", "typ", "==", "'property'", ":", "f", "=", "_create_delegator_property", "(", "name", ")", "else", ":", "f", "=", "_create_delegator_method", "(", "name", ")", "# don't overwrite existing methods/properties", "if", "overwrite", "or", "not", "hasattr", "(", "cls", ",", "name", ")", ":", "setattr", "(", "cls", ",", "name", ",", "f", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_evaluate_standard
standard evaluation
pandas/core/computation/expressions.py
def _evaluate_standard(op, op_str, a, b, **eval_kwargs): """ standard evaluation """ if _TEST_MODE: _store_test_result(False) with np.errstate(all='ignore'): return op(a, b)
def _evaluate_standard(op, op_str, a, b, **eval_kwargs): """ standard evaluation """ if _TEST_MODE: _store_test_result(False) with np.errstate(all='ignore'): return op(a, b)
[ "standard", "evaluation" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/expressions.py#L64-L69
[ "def", "_evaluate_standard", "(", "op", ",", "op_str", ",", "a", ",", "b", ",", "*", "*", "eval_kwargs", ")", ":", "if", "_TEST_MODE", ":", "_store_test_result", "(", "False", ")", "with", "np", ".", "errstate", "(", "all", "=", "'ignore'", ")", ":", "return", "op", "(", "a", ",", "b", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_can_use_numexpr
return a boolean if we WILL be using numexpr
pandas/core/computation/expressions.py
def _can_use_numexpr(op, op_str, a, b, dtype_check): """ return a boolean if we WILL be using numexpr """ if op_str is not None: # required min elements (otherwise we are adding overhead) if np.prod(a.shape) > _MIN_ELEMENTS: # check for dtype compatibility dtypes = set() for o in [a, b]: if hasattr(o, 'get_dtype_counts'): s = o.get_dtype_counts() if len(s) > 1: return False dtypes |= set(s.index) elif isinstance(o, np.ndarray): dtypes |= {o.dtype.name} # allowed are a superset if not len(dtypes) or _ALLOWED_DTYPES[dtype_check] >= dtypes: return True return False
def _can_use_numexpr(op, op_str, a, b, dtype_check): """ return a boolean if we WILL be using numexpr """ if op_str is not None: # required min elements (otherwise we are adding overhead) if np.prod(a.shape) > _MIN_ELEMENTS: # check for dtype compatibility dtypes = set() for o in [a, b]: if hasattr(o, 'get_dtype_counts'): s = o.get_dtype_counts() if len(s) > 1: return False dtypes |= set(s.index) elif isinstance(o, np.ndarray): dtypes |= {o.dtype.name} # allowed are a superset if not len(dtypes) or _ALLOWED_DTYPES[dtype_check] >= dtypes: return True return False
[ "return", "a", "boolean", "if", "we", "WILL", "be", "using", "numexpr" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/expressions.py#L72-L94
[ "def", "_can_use_numexpr", "(", "op", ",", "op_str", ",", "a", ",", "b", ",", "dtype_check", ")", ":", "if", "op_str", "is", "not", "None", ":", "# required min elements (otherwise we are adding overhead)", "if", "np", ".", "prod", "(", "a", ".", "shape", ")", ">", "_MIN_ELEMENTS", ":", "# check for dtype compatibility", "dtypes", "=", "set", "(", ")", "for", "o", "in", "[", "a", ",", "b", "]", ":", "if", "hasattr", "(", "o", ",", "'get_dtype_counts'", ")", ":", "s", "=", "o", ".", "get_dtype_counts", "(", ")", "if", "len", "(", "s", ")", ">", "1", ":", "return", "False", "dtypes", "|=", "set", "(", "s", ".", "index", ")", "elif", "isinstance", "(", "o", ",", "np", ".", "ndarray", ")", ":", "dtypes", "|=", "{", "o", ".", "dtype", ".", "name", "}", "# allowed are a superset", "if", "not", "len", "(", "dtypes", ")", "or", "_ALLOWED_DTYPES", "[", "dtype_check", "]", ">=", "dtypes", ":", "return", "True", "return", "False" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
evaluate
evaluate and return the expression of the op on a and b Parameters ---------- op : the actual operand op_str: the string version of the op a : left operand b : right operand use_numexpr : whether to try to use numexpr (default True)
pandas/core/computation/expressions.py
def evaluate(op, op_str, a, b, use_numexpr=True, **eval_kwargs): """ evaluate and return the expression of the op on a and b Parameters ---------- op : the actual operand op_str: the string version of the op a : left operand b : right operand use_numexpr : whether to try to use numexpr (default True) """ use_numexpr = use_numexpr and _bool_arith_check(op_str, a, b) if use_numexpr: return _evaluate(op, op_str, a, b, **eval_kwargs) return _evaluate_standard(op, op_str, a, b)
def evaluate(op, op_str, a, b, use_numexpr=True, **eval_kwargs): """ evaluate and return the expression of the op on a and b Parameters ---------- op : the actual operand op_str: the string version of the op a : left operand b : right operand use_numexpr : whether to try to use numexpr (default True) """ use_numexpr = use_numexpr and _bool_arith_check(op_str, a, b) if use_numexpr: return _evaluate(op, op_str, a, b, **eval_kwargs) return _evaluate_standard(op, op_str, a, b)
[ "evaluate", "and", "return", "the", "expression", "of", "the", "op", "on", "a", "and", "b" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/expressions.py#L193-L210
[ "def", "evaluate", "(", "op", ",", "op_str", ",", "a", ",", "b", ",", "use_numexpr", "=", "True", ",", "*", "*", "eval_kwargs", ")", ":", "use_numexpr", "=", "use_numexpr", "and", "_bool_arith_check", "(", "op_str", ",", "a", ",", "b", ")", "if", "use_numexpr", ":", "return", "_evaluate", "(", "op", ",", "op_str", ",", "a", ",", "b", ",", "*", "*", "eval_kwargs", ")", "return", "_evaluate_standard", "(", "op", ",", "op_str", ",", "a", ",", "b", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
where
evaluate the where condition cond on a and b Parameters ---------- cond : a boolean array a : return if cond is True b : return if cond is False use_numexpr : whether to try to use numexpr (default True)
pandas/core/computation/expressions.py
def where(cond, a, b, use_numexpr=True): """ evaluate the where condition cond on a and b Parameters ---------- cond : a boolean array a : return if cond is True b : return if cond is False use_numexpr : whether to try to use numexpr (default True) """ if use_numexpr: return _where(cond, a, b) return _where_standard(cond, a, b)
def where(cond, a, b, use_numexpr=True): """ evaluate the where condition cond on a and b Parameters ---------- cond : a boolean array a : return if cond is True b : return if cond is False use_numexpr : whether to try to use numexpr (default True) """ if use_numexpr: return _where(cond, a, b) return _where_standard(cond, a, b)
[ "evaluate", "the", "where", "condition", "cond", "on", "a", "and", "b" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/computation/expressions.py#L213-L227
[ "def", "where", "(", "cond", ",", "a", ",", "b", ",", "use_numexpr", "=", "True", ")", ":", "if", "use_numexpr", ":", "return", "_where", "(", "cond", ",", "a", ",", "b", ")", "return", "_where_standard", "(", "cond", ",", "a", ",", "b", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ExcelFormatter.write
writer : string or ExcelWriter object File path or existing ExcelWriter sheet_name : string, default 'Sheet1' Name of sheet which will contain DataFrame startrow : upper left cell row to dump data frame startcol : upper left cell column to dump data frame freeze_panes : tuple of integer (length 2), default None Specifies the one-based bottommost row and rightmost column that is to be frozen engine : string, default None write engine to use if writer is a path - you can also set this via the options ``io.excel.xlsx.writer``, ``io.excel.xls.writer``, and ``io.excel.xlsm.writer``.
pandas/io/formats/excel.py
def write(self, writer, sheet_name='Sheet1', startrow=0, startcol=0, freeze_panes=None, engine=None): """ writer : string or ExcelWriter object File path or existing ExcelWriter sheet_name : string, default 'Sheet1' Name of sheet which will contain DataFrame startrow : upper left cell row to dump data frame startcol : upper left cell column to dump data frame freeze_panes : tuple of integer (length 2), default None Specifies the one-based bottommost row and rightmost column that is to be frozen engine : string, default None write engine to use if writer is a path - you can also set this via the options ``io.excel.xlsx.writer``, ``io.excel.xls.writer``, and ``io.excel.xlsm.writer``. """ from pandas.io.excel import ExcelWriter from pandas.io.common import _stringify_path if isinstance(writer, ExcelWriter): need_save = False else: writer = ExcelWriter(_stringify_path(writer), engine=engine) need_save = True formatted_cells = self.get_formatted_cells() writer.write_cells(formatted_cells, sheet_name, startrow=startrow, startcol=startcol, freeze_panes=freeze_panes) if need_save: writer.save()
def write(self, writer, sheet_name='Sheet1', startrow=0, startcol=0, freeze_panes=None, engine=None): """ writer : string or ExcelWriter object File path or existing ExcelWriter sheet_name : string, default 'Sheet1' Name of sheet which will contain DataFrame startrow : upper left cell row to dump data frame startcol : upper left cell column to dump data frame freeze_panes : tuple of integer (length 2), default None Specifies the one-based bottommost row and rightmost column that is to be frozen engine : string, default None write engine to use if writer is a path - you can also set this via the options ``io.excel.xlsx.writer``, ``io.excel.xls.writer``, and ``io.excel.xlsm.writer``. """ from pandas.io.excel import ExcelWriter from pandas.io.common import _stringify_path if isinstance(writer, ExcelWriter): need_save = False else: writer = ExcelWriter(_stringify_path(writer), engine=engine) need_save = True formatted_cells = self.get_formatted_cells() writer.write_cells(formatted_cells, sheet_name, startrow=startrow, startcol=startcol, freeze_panes=freeze_panes) if need_save: writer.save()
[ "writer", ":", "string", "or", "ExcelWriter", "object", "File", "path", "or", "existing", "ExcelWriter", "sheet_name", ":", "string", "default", "Sheet1", "Name", "of", "sheet", "which", "will", "contain", "DataFrame", "startrow", ":", "upper", "left", "cell", "row", "to", "dump", "data", "frame", "startcol", ":", "upper", "left", "cell", "column", "to", "dump", "data", "frame", "freeze_panes", ":", "tuple", "of", "integer", "(", "length", "2", ")", "default", "None", "Specifies", "the", "one", "-", "based", "bottommost", "row", "and", "rightmost", "column", "that", "is", "to", "be", "frozen", "engine", ":", "string", "default", "None", "write", "engine", "to", "use", "if", "writer", "is", "a", "path", "-", "you", "can", "also", "set", "this", "via", "the", "options", "io", ".", "excel", ".", "xlsx", ".", "writer", "io", ".", "excel", ".", "xls", ".", "writer", "and", "io", ".", "excel", ".", "xlsm", ".", "writer", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/formats/excel.py#L629-L662
[ "def", "write", "(", "self", ",", "writer", ",", "sheet_name", "=", "'Sheet1'", ",", "startrow", "=", "0", ",", "startcol", "=", "0", ",", "freeze_panes", "=", "None", ",", "engine", "=", "None", ")", ":", "from", "pandas", ".", "io", ".", "excel", "import", "ExcelWriter", "from", "pandas", ".", "io", ".", "common", "import", "_stringify_path", "if", "isinstance", "(", "writer", ",", "ExcelWriter", ")", ":", "need_save", "=", "False", "else", ":", "writer", "=", "ExcelWriter", "(", "_stringify_path", "(", "writer", ")", ",", "engine", "=", "engine", ")", "need_save", "=", "True", "formatted_cells", "=", "self", ".", "get_formatted_cells", "(", ")", "writer", ".", "write_cells", "(", "formatted_cells", ",", "sheet_name", ",", "startrow", "=", "startrow", ",", "startcol", "=", "startcol", ",", "freeze_panes", "=", "freeze_panes", ")", "if", "need_save", ":", "writer", ".", "save", "(", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
to_feather
Write a DataFrame to the feather-format Parameters ---------- df : DataFrame path : string file path, or file-like object
pandas/io/feather_format.py
def to_feather(df, path): """ Write a DataFrame to the feather-format Parameters ---------- df : DataFrame path : string file path, or file-like object """ path = _stringify_path(path) if not isinstance(df, DataFrame): raise ValueError("feather only support IO with DataFrames") feather = _try_import()[0] valid_types = {'string', 'unicode'} # validate index # -------------- # validate that we have only a default index # raise on anything else as we don't serialize the index if not isinstance(df.index, Int64Index): raise ValueError("feather does not support serializing {} " "for the index; you can .reset_index()" "to make the index into column(s)".format( type(df.index))) if not df.index.equals(RangeIndex.from_range(range(len(df)))): raise ValueError("feather does not support serializing a " "non-default index for the index; you " "can .reset_index() to make the index " "into column(s)") if df.index.name is not None: raise ValueError("feather does not serialize index meta-data on a " "default index") # validate columns # ---------------- # must have value column names (strings only) if df.columns.inferred_type not in valid_types: raise ValueError("feather must have string column names") feather.write_feather(df, path)
def to_feather(df, path): """ Write a DataFrame to the feather-format Parameters ---------- df : DataFrame path : string file path, or file-like object """ path = _stringify_path(path) if not isinstance(df, DataFrame): raise ValueError("feather only support IO with DataFrames") feather = _try_import()[0] valid_types = {'string', 'unicode'} # validate index # -------------- # validate that we have only a default index # raise on anything else as we don't serialize the index if not isinstance(df.index, Int64Index): raise ValueError("feather does not support serializing {} " "for the index; you can .reset_index()" "to make the index into column(s)".format( type(df.index))) if not df.index.equals(RangeIndex.from_range(range(len(df)))): raise ValueError("feather does not support serializing a " "non-default index for the index; you " "can .reset_index() to make the index " "into column(s)") if df.index.name is not None: raise ValueError("feather does not serialize index meta-data on a " "default index") # validate columns # ---------------- # must have value column names (strings only) if df.columns.inferred_type not in valid_types: raise ValueError("feather must have string column names") feather.write_feather(df, path)
[ "Write", "a", "DataFrame", "to", "the", "feather", "-", "format" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/feather_format.py#L36-L82
[ "def", "to_feather", "(", "df", ",", "path", ")", ":", "path", "=", "_stringify_path", "(", "path", ")", "if", "not", "isinstance", "(", "df", ",", "DataFrame", ")", ":", "raise", "ValueError", "(", "\"feather only support IO with DataFrames\"", ")", "feather", "=", "_try_import", "(", ")", "[", "0", "]", "valid_types", "=", "{", "'string'", ",", "'unicode'", "}", "# validate index", "# --------------", "# validate that we have only a default index", "# raise on anything else as we don't serialize the index", "if", "not", "isinstance", "(", "df", ".", "index", ",", "Int64Index", ")", ":", "raise", "ValueError", "(", "\"feather does not support serializing {} \"", "\"for the index; you can .reset_index()\"", "\"to make the index into column(s)\"", ".", "format", "(", "type", "(", "df", ".", "index", ")", ")", ")", "if", "not", "df", ".", "index", ".", "equals", "(", "RangeIndex", ".", "from_range", "(", "range", "(", "len", "(", "df", ")", ")", ")", ")", ":", "raise", "ValueError", "(", "\"feather does not support serializing a \"", "\"non-default index for the index; you \"", "\"can .reset_index() to make the index \"", "\"into column(s)\"", ")", "if", "df", ".", "index", ".", "name", "is", "not", "None", ":", "raise", "ValueError", "(", "\"feather does not serialize index meta-data on a \"", "\"default index\"", ")", "# validate columns", "# ----------------", "# must have value column names (strings only)", "if", "df", ".", "columns", ".", "inferred_type", "not", "in", "valid_types", ":", "raise", "ValueError", "(", "\"feather must have string column names\"", ")", "feather", ".", "write_feather", "(", "df", ",", "path", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
read_feather
Load a feather-format object from the file path .. versionadded 0.20.0 Parameters ---------- path : string file path, or file-like object columns : sequence, default None If not provided, all columns are read .. versionadded 0.24.0 nthreads : int, default 1 Number of CPU threads to use when reading to pandas.DataFrame .. versionadded 0.21.0 .. deprecated 0.24.0 use_threads : bool, default True Whether to parallelize reading using multiple threads .. versionadded 0.24.0 Returns ------- type of object stored in file
pandas/io/feather_format.py
def read_feather(path, columns=None, use_threads=True): """ Load a feather-format object from the file path .. versionadded 0.20.0 Parameters ---------- path : string file path, or file-like object columns : sequence, default None If not provided, all columns are read .. versionadded 0.24.0 nthreads : int, default 1 Number of CPU threads to use when reading to pandas.DataFrame .. versionadded 0.21.0 .. deprecated 0.24.0 use_threads : bool, default True Whether to parallelize reading using multiple threads .. versionadded 0.24.0 Returns ------- type of object stored in file """ feather, pyarrow = _try_import() path = _stringify_path(path) if LooseVersion(pyarrow.__version__) < LooseVersion('0.11.0'): int_use_threads = int(use_threads) if int_use_threads < 1: int_use_threads = 1 return feather.read_feather(path, columns=columns, nthreads=int_use_threads) return feather.read_feather(path, columns=columns, use_threads=bool(use_threads))
def read_feather(path, columns=None, use_threads=True): """ Load a feather-format object from the file path .. versionadded 0.20.0 Parameters ---------- path : string file path, or file-like object columns : sequence, default None If not provided, all columns are read .. versionadded 0.24.0 nthreads : int, default 1 Number of CPU threads to use when reading to pandas.DataFrame .. versionadded 0.21.0 .. deprecated 0.24.0 use_threads : bool, default True Whether to parallelize reading using multiple threads .. versionadded 0.24.0 Returns ------- type of object stored in file """ feather, pyarrow = _try_import() path = _stringify_path(path) if LooseVersion(pyarrow.__version__) < LooseVersion('0.11.0'): int_use_threads = int(use_threads) if int_use_threads < 1: int_use_threads = 1 return feather.read_feather(path, columns=columns, nthreads=int_use_threads) return feather.read_feather(path, columns=columns, use_threads=bool(use_threads))
[ "Load", "a", "feather", "-", "format", "object", "from", "the", "file", "path" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/io/feather_format.py#L86-L125
[ "def", "read_feather", "(", "path", ",", "columns", "=", "None", ",", "use_threads", "=", "True", ")", ":", "feather", ",", "pyarrow", "=", "_try_import", "(", ")", "path", "=", "_stringify_path", "(", "path", ")", "if", "LooseVersion", "(", "pyarrow", ".", "__version__", ")", "<", "LooseVersion", "(", "'0.11.0'", ")", ":", "int_use_threads", "=", "int", "(", "use_threads", ")", "if", "int_use_threads", "<", "1", ":", "int_use_threads", "=", "1", "return", "feather", ".", "read_feather", "(", "path", ",", "columns", "=", "columns", ",", "nthreads", "=", "int_use_threads", ")", "return", "feather", ".", "read_feather", "(", "path", ",", "columns", "=", "columns", ",", "use_threads", "=", "bool", "(", "use_threads", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
generate_regular_range
Generate a range of dates with the spans between dates described by the given `freq` DateOffset. Parameters ---------- start : Timestamp or None first point of produced date range end : Timestamp or None last point of produced date range periods : int number of periods in produced date range freq : DateOffset describes space between dates in produced date range Returns ------- ndarray[np.int64] representing nanosecond unix timestamps
pandas/core/arrays/_ranges.py
def generate_regular_range(start, end, periods, freq): """ Generate a range of dates with the spans between dates described by the given `freq` DateOffset. Parameters ---------- start : Timestamp or None first point of produced date range end : Timestamp or None last point of produced date range periods : int number of periods in produced date range freq : DateOffset describes space between dates in produced date range Returns ------- ndarray[np.int64] representing nanosecond unix timestamps """ if isinstance(freq, Tick): stride = freq.nanos if periods is None: b = Timestamp(start).value # cannot just use e = Timestamp(end) + 1 because arange breaks when # stride is too large, see GH10887 e = (b + (Timestamp(end).value - b) // stride * stride + stride // 2 + 1) # end.tz == start.tz by this point due to _generate implementation tz = start.tz elif start is not None: b = Timestamp(start).value e = _generate_range_overflow_safe(b, periods, stride, side='start') tz = start.tz elif end is not None: e = Timestamp(end).value + stride b = _generate_range_overflow_safe(e, periods, stride, side='end') tz = end.tz else: raise ValueError("at least 'start' or 'end' should be specified " "if a 'period' is given.") with np.errstate(over="raise"): # If the range is sufficiently large, np.arange may overflow # and incorrectly return an empty array if not caught. try: values = np.arange(b, e, stride, dtype=np.int64) except FloatingPointError: xdr = [b] while xdr[-1] != e: xdr.append(xdr[-1] + stride) values = np.array(xdr[:-1], dtype=np.int64) else: tz = None # start and end should have the same timezone by this point if start is not None: tz = start.tz elif end is not None: tz = end.tz xdr = generate_range(start=start, end=end, periods=periods, offset=freq) values = np.array([x.value for x in xdr], dtype=np.int64) return values, tz
def generate_regular_range(start, end, periods, freq): """ Generate a range of dates with the spans between dates described by the given `freq` DateOffset. Parameters ---------- start : Timestamp or None first point of produced date range end : Timestamp or None last point of produced date range periods : int number of periods in produced date range freq : DateOffset describes space between dates in produced date range Returns ------- ndarray[np.int64] representing nanosecond unix timestamps """ if isinstance(freq, Tick): stride = freq.nanos if periods is None: b = Timestamp(start).value # cannot just use e = Timestamp(end) + 1 because arange breaks when # stride is too large, see GH10887 e = (b + (Timestamp(end).value - b) // stride * stride + stride // 2 + 1) # end.tz == start.tz by this point due to _generate implementation tz = start.tz elif start is not None: b = Timestamp(start).value e = _generate_range_overflow_safe(b, periods, stride, side='start') tz = start.tz elif end is not None: e = Timestamp(end).value + stride b = _generate_range_overflow_safe(e, periods, stride, side='end') tz = end.tz else: raise ValueError("at least 'start' or 'end' should be specified " "if a 'period' is given.") with np.errstate(over="raise"): # If the range is sufficiently large, np.arange may overflow # and incorrectly return an empty array if not caught. try: values = np.arange(b, e, stride, dtype=np.int64) except FloatingPointError: xdr = [b] while xdr[-1] != e: xdr.append(xdr[-1] + stride) values = np.array(xdr[:-1], dtype=np.int64) else: tz = None # start and end should have the same timezone by this point if start is not None: tz = start.tz elif end is not None: tz = end.tz xdr = generate_range(start=start, end=end, periods=periods, offset=freq) values = np.array([x.value for x in xdr], dtype=np.int64) return values, tz
[ "Generate", "a", "range", "of", "dates", "with", "the", "spans", "between", "dates", "described", "by", "the", "given", "freq", "DateOffset", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/arrays/_ranges.py#L13-L79
[ "def", "generate_regular_range", "(", "start", ",", "end", ",", "periods", ",", "freq", ")", ":", "if", "isinstance", "(", "freq", ",", "Tick", ")", ":", "stride", "=", "freq", ".", "nanos", "if", "periods", "is", "None", ":", "b", "=", "Timestamp", "(", "start", ")", ".", "value", "# cannot just use e = Timestamp(end) + 1 because arange breaks when", "# stride is too large, see GH10887", "e", "=", "(", "b", "+", "(", "Timestamp", "(", "end", ")", ".", "value", "-", "b", ")", "//", "stride", "*", "stride", "+", "stride", "//", "2", "+", "1", ")", "# end.tz == start.tz by this point due to _generate implementation", "tz", "=", "start", ".", "tz", "elif", "start", "is", "not", "None", ":", "b", "=", "Timestamp", "(", "start", ")", ".", "value", "e", "=", "_generate_range_overflow_safe", "(", "b", ",", "periods", ",", "stride", ",", "side", "=", "'start'", ")", "tz", "=", "start", ".", "tz", "elif", "end", "is", "not", "None", ":", "e", "=", "Timestamp", "(", "end", ")", ".", "value", "+", "stride", "b", "=", "_generate_range_overflow_safe", "(", "e", ",", "periods", ",", "stride", ",", "side", "=", "'end'", ")", "tz", "=", "end", ".", "tz", "else", ":", "raise", "ValueError", "(", "\"at least 'start' or 'end' should be specified \"", "\"if a 'period' is given.\"", ")", "with", "np", ".", "errstate", "(", "over", "=", "\"raise\"", ")", ":", "# If the range is sufficiently large, np.arange may overflow", "# and incorrectly return an empty array if not caught.", "try", ":", "values", "=", "np", ".", "arange", "(", "b", ",", "e", ",", "stride", ",", "dtype", "=", "np", ".", "int64", ")", "except", "FloatingPointError", ":", "xdr", "=", "[", "b", "]", "while", "xdr", "[", "-", "1", "]", "!=", "e", ":", "xdr", ".", "append", "(", "xdr", "[", "-", "1", "]", "+", "stride", ")", "values", "=", "np", ".", "array", "(", "xdr", "[", ":", "-", "1", "]", ",", "dtype", "=", "np", ".", "int64", ")", "else", ":", "tz", "=", "None", "# start and end should have the same timezone by this point", "if", "start", "is", "not", "None", ":", "tz", "=", "start", ".", "tz", "elif", "end", "is", "not", "None", ":", "tz", "=", "end", ".", "tz", "xdr", "=", "generate_range", "(", "start", "=", "start", ",", "end", "=", "end", ",", "periods", "=", "periods", ",", "offset", "=", "freq", ")", "values", "=", "np", ".", "array", "(", "[", "x", ".", "value", "for", "x", "in", "xdr", "]", ",", "dtype", "=", "np", ".", "int64", ")", "return", "values", ",", "tz" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_generate_range_overflow_safe
Calculate the second endpoint for passing to np.arange, checking to avoid an integer overflow. Catch OverflowError and re-raise as OutOfBoundsDatetime. Parameters ---------- endpoint : int nanosecond timestamp of the known endpoint of the desired range periods : int number of periods in the desired range stride : int nanoseconds between periods in the desired range side : {'start', 'end'} which end of the range `endpoint` refers to Returns ------- other_end : int Raises ------ OutOfBoundsDatetime
pandas/core/arrays/_ranges.py
def _generate_range_overflow_safe(endpoint, periods, stride, side='start'): """ Calculate the second endpoint for passing to np.arange, checking to avoid an integer overflow. Catch OverflowError and re-raise as OutOfBoundsDatetime. Parameters ---------- endpoint : int nanosecond timestamp of the known endpoint of the desired range periods : int number of periods in the desired range stride : int nanoseconds between periods in the desired range side : {'start', 'end'} which end of the range `endpoint` refers to Returns ------- other_end : int Raises ------ OutOfBoundsDatetime """ # GH#14187 raise instead of incorrectly wrapping around assert side in ['start', 'end'] i64max = np.uint64(np.iinfo(np.int64).max) msg = ('Cannot generate range with {side}={endpoint} and ' 'periods={periods}' .format(side=side, endpoint=endpoint, periods=periods)) with np.errstate(over="raise"): # if periods * strides cannot be multiplied within the *uint64* bounds, # we cannot salvage the operation by recursing, so raise try: addend = np.uint64(periods) * np.uint64(np.abs(stride)) except FloatingPointError: raise OutOfBoundsDatetime(msg) if np.abs(addend) <= i64max: # relatively easy case without casting concerns return _generate_range_overflow_safe_signed( endpoint, periods, stride, side) elif ((endpoint > 0 and side == 'start' and stride > 0) or (endpoint < 0 and side == 'end' and stride > 0)): # no chance of not-overflowing raise OutOfBoundsDatetime(msg) elif (side == 'end' and endpoint > i64max and endpoint - stride <= i64max): # in _generate_regular_range we added `stride` thereby overflowing # the bounds. Adjust to fix this. return _generate_range_overflow_safe(endpoint - stride, periods - 1, stride, side) # split into smaller pieces mid_periods = periods // 2 remaining = periods - mid_periods assert 0 < remaining < periods, (remaining, periods, endpoint, stride) midpoint = _generate_range_overflow_safe(endpoint, mid_periods, stride, side) return _generate_range_overflow_safe(midpoint, remaining, stride, side)
def _generate_range_overflow_safe(endpoint, periods, stride, side='start'): """ Calculate the second endpoint for passing to np.arange, checking to avoid an integer overflow. Catch OverflowError and re-raise as OutOfBoundsDatetime. Parameters ---------- endpoint : int nanosecond timestamp of the known endpoint of the desired range periods : int number of periods in the desired range stride : int nanoseconds between periods in the desired range side : {'start', 'end'} which end of the range `endpoint` refers to Returns ------- other_end : int Raises ------ OutOfBoundsDatetime """ # GH#14187 raise instead of incorrectly wrapping around assert side in ['start', 'end'] i64max = np.uint64(np.iinfo(np.int64).max) msg = ('Cannot generate range with {side}={endpoint} and ' 'periods={periods}' .format(side=side, endpoint=endpoint, periods=periods)) with np.errstate(over="raise"): # if periods * strides cannot be multiplied within the *uint64* bounds, # we cannot salvage the operation by recursing, so raise try: addend = np.uint64(periods) * np.uint64(np.abs(stride)) except FloatingPointError: raise OutOfBoundsDatetime(msg) if np.abs(addend) <= i64max: # relatively easy case without casting concerns return _generate_range_overflow_safe_signed( endpoint, periods, stride, side) elif ((endpoint > 0 and side == 'start' and stride > 0) or (endpoint < 0 and side == 'end' and stride > 0)): # no chance of not-overflowing raise OutOfBoundsDatetime(msg) elif (side == 'end' and endpoint > i64max and endpoint - stride <= i64max): # in _generate_regular_range we added `stride` thereby overflowing # the bounds. Adjust to fix this. return _generate_range_overflow_safe(endpoint - stride, periods - 1, stride, side) # split into smaller pieces mid_periods = periods // 2 remaining = periods - mid_periods assert 0 < remaining < periods, (remaining, periods, endpoint, stride) midpoint = _generate_range_overflow_safe(endpoint, mid_periods, stride, side) return _generate_range_overflow_safe(midpoint, remaining, stride, side)
[ "Calculate", "the", "second", "endpoint", "for", "passing", "to", "np", ".", "arange", "checking", "to", "avoid", "an", "integer", "overflow", ".", "Catch", "OverflowError", "and", "re", "-", "raise", "as", "OutOfBoundsDatetime", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/arrays/_ranges.py#L82-L146
[ "def", "_generate_range_overflow_safe", "(", "endpoint", ",", "periods", ",", "stride", ",", "side", "=", "'start'", ")", ":", "# GH#14187 raise instead of incorrectly wrapping around", "assert", "side", "in", "[", "'start'", ",", "'end'", "]", "i64max", "=", "np", ".", "uint64", "(", "np", ".", "iinfo", "(", "np", ".", "int64", ")", ".", "max", ")", "msg", "=", "(", "'Cannot generate range with {side}={endpoint} and '", "'periods={periods}'", ".", "format", "(", "side", "=", "side", ",", "endpoint", "=", "endpoint", ",", "periods", "=", "periods", ")", ")", "with", "np", ".", "errstate", "(", "over", "=", "\"raise\"", ")", ":", "# if periods * strides cannot be multiplied within the *uint64* bounds,", "# we cannot salvage the operation by recursing, so raise", "try", ":", "addend", "=", "np", ".", "uint64", "(", "periods", ")", "*", "np", ".", "uint64", "(", "np", ".", "abs", "(", "stride", ")", ")", "except", "FloatingPointError", ":", "raise", "OutOfBoundsDatetime", "(", "msg", ")", "if", "np", ".", "abs", "(", "addend", ")", "<=", "i64max", ":", "# relatively easy case without casting concerns", "return", "_generate_range_overflow_safe_signed", "(", "endpoint", ",", "periods", ",", "stride", ",", "side", ")", "elif", "(", "(", "endpoint", ">", "0", "and", "side", "==", "'start'", "and", "stride", ">", "0", ")", "or", "(", "endpoint", "<", "0", "and", "side", "==", "'end'", "and", "stride", ">", "0", ")", ")", ":", "# no chance of not-overflowing", "raise", "OutOfBoundsDatetime", "(", "msg", ")", "elif", "(", "side", "==", "'end'", "and", "endpoint", ">", "i64max", "and", "endpoint", "-", "stride", "<=", "i64max", ")", ":", "# in _generate_regular_range we added `stride` thereby overflowing", "# the bounds. Adjust to fix this.", "return", "_generate_range_overflow_safe", "(", "endpoint", "-", "stride", ",", "periods", "-", "1", ",", "stride", ",", "side", ")", "# split into smaller pieces", "mid_periods", "=", "periods", "//", "2", "remaining", "=", "periods", "-", "mid_periods", "assert", "0", "<", "remaining", "<", "periods", ",", "(", "remaining", ",", "periods", ",", "endpoint", ",", "stride", ")", "midpoint", "=", "_generate_range_overflow_safe", "(", "endpoint", ",", "mid_periods", ",", "stride", ",", "side", ")", "return", "_generate_range_overflow_safe", "(", "midpoint", ",", "remaining", ",", "stride", ",", "side", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_generate_range_overflow_safe_signed
A special case for _generate_range_overflow_safe where `periods * stride` can be calculated without overflowing int64 bounds.
pandas/core/arrays/_ranges.py
def _generate_range_overflow_safe_signed(endpoint, periods, stride, side): """ A special case for _generate_range_overflow_safe where `periods * stride` can be calculated without overflowing int64 bounds. """ assert side in ['start', 'end'] if side == 'end': stride *= -1 with np.errstate(over="raise"): addend = np.int64(periods) * np.int64(stride) try: # easy case with no overflows return np.int64(endpoint) + addend except (FloatingPointError, OverflowError): # with endpoint negative and addend positive we risk # FloatingPointError; with reversed signed we risk OverflowError pass # if stride and endpoint had opposite signs, then endpoint + addend # should never overflow. so they must have the same signs assert (stride > 0 and endpoint >= 0) or (stride < 0 and endpoint <= 0) if stride > 0: # watch out for very special case in which we just slightly # exceed implementation bounds, but when passing the result to # np.arange will get a result slightly within the bounds assert endpoint >= 0 result = np.uint64(endpoint) + np.uint64(addend) i64max = np.uint64(np.iinfo(np.int64).max) assert result > i64max if result <= i64max + np.uint64(stride): return result raise OutOfBoundsDatetime('Cannot generate range with ' '{side}={endpoint} and ' 'periods={periods}' .format(side=side, endpoint=endpoint, periods=periods))
def _generate_range_overflow_safe_signed(endpoint, periods, stride, side): """ A special case for _generate_range_overflow_safe where `periods * stride` can be calculated without overflowing int64 bounds. """ assert side in ['start', 'end'] if side == 'end': stride *= -1 with np.errstate(over="raise"): addend = np.int64(periods) * np.int64(stride) try: # easy case with no overflows return np.int64(endpoint) + addend except (FloatingPointError, OverflowError): # with endpoint negative and addend positive we risk # FloatingPointError; with reversed signed we risk OverflowError pass # if stride and endpoint had opposite signs, then endpoint + addend # should never overflow. so they must have the same signs assert (stride > 0 and endpoint >= 0) or (stride < 0 and endpoint <= 0) if stride > 0: # watch out for very special case in which we just slightly # exceed implementation bounds, but when passing the result to # np.arange will get a result slightly within the bounds assert endpoint >= 0 result = np.uint64(endpoint) + np.uint64(addend) i64max = np.uint64(np.iinfo(np.int64).max) assert result > i64max if result <= i64max + np.uint64(stride): return result raise OutOfBoundsDatetime('Cannot generate range with ' '{side}={endpoint} and ' 'periods={periods}' .format(side=side, endpoint=endpoint, periods=periods))
[ "A", "special", "case", "for", "_generate_range_overflow_safe", "where", "periods", "*", "stride", "can", "be", "calculated", "without", "overflowing", "int64", "bounds", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/arrays/_ranges.py#L149-L187
[ "def", "_generate_range_overflow_safe_signed", "(", "endpoint", ",", "periods", ",", "stride", ",", "side", ")", ":", "assert", "side", "in", "[", "'start'", ",", "'end'", "]", "if", "side", "==", "'end'", ":", "stride", "*=", "-", "1", "with", "np", ".", "errstate", "(", "over", "=", "\"raise\"", ")", ":", "addend", "=", "np", ".", "int64", "(", "periods", ")", "*", "np", ".", "int64", "(", "stride", ")", "try", ":", "# easy case with no overflows", "return", "np", ".", "int64", "(", "endpoint", ")", "+", "addend", "except", "(", "FloatingPointError", ",", "OverflowError", ")", ":", "# with endpoint negative and addend positive we risk", "# FloatingPointError; with reversed signed we risk OverflowError", "pass", "# if stride and endpoint had opposite signs, then endpoint + addend", "# should never overflow. so they must have the same signs", "assert", "(", "stride", ">", "0", "and", "endpoint", ">=", "0", ")", "or", "(", "stride", "<", "0", "and", "endpoint", "<=", "0", ")", "if", "stride", ">", "0", ":", "# watch out for very special case in which we just slightly", "# exceed implementation bounds, but when passing the result to", "# np.arange will get a result slightly within the bounds", "assert", "endpoint", ">=", "0", "result", "=", "np", ".", "uint64", "(", "endpoint", ")", "+", "np", ".", "uint64", "(", "addend", ")", "i64max", "=", "np", ".", "uint64", "(", "np", ".", "iinfo", "(", "np", ".", "int64", ")", ".", "max", ")", "assert", "result", ">", "i64max", "if", "result", "<=", "i64max", "+", "np", ".", "uint64", "(", "stride", ")", ":", "return", "result", "raise", "OutOfBoundsDatetime", "(", "'Cannot generate range with '", "'{side}={endpoint} and '", "'periods={periods}'", ".", "format", "(", "side", "=", "side", ",", "endpoint", "=", "endpoint", ",", "periods", "=", "periods", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
set_locale
Context manager for temporarily setting a locale. Parameters ---------- new_locale : str or tuple A string of the form <language_country>.<encoding>. For example to set the current locale to US English with a UTF8 encoding, you would pass "en_US.UTF-8". lc_var : int, default `locale.LC_ALL` The category of the locale being set. Notes ----- This is useful when you want to run a particular block of code under a particular locale, without globally setting the locale. This probably isn't thread-safe.
pandas/_config/localization.py
def set_locale(new_locale, lc_var=locale.LC_ALL): """ Context manager for temporarily setting a locale. Parameters ---------- new_locale : str or tuple A string of the form <language_country>.<encoding>. For example to set the current locale to US English with a UTF8 encoding, you would pass "en_US.UTF-8". lc_var : int, default `locale.LC_ALL` The category of the locale being set. Notes ----- This is useful when you want to run a particular block of code under a particular locale, without globally setting the locale. This probably isn't thread-safe. """ current_locale = locale.getlocale() try: locale.setlocale(lc_var, new_locale) normalized_locale = locale.getlocale() if all(x is not None for x in normalized_locale): yield '.'.join(normalized_locale) else: yield new_locale finally: locale.setlocale(lc_var, current_locale)
def set_locale(new_locale, lc_var=locale.LC_ALL): """ Context manager for temporarily setting a locale. Parameters ---------- new_locale : str or tuple A string of the form <language_country>.<encoding>. For example to set the current locale to US English with a UTF8 encoding, you would pass "en_US.UTF-8". lc_var : int, default `locale.LC_ALL` The category of the locale being set. Notes ----- This is useful when you want to run a particular block of code under a particular locale, without globally setting the locale. This probably isn't thread-safe. """ current_locale = locale.getlocale() try: locale.setlocale(lc_var, new_locale) normalized_locale = locale.getlocale() if all(x is not None for x in normalized_locale): yield '.'.join(normalized_locale) else: yield new_locale finally: locale.setlocale(lc_var, current_locale)
[ "Context", "manager", "for", "temporarily", "setting", "a", "locale", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/_config/localization.py#L15-L44
[ "def", "set_locale", "(", "new_locale", ",", "lc_var", "=", "locale", ".", "LC_ALL", ")", ":", "current_locale", "=", "locale", ".", "getlocale", "(", ")", "try", ":", "locale", ".", "setlocale", "(", "lc_var", ",", "new_locale", ")", "normalized_locale", "=", "locale", ".", "getlocale", "(", ")", "if", "all", "(", "x", "is", "not", "None", "for", "x", "in", "normalized_locale", ")", ":", "yield", "'.'", ".", "join", "(", "normalized_locale", ")", "else", ":", "yield", "new_locale", "finally", ":", "locale", ".", "setlocale", "(", "lc_var", ",", "current_locale", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
can_set_locale
Check to see if we can set a locale, and subsequently get the locale, without raising an Exception. Parameters ---------- lc : str The locale to attempt to set. lc_var : int, default `locale.LC_ALL` The category of the locale being set. Returns ------- is_valid : bool Whether the passed locale can be set
pandas/_config/localization.py
def can_set_locale(lc, lc_var=locale.LC_ALL): """ Check to see if we can set a locale, and subsequently get the locale, without raising an Exception. Parameters ---------- lc : str The locale to attempt to set. lc_var : int, default `locale.LC_ALL` The category of the locale being set. Returns ------- is_valid : bool Whether the passed locale can be set """ try: with set_locale(lc, lc_var=lc_var): pass except (ValueError, locale.Error): # horrible name for a Exception subclass return False else: return True
def can_set_locale(lc, lc_var=locale.LC_ALL): """ Check to see if we can set a locale, and subsequently get the locale, without raising an Exception. Parameters ---------- lc : str The locale to attempt to set. lc_var : int, default `locale.LC_ALL` The category of the locale being set. Returns ------- is_valid : bool Whether the passed locale can be set """ try: with set_locale(lc, lc_var=lc_var): pass except (ValueError, locale.Error): # horrible name for a Exception subclass return False else: return True
[ "Check", "to", "see", "if", "we", "can", "set", "a", "locale", "and", "subsequently", "get", "the", "locale", "without", "raising", "an", "Exception", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/_config/localization.py#L47-L72
[ "def", "can_set_locale", "(", "lc", ",", "lc_var", "=", "locale", ".", "LC_ALL", ")", ":", "try", ":", "with", "set_locale", "(", "lc", ",", "lc_var", "=", "lc_var", ")", ":", "pass", "except", "(", "ValueError", ",", "locale", ".", "Error", ")", ":", "# horrible name for a Exception subclass", "return", "False", "else", ":", "return", "True" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
_valid_locales
Return a list of normalized locales that do not throw an ``Exception`` when set. Parameters ---------- locales : str A string where each locale is separated by a newline. normalize : bool Whether to call ``locale.normalize`` on each locale. Returns ------- valid_locales : list A list of valid locales.
pandas/_config/localization.py
def _valid_locales(locales, normalize): """ Return a list of normalized locales that do not throw an ``Exception`` when set. Parameters ---------- locales : str A string where each locale is separated by a newline. normalize : bool Whether to call ``locale.normalize`` on each locale. Returns ------- valid_locales : list A list of valid locales. """ if normalize: normalizer = lambda x: locale.normalize(x.strip()) else: normalizer = lambda x: x.strip() return list(filter(can_set_locale, map(normalizer, locales)))
def _valid_locales(locales, normalize): """ Return a list of normalized locales that do not throw an ``Exception`` when set. Parameters ---------- locales : str A string where each locale is separated by a newline. normalize : bool Whether to call ``locale.normalize`` on each locale. Returns ------- valid_locales : list A list of valid locales. """ if normalize: normalizer = lambda x: locale.normalize(x.strip()) else: normalizer = lambda x: x.strip() return list(filter(can_set_locale, map(normalizer, locales)))
[ "Return", "a", "list", "of", "normalized", "locales", "that", "do", "not", "throw", "an", "Exception", "when", "set", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/_config/localization.py#L75-L97
[ "def", "_valid_locales", "(", "locales", ",", "normalize", ")", ":", "if", "normalize", ":", "normalizer", "=", "lambda", "x", ":", "locale", ".", "normalize", "(", "x", ".", "strip", "(", ")", ")", "else", ":", "normalizer", "=", "lambda", "x", ":", "x", ".", "strip", "(", ")", "return", "list", "(", "filter", "(", "can_set_locale", ",", "map", "(", "normalizer", ",", "locales", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
get_locales
Get all the locales that are available on the system. Parameters ---------- prefix : str If not ``None`` then return only those locales with the prefix provided. For example to get all English language locales (those that start with ``"en"``), pass ``prefix="en"``. normalize : bool Call ``locale.normalize`` on the resulting list of available locales. If ``True``, only locales that can be set without throwing an ``Exception`` are returned. locale_getter : callable The function to use to retrieve the current locales. This should return a string with each locale separated by a newline character. Returns ------- locales : list of strings A list of locale strings that can be set with ``locale.setlocale()``. For example:: locale.setlocale(locale.LC_ALL, locale_string) On error will return None (no locale available, e.g. Windows)
pandas/_config/localization.py
def get_locales(prefix=None, normalize=True, locale_getter=_default_locale_getter): """ Get all the locales that are available on the system. Parameters ---------- prefix : str If not ``None`` then return only those locales with the prefix provided. For example to get all English language locales (those that start with ``"en"``), pass ``prefix="en"``. normalize : bool Call ``locale.normalize`` on the resulting list of available locales. If ``True``, only locales that can be set without throwing an ``Exception`` are returned. locale_getter : callable The function to use to retrieve the current locales. This should return a string with each locale separated by a newline character. Returns ------- locales : list of strings A list of locale strings that can be set with ``locale.setlocale()``. For example:: locale.setlocale(locale.LC_ALL, locale_string) On error will return None (no locale available, e.g. Windows) """ try: raw_locales = locale_getter() except Exception: return None try: # raw_locales is "\n" separated list of locales # it may contain non-decodable parts, so split # extract what we can and then rejoin. raw_locales = raw_locales.split(b'\n') out_locales = [] for x in raw_locales: out_locales.append(str( x, encoding=options.display.encoding)) except TypeError: pass if prefix is None: return _valid_locales(out_locales, normalize) pattern = re.compile('{prefix}.*'.format(prefix=prefix)) found = pattern.findall('\n'.join(out_locales)) return _valid_locales(found, normalize)
def get_locales(prefix=None, normalize=True, locale_getter=_default_locale_getter): """ Get all the locales that are available on the system. Parameters ---------- prefix : str If not ``None`` then return only those locales with the prefix provided. For example to get all English language locales (those that start with ``"en"``), pass ``prefix="en"``. normalize : bool Call ``locale.normalize`` on the resulting list of available locales. If ``True``, only locales that can be set without throwing an ``Exception`` are returned. locale_getter : callable The function to use to retrieve the current locales. This should return a string with each locale separated by a newline character. Returns ------- locales : list of strings A list of locale strings that can be set with ``locale.setlocale()``. For example:: locale.setlocale(locale.LC_ALL, locale_string) On error will return None (no locale available, e.g. Windows) """ try: raw_locales = locale_getter() except Exception: return None try: # raw_locales is "\n" separated list of locales # it may contain non-decodable parts, so split # extract what we can and then rejoin. raw_locales = raw_locales.split(b'\n') out_locales = [] for x in raw_locales: out_locales.append(str( x, encoding=options.display.encoding)) except TypeError: pass if prefix is None: return _valid_locales(out_locales, normalize) pattern = re.compile('{prefix}.*'.format(prefix=prefix)) found = pattern.findall('\n'.join(out_locales)) return _valid_locales(found, normalize)
[ "Get", "all", "the", "locales", "that", "are", "available", "on", "the", "system", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/_config/localization.py#L109-L162
[ "def", "get_locales", "(", "prefix", "=", "None", ",", "normalize", "=", "True", ",", "locale_getter", "=", "_default_locale_getter", ")", ":", "try", ":", "raw_locales", "=", "locale_getter", "(", ")", "except", "Exception", ":", "return", "None", "try", ":", "# raw_locales is \"\\n\" separated list of locales", "# it may contain non-decodable parts, so split", "# extract what we can and then rejoin.", "raw_locales", "=", "raw_locales", ".", "split", "(", "b'\\n'", ")", "out_locales", "=", "[", "]", "for", "x", "in", "raw_locales", ":", "out_locales", ".", "append", "(", "str", "(", "x", ",", "encoding", "=", "options", ".", "display", ".", "encoding", ")", ")", "except", "TypeError", ":", "pass", "if", "prefix", "is", "None", ":", "return", "_valid_locales", "(", "out_locales", ",", "normalize", ")", "pattern", "=", "re", ".", "compile", "(", "'{prefix}.*'", ".", "format", "(", "prefix", "=", "prefix", ")", ")", "found", "=", "pattern", ".", "findall", "(", "'\\n'", ".", "join", "(", "out_locales", ")", ")", "return", "_valid_locales", "(", "found", ",", "normalize", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ensure_float
Ensure that an array object has a float dtype if possible. Parameters ---------- arr : array-like The array whose data type we want to enforce as float. Returns ------- float_arr : The original array cast to the float dtype if possible. Otherwise, the original array is returned.
pandas/core/dtypes/common.py
def ensure_float(arr): """ Ensure that an array object has a float dtype if possible. Parameters ---------- arr : array-like The array whose data type we want to enforce as float. Returns ------- float_arr : The original array cast to the float dtype if possible. Otherwise, the original array is returned. """ if issubclass(arr.dtype.type, (np.integer, np.bool_)): arr = arr.astype(float) return arr
def ensure_float(arr): """ Ensure that an array object has a float dtype if possible. Parameters ---------- arr : array-like The array whose data type we want to enforce as float. Returns ------- float_arr : The original array cast to the float dtype if possible. Otherwise, the original array is returned. """ if issubclass(arr.dtype.type, (np.integer, np.bool_)): arr = arr.astype(float) return arr
[ "Ensure", "that", "an", "array", "object", "has", "a", "float", "dtype", "if", "possible", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L40-L57
[ "def", "ensure_float", "(", "arr", ")", ":", "if", "issubclass", "(", "arr", ".", "dtype", ".", "type", ",", "(", "np", ".", "integer", ",", "np", ".", "bool_", ")", ")", ":", "arr", "=", "arr", ".", "astype", "(", "float", ")", "return", "arr" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ensure_categorical
Ensure that an array-like object is a Categorical (if not already). Parameters ---------- arr : array-like The array that we want to convert into a Categorical. Returns ------- cat_arr : The original array cast as a Categorical. If it already is a Categorical, we return as is.
pandas/core/dtypes/common.py
def ensure_categorical(arr): """ Ensure that an array-like object is a Categorical (if not already). Parameters ---------- arr : array-like The array that we want to convert into a Categorical. Returns ------- cat_arr : The original array cast as a Categorical. If it already is a Categorical, we return as is. """ if not is_categorical(arr): from pandas import Categorical arr = Categorical(arr) return arr
def ensure_categorical(arr): """ Ensure that an array-like object is a Categorical (if not already). Parameters ---------- arr : array-like The array that we want to convert into a Categorical. Returns ------- cat_arr : The original array cast as a Categorical. If it already is a Categorical, we return as is. """ if not is_categorical(arr): from pandas import Categorical arr = Categorical(arr) return arr
[ "Ensure", "that", "an", "array", "-", "like", "object", "is", "a", "Categorical", "(", "if", "not", "already", ")", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L69-L87
[ "def", "ensure_categorical", "(", "arr", ")", ":", "if", "not", "is_categorical", "(", "arr", ")", ":", "from", "pandas", "import", "Categorical", "arr", "=", "Categorical", "(", "arr", ")", "return", "arr" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
ensure_int64_or_float64
Ensure that an dtype array of some integer dtype has an int64 dtype if possible If it's not possible, potentially because of overflow, convert the array to float64 instead. Parameters ---------- arr : array-like The array whose data type we want to enforce. copy: boolean Whether to copy the original array or reuse it in place, if possible. Returns ------- out_arr : The input array cast as int64 if possible without overflow. Otherwise the input array cast to float64.
pandas/core/dtypes/common.py
def ensure_int64_or_float64(arr, copy=False): """ Ensure that an dtype array of some integer dtype has an int64 dtype if possible If it's not possible, potentially because of overflow, convert the array to float64 instead. Parameters ---------- arr : array-like The array whose data type we want to enforce. copy: boolean Whether to copy the original array or reuse it in place, if possible. Returns ------- out_arr : The input array cast as int64 if possible without overflow. Otherwise the input array cast to float64. """ try: return arr.astype('int64', copy=copy, casting='safe') except TypeError: return arr.astype('float64', copy=copy)
def ensure_int64_or_float64(arr, copy=False): """ Ensure that an dtype array of some integer dtype has an int64 dtype if possible If it's not possible, potentially because of overflow, convert the array to float64 instead. Parameters ---------- arr : array-like The array whose data type we want to enforce. copy: boolean Whether to copy the original array or reuse it in place, if possible. Returns ------- out_arr : The input array cast as int64 if possible without overflow. Otherwise the input array cast to float64. """ try: return arr.astype('int64', copy=copy, casting='safe') except TypeError: return arr.astype('float64', copy=copy)
[ "Ensure", "that", "an", "dtype", "array", "of", "some", "integer", "dtype", "has", "an", "int64", "dtype", "if", "possible", "If", "it", "s", "not", "possible", "potentially", "because", "of", "overflow", "convert", "the", "array", "to", "float64", "instead", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L90-L114
[ "def", "ensure_int64_or_float64", "(", "arr", ",", "copy", "=", "False", ")", ":", "try", ":", "return", "arr", ".", "astype", "(", "'int64'", ",", "copy", "=", "copy", ",", "casting", "=", "'safe'", ")", "except", "TypeError", ":", "return", "arr", ".", "astype", "(", "'float64'", ",", "copy", "=", "copy", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
classes_and_not_datetimelike
evaluate if the tipo is a subclass of the klasses and not a datetimelike
pandas/core/dtypes/common.py
def classes_and_not_datetimelike(*klasses): """ evaluate if the tipo is a subclass of the klasses and not a datetimelike """ return lambda tipo: (issubclass(tipo, klasses) and not issubclass(tipo, (np.datetime64, np.timedelta64)))
def classes_and_not_datetimelike(*klasses): """ evaluate if the tipo is a subclass of the klasses and not a datetimelike """ return lambda tipo: (issubclass(tipo, klasses) and not issubclass(tipo, (np.datetime64, np.timedelta64)))
[ "evaluate", "if", "the", "tipo", "is", "a", "subclass", "of", "the", "klasses", "and", "not", "a", "datetimelike" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L122-L128
[ "def", "classes_and_not_datetimelike", "(", "*", "klasses", ")", ":", "return", "lambda", "tipo", ":", "(", "issubclass", "(", "tipo", ",", "klasses", ")", "and", "not", "issubclass", "(", "tipo", ",", "(", "np", ".", "datetime64", ",", "np", ".", "timedelta64", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_sparse
Check whether an array-like is a 1-D pandas sparse array. Check that the one-dimensional array-like is a pandas sparse array. Returns True if it is a pandas sparse array, not another type of sparse array. Parameters ---------- arr : array-like Array-like to check. Returns ------- bool Whether or not the array-like is a pandas sparse array. See Also -------- DataFrame.to_sparse : Convert DataFrame to a SparseDataFrame. Series.to_sparse : Convert Series to SparseSeries. Series.to_dense : Return dense representation of a Series. Examples -------- Returns `True` if the parameter is a 1-D pandas sparse array. >>> is_sparse(pd.SparseArray([0, 0, 1, 0])) True >>> is_sparse(pd.SparseSeries([0, 0, 1, 0])) True Returns `False` if the parameter is not sparse. >>> is_sparse(np.array([0, 0, 1, 0])) False >>> is_sparse(pd.Series([0, 1, 0, 0])) False Returns `False` if the parameter is not a pandas sparse array. >>> from scipy.sparse import bsr_matrix >>> is_sparse(bsr_matrix([0, 1, 0, 0])) False Returns `False` if the parameter has more than one dimension. >>> df = pd.SparseDataFrame([389., 24., 80.5, np.nan], columns=['max_speed'], index=['falcon', 'parrot', 'lion', 'monkey']) >>> is_sparse(df) False >>> is_sparse(df.max_speed) True
pandas/core/dtypes/common.py
def is_sparse(arr): """ Check whether an array-like is a 1-D pandas sparse array. Check that the one-dimensional array-like is a pandas sparse array. Returns True if it is a pandas sparse array, not another type of sparse array. Parameters ---------- arr : array-like Array-like to check. Returns ------- bool Whether or not the array-like is a pandas sparse array. See Also -------- DataFrame.to_sparse : Convert DataFrame to a SparseDataFrame. Series.to_sparse : Convert Series to SparseSeries. Series.to_dense : Return dense representation of a Series. Examples -------- Returns `True` if the parameter is a 1-D pandas sparse array. >>> is_sparse(pd.SparseArray([0, 0, 1, 0])) True >>> is_sparse(pd.SparseSeries([0, 0, 1, 0])) True Returns `False` if the parameter is not sparse. >>> is_sparse(np.array([0, 0, 1, 0])) False >>> is_sparse(pd.Series([0, 1, 0, 0])) False Returns `False` if the parameter is not a pandas sparse array. >>> from scipy.sparse import bsr_matrix >>> is_sparse(bsr_matrix([0, 1, 0, 0])) False Returns `False` if the parameter has more than one dimension. >>> df = pd.SparseDataFrame([389., 24., 80.5, np.nan], columns=['max_speed'], index=['falcon', 'parrot', 'lion', 'monkey']) >>> is_sparse(df) False >>> is_sparse(df.max_speed) True """ from pandas.core.arrays.sparse import SparseDtype dtype = getattr(arr, 'dtype', arr) return isinstance(dtype, SparseDtype)
def is_sparse(arr): """ Check whether an array-like is a 1-D pandas sparse array. Check that the one-dimensional array-like is a pandas sparse array. Returns True if it is a pandas sparse array, not another type of sparse array. Parameters ---------- arr : array-like Array-like to check. Returns ------- bool Whether or not the array-like is a pandas sparse array. See Also -------- DataFrame.to_sparse : Convert DataFrame to a SparseDataFrame. Series.to_sparse : Convert Series to SparseSeries. Series.to_dense : Return dense representation of a Series. Examples -------- Returns `True` if the parameter is a 1-D pandas sparse array. >>> is_sparse(pd.SparseArray([0, 0, 1, 0])) True >>> is_sparse(pd.SparseSeries([0, 0, 1, 0])) True Returns `False` if the parameter is not sparse. >>> is_sparse(np.array([0, 0, 1, 0])) False >>> is_sparse(pd.Series([0, 1, 0, 0])) False Returns `False` if the parameter is not a pandas sparse array. >>> from scipy.sparse import bsr_matrix >>> is_sparse(bsr_matrix([0, 1, 0, 0])) False Returns `False` if the parameter has more than one dimension. >>> df = pd.SparseDataFrame([389., 24., 80.5, np.nan], columns=['max_speed'], index=['falcon', 'parrot', 'lion', 'monkey']) >>> is_sparse(df) False >>> is_sparse(df.max_speed) True """ from pandas.core.arrays.sparse import SparseDtype dtype = getattr(arr, 'dtype', arr) return isinstance(dtype, SparseDtype)
[ "Check", "whether", "an", "array", "-", "like", "is", "a", "1", "-", "D", "pandas", "sparse", "array", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L161-L220
[ "def", "is_sparse", "(", "arr", ")", ":", "from", "pandas", ".", "core", ".", "arrays", ".", "sparse", "import", "SparseDtype", "dtype", "=", "getattr", "(", "arr", ",", "'dtype'", ",", "arr", ")", "return", "isinstance", "(", "dtype", ",", "SparseDtype", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_scipy_sparse
Check whether an array-like is a scipy.sparse.spmatrix instance. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a scipy.sparse.spmatrix instance. Notes ----- If scipy is not installed, this function will always return False. Examples -------- >>> from scipy.sparse import bsr_matrix >>> is_scipy_sparse(bsr_matrix([1, 2, 3])) True >>> is_scipy_sparse(pd.SparseArray([1, 2, 3])) False >>> is_scipy_sparse(pd.SparseSeries([1, 2, 3])) False
pandas/core/dtypes/common.py
def is_scipy_sparse(arr): """ Check whether an array-like is a scipy.sparse.spmatrix instance. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a scipy.sparse.spmatrix instance. Notes ----- If scipy is not installed, this function will always return False. Examples -------- >>> from scipy.sparse import bsr_matrix >>> is_scipy_sparse(bsr_matrix([1, 2, 3])) True >>> is_scipy_sparse(pd.SparseArray([1, 2, 3])) False >>> is_scipy_sparse(pd.SparseSeries([1, 2, 3])) False """ global _is_scipy_sparse if _is_scipy_sparse is None: try: from scipy.sparse import issparse as _is_scipy_sparse except ImportError: _is_scipy_sparse = lambda _: False return _is_scipy_sparse(arr)
def is_scipy_sparse(arr): """ Check whether an array-like is a scipy.sparse.spmatrix instance. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a scipy.sparse.spmatrix instance. Notes ----- If scipy is not installed, this function will always return False. Examples -------- >>> from scipy.sparse import bsr_matrix >>> is_scipy_sparse(bsr_matrix([1, 2, 3])) True >>> is_scipy_sparse(pd.SparseArray([1, 2, 3])) False >>> is_scipy_sparse(pd.SparseSeries([1, 2, 3])) False """ global _is_scipy_sparse if _is_scipy_sparse is None: try: from scipy.sparse import issparse as _is_scipy_sparse except ImportError: _is_scipy_sparse = lambda _: False return _is_scipy_sparse(arr)
[ "Check", "whether", "an", "array", "-", "like", "is", "a", "scipy", ".", "sparse", ".", "spmatrix", "instance", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L223-L260
[ "def", "is_scipy_sparse", "(", "arr", ")", ":", "global", "_is_scipy_sparse", "if", "_is_scipy_sparse", "is", "None", ":", "try", ":", "from", "scipy", ".", "sparse", "import", "issparse", "as", "_is_scipy_sparse", "except", "ImportError", ":", "_is_scipy_sparse", "=", "lambda", "_", ":", "False", "return", "_is_scipy_sparse", "(", "arr", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_offsetlike
Check if obj or all elements of list-like is DateOffset Parameters ---------- arr_or_obj : object Returns ------- boolean Whether the object is a DateOffset or listlike of DatetOffsets Examples -------- >>> is_offsetlike(pd.DateOffset(days=1)) True >>> is_offsetlike('offset') False >>> is_offsetlike([pd.offsets.Minute(4), pd.offsets.MonthEnd()]) True >>> is_offsetlike(np.array([pd.DateOffset(months=3), pd.Timestamp.now()])) False
pandas/core/dtypes/common.py
def is_offsetlike(arr_or_obj): """ Check if obj or all elements of list-like is DateOffset Parameters ---------- arr_or_obj : object Returns ------- boolean Whether the object is a DateOffset or listlike of DatetOffsets Examples -------- >>> is_offsetlike(pd.DateOffset(days=1)) True >>> is_offsetlike('offset') False >>> is_offsetlike([pd.offsets.Minute(4), pd.offsets.MonthEnd()]) True >>> is_offsetlike(np.array([pd.DateOffset(months=3), pd.Timestamp.now()])) False """ if isinstance(arr_or_obj, ABCDateOffset): return True elif (is_list_like(arr_or_obj) and len(arr_or_obj) and is_object_dtype(arr_or_obj)): return all(isinstance(x, ABCDateOffset) for x in arr_or_obj) return False
def is_offsetlike(arr_or_obj): """ Check if obj or all elements of list-like is DateOffset Parameters ---------- arr_or_obj : object Returns ------- boolean Whether the object is a DateOffset or listlike of DatetOffsets Examples -------- >>> is_offsetlike(pd.DateOffset(days=1)) True >>> is_offsetlike('offset') False >>> is_offsetlike([pd.offsets.Minute(4), pd.offsets.MonthEnd()]) True >>> is_offsetlike(np.array([pd.DateOffset(months=3), pd.Timestamp.now()])) False """ if isinstance(arr_or_obj, ABCDateOffset): return True elif (is_list_like(arr_or_obj) and len(arr_or_obj) and is_object_dtype(arr_or_obj)): return all(isinstance(x, ABCDateOffset) for x in arr_or_obj) return False
[ "Check", "if", "obj", "or", "all", "elements", "of", "list", "-", "like", "is", "DateOffset" ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L343-L372
[ "def", "is_offsetlike", "(", "arr_or_obj", ")", ":", "if", "isinstance", "(", "arr_or_obj", ",", "ABCDateOffset", ")", ":", "return", "True", "elif", "(", "is_list_like", "(", "arr_or_obj", ")", "and", "len", "(", "arr_or_obj", ")", "and", "is_object_dtype", "(", "arr_or_obj", ")", ")", ":", "return", "all", "(", "isinstance", "(", "x", ",", "ABCDateOffset", ")", "for", "x", "in", "arr_or_obj", ")", "return", "False" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_period
Check whether an array-like is a periodical index. .. deprecated:: 0.24.0 Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a periodical index. Examples -------- >>> is_period([1, 2, 3]) False >>> is_period(pd.Index([1, 2, 3])) False >>> is_period(pd.PeriodIndex(["2017-01-01"], freq="D")) True
pandas/core/dtypes/common.py
def is_period(arr): """ Check whether an array-like is a periodical index. .. deprecated:: 0.24.0 Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a periodical index. Examples -------- >>> is_period([1, 2, 3]) False >>> is_period(pd.Index([1, 2, 3])) False >>> is_period(pd.PeriodIndex(["2017-01-01"], freq="D")) True """ warnings.warn("'is_period' is deprecated and will be removed in a future " "version. Use 'is_period_dtype' or is_period_arraylike' " "instead.", FutureWarning, stacklevel=2) return isinstance(arr, ABCPeriodIndex) or is_period_arraylike(arr)
def is_period(arr): """ Check whether an array-like is a periodical index. .. deprecated:: 0.24.0 Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a periodical index. Examples -------- >>> is_period([1, 2, 3]) False >>> is_period(pd.Index([1, 2, 3])) False >>> is_period(pd.PeriodIndex(["2017-01-01"], freq="D")) True """ warnings.warn("'is_period' is deprecated and will be removed in a future " "version. Use 'is_period_dtype' or is_period_arraylike' " "instead.", FutureWarning, stacklevel=2) return isinstance(arr, ABCPeriodIndex) or is_period_arraylike(arr)
[ "Check", "whether", "an", "array", "-", "like", "is", "a", "periodical", "index", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L375-L405
[ "def", "is_period", "(", "arr", ")", ":", "warnings", ".", "warn", "(", "\"'is_period' is deprecated and will be removed in a future \"", "\"version. Use 'is_period_dtype' or is_period_arraylike' \"", "\"instead.\"", ",", "FutureWarning", ",", "stacklevel", "=", "2", ")", "return", "isinstance", "(", "arr", ",", "ABCPeriodIndex", ")", "or", "is_period_arraylike", "(", "arr", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_string_dtype
Check whether the provided array or dtype is of the string dtype. Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype is of the string dtype. Examples -------- >>> is_string_dtype(str) True >>> is_string_dtype(object) True >>> is_string_dtype(int) False >>> >>> is_string_dtype(np.array(['a', 'b'])) True >>> is_string_dtype(pd.Series([1, 2])) False
pandas/core/dtypes/common.py
def is_string_dtype(arr_or_dtype): """ Check whether the provided array or dtype is of the string dtype. Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype is of the string dtype. Examples -------- >>> is_string_dtype(str) True >>> is_string_dtype(object) True >>> is_string_dtype(int) False >>> >>> is_string_dtype(np.array(['a', 'b'])) True >>> is_string_dtype(pd.Series([1, 2])) False """ # TODO: gh-15585: consider making the checks stricter. def condition(dtype): return dtype.kind in ('O', 'S', 'U') and not is_period_dtype(dtype) return _is_dtype(arr_or_dtype, condition)
def is_string_dtype(arr_or_dtype): """ Check whether the provided array or dtype is of the string dtype. Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype is of the string dtype. Examples -------- >>> is_string_dtype(str) True >>> is_string_dtype(object) True >>> is_string_dtype(int) False >>> >>> is_string_dtype(np.array(['a', 'b'])) True >>> is_string_dtype(pd.Series([1, 2])) False """ # TODO: gh-15585: consider making the checks stricter. def condition(dtype): return dtype.kind in ('O', 'S', 'U') and not is_period_dtype(dtype) return _is_dtype(arr_or_dtype, condition)
[ "Check", "whether", "the", "provided", "array", "or", "dtype", "is", "of", "the", "string", "dtype", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L611-L643
[ "def", "is_string_dtype", "(", "arr_or_dtype", ")", ":", "# TODO: gh-15585: consider making the checks stricter.", "def", "condition", "(", "dtype", ")", ":", "return", "dtype", ".", "kind", "in", "(", "'O'", ",", "'S'", ",", "'U'", ")", "and", "not", "is_period_dtype", "(", "dtype", ")", "return", "_is_dtype", "(", "arr_or_dtype", ",", "condition", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_period_arraylike
Check whether an array-like is a periodical array-like or PeriodIndex. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a periodical array-like or PeriodIndex instance. Examples -------- >>> is_period_arraylike([1, 2, 3]) False >>> is_period_arraylike(pd.Index([1, 2, 3])) False >>> is_period_arraylike(pd.PeriodIndex(["2017-01-01"], freq="D")) True
pandas/core/dtypes/common.py
def is_period_arraylike(arr): """ Check whether an array-like is a periodical array-like or PeriodIndex. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a periodical array-like or PeriodIndex instance. Examples -------- >>> is_period_arraylike([1, 2, 3]) False >>> is_period_arraylike(pd.Index([1, 2, 3])) False >>> is_period_arraylike(pd.PeriodIndex(["2017-01-01"], freq="D")) True """ if isinstance(arr, (ABCPeriodIndex, ABCPeriodArray)): return True elif isinstance(arr, (np.ndarray, ABCSeries)): return is_period_dtype(arr.dtype) return getattr(arr, 'inferred_type', None) == 'period'
def is_period_arraylike(arr): """ Check whether an array-like is a periodical array-like or PeriodIndex. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a periodical array-like or PeriodIndex instance. Examples -------- >>> is_period_arraylike([1, 2, 3]) False >>> is_period_arraylike(pd.Index([1, 2, 3])) False >>> is_period_arraylike(pd.PeriodIndex(["2017-01-01"], freq="D")) True """ if isinstance(arr, (ABCPeriodIndex, ABCPeriodArray)): return True elif isinstance(arr, (np.ndarray, ABCSeries)): return is_period_dtype(arr.dtype) return getattr(arr, 'inferred_type', None) == 'period'
[ "Check", "whether", "an", "array", "-", "like", "is", "a", "periodical", "array", "-", "like", "or", "PeriodIndex", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L646-L675
[ "def", "is_period_arraylike", "(", "arr", ")", ":", "if", "isinstance", "(", "arr", ",", "(", "ABCPeriodIndex", ",", "ABCPeriodArray", ")", ")", ":", "return", "True", "elif", "isinstance", "(", "arr", ",", "(", "np", ".", "ndarray", ",", "ABCSeries", ")", ")", ":", "return", "is_period_dtype", "(", "arr", ".", "dtype", ")", "return", "getattr", "(", "arr", ",", "'inferred_type'", ",", "None", ")", "==", "'period'" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_datetime_arraylike
Check whether an array-like is a datetime array-like or DatetimeIndex. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a datetime array-like or DatetimeIndex. Examples -------- >>> is_datetime_arraylike([1, 2, 3]) False >>> is_datetime_arraylike(pd.Index([1, 2, 3])) False >>> is_datetime_arraylike(pd.DatetimeIndex([1, 2, 3])) True
pandas/core/dtypes/common.py
def is_datetime_arraylike(arr): """ Check whether an array-like is a datetime array-like or DatetimeIndex. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a datetime array-like or DatetimeIndex. Examples -------- >>> is_datetime_arraylike([1, 2, 3]) False >>> is_datetime_arraylike(pd.Index([1, 2, 3])) False >>> is_datetime_arraylike(pd.DatetimeIndex([1, 2, 3])) True """ if isinstance(arr, ABCDatetimeIndex): return True elif isinstance(arr, (np.ndarray, ABCSeries)): return (is_object_dtype(arr.dtype) and lib.infer_dtype(arr, skipna=False) == 'datetime') return getattr(arr, 'inferred_type', None) == 'datetime'
def is_datetime_arraylike(arr): """ Check whether an array-like is a datetime array-like or DatetimeIndex. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a datetime array-like or DatetimeIndex. Examples -------- >>> is_datetime_arraylike([1, 2, 3]) False >>> is_datetime_arraylike(pd.Index([1, 2, 3])) False >>> is_datetime_arraylike(pd.DatetimeIndex([1, 2, 3])) True """ if isinstance(arr, ABCDatetimeIndex): return True elif isinstance(arr, (np.ndarray, ABCSeries)): return (is_object_dtype(arr.dtype) and lib.infer_dtype(arr, skipna=False) == 'datetime') return getattr(arr, 'inferred_type', None) == 'datetime'
[ "Check", "whether", "an", "array", "-", "like", "is", "a", "datetime", "array", "-", "like", "or", "DatetimeIndex", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L678-L708
[ "def", "is_datetime_arraylike", "(", "arr", ")", ":", "if", "isinstance", "(", "arr", ",", "ABCDatetimeIndex", ")", ":", "return", "True", "elif", "isinstance", "(", "arr", ",", "(", "np", ".", "ndarray", ",", "ABCSeries", ")", ")", ":", "return", "(", "is_object_dtype", "(", "arr", ".", "dtype", ")", "and", "lib", ".", "infer_dtype", "(", "arr", ",", "skipna", "=", "False", ")", "==", "'datetime'", ")", "return", "getattr", "(", "arr", ",", "'inferred_type'", ",", "None", ")", "==", "'datetime'" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_datetimelike
Check whether an array-like is a datetime-like array-like. Acceptable datetime-like objects are (but not limited to) datetime indices, periodic indices, and timedelta indices. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a datetime-like array-like. Examples -------- >>> is_datetimelike([1, 2, 3]) False >>> is_datetimelike(pd.Index([1, 2, 3])) False >>> is_datetimelike(pd.DatetimeIndex([1, 2, 3])) True >>> is_datetimelike(pd.DatetimeIndex([1, 2, 3], tz="US/Eastern")) True >>> is_datetimelike(pd.PeriodIndex([], freq="A")) True >>> is_datetimelike(np.array([], dtype=np.datetime64)) True >>> is_datetimelike(pd.Series([], dtype="timedelta64[ns]")) True >>> >>> dtype = DatetimeTZDtype("ns", tz="US/Eastern") >>> s = pd.Series([], dtype=dtype) >>> is_datetimelike(s) True
pandas/core/dtypes/common.py
def is_datetimelike(arr): """ Check whether an array-like is a datetime-like array-like. Acceptable datetime-like objects are (but not limited to) datetime indices, periodic indices, and timedelta indices. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a datetime-like array-like. Examples -------- >>> is_datetimelike([1, 2, 3]) False >>> is_datetimelike(pd.Index([1, 2, 3])) False >>> is_datetimelike(pd.DatetimeIndex([1, 2, 3])) True >>> is_datetimelike(pd.DatetimeIndex([1, 2, 3], tz="US/Eastern")) True >>> is_datetimelike(pd.PeriodIndex([], freq="A")) True >>> is_datetimelike(np.array([], dtype=np.datetime64)) True >>> is_datetimelike(pd.Series([], dtype="timedelta64[ns]")) True >>> >>> dtype = DatetimeTZDtype("ns", tz="US/Eastern") >>> s = pd.Series([], dtype=dtype) >>> is_datetimelike(s) True """ return (is_datetime64_dtype(arr) or is_datetime64tz_dtype(arr) or is_timedelta64_dtype(arr) or isinstance(arr, ABCPeriodIndex))
def is_datetimelike(arr): """ Check whether an array-like is a datetime-like array-like. Acceptable datetime-like objects are (but not limited to) datetime indices, periodic indices, and timedelta indices. Parameters ---------- arr : array-like The array-like to check. Returns ------- boolean Whether or not the array-like is a datetime-like array-like. Examples -------- >>> is_datetimelike([1, 2, 3]) False >>> is_datetimelike(pd.Index([1, 2, 3])) False >>> is_datetimelike(pd.DatetimeIndex([1, 2, 3])) True >>> is_datetimelike(pd.DatetimeIndex([1, 2, 3], tz="US/Eastern")) True >>> is_datetimelike(pd.PeriodIndex([], freq="A")) True >>> is_datetimelike(np.array([], dtype=np.datetime64)) True >>> is_datetimelike(pd.Series([], dtype="timedelta64[ns]")) True >>> >>> dtype = DatetimeTZDtype("ns", tz="US/Eastern") >>> s = pd.Series([], dtype=dtype) >>> is_datetimelike(s) True """ return (is_datetime64_dtype(arr) or is_datetime64tz_dtype(arr) or is_timedelta64_dtype(arr) or isinstance(arr, ABCPeriodIndex))
[ "Check", "whether", "an", "array", "-", "like", "is", "a", "datetime", "-", "like", "array", "-", "like", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L711-L753
[ "def", "is_datetimelike", "(", "arr", ")", ":", "return", "(", "is_datetime64_dtype", "(", "arr", ")", "or", "is_datetime64tz_dtype", "(", "arr", ")", "or", "is_timedelta64_dtype", "(", "arr", ")", "or", "isinstance", "(", "arr", ",", "ABCPeriodIndex", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_dtype_equal
Check if two dtypes are equal. Parameters ---------- source : The first dtype to compare target : The second dtype to compare Returns ---------- boolean Whether or not the two dtypes are equal. Examples -------- >>> is_dtype_equal(int, float) False >>> is_dtype_equal("int", int) True >>> is_dtype_equal(object, "category") False >>> is_dtype_equal(CategoricalDtype(), "category") True >>> is_dtype_equal(DatetimeTZDtype(), "datetime64") False
pandas/core/dtypes/common.py
def is_dtype_equal(source, target): """ Check if two dtypes are equal. Parameters ---------- source : The first dtype to compare target : The second dtype to compare Returns ---------- boolean Whether or not the two dtypes are equal. Examples -------- >>> is_dtype_equal(int, float) False >>> is_dtype_equal("int", int) True >>> is_dtype_equal(object, "category") False >>> is_dtype_equal(CategoricalDtype(), "category") True >>> is_dtype_equal(DatetimeTZDtype(), "datetime64") False """ try: source = _get_dtype(source) target = _get_dtype(target) return source == target except (TypeError, AttributeError): # invalid comparison # object == category will hit this return False
def is_dtype_equal(source, target): """ Check if two dtypes are equal. Parameters ---------- source : The first dtype to compare target : The second dtype to compare Returns ---------- boolean Whether or not the two dtypes are equal. Examples -------- >>> is_dtype_equal(int, float) False >>> is_dtype_equal("int", int) True >>> is_dtype_equal(object, "category") False >>> is_dtype_equal(CategoricalDtype(), "category") True >>> is_dtype_equal(DatetimeTZDtype(), "datetime64") False """ try: source = _get_dtype(source) target = _get_dtype(target) return source == target except (TypeError, AttributeError): # invalid comparison # object == category will hit this return False
[ "Check", "if", "two", "dtypes", "are", "equal", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L756-L792
[ "def", "is_dtype_equal", "(", "source", ",", "target", ")", ":", "try", ":", "source", "=", "_get_dtype", "(", "source", ")", "target", "=", "_get_dtype", "(", "target", ")", "return", "source", "==", "target", "except", "(", "TypeError", ",", "AttributeError", ")", ":", "# invalid comparison", "# object == category will hit this", "return", "False" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_dtype_union_equal
Check whether two arrays have compatible dtypes to do a union. numpy types are checked with ``is_dtype_equal``. Extension types are checked separately. Parameters ---------- source : The first dtype to compare target : The second dtype to compare Returns ---------- boolean Whether or not the two dtypes are equal. >>> is_dtype_equal("int", int) True >>> is_dtype_equal(CategoricalDtype(['a', 'b'], ... CategoricalDtype(['b', 'c'])) True >>> is_dtype_equal(CategoricalDtype(['a', 'b'], ... CategoricalDtype(['b', 'c'], ordered=True)) False
pandas/core/dtypes/common.py
def is_dtype_union_equal(source, target): """ Check whether two arrays have compatible dtypes to do a union. numpy types are checked with ``is_dtype_equal``. Extension types are checked separately. Parameters ---------- source : The first dtype to compare target : The second dtype to compare Returns ---------- boolean Whether or not the two dtypes are equal. >>> is_dtype_equal("int", int) True >>> is_dtype_equal(CategoricalDtype(['a', 'b'], ... CategoricalDtype(['b', 'c'])) True >>> is_dtype_equal(CategoricalDtype(['a', 'b'], ... CategoricalDtype(['b', 'c'], ordered=True)) False """ source = _get_dtype(source) target = _get_dtype(target) if is_categorical_dtype(source) and is_categorical_dtype(target): # ordered False for both return source.ordered is target.ordered return is_dtype_equal(source, target)
def is_dtype_union_equal(source, target): """ Check whether two arrays have compatible dtypes to do a union. numpy types are checked with ``is_dtype_equal``. Extension types are checked separately. Parameters ---------- source : The first dtype to compare target : The second dtype to compare Returns ---------- boolean Whether or not the two dtypes are equal. >>> is_dtype_equal("int", int) True >>> is_dtype_equal(CategoricalDtype(['a', 'b'], ... CategoricalDtype(['b', 'c'])) True >>> is_dtype_equal(CategoricalDtype(['a', 'b'], ... CategoricalDtype(['b', 'c'], ordered=True)) False """ source = _get_dtype(source) target = _get_dtype(target) if is_categorical_dtype(source) and is_categorical_dtype(target): # ordered False for both return source.ordered is target.ordered return is_dtype_equal(source, target)
[ "Check", "whether", "two", "arrays", "have", "compatible", "dtypes", "to", "do", "a", "union", ".", "numpy", "types", "are", "checked", "with", "is_dtype_equal", ".", "Extension", "types", "are", "checked", "separately", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L795-L827
[ "def", "is_dtype_union_equal", "(", "source", ",", "target", ")", ":", "source", "=", "_get_dtype", "(", "source", ")", "target", "=", "_get_dtype", "(", "target", ")", "if", "is_categorical_dtype", "(", "source", ")", "and", "is_categorical_dtype", "(", "target", ")", ":", "# ordered False for both", "return", "source", ".", "ordered", "is", "target", ".", "ordered", "return", "is_dtype_equal", "(", "source", ",", "target", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_datetime64_ns_dtype
Check whether the provided array or dtype is of the datetime64[ns] dtype. Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype is of the datetime64[ns] dtype. Examples -------- >>> is_datetime64_ns_dtype(str) False >>> is_datetime64_ns_dtype(int) False >>> is_datetime64_ns_dtype(np.datetime64) # no unit False >>> is_datetime64_ns_dtype(DatetimeTZDtype("ns", "US/Eastern")) True >>> is_datetime64_ns_dtype(np.array(['a', 'b'])) False >>> is_datetime64_ns_dtype(np.array([1, 2])) False >>> is_datetime64_ns_dtype(np.array([], dtype=np.datetime64)) # no unit False >>> is_datetime64_ns_dtype(np.array([], dtype="datetime64[ps]")) # wrong unit False >>> is_datetime64_ns_dtype(pd.DatetimeIndex([1, 2, 3], dtype=np.datetime64)) # has 'ns' unit True
pandas/core/dtypes/common.py
def is_datetime64_ns_dtype(arr_or_dtype): """ Check whether the provided array or dtype is of the datetime64[ns] dtype. Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype is of the datetime64[ns] dtype. Examples -------- >>> is_datetime64_ns_dtype(str) False >>> is_datetime64_ns_dtype(int) False >>> is_datetime64_ns_dtype(np.datetime64) # no unit False >>> is_datetime64_ns_dtype(DatetimeTZDtype("ns", "US/Eastern")) True >>> is_datetime64_ns_dtype(np.array(['a', 'b'])) False >>> is_datetime64_ns_dtype(np.array([1, 2])) False >>> is_datetime64_ns_dtype(np.array([], dtype=np.datetime64)) # no unit False >>> is_datetime64_ns_dtype(np.array([], dtype="datetime64[ps]")) # wrong unit False >>> is_datetime64_ns_dtype(pd.DatetimeIndex([1, 2, 3], dtype=np.datetime64)) # has 'ns' unit True """ if arr_or_dtype is None: return False try: tipo = _get_dtype(arr_or_dtype) except TypeError: if is_datetime64tz_dtype(arr_or_dtype): tipo = _get_dtype(arr_or_dtype.dtype) else: return False return tipo == _NS_DTYPE or getattr(tipo, 'base', None) == _NS_DTYPE
def is_datetime64_ns_dtype(arr_or_dtype): """ Check whether the provided array or dtype is of the datetime64[ns] dtype. Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype is of the datetime64[ns] dtype. Examples -------- >>> is_datetime64_ns_dtype(str) False >>> is_datetime64_ns_dtype(int) False >>> is_datetime64_ns_dtype(np.datetime64) # no unit False >>> is_datetime64_ns_dtype(DatetimeTZDtype("ns", "US/Eastern")) True >>> is_datetime64_ns_dtype(np.array(['a', 'b'])) False >>> is_datetime64_ns_dtype(np.array([1, 2])) False >>> is_datetime64_ns_dtype(np.array([], dtype=np.datetime64)) # no unit False >>> is_datetime64_ns_dtype(np.array([], dtype="datetime64[ps]")) # wrong unit False >>> is_datetime64_ns_dtype(pd.DatetimeIndex([1, 2, 3], dtype=np.datetime64)) # has 'ns' unit True """ if arr_or_dtype is None: return False try: tipo = _get_dtype(arr_or_dtype) except TypeError: if is_datetime64tz_dtype(arr_or_dtype): tipo = _get_dtype(arr_or_dtype.dtype) else: return False return tipo == _NS_DTYPE or getattr(tipo, 'base', None) == _NS_DTYPE
[ "Check", "whether", "the", "provided", "array", "or", "dtype", "is", "of", "the", "datetime64", "[", "ns", "]", "dtype", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L1135-L1182
[ "def", "is_datetime64_ns_dtype", "(", "arr_or_dtype", ")", ":", "if", "arr_or_dtype", "is", "None", ":", "return", "False", "try", ":", "tipo", "=", "_get_dtype", "(", "arr_or_dtype", ")", "except", "TypeError", ":", "if", "is_datetime64tz_dtype", "(", "arr_or_dtype", ")", ":", "tipo", "=", "_get_dtype", "(", "arr_or_dtype", ".", "dtype", ")", "else", ":", "return", "False", "return", "tipo", "==", "_NS_DTYPE", "or", "getattr", "(", "tipo", ",", "'base'", ",", "None", ")", "==", "_NS_DTYPE" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_numeric_v_string_like
Check if we are comparing a string-like object to a numeric ndarray. NumPy doesn't like to compare such objects, especially numeric arrays and scalar string-likes. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a string-like object to a numeric array. Examples -------- >>> is_numeric_v_string_like(1, 1) False >>> is_numeric_v_string_like("foo", "foo") False >>> is_numeric_v_string_like(1, "foo") # non-array numeric False >>> is_numeric_v_string_like(np.array([1]), "foo") True >>> is_numeric_v_string_like("foo", np.array([1])) # symmetric check True >>> is_numeric_v_string_like(np.array([1, 2]), np.array(["foo"])) True >>> is_numeric_v_string_like(np.array(["foo"]), np.array([1, 2])) True >>> is_numeric_v_string_like(np.array([1]), np.array([2])) False >>> is_numeric_v_string_like(np.array(["foo"]), np.array(["foo"])) False
pandas/core/dtypes/common.py
def is_numeric_v_string_like(a, b): """ Check if we are comparing a string-like object to a numeric ndarray. NumPy doesn't like to compare such objects, especially numeric arrays and scalar string-likes. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a string-like object to a numeric array. Examples -------- >>> is_numeric_v_string_like(1, 1) False >>> is_numeric_v_string_like("foo", "foo") False >>> is_numeric_v_string_like(1, "foo") # non-array numeric False >>> is_numeric_v_string_like(np.array([1]), "foo") True >>> is_numeric_v_string_like("foo", np.array([1])) # symmetric check True >>> is_numeric_v_string_like(np.array([1, 2]), np.array(["foo"])) True >>> is_numeric_v_string_like(np.array(["foo"]), np.array([1, 2])) True >>> is_numeric_v_string_like(np.array([1]), np.array([2])) False >>> is_numeric_v_string_like(np.array(["foo"]), np.array(["foo"])) False """ is_a_array = isinstance(a, np.ndarray) is_b_array = isinstance(b, np.ndarray) is_a_numeric_array = is_a_array and is_numeric_dtype(a) is_b_numeric_array = is_b_array and is_numeric_dtype(b) is_a_string_array = is_a_array and is_string_like_dtype(a) is_b_string_array = is_b_array and is_string_like_dtype(b) is_a_scalar_string_like = not is_a_array and is_string_like(a) is_b_scalar_string_like = not is_b_array and is_string_like(b) return ((is_a_numeric_array and is_b_scalar_string_like) or (is_b_numeric_array and is_a_scalar_string_like) or (is_a_numeric_array and is_b_string_array) or (is_b_numeric_array and is_a_string_array))
def is_numeric_v_string_like(a, b): """ Check if we are comparing a string-like object to a numeric ndarray. NumPy doesn't like to compare such objects, especially numeric arrays and scalar string-likes. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a string-like object to a numeric array. Examples -------- >>> is_numeric_v_string_like(1, 1) False >>> is_numeric_v_string_like("foo", "foo") False >>> is_numeric_v_string_like(1, "foo") # non-array numeric False >>> is_numeric_v_string_like(np.array([1]), "foo") True >>> is_numeric_v_string_like("foo", np.array([1])) # symmetric check True >>> is_numeric_v_string_like(np.array([1, 2]), np.array(["foo"])) True >>> is_numeric_v_string_like(np.array(["foo"]), np.array([1, 2])) True >>> is_numeric_v_string_like(np.array([1]), np.array([2])) False >>> is_numeric_v_string_like(np.array(["foo"]), np.array(["foo"])) False """ is_a_array = isinstance(a, np.ndarray) is_b_array = isinstance(b, np.ndarray) is_a_numeric_array = is_a_array and is_numeric_dtype(a) is_b_numeric_array = is_b_array and is_numeric_dtype(b) is_a_string_array = is_a_array and is_string_like_dtype(a) is_b_string_array = is_b_array and is_string_like_dtype(b) is_a_scalar_string_like = not is_a_array and is_string_like(a) is_b_scalar_string_like = not is_b_array and is_string_like(b) return ((is_a_numeric_array and is_b_scalar_string_like) or (is_b_numeric_array and is_a_scalar_string_like) or (is_a_numeric_array and is_b_string_array) or (is_b_numeric_array and is_a_string_array))
[ "Check", "if", "we", "are", "comparing", "a", "string", "-", "like", "object", "to", "a", "numeric", "ndarray", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L1280-L1335
[ "def", "is_numeric_v_string_like", "(", "a", ",", "b", ")", ":", "is_a_array", "=", "isinstance", "(", "a", ",", "np", ".", "ndarray", ")", "is_b_array", "=", "isinstance", "(", "b", ",", "np", ".", "ndarray", ")", "is_a_numeric_array", "=", "is_a_array", "and", "is_numeric_dtype", "(", "a", ")", "is_b_numeric_array", "=", "is_b_array", "and", "is_numeric_dtype", "(", "b", ")", "is_a_string_array", "=", "is_a_array", "and", "is_string_like_dtype", "(", "a", ")", "is_b_string_array", "=", "is_b_array", "and", "is_string_like_dtype", "(", "b", ")", "is_a_scalar_string_like", "=", "not", "is_a_array", "and", "is_string_like", "(", "a", ")", "is_b_scalar_string_like", "=", "not", "is_b_array", "and", "is_string_like", "(", "b", ")", "return", "(", "(", "is_a_numeric_array", "and", "is_b_scalar_string_like", ")", "or", "(", "is_b_numeric_array", "and", "is_a_scalar_string_like", ")", "or", "(", "is_a_numeric_array", "and", "is_b_string_array", ")", "or", "(", "is_b_numeric_array", "and", "is_a_string_array", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_datetimelike_v_numeric
Check if we are comparing a datetime-like object to a numeric object. By "numeric," we mean an object that is either of an int or float dtype. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a datetime-like to a numeric object. Examples -------- >>> dt = np.datetime64(pd.datetime(2017, 1, 1)) >>> >>> is_datetimelike_v_numeric(1, 1) False >>> is_datetimelike_v_numeric(dt, dt) False >>> is_datetimelike_v_numeric(1, dt) True >>> is_datetimelike_v_numeric(dt, 1) # symmetric check True >>> is_datetimelike_v_numeric(np.array([dt]), 1) True >>> is_datetimelike_v_numeric(np.array([1]), dt) True >>> is_datetimelike_v_numeric(np.array([dt]), np.array([1])) True >>> is_datetimelike_v_numeric(np.array([1]), np.array([2])) False >>> is_datetimelike_v_numeric(np.array([dt]), np.array([dt])) False
pandas/core/dtypes/common.py
def is_datetimelike_v_numeric(a, b): """ Check if we are comparing a datetime-like object to a numeric object. By "numeric," we mean an object that is either of an int or float dtype. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a datetime-like to a numeric object. Examples -------- >>> dt = np.datetime64(pd.datetime(2017, 1, 1)) >>> >>> is_datetimelike_v_numeric(1, 1) False >>> is_datetimelike_v_numeric(dt, dt) False >>> is_datetimelike_v_numeric(1, dt) True >>> is_datetimelike_v_numeric(dt, 1) # symmetric check True >>> is_datetimelike_v_numeric(np.array([dt]), 1) True >>> is_datetimelike_v_numeric(np.array([1]), dt) True >>> is_datetimelike_v_numeric(np.array([dt]), np.array([1])) True >>> is_datetimelike_v_numeric(np.array([1]), np.array([2])) False >>> is_datetimelike_v_numeric(np.array([dt]), np.array([dt])) False """ if not hasattr(a, 'dtype'): a = np.asarray(a) if not hasattr(b, 'dtype'): b = np.asarray(b) def is_numeric(x): """ Check if an object has a numeric dtype (i.e. integer or float). """ return is_integer_dtype(x) or is_float_dtype(x) is_datetimelike = needs_i8_conversion return ((is_datetimelike(a) and is_numeric(b)) or (is_datetimelike(b) and is_numeric(a)))
def is_datetimelike_v_numeric(a, b): """ Check if we are comparing a datetime-like object to a numeric object. By "numeric," we mean an object that is either of an int or float dtype. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a datetime-like to a numeric object. Examples -------- >>> dt = np.datetime64(pd.datetime(2017, 1, 1)) >>> >>> is_datetimelike_v_numeric(1, 1) False >>> is_datetimelike_v_numeric(dt, dt) False >>> is_datetimelike_v_numeric(1, dt) True >>> is_datetimelike_v_numeric(dt, 1) # symmetric check True >>> is_datetimelike_v_numeric(np.array([dt]), 1) True >>> is_datetimelike_v_numeric(np.array([1]), dt) True >>> is_datetimelike_v_numeric(np.array([dt]), np.array([1])) True >>> is_datetimelike_v_numeric(np.array([1]), np.array([2])) False >>> is_datetimelike_v_numeric(np.array([dt]), np.array([dt])) False """ if not hasattr(a, 'dtype'): a = np.asarray(a) if not hasattr(b, 'dtype'): b = np.asarray(b) def is_numeric(x): """ Check if an object has a numeric dtype (i.e. integer or float). """ return is_integer_dtype(x) or is_float_dtype(x) is_datetimelike = needs_i8_conversion return ((is_datetimelike(a) and is_numeric(b)) or (is_datetimelike(b) and is_numeric(a)))
[ "Check", "if", "we", "are", "comparing", "a", "datetime", "-", "like", "object", "to", "a", "numeric", "object", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L1338-L1393
[ "def", "is_datetimelike_v_numeric", "(", "a", ",", "b", ")", ":", "if", "not", "hasattr", "(", "a", ",", "'dtype'", ")", ":", "a", "=", "np", ".", "asarray", "(", "a", ")", "if", "not", "hasattr", "(", "b", ",", "'dtype'", ")", ":", "b", "=", "np", ".", "asarray", "(", "b", ")", "def", "is_numeric", "(", "x", ")", ":", "\"\"\"\n Check if an object has a numeric dtype (i.e. integer or float).\n \"\"\"", "return", "is_integer_dtype", "(", "x", ")", "or", "is_float_dtype", "(", "x", ")", "is_datetimelike", "=", "needs_i8_conversion", "return", "(", "(", "is_datetimelike", "(", "a", ")", "and", "is_numeric", "(", "b", ")", ")", "or", "(", "is_datetimelike", "(", "b", ")", "and", "is_numeric", "(", "a", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
is_datetimelike_v_object
Check if we are comparing a datetime-like object to an object instance. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a datetime-like to an object instance. Examples -------- >>> obj = object() >>> dt = np.datetime64(pd.datetime(2017, 1, 1)) >>> >>> is_datetimelike_v_object(obj, obj) False >>> is_datetimelike_v_object(dt, dt) False >>> is_datetimelike_v_object(obj, dt) True >>> is_datetimelike_v_object(dt, obj) # symmetric check True >>> is_datetimelike_v_object(np.array([dt]), obj) True >>> is_datetimelike_v_object(np.array([obj]), dt) True >>> is_datetimelike_v_object(np.array([dt]), np.array([obj])) True >>> is_datetimelike_v_object(np.array([obj]), np.array([obj])) False >>> is_datetimelike_v_object(np.array([dt]), np.array([1])) False >>> is_datetimelike_v_object(np.array([dt]), np.array([dt])) False
pandas/core/dtypes/common.py
def is_datetimelike_v_object(a, b): """ Check if we are comparing a datetime-like object to an object instance. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a datetime-like to an object instance. Examples -------- >>> obj = object() >>> dt = np.datetime64(pd.datetime(2017, 1, 1)) >>> >>> is_datetimelike_v_object(obj, obj) False >>> is_datetimelike_v_object(dt, dt) False >>> is_datetimelike_v_object(obj, dt) True >>> is_datetimelike_v_object(dt, obj) # symmetric check True >>> is_datetimelike_v_object(np.array([dt]), obj) True >>> is_datetimelike_v_object(np.array([obj]), dt) True >>> is_datetimelike_v_object(np.array([dt]), np.array([obj])) True >>> is_datetimelike_v_object(np.array([obj]), np.array([obj])) False >>> is_datetimelike_v_object(np.array([dt]), np.array([1])) False >>> is_datetimelike_v_object(np.array([dt]), np.array([dt])) False """ if not hasattr(a, 'dtype'): a = np.asarray(a) if not hasattr(b, 'dtype'): b = np.asarray(b) is_datetimelike = needs_i8_conversion return ((is_datetimelike(a) and is_object_dtype(b)) or (is_datetimelike(b) and is_object_dtype(a)))
def is_datetimelike_v_object(a, b): """ Check if we are comparing a datetime-like object to an object instance. Parameters ---------- a : array-like, scalar The first object to check. b : array-like, scalar The second object to check. Returns ------- boolean Whether we return a comparing a datetime-like to an object instance. Examples -------- >>> obj = object() >>> dt = np.datetime64(pd.datetime(2017, 1, 1)) >>> >>> is_datetimelike_v_object(obj, obj) False >>> is_datetimelike_v_object(dt, dt) False >>> is_datetimelike_v_object(obj, dt) True >>> is_datetimelike_v_object(dt, obj) # symmetric check True >>> is_datetimelike_v_object(np.array([dt]), obj) True >>> is_datetimelike_v_object(np.array([obj]), dt) True >>> is_datetimelike_v_object(np.array([dt]), np.array([obj])) True >>> is_datetimelike_v_object(np.array([obj]), np.array([obj])) False >>> is_datetimelike_v_object(np.array([dt]), np.array([1])) False >>> is_datetimelike_v_object(np.array([dt]), np.array([dt])) False """ if not hasattr(a, 'dtype'): a = np.asarray(a) if not hasattr(b, 'dtype'): b = np.asarray(b) is_datetimelike = needs_i8_conversion return ((is_datetimelike(a) and is_object_dtype(b)) or (is_datetimelike(b) and is_object_dtype(a)))
[ "Check", "if", "we", "are", "comparing", "a", "datetime", "-", "like", "object", "to", "an", "object", "instance", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L1396-L1446
[ "def", "is_datetimelike_v_object", "(", "a", ",", "b", ")", ":", "if", "not", "hasattr", "(", "a", ",", "'dtype'", ")", ":", "a", "=", "np", ".", "asarray", "(", "a", ")", "if", "not", "hasattr", "(", "b", ",", "'dtype'", ")", ":", "b", "=", "np", ".", "asarray", "(", "b", ")", "is_datetimelike", "=", "needs_i8_conversion", "return", "(", "(", "is_datetimelike", "(", "a", ")", "and", "is_object_dtype", "(", "b", ")", ")", "or", "(", "is_datetimelike", "(", "b", ")", "and", "is_object_dtype", "(", "a", ")", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037
train
needs_i8_conversion
Check whether the array or dtype should be converted to int64. An array-like or dtype "needs" such a conversion if the array-like or dtype is of a datetime-like dtype Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype should be converted to int64. Examples -------- >>> needs_i8_conversion(str) False >>> needs_i8_conversion(np.int64) False >>> needs_i8_conversion(np.datetime64) True >>> needs_i8_conversion(np.array(['a', 'b'])) False >>> needs_i8_conversion(pd.Series([1, 2])) False >>> needs_i8_conversion(pd.Series([], dtype="timedelta64[ns]")) True >>> needs_i8_conversion(pd.DatetimeIndex([1, 2, 3], tz="US/Eastern")) True
pandas/core/dtypes/common.py
def needs_i8_conversion(arr_or_dtype): """ Check whether the array or dtype should be converted to int64. An array-like or dtype "needs" such a conversion if the array-like or dtype is of a datetime-like dtype Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype should be converted to int64. Examples -------- >>> needs_i8_conversion(str) False >>> needs_i8_conversion(np.int64) False >>> needs_i8_conversion(np.datetime64) True >>> needs_i8_conversion(np.array(['a', 'b'])) False >>> needs_i8_conversion(pd.Series([1, 2])) False >>> needs_i8_conversion(pd.Series([], dtype="timedelta64[ns]")) True >>> needs_i8_conversion(pd.DatetimeIndex([1, 2, 3], tz="US/Eastern")) True """ if arr_or_dtype is None: return False return (is_datetime_or_timedelta_dtype(arr_or_dtype) or is_datetime64tz_dtype(arr_or_dtype) or is_period_dtype(arr_or_dtype))
def needs_i8_conversion(arr_or_dtype): """ Check whether the array or dtype should be converted to int64. An array-like or dtype "needs" such a conversion if the array-like or dtype is of a datetime-like dtype Parameters ---------- arr_or_dtype : array-like The array or dtype to check. Returns ------- boolean Whether or not the array or dtype should be converted to int64. Examples -------- >>> needs_i8_conversion(str) False >>> needs_i8_conversion(np.int64) False >>> needs_i8_conversion(np.datetime64) True >>> needs_i8_conversion(np.array(['a', 'b'])) False >>> needs_i8_conversion(pd.Series([1, 2])) False >>> needs_i8_conversion(pd.Series([], dtype="timedelta64[ns]")) True >>> needs_i8_conversion(pd.DatetimeIndex([1, 2, 3], tz="US/Eastern")) True """ if arr_or_dtype is None: return False return (is_datetime_or_timedelta_dtype(arr_or_dtype) or is_datetime64tz_dtype(arr_or_dtype) or is_period_dtype(arr_or_dtype))
[ "Check", "whether", "the", "array", "or", "dtype", "should", "be", "converted", "to", "int64", "." ]
pandas-dev/pandas
python
https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/common.py#L1449-L1488
[ "def", "needs_i8_conversion", "(", "arr_or_dtype", ")", ":", "if", "arr_or_dtype", "is", "None", ":", "return", "False", "return", "(", "is_datetime_or_timedelta_dtype", "(", "arr_or_dtype", ")", "or", "is_datetime64tz_dtype", "(", "arr_or_dtype", ")", "or", "is_period_dtype", "(", "arr_or_dtype", ")", ")" ]
9feb3ad92cc0397a04b665803a49299ee7aa1037