repository_name
stringlengths
7
55
func_path_in_repository
stringlengths
4
223
func_name
stringlengths
1
134
whole_func_string
stringlengths
75
104k
language
stringclasses
1 value
func_code_string
stringlengths
75
104k
func_code_tokens
listlengths
19
28.4k
func_documentation_string
stringlengths
1
46.9k
func_documentation_tokens
listlengths
1
1.97k
split_name
stringclasses
1 value
func_code_url
stringlengths
87
315
libyal/dtfabric
dtfabric/runtime/data_maps.py
SemanticDataTypeMap.MapByteStream
def MapByteStream(self, byte_stream, **unused_kwargs): # pylint: disable=redundant-returns-doc """Maps the data type on a byte stream. Args: byte_stream (bytes): byte stream. Returns: object: mapped value. Raises: MappingError: if the data type definition cannot be mapped on the byte stream. """ raise errors.MappingError( 'Unable to map {0:s} data type to byte stream'.format( self._data_type_definition.TYPE_INDICATOR))
python
def MapByteStream(self, byte_stream, **unused_kwargs): # pylint: disable=redundant-returns-doc """Maps the data type on a byte stream. Args: byte_stream (bytes): byte stream. Returns: object: mapped value. Raises: MappingError: if the data type definition cannot be mapped on the byte stream. """ raise errors.MappingError( 'Unable to map {0:s} data type to byte stream'.format( self._data_type_definition.TYPE_INDICATOR))
[ "def", "MapByteStream", "(", "self", ",", "byte_stream", ",", "*", "*", "unused_kwargs", ")", ":", "# pylint: disable=redundant-returns-doc", "raise", "errors", ".", "MappingError", "(", "'Unable to map {0:s} data type to byte stream'", ".", "format", "(", "self", ".", "_data_type_definition", ".", "TYPE_INDICATOR", ")", ")" ]
Maps the data type on a byte stream. Args: byte_stream (bytes): byte stream. Returns: object: mapped value. Raises: MappingError: if the data type definition cannot be mapped on the byte stream.
[ "Maps", "the", "data", "type", "on", "a", "byte", "stream", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/data_maps.py#L1961-L1976
libyal/dtfabric
dtfabric/runtime/data_maps.py
EnumerationMap.GetName
def GetName(self, number): """Retrieves the name of an enumeration value by number. Args: number (int): number. Returns: str: name of the enumeration value or None if no corresponding enumeration value was found. """ value = self._data_type_definition.values_per_number.get(number, None) if not value: return None return value.name
python
def GetName(self, number): """Retrieves the name of an enumeration value by number. Args: number (int): number. Returns: str: name of the enumeration value or None if no corresponding enumeration value was found. """ value = self._data_type_definition.values_per_number.get(number, None) if not value: return None return value.name
[ "def", "GetName", "(", "self", ",", "number", ")", ":", "value", "=", "self", ".", "_data_type_definition", ".", "values_per_number", ".", "get", "(", "number", ",", "None", ")", "if", "not", "value", ":", "return", "None", "return", "value", ".", "name" ]
Retrieves the name of an enumeration value by number. Args: number (int): number. Returns: str: name of the enumeration value or None if no corresponding enumeration value was found.
[ "Retrieves", "the", "name", "of", "an", "enumeration", "value", "by", "number", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/data_maps.py#L1986-L2000
libyal/dtfabric
dtfabric/runtime/data_maps.py
DataTypeMapFactory.CreateDataTypeMap
def CreateDataTypeMap(self, definition_name): """Creates a specific data type map by name. Args: definition_name (str): name of the data type definition. Returns: DataTypeMap: data type map or None if the date type definition is not available. """ data_type_definition = self._definitions_registry.GetDefinitionByName( definition_name) if not data_type_definition: return None return DataTypeMapFactory.CreateDataTypeMapByType(data_type_definition)
python
def CreateDataTypeMap(self, definition_name): """Creates a specific data type map by name. Args: definition_name (str): name of the data type definition. Returns: DataTypeMap: data type map or None if the date type definition is not available. """ data_type_definition = self._definitions_registry.GetDefinitionByName( definition_name) if not data_type_definition: return None return DataTypeMapFactory.CreateDataTypeMapByType(data_type_definition)
[ "def", "CreateDataTypeMap", "(", "self", ",", "definition_name", ")", ":", "data_type_definition", "=", "self", ".", "_definitions_registry", ".", "GetDefinitionByName", "(", "definition_name", ")", "if", "not", "data_type_definition", ":", "return", "None", "return", "DataTypeMapFactory", ".", "CreateDataTypeMapByType", "(", "data_type_definition", ")" ]
Creates a specific data type map by name. Args: definition_name (str): name of the data type definition. Returns: DataTypeMap: data type map or None if the date type definition is not available.
[ "Creates", "a", "specific", "data", "type", "map", "by", "name", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/data_maps.py#L2032-L2047
libyal/dtfabric
dtfabric/runtime/data_maps.py
DataTypeMapFactory.CreateDataTypeMapByType
def CreateDataTypeMapByType(cls, data_type_definition): """Creates a specific data type map by type indicator. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: DataTypeMap: data type map or None if the date type definition is not available. """ data_type_map_class = cls._MAP_PER_DEFINITION.get( data_type_definition.TYPE_INDICATOR, None) if not data_type_map_class: return None return data_type_map_class(data_type_definition)
python
def CreateDataTypeMapByType(cls, data_type_definition): """Creates a specific data type map by type indicator. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: DataTypeMap: data type map or None if the date type definition is not available. """ data_type_map_class = cls._MAP_PER_DEFINITION.get( data_type_definition.TYPE_INDICATOR, None) if not data_type_map_class: return None return data_type_map_class(data_type_definition)
[ "def", "CreateDataTypeMapByType", "(", "cls", ",", "data_type_definition", ")", ":", "data_type_map_class", "=", "cls", ".", "_MAP_PER_DEFINITION", ".", "get", "(", "data_type_definition", ".", "TYPE_INDICATOR", ",", "None", ")", "if", "not", "data_type_map_class", ":", "return", "None", "return", "data_type_map_class", "(", "data_type_definition", ")" ]
Creates a specific data type map by type indicator. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: DataTypeMap: data type map or None if the date type definition is not available.
[ "Creates", "a", "specific", "data", "type", "map", "by", "type", "indicator", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/data_maps.py#L2050-L2065
libyal/dtfabric
dtfabric/data_types.py
FixedSizeDataTypeDefinition.GetByteSize
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if self.size == definitions.SIZE_NATIVE or self.units != 'bytes': return None return self.size
python
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if self.size == definitions.SIZE_NATIVE or self.units != 'bytes': return None return self.size
[ "def", "GetByteSize", "(", "self", ")", ":", "if", "self", ".", "size", "==", "definitions", ".", "SIZE_NATIVE", "or", "self", ".", "units", "!=", "'bytes'", ":", "return", "None", "return", "self", ".", "size" ]
Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined.
[ "Retrieves", "the", "byte", "size", "of", "the", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L119-L128
libyal/dtfabric
dtfabric/data_types.py
ElementSequenceDataTypeDefinition.GetByteSize
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if not self.element_data_type_definition: return None if self.elements_data_size: return self.elements_data_size if not self.number_of_elements: return None element_byte_size = self.element_data_type_definition.GetByteSize() if not element_byte_size: return None return element_byte_size * self.number_of_elements
python
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if not self.element_data_type_definition: return None if self.elements_data_size: return self.elements_data_size if not self.number_of_elements: return None element_byte_size = self.element_data_type_definition.GetByteSize() if not element_byte_size: return None return element_byte_size * self.number_of_elements
[ "def", "GetByteSize", "(", "self", ")", ":", "if", "not", "self", ".", "element_data_type_definition", ":", "return", "None", "if", "self", ".", "elements_data_size", ":", "return", "self", ".", "elements_data_size", "if", "not", "self", ".", "number_of_elements", ":", "return", "None", "element_byte_size", "=", "self", ".", "element_data_type_definition", ".", "GetByteSize", "(", ")", "if", "not", "element_byte_size", ":", "return", "None", "return", "element_byte_size", "*", "self", ".", "number_of_elements" ]
Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined.
[ "Retrieves", "the", "byte", "size", "of", "the", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L306-L325
libyal/dtfabric
dtfabric/data_types.py
DataTypeDefinitionWithMembers.AddMemberDefinition
def AddMemberDefinition(self, member_definition): """Adds a member definition. Args: member_definition (DataTypeDefinition): member data type definition. """ self._byte_size = None self.members.append(member_definition) if self.sections: section_definition = self.sections[-1] section_definition.members.append(member_definition)
python
def AddMemberDefinition(self, member_definition): """Adds a member definition. Args: member_definition (DataTypeDefinition): member data type definition. """ self._byte_size = None self.members.append(member_definition) if self.sections: section_definition = self.sections[-1] section_definition.members.append(member_definition)
[ "def", "AddMemberDefinition", "(", "self", ",", "member_definition", ")", ":", "self", ".", "_byte_size", "=", "None", "self", ".", "members", ".", "append", "(", "member_definition", ")", "if", "self", ".", "sections", ":", "section_definition", "=", "self", ".", "sections", "[", "-", "1", "]", "section_definition", ".", "members", ".", "append", "(", "member_definition", ")" ]
Adds a member definition. Args: member_definition (DataTypeDefinition): member data type definition.
[ "Adds", "a", "member", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L398-L409
libyal/dtfabric
dtfabric/data_types.py
MemberDataTypeDefinition.IsComposite
def IsComposite(self): """Determines if the data type is composite. A composite data type consists of other data types. Returns: bool: True if the data type is composite, False otherwise. """ return bool(self.condition) or ( self.member_data_type_definition and self.member_data_type_definition.IsComposite())
python
def IsComposite(self): """Determines if the data type is composite. A composite data type consists of other data types. Returns: bool: True if the data type is composite, False otherwise. """ return bool(self.condition) or ( self.member_data_type_definition and self.member_data_type_definition.IsComposite())
[ "def", "IsComposite", "(", "self", ")", ":", "return", "bool", "(", "self", ".", "condition", ")", "or", "(", "self", ".", "member_data_type_definition", "and", "self", ".", "member_data_type_definition", ".", "IsComposite", "(", ")", ")" ]
Determines if the data type is composite. A composite data type consists of other data types. Returns: bool: True if the data type is composite, False otherwise.
[ "Determines", "if", "the", "data", "type", "is", "composite", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L476-L486
libyal/dtfabric
dtfabric/data_types.py
StructureDefinition.GetByteSize
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if self._byte_size is None and self.members: self._byte_size = 0 for member_definition in self.members: if isinstance(member_definition, PaddingDefinition): _, byte_size = divmod( self._byte_size, member_definition.alignment_size) if byte_size > 0: byte_size = member_definition.alignment_size - byte_size else: byte_size = member_definition.GetByteSize() if byte_size is None: self._byte_size = None break self._byte_size += byte_size return self._byte_size
python
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if self._byte_size is None and self.members: self._byte_size = 0 for member_definition in self.members: if isinstance(member_definition, PaddingDefinition): _, byte_size = divmod( self._byte_size, member_definition.alignment_size) if byte_size > 0: byte_size = member_definition.alignment_size - byte_size else: byte_size = member_definition.GetByteSize() if byte_size is None: self._byte_size = None break self._byte_size += byte_size return self._byte_size
[ "def", "GetByteSize", "(", "self", ")", ":", "if", "self", ".", "_byte_size", "is", "None", "and", "self", ".", "members", ":", "self", ".", "_byte_size", "=", "0", "for", "member_definition", "in", "self", ".", "members", ":", "if", "isinstance", "(", "member_definition", ",", "PaddingDefinition", ")", ":", "_", ",", "byte_size", "=", "divmod", "(", "self", ".", "_byte_size", ",", "member_definition", ".", "alignment_size", ")", "if", "byte_size", ">", "0", ":", "byte_size", "=", "member_definition", ".", "alignment_size", "-", "byte_size", "else", ":", "byte_size", "=", "member_definition", ".", "GetByteSize", "(", ")", "if", "byte_size", "is", "None", ":", "self", ".", "_byte_size", "=", "None", "break", "self", ".", "_byte_size", "+=", "byte_size", "return", "self", ".", "_byte_size" ]
Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined.
[ "Retrieves", "the", "byte", "size", "of", "the", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L531-L554
libyal/dtfabric
dtfabric/data_types.py
UnionDefinition.GetByteSize
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if self._byte_size is None and self.members: self._byte_size = 0 for member_definition in self.members: byte_size = member_definition.GetByteSize() if byte_size is None: self._byte_size = None break self._byte_size = max(self._byte_size, byte_size) return self._byte_size
python
def GetByteSize(self): """Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined. """ if self._byte_size is None and self.members: self._byte_size = 0 for member_definition in self.members: byte_size = member_definition.GetByteSize() if byte_size is None: self._byte_size = None break self._byte_size = max(self._byte_size, byte_size) return self._byte_size
[ "def", "GetByteSize", "(", "self", ")", ":", "if", "self", ".", "_byte_size", "is", "None", "and", "self", ".", "members", ":", "self", ".", "_byte_size", "=", "0", "for", "member_definition", "in", "self", ".", "members", ":", "byte_size", "=", "member_definition", ".", "GetByteSize", "(", ")", "if", "byte_size", "is", "None", ":", "self", ".", "_byte_size", "=", "None", "break", "self", ".", "_byte_size", "=", "max", "(", "self", ".", "_byte_size", ",", "byte_size", ")", "return", "self", ".", "_byte_size" ]
Retrieves the byte size of the data type definition. Returns: int: data type size in bytes or None if size cannot be determined.
[ "Retrieves", "the", "byte", "size", "of", "the", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L562-L578
libyal/dtfabric
dtfabric/data_types.py
EnumerationDefinition.AddValue
def AddValue(self, name, number, aliases=None, description=None): """Adds an enumeration value. Args: name (str): name. number (int): number. aliases (Optional[list[str]]): aliases. description (Optional[str]): description. Raises: KeyError: if the enumeration value already exists. """ if name in self.values_per_name: raise KeyError('Value with name: {0:s} already exists.'.format(name)) if number in self.values_per_number: raise KeyError('Value with number: {0!s} already exists.'.format(number)) for alias in aliases or []: if alias in self.values_per_alias: raise KeyError('Value with alias: {0:s} already exists.'.format(alias)) enumeration_value = EnumerationValue( name, number, aliases=aliases, description=description) self.values.append(enumeration_value) self.values_per_name[name] = enumeration_value self.values_per_number[number] = enumeration_value for alias in aliases or []: self.values_per_alias[alias] = enumeration_value
python
def AddValue(self, name, number, aliases=None, description=None): """Adds an enumeration value. Args: name (str): name. number (int): number. aliases (Optional[list[str]]): aliases. description (Optional[str]): description. Raises: KeyError: if the enumeration value already exists. """ if name in self.values_per_name: raise KeyError('Value with name: {0:s} already exists.'.format(name)) if number in self.values_per_number: raise KeyError('Value with number: {0!s} already exists.'.format(number)) for alias in aliases or []: if alias in self.values_per_alias: raise KeyError('Value with alias: {0:s} already exists.'.format(alias)) enumeration_value = EnumerationValue( name, number, aliases=aliases, description=description) self.values.append(enumeration_value) self.values_per_name[name] = enumeration_value self.values_per_number[number] = enumeration_value for alias in aliases or []: self.values_per_alias[alias] = enumeration_value
[ "def", "AddValue", "(", "self", ",", "name", ",", "number", ",", "aliases", "=", "None", ",", "description", "=", "None", ")", ":", "if", "name", "in", "self", ".", "values_per_name", ":", "raise", "KeyError", "(", "'Value with name: {0:s} already exists.'", ".", "format", "(", "name", ")", ")", "if", "number", "in", "self", ".", "values_per_number", ":", "raise", "KeyError", "(", "'Value with number: {0!s} already exists.'", ".", "format", "(", "number", ")", ")", "for", "alias", "in", "aliases", "or", "[", "]", ":", "if", "alias", "in", "self", ".", "values_per_alias", ":", "raise", "KeyError", "(", "'Value with alias: {0:s} already exists.'", ".", "format", "(", "alias", ")", ")", "enumeration_value", "=", "EnumerationValue", "(", "name", ",", "number", ",", "aliases", "=", "aliases", ",", "description", "=", "description", ")", "self", ".", "values", ".", "append", "(", "enumeration_value", ")", "self", ".", "values_per_name", "[", "name", "]", "=", "enumeration_value", "self", ".", "values_per_number", "[", "number", "]", "=", "enumeration_value", "for", "alias", "in", "aliases", "or", "[", "]", ":", "self", ".", "values_per_alias", "[", "alias", "]", "=", "enumeration_value" ]
Adds an enumeration value. Args: name (str): name. number (int): number. aliases (Optional[list[str]]): aliases. description (Optional[str]): description. Raises: KeyError: if the enumeration value already exists.
[ "Adds", "an", "enumeration", "value", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L677-L707
libyal/dtfabric
dtfabric/data_types.py
StructureFamilyDefinition.AddMemberDefinition
def AddMemberDefinition(self, member_definition): """Adds a member definition. Args: member_definition (DataTypeDefinition): member data type definition. """ self.members.append(member_definition) member_definition.family_definition = self
python
def AddMemberDefinition(self, member_definition): """Adds a member definition. Args: member_definition (DataTypeDefinition): member data type definition. """ self.members.append(member_definition) member_definition.family_definition = self
[ "def", "AddMemberDefinition", "(", "self", ",", "member_definition", ")", ":", "self", ".", "members", ".", "append", "(", "member_definition", ")", "member_definition", ".", "family_definition", "=", "self" ]
Adds a member definition. Args: member_definition (DataTypeDefinition): member data type definition.
[ "Adds", "a", "member", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/data_types.py#L771-L778
dlanger/inlinestyler
inlinestyler/cssselect.py
parse_series
def parse_series(s): """ Parses things like '1n+2', or 'an+b' generally, returning (a, b) """ if isinstance(s, Element): s = s._format_element() if not s or s == '*': # Happens when there's nothing, which the CSS parser thinks of as * return (0, 0) if isinstance(s, int): # Happens when you just get a number return (0, s) if s == 'odd': return (2, 1) elif s == 'even': return (2, 0) elif s == 'n': return (1, 0) if 'n' not in s: # Just a b return (0, int(s)) a, b = s.split('n', 1) if not a: a = 1 elif a == '-' or a == '+': a = int(a+'1') else: a = int(a) if not b: b = 0 elif b == '-' or b == '+': b = int(b+'1') else: b = int(b) return (a, b)
python
def parse_series(s): """ Parses things like '1n+2', or 'an+b' generally, returning (a, b) """ if isinstance(s, Element): s = s._format_element() if not s or s == '*': # Happens when there's nothing, which the CSS parser thinks of as * return (0, 0) if isinstance(s, int): # Happens when you just get a number return (0, s) if s == 'odd': return (2, 1) elif s == 'even': return (2, 0) elif s == 'n': return (1, 0) if 'n' not in s: # Just a b return (0, int(s)) a, b = s.split('n', 1) if not a: a = 1 elif a == '-' or a == '+': a = int(a+'1') else: a = int(a) if not b: b = 0 elif b == '-' or b == '+': b = int(b+'1') else: b = int(b) return (a, b)
[ "def", "parse_series", "(", "s", ")", ":", "if", "isinstance", "(", "s", ",", "Element", ")", ":", "s", "=", "s", ".", "_format_element", "(", ")", "if", "not", "s", "or", "s", "==", "'*'", ":", "# Happens when there's nothing, which the CSS parser thinks of as *", "return", "(", "0", ",", "0", ")", "if", "isinstance", "(", "s", ",", "int", ")", ":", "# Happens when you just get a number", "return", "(", "0", ",", "s", ")", "if", "s", "==", "'odd'", ":", "return", "(", "2", ",", "1", ")", "elif", "s", "==", "'even'", ":", "return", "(", "2", ",", "0", ")", "elif", "s", "==", "'n'", ":", "return", "(", "1", ",", "0", ")", "if", "'n'", "not", "in", "s", ":", "# Just a b", "return", "(", "0", ",", "int", "(", "s", ")", ")", "a", ",", "b", "=", "s", ".", "split", "(", "'n'", ",", "1", ")", "if", "not", "a", ":", "a", "=", "1", "elif", "a", "==", "'-'", "or", "a", "==", "'+'", ":", "a", "=", "int", "(", "a", "+", "'1'", ")", "else", ":", "a", "=", "int", "(", "a", ")", "if", "not", "b", ":", "b", "=", "0", "elif", "b", "==", "'-'", "or", "b", "==", "'+'", ":", "b", "=", "int", "(", "b", "+", "'1'", ")", "else", ":", "b", "=", "int", "(", "b", ")", "return", "(", "a", ",", "b", ")" ]
Parses things like '1n+2', or 'an+b' generally, returning (a, b)
[ "Parses", "things", "like", "1n", "+", "2", "or", "an", "+", "b", "generally", "returning", "(", "a", "b", ")" ]
train
https://github.com/dlanger/inlinestyler/blob/335c4fbab892f0ed67466a6beaea6a91f395ad12/inlinestyler/cssselect.py#L776-L810
dlanger/inlinestyler
inlinestyler/cssselect.py
XPathExpr.add_star_prefix
def add_star_prefix(self): """ Adds a /* prefix if there is no prefix. This is when you need to keep context's constrained to a single parent. """ if self.path: self.path += '*/' else: self.path = '*/' self.star_prefix = True
python
def add_star_prefix(self): """ Adds a /* prefix if there is no prefix. This is when you need to keep context's constrained to a single parent. """ if self.path: self.path += '*/' else: self.path = '*/' self.star_prefix = True
[ "def", "add_star_prefix", "(", "self", ")", ":", "if", "self", ".", "path", ":", "self", ".", "path", "+=", "'*/'", "else", ":", "self", ".", "path", "=", "'*/'", "self", ".", "star_prefix", "=", "True" ]
Adds a /* prefix if there is no prefix. This is when you need to keep context's constrained to a single parent.
[ "Adds", "a", "/", "*", "prefix", "if", "there", "is", "no", "prefix", ".", "This", "is", "when", "you", "need", "to", "keep", "context", "s", "constrained", "to", "a", "single", "parent", "." ]
train
https://github.com/dlanger/inlinestyler/blob/335c4fbab892f0ed67466a6beaea6a91f395ad12/inlinestyler/cssselect.py#L580-L589
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadBooleanDataTypeDefinition
def _ReadBooleanDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a boolean data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: BooleanDataTypeDefinition: boolean data type definition. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.BooleanDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_BOOLEAN, is_member=is_member, supported_size_values=(1, 2, 4))
python
def _ReadBooleanDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a boolean data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: BooleanDataTypeDefinition: boolean data type definition. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.BooleanDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_BOOLEAN, is_member=is_member, supported_size_values=(1, 2, 4))
[ "def", "_ReadBooleanDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "return", "self", ".", "_ReadFixedSizeDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "BooleanDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_ATTRIBUTES_BOOLEAN", ",", "is_member", "=", "is_member", ",", "supported_size_values", "=", "(", "1", ",", "2", ",", "4", ")", ")" ]
Reads a boolean data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: BooleanDataTypeDefinition: boolean data type definition.
[ "Reads", "a", "boolean", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L102-L122
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadCharacterDataTypeDefinition
def _ReadCharacterDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a character data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: CharacterDataTypeDefinition: character data type definition. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.CharacterDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE, is_member=is_member, supported_size_values=(1, 2, 4))
python
def _ReadCharacterDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a character data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: CharacterDataTypeDefinition: character data type definition. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.CharacterDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE, is_member=is_member, supported_size_values=(1, 2, 4))
[ "def", "_ReadCharacterDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "return", "self", ".", "_ReadFixedSizeDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "CharacterDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE", ",", "is_member", "=", "is_member", ",", "supported_size_values", "=", "(", "1", ",", "2", ",", "4", ")", ")" ]
Reads a character data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: CharacterDataTypeDefinition: character data type definition.
[ "Reads", "a", "character", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L124-L144
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadConstantDataTypeDefinition
def _ReadConstantDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a constant data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: ConstantDataTypeDefinition: constant data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) value = definition_values.get('value', None) if value is None: error_message = 'missing value' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadSemanticDataTypeDefinition( definitions_registry, definition_values, data_types.ConstantDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_CONSTANT) definition_object.value = value return definition_object
python
def _ReadConstantDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a constant data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: ConstantDataTypeDefinition: constant data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) value = definition_values.get('value', None) if value is None: error_message = 'missing value' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadSemanticDataTypeDefinition( definitions_registry, definition_values, data_types.ConstantDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_CONSTANT) definition_object.value = value return definition_object
[ "def", "_ReadConstantDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "error_message", "=", "'data type not supported as member'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "value", "=", "definition_values", ".", "get", "(", "'value'", ",", "None", ")", "if", "value", "is", "None", ":", "error_message", "=", "'missing value'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", "=", "self", ".", "_ReadSemanticDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "ConstantDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_DEFINITION_VALUES_CONSTANT", ")", "definition_object", ".", "value", "=", "value", "return", "definition_object" ]
Reads a constant data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: ConstantDataTypeDefinition: constant data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "constant", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L146-L180
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadDataTypeDefinition
def _ReadDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads a data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: DataTypeDefinition: data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ aliases = definition_values.get('aliases', None) description = definition_values.get('description', None) urls = definition_values.get('urls', None) unsupported_definition_values = set(definition_values.keys()).difference( supported_definition_values) if unsupported_definition_values: error_message = 'unsupported definition values: {0:s}'.format( ', '.join(unsupported_definition_values)) raise errors.DefinitionReaderError(definition_name, error_message) return data_type_definition_class( definition_name, aliases=aliases, description=description, urls=urls)
python
def _ReadDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads a data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: DataTypeDefinition: data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ aliases = definition_values.get('aliases', None) description = definition_values.get('description', None) urls = definition_values.get('urls', None) unsupported_definition_values = set(definition_values.keys()).difference( supported_definition_values) if unsupported_definition_values: error_message = 'unsupported definition values: {0:s}'.format( ', '.join(unsupported_definition_values)) raise errors.DefinitionReaderError(definition_name, error_message) return data_type_definition_class( definition_name, aliases=aliases, description=description, urls=urls)
[ "def", "_ReadDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")", ":", "aliases", "=", "definition_values", ".", "get", "(", "'aliases'", ",", "None", ")", "description", "=", "definition_values", ".", "get", "(", "'description'", ",", "None", ")", "urls", "=", "definition_values", ".", "get", "(", "'urls'", ",", "None", ")", "unsupported_definition_values", "=", "set", "(", "definition_values", ".", "keys", "(", ")", ")", ".", "difference", "(", "supported_definition_values", ")", "if", "unsupported_definition_values", ":", "error_message", "=", "'unsupported definition values: {0:s}'", ".", "format", "(", "', '", ".", "join", "(", "unsupported_definition_values", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "return", "data_type_definition_class", "(", "definition_name", ",", "aliases", "=", "aliases", ",", "description", "=", "description", ",", "urls", "=", "urls", ")" ]
Reads a data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: DataTypeDefinition: data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L184-L217
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadDataTypeDefinitionWithMembers
def _ReadDataTypeDefinitionWithMembers( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supports_conditions=False): """Reads a data type definition with members. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supports_conditions (Optional[bool]): True if conditions are supported by the data type definition. Returns: StringDefinition: string data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ members = definition_values.get('members', None) if not members: error_message = 'missing members' raise errors.DefinitionReaderError(definition_name, error_message) supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_STORAGE_DATA_TYPE_WITH_MEMBERS) definition_object = self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values) attributes = definition_values.get('attributes', None) if attributes: unsupported_attributes = set(attributes.keys()).difference( self._SUPPORTED_ATTRIBUTES_STORAGE_DATA_TYPE) if unsupported_attributes: error_message = 'unsupported attributes: {0:s}'.format( ', '.join(unsupported_attributes)) raise errors.DefinitionReaderError(definition_name, error_message) byte_order = attributes.get('byte_order', definitions.BYTE_ORDER_NATIVE) if byte_order not in definitions.BYTE_ORDERS: error_message = 'unsupported byte-order attribute: {0!s}'.format( byte_order) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.byte_order = byte_order for member in members: section = member.get('section', None) if section: member_section_definition = data_types.MemberSectionDefinition(section) definition_object.AddSectionDefinition(member_section_definition) else: member_data_type_definition = self._ReadMemberDataTypeDefinitionMember( definitions_registry, member, definition_object.name, supports_conditions=supports_conditions) definition_object.AddMemberDefinition(member_data_type_definition) return definition_object
python
def _ReadDataTypeDefinitionWithMembers( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supports_conditions=False): """Reads a data type definition with members. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supports_conditions (Optional[bool]): True if conditions are supported by the data type definition. Returns: StringDefinition: string data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ members = definition_values.get('members', None) if not members: error_message = 'missing members' raise errors.DefinitionReaderError(definition_name, error_message) supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_STORAGE_DATA_TYPE_WITH_MEMBERS) definition_object = self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values) attributes = definition_values.get('attributes', None) if attributes: unsupported_attributes = set(attributes.keys()).difference( self._SUPPORTED_ATTRIBUTES_STORAGE_DATA_TYPE) if unsupported_attributes: error_message = 'unsupported attributes: {0:s}'.format( ', '.join(unsupported_attributes)) raise errors.DefinitionReaderError(definition_name, error_message) byte_order = attributes.get('byte_order', definitions.BYTE_ORDER_NATIVE) if byte_order not in definitions.BYTE_ORDERS: error_message = 'unsupported byte-order attribute: {0!s}'.format( byte_order) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.byte_order = byte_order for member in members: section = member.get('section', None) if section: member_section_definition = data_types.MemberSectionDefinition(section) definition_object.AddSectionDefinition(member_section_definition) else: member_data_type_definition = self._ReadMemberDataTypeDefinitionMember( definitions_registry, member, definition_object.name, supports_conditions=supports_conditions) definition_object.AddMemberDefinition(member_data_type_definition) return definition_object
[ "def", "_ReadDataTypeDefinitionWithMembers", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supports_conditions", "=", "False", ")", ":", "members", "=", "definition_values", ".", "get", "(", "'members'", ",", "None", ")", "if", "not", "members", ":", "error_message", "=", "'missing members'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_STORAGE_DATA_TYPE_WITH_MEMBERS", ")", "definition_object", "=", "self", ".", "_ReadDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")", "attributes", "=", "definition_values", ".", "get", "(", "'attributes'", ",", "None", ")", "if", "attributes", ":", "unsupported_attributes", "=", "set", "(", "attributes", ".", "keys", "(", ")", ")", ".", "difference", "(", "self", ".", "_SUPPORTED_ATTRIBUTES_STORAGE_DATA_TYPE", ")", "if", "unsupported_attributes", ":", "error_message", "=", "'unsupported attributes: {0:s}'", ".", "format", "(", "', '", ".", "join", "(", "unsupported_attributes", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "byte_order", "=", "attributes", ".", "get", "(", "'byte_order'", ",", "definitions", ".", "BYTE_ORDER_NATIVE", ")", "if", "byte_order", "not", "in", "definitions", ".", "BYTE_ORDERS", ":", "error_message", "=", "'unsupported byte-order attribute: {0!s}'", ".", "format", "(", "byte_order", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "byte_order", "=", "byte_order", "for", "member", "in", "members", ":", "section", "=", "member", ".", "get", "(", "'section'", ",", "None", ")", "if", "section", ":", "member_section_definition", "=", "data_types", ".", "MemberSectionDefinition", "(", "section", ")", "definition_object", ".", "AddSectionDefinition", "(", "member_section_definition", ")", "else", ":", "member_data_type_definition", "=", "self", ".", "_ReadMemberDataTypeDefinitionMember", "(", "definitions_registry", ",", "member", ",", "definition_object", ".", "name", ",", "supports_conditions", "=", "supports_conditions", ")", "definition_object", ".", "AddMemberDefinition", "(", "member_data_type_definition", ")", "return", "definition_object" ]
Reads a data type definition with members. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supports_conditions (Optional[bool]): True if conditions are supported by the data type definition. Returns: StringDefinition: string data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "data", "type", "definition", "with", "members", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L219-L279
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadEnumerationDataTypeDefinition
def _ReadEnumerationDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an enumeration data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: EnumerationDataTypeDefinition: enumeration data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) values = definition_values.get('values') if not values: error_message = 'missing values' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadSemanticDataTypeDefinition( definitions_registry, definition_values, data_types.EnumerationDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_ENUMERATION) last_name = None for enumeration_value in values: aliases = enumeration_value.get('aliases', None) description = enumeration_value.get('description', None) name = enumeration_value.get('name', None) number = enumeration_value.get('number', None) if not name or number is None: if last_name: error_location = 'after: {0:s}'.format(last_name) else: error_location = 'at start' error_message = '{0:s} missing name or number'.format(error_location) raise errors.DefinitionReaderError(definition_name, error_message) else: try: definition_object.AddValue( name, number, aliases=aliases, description=description) except KeyError as exception: error_message = '{0!s}'.format(exception) raise errors.DefinitionReaderError(definition_name, error_message) last_name = name return definition_object
python
def _ReadEnumerationDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an enumeration data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: EnumerationDataTypeDefinition: enumeration data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) values = definition_values.get('values') if not values: error_message = 'missing values' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadSemanticDataTypeDefinition( definitions_registry, definition_values, data_types.EnumerationDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_ENUMERATION) last_name = None for enumeration_value in values: aliases = enumeration_value.get('aliases', None) description = enumeration_value.get('description', None) name = enumeration_value.get('name', None) number = enumeration_value.get('number', None) if not name or number is None: if last_name: error_location = 'after: {0:s}'.format(last_name) else: error_location = 'at start' error_message = '{0:s} missing name or number'.format(error_location) raise errors.DefinitionReaderError(definition_name, error_message) else: try: definition_object.AddValue( name, number, aliases=aliases, description=description) except KeyError as exception: error_message = '{0!s}'.format(exception) raise errors.DefinitionReaderError(definition_name, error_message) last_name = name return definition_object
[ "def", "_ReadEnumerationDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "error_message", "=", "'data type not supported as member'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "values", "=", "definition_values", ".", "get", "(", "'values'", ")", "if", "not", "values", ":", "error_message", "=", "'missing values'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", "=", "self", ".", "_ReadSemanticDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "EnumerationDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_DEFINITION_VALUES_ENUMERATION", ")", "last_name", "=", "None", "for", "enumeration_value", "in", "values", ":", "aliases", "=", "enumeration_value", ".", "get", "(", "'aliases'", ",", "None", ")", "description", "=", "enumeration_value", ".", "get", "(", "'description'", ",", "None", ")", "name", "=", "enumeration_value", ".", "get", "(", "'name'", ",", "None", ")", "number", "=", "enumeration_value", ".", "get", "(", "'number'", ",", "None", ")", "if", "not", "name", "or", "number", "is", "None", ":", "if", "last_name", ":", "error_location", "=", "'after: {0:s}'", ".", "format", "(", "last_name", ")", "else", ":", "error_location", "=", "'at start'", "error_message", "=", "'{0:s} missing name or number'", ".", "format", "(", "error_location", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "else", ":", "try", ":", "definition_object", ".", "AddValue", "(", "name", ",", "number", ",", "aliases", "=", "aliases", ",", "description", "=", "description", ")", "except", "KeyError", "as", "exception", ":", "error_message", "=", "'{0!s}'", ".", "format", "(", "exception", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "last_name", "=", "name", "return", "definition_object" ]
Reads an enumeration data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: EnumerationDataTypeDefinition: enumeration data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "an", "enumeration", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L281-L341
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadElementSequenceDataTypeDefinition
def _ReadElementSequenceDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads an element sequence data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: SequenceDefinition: sequence data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ unsupported_definition_values = set(definition_values.keys()).difference( supported_definition_values) if unsupported_definition_values: error_message = 'unsupported definition values: {0:s}'.format( ', '.join(unsupported_definition_values)) raise errors.DefinitionReaderError(definition_name, error_message) element_data_type = definition_values.get('element_data_type', None) if not element_data_type: error_message = 'missing element data type' raise errors.DefinitionReaderError(definition_name, error_message) elements_data_size = definition_values.get('elements_data_size', None) elements_terminator = definition_values.get('elements_terminator', None) number_of_elements = definition_values.get('number_of_elements', None) size_values = (elements_data_size, elements_terminator, number_of_elements) size_values = [value for value in size_values if value is not None] if not size_values: error_message = ( 'missing element data size, elements terminator and number of ' 'elements') raise errors.DefinitionReaderError(definition_name, error_message) if len(size_values) > 1: error_message = ( 'element data size, elements terminator and number of elements ' 'not allowed to be set at the same time') raise errors.DefinitionReaderError(definition_name, error_message) element_data_type_definition = definitions_registry.GetDefinitionByName( element_data_type) if not element_data_type_definition: error_message = 'undefined element data type: {0:s}.'.format( element_data_type) raise errors.DefinitionReaderError(definition_name, error_message) element_byte_size = element_data_type_definition.GetByteSize() element_type_indicator = element_data_type_definition.TYPE_INDICATOR if not element_byte_size and element_type_indicator != ( definitions.TYPE_INDICATOR_STRING): error_message = ( 'unsupported variable size element data type: {0:s}'.format( element_data_type)) raise errors.DefinitionReaderError(definition_name, error_message) aliases = definition_values.get('aliases', None) description = definition_values.get('description', None) urls = definition_values.get('urls', None) definition_object = data_type_definition_class( definition_name, element_data_type_definition, aliases=aliases, data_type=element_data_type, description=description, urls=urls) if elements_data_size is not None: try: definition_object.elements_data_size = int(elements_data_size) except ValueError: definition_object.elements_data_size_expression = elements_data_size elif elements_terminator is not None: if isinstance(elements_terminator, py2to3.UNICODE_TYPE): elements_terminator = elements_terminator.encode('ascii') definition_object.elements_terminator = elements_terminator elif number_of_elements is not None: try: definition_object.number_of_elements = int(number_of_elements) except ValueError: definition_object.number_of_elements_expression = number_of_elements return definition_object
python
def _ReadElementSequenceDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads an element sequence data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: SequenceDefinition: sequence data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ unsupported_definition_values = set(definition_values.keys()).difference( supported_definition_values) if unsupported_definition_values: error_message = 'unsupported definition values: {0:s}'.format( ', '.join(unsupported_definition_values)) raise errors.DefinitionReaderError(definition_name, error_message) element_data_type = definition_values.get('element_data_type', None) if not element_data_type: error_message = 'missing element data type' raise errors.DefinitionReaderError(definition_name, error_message) elements_data_size = definition_values.get('elements_data_size', None) elements_terminator = definition_values.get('elements_terminator', None) number_of_elements = definition_values.get('number_of_elements', None) size_values = (elements_data_size, elements_terminator, number_of_elements) size_values = [value for value in size_values if value is not None] if not size_values: error_message = ( 'missing element data size, elements terminator and number of ' 'elements') raise errors.DefinitionReaderError(definition_name, error_message) if len(size_values) > 1: error_message = ( 'element data size, elements terminator and number of elements ' 'not allowed to be set at the same time') raise errors.DefinitionReaderError(definition_name, error_message) element_data_type_definition = definitions_registry.GetDefinitionByName( element_data_type) if not element_data_type_definition: error_message = 'undefined element data type: {0:s}.'.format( element_data_type) raise errors.DefinitionReaderError(definition_name, error_message) element_byte_size = element_data_type_definition.GetByteSize() element_type_indicator = element_data_type_definition.TYPE_INDICATOR if not element_byte_size and element_type_indicator != ( definitions.TYPE_INDICATOR_STRING): error_message = ( 'unsupported variable size element data type: {0:s}'.format( element_data_type)) raise errors.DefinitionReaderError(definition_name, error_message) aliases = definition_values.get('aliases', None) description = definition_values.get('description', None) urls = definition_values.get('urls', None) definition_object = data_type_definition_class( definition_name, element_data_type_definition, aliases=aliases, data_type=element_data_type, description=description, urls=urls) if elements_data_size is not None: try: definition_object.elements_data_size = int(elements_data_size) except ValueError: definition_object.elements_data_size_expression = elements_data_size elif elements_terminator is not None: if isinstance(elements_terminator, py2to3.UNICODE_TYPE): elements_terminator = elements_terminator.encode('ascii') definition_object.elements_terminator = elements_terminator elif number_of_elements is not None: try: definition_object.number_of_elements = int(number_of_elements) except ValueError: definition_object.number_of_elements_expression = number_of_elements return definition_object
[ "def", "_ReadElementSequenceDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")", ":", "unsupported_definition_values", "=", "set", "(", "definition_values", ".", "keys", "(", ")", ")", ".", "difference", "(", "supported_definition_values", ")", "if", "unsupported_definition_values", ":", "error_message", "=", "'unsupported definition values: {0:s}'", ".", "format", "(", "', '", ".", "join", "(", "unsupported_definition_values", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "element_data_type", "=", "definition_values", ".", "get", "(", "'element_data_type'", ",", "None", ")", "if", "not", "element_data_type", ":", "error_message", "=", "'missing element data type'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "elements_data_size", "=", "definition_values", ".", "get", "(", "'elements_data_size'", ",", "None", ")", "elements_terminator", "=", "definition_values", ".", "get", "(", "'elements_terminator'", ",", "None", ")", "number_of_elements", "=", "definition_values", ".", "get", "(", "'number_of_elements'", ",", "None", ")", "size_values", "=", "(", "elements_data_size", ",", "elements_terminator", ",", "number_of_elements", ")", "size_values", "=", "[", "value", "for", "value", "in", "size_values", "if", "value", "is", "not", "None", "]", "if", "not", "size_values", ":", "error_message", "=", "(", "'missing element data size, elements terminator and number of '", "'elements'", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "len", "(", "size_values", ")", ">", "1", ":", "error_message", "=", "(", "'element data size, elements terminator and number of elements '", "'not allowed to be set at the same time'", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "element_data_type_definition", "=", "definitions_registry", ".", "GetDefinitionByName", "(", "element_data_type", ")", "if", "not", "element_data_type_definition", ":", "error_message", "=", "'undefined element data type: {0:s}.'", ".", "format", "(", "element_data_type", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "element_byte_size", "=", "element_data_type_definition", ".", "GetByteSize", "(", ")", "element_type_indicator", "=", "element_data_type_definition", ".", "TYPE_INDICATOR", "if", "not", "element_byte_size", "and", "element_type_indicator", "!=", "(", "definitions", ".", "TYPE_INDICATOR_STRING", ")", ":", "error_message", "=", "(", "'unsupported variable size element data type: {0:s}'", ".", "format", "(", "element_data_type", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "aliases", "=", "definition_values", ".", "get", "(", "'aliases'", ",", "None", ")", "description", "=", "definition_values", ".", "get", "(", "'description'", ",", "None", ")", "urls", "=", "definition_values", ".", "get", "(", "'urls'", ",", "None", ")", "definition_object", "=", "data_type_definition_class", "(", "definition_name", ",", "element_data_type_definition", ",", "aliases", "=", "aliases", ",", "data_type", "=", "element_data_type", ",", "description", "=", "description", ",", "urls", "=", "urls", ")", "if", "elements_data_size", "is", "not", "None", ":", "try", ":", "definition_object", ".", "elements_data_size", "=", "int", "(", "elements_data_size", ")", "except", "ValueError", ":", "definition_object", ".", "elements_data_size_expression", "=", "elements_data_size", "elif", "elements_terminator", "is", "not", "None", ":", "if", "isinstance", "(", "elements_terminator", ",", "py2to3", ".", "UNICODE_TYPE", ")", ":", "elements_terminator", "=", "elements_terminator", ".", "encode", "(", "'ascii'", ")", "definition_object", ".", "elements_terminator", "=", "elements_terminator", "elif", "number_of_elements", "is", "not", "None", ":", "try", ":", "definition_object", ".", "number_of_elements", "=", "int", "(", "number_of_elements", ")", "except", "ValueError", ":", "definition_object", ".", "number_of_elements_expression", "=", "number_of_elements", "return", "definition_object" ]
Reads an element sequence data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: SequenceDefinition: sequence data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "an", "element", "sequence", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L343-L437
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadFixedSizeDataTypeDefinition
def _ReadFixedSizeDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_attributes, default_size=definitions.SIZE_NATIVE, default_units='bytes', is_member=False, supported_size_values=None): """Reads a fixed-size data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_attributes (set[str]): names of the supported attributes. default_size (Optional[int]): default size. default_units (Optional[str]): default units. is_member (Optional[bool]): True if the data type definition is a member data type definition. supported_size_values (Optional[tuple[int]]): supported size values, or None if not set. Returns: FixedSizeDataTypeDefinition: fixed-size data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ definition_object = self._ReadStorageDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_attributes, is_member=is_member) attributes = definition_values.get('attributes', None) if attributes: size = attributes.get('size', default_size) if size != definitions.SIZE_NATIVE: try: int(size) except ValueError: error_message = 'unuspported size attribute: {0!s}'.format(size) raise errors.DefinitionReaderError(definition_name, error_message) if supported_size_values and size not in supported_size_values: error_message = 'unuspported size value: {0!s}'.format(size) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.size = size definition_object.units = attributes.get('units', default_units) return definition_object
python
def _ReadFixedSizeDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_attributes, default_size=definitions.SIZE_NATIVE, default_units='bytes', is_member=False, supported_size_values=None): """Reads a fixed-size data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_attributes (set[str]): names of the supported attributes. default_size (Optional[int]): default size. default_units (Optional[str]): default units. is_member (Optional[bool]): True if the data type definition is a member data type definition. supported_size_values (Optional[tuple[int]]): supported size values, or None if not set. Returns: FixedSizeDataTypeDefinition: fixed-size data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ definition_object = self._ReadStorageDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_attributes, is_member=is_member) attributes = definition_values.get('attributes', None) if attributes: size = attributes.get('size', default_size) if size != definitions.SIZE_NATIVE: try: int(size) except ValueError: error_message = 'unuspported size attribute: {0!s}'.format(size) raise errors.DefinitionReaderError(definition_name, error_message) if supported_size_values and size not in supported_size_values: error_message = 'unuspported size value: {0!s}'.format(size) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.size = size definition_object.units = attributes.get('units', default_units) return definition_object
[ "def", "_ReadFixedSizeDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_attributes", ",", "default_size", "=", "definitions", ".", "SIZE_NATIVE", ",", "default_units", "=", "'bytes'", ",", "is_member", "=", "False", ",", "supported_size_values", "=", "None", ")", ":", "definition_object", "=", "self", ".", "_ReadStorageDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_attributes", ",", "is_member", "=", "is_member", ")", "attributes", "=", "definition_values", ".", "get", "(", "'attributes'", ",", "None", ")", "if", "attributes", ":", "size", "=", "attributes", ".", "get", "(", "'size'", ",", "default_size", ")", "if", "size", "!=", "definitions", ".", "SIZE_NATIVE", ":", "try", ":", "int", "(", "size", ")", "except", "ValueError", ":", "error_message", "=", "'unuspported size attribute: {0!s}'", ".", "format", "(", "size", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "supported_size_values", "and", "size", "not", "in", "supported_size_values", ":", "error_message", "=", "'unuspported size value: {0!s}'", ".", "format", "(", "size", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "size", "=", "size", "definition_object", ".", "units", "=", "attributes", ".", "get", "(", "'units'", ",", "default_units", ")", "return", "definition_object" ]
Reads a fixed-size data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_attributes (set[str]): names of the supported attributes. default_size (Optional[int]): default size. default_units (Optional[str]): default units. is_member (Optional[bool]): True if the data type definition is a member data type definition. supported_size_values (Optional[tuple[int]]): supported size values, or None if not set. Returns: FixedSizeDataTypeDefinition: fixed-size data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "fixed", "-", "size", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L439-L488
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadFloatingPointDataTypeDefinition
def _ReadFloatingPointDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a floating-point data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: FloatingPointDefinition: floating-point data type definition. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.FloatingPointDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE, is_member=is_member, supported_size_values=(4, 8))
python
def _ReadFloatingPointDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a floating-point data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: FloatingPointDefinition: floating-point data type definition. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.FloatingPointDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE, is_member=is_member, supported_size_values=(4, 8))
[ "def", "_ReadFloatingPointDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "return", "self", ".", "_ReadFixedSizeDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "FloatingPointDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE", ",", "is_member", "=", "is_member", ",", "supported_size_values", "=", "(", "4", ",", "8", ")", ")" ]
Reads a floating-point data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: FloatingPointDefinition: floating-point data type definition.
[ "Reads", "a", "floating", "-", "point", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L490-L510
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadFormatDataTypeDefinition
def _ReadFormatDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a format data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: FormatDefinition: format definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadLayoutDataTypeDefinition( definitions_registry, definition_values, data_types.FormatDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_FORMAT) # TODO: disabled for now # layout = definition_values.get('layout', None) # if layout is None: # error_message = 'missing layout' # raise errors.DefinitionReaderError(definition_name, error_message) definition_object.metadata = definition_values.get('metadata', {}) attributes = definition_values.get('attributes', None) if attributes: unsupported_attributes = set(attributes.keys()).difference( self._SUPPORTED_ATTRIBUTES_FORMAT) if unsupported_attributes: error_message = 'unsupported attributes: {0:s}'.format( ', '.join(unsupported_attributes)) raise errors.DefinitionReaderError(definition_name, error_message) byte_order = attributes.get('byte_order', definitions.BYTE_ORDER_NATIVE) if byte_order not in definitions.BYTE_ORDERS: error_message = 'unsupported byte-order attribute: {0!s}'.format( byte_order) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.byte_order = byte_order return definition_object
python
def _ReadFormatDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a format data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: FormatDefinition: format definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadLayoutDataTypeDefinition( definitions_registry, definition_values, data_types.FormatDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_FORMAT) # TODO: disabled for now # layout = definition_values.get('layout', None) # if layout is None: # error_message = 'missing layout' # raise errors.DefinitionReaderError(definition_name, error_message) definition_object.metadata = definition_values.get('metadata', {}) attributes = definition_values.get('attributes', None) if attributes: unsupported_attributes = set(attributes.keys()).difference( self._SUPPORTED_ATTRIBUTES_FORMAT) if unsupported_attributes: error_message = 'unsupported attributes: {0:s}'.format( ', '.join(unsupported_attributes)) raise errors.DefinitionReaderError(definition_name, error_message) byte_order = attributes.get('byte_order', definitions.BYTE_ORDER_NATIVE) if byte_order not in definitions.BYTE_ORDERS: error_message = 'unsupported byte-order attribute: {0!s}'.format( byte_order) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.byte_order = byte_order return definition_object
[ "def", "_ReadFormatDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "error_message", "=", "'data type not supported as member'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", "=", "self", ".", "_ReadLayoutDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "FormatDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_DEFINITION_VALUES_FORMAT", ")", "# TODO: disabled for now", "# layout = definition_values.get('layout', None)", "# if layout is None:", "# error_message = 'missing layout'", "# raise errors.DefinitionReaderError(definition_name, error_message)", "definition_object", ".", "metadata", "=", "definition_values", ".", "get", "(", "'metadata'", ",", "{", "}", ")", "attributes", "=", "definition_values", ".", "get", "(", "'attributes'", ",", "None", ")", "if", "attributes", ":", "unsupported_attributes", "=", "set", "(", "attributes", ".", "keys", "(", ")", ")", ".", "difference", "(", "self", ".", "_SUPPORTED_ATTRIBUTES_FORMAT", ")", "if", "unsupported_attributes", ":", "error_message", "=", "'unsupported attributes: {0:s}'", ".", "format", "(", "', '", ".", "join", "(", "unsupported_attributes", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "byte_order", "=", "attributes", ".", "get", "(", "'byte_order'", ",", "definitions", ".", "BYTE_ORDER_NATIVE", ")", "if", "byte_order", "not", "in", "definitions", ".", "BYTE_ORDERS", ":", "error_message", "=", "'unsupported byte-order attribute: {0!s}'", ".", "format", "(", "byte_order", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "byte_order", "=", "byte_order", "return", "definition_object" ]
Reads a format data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: FormatDefinition: format definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "format", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L512-L565
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadIntegerDataTypeDefinition
def _ReadIntegerDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an integer data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: IntegerDataTypeDefinition: integer data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ definition_object = self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.IntegerDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_INTEGER, is_member=is_member, supported_size_values=(1, 2, 4, 8)) attributes = definition_values.get('attributes', None) if attributes: format_attribute = attributes.get('format', definitions.FORMAT_SIGNED) if format_attribute not in self._INTEGER_FORMAT_ATTRIBUTES: error_message = 'unsupported format attribute: {0!s}'.format( format_attribute) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.format = format_attribute return definition_object
python
def _ReadIntegerDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an integer data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: IntegerDataTypeDefinition: integer data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ definition_object = self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.IntegerDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_INTEGER, is_member=is_member, supported_size_values=(1, 2, 4, 8)) attributes = definition_values.get('attributes', None) if attributes: format_attribute = attributes.get('format', definitions.FORMAT_SIGNED) if format_attribute not in self._INTEGER_FORMAT_ATTRIBUTES: error_message = 'unsupported format attribute: {0!s}'.format( format_attribute) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.format = format_attribute return definition_object
[ "def", "_ReadIntegerDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "definition_object", "=", "self", ".", "_ReadFixedSizeDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "IntegerDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_ATTRIBUTES_INTEGER", ",", "is_member", "=", "is_member", ",", "supported_size_values", "=", "(", "1", ",", "2", ",", "4", ",", "8", ")", ")", "attributes", "=", "definition_values", ".", "get", "(", "'attributes'", ",", "None", ")", "if", "attributes", ":", "format_attribute", "=", "attributes", ".", "get", "(", "'format'", ",", "definitions", ".", "FORMAT_SIGNED", ")", "if", "format_attribute", "not", "in", "self", ".", "_INTEGER_FORMAT_ATTRIBUTES", ":", "error_message", "=", "'unsupported format attribute: {0!s}'", ".", "format", "(", "format_attribute", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "format", "=", "format_attribute", "return", "definition_object" ]
Reads an integer data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: IntegerDataTypeDefinition: integer data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "an", "integer", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L567-L603
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadLayoutDataTypeDefinition
def _ReadLayoutDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads a layout data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: LayoutDataTypeDefinition: layout data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values)
python
def _ReadLayoutDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads a layout data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: LayoutDataTypeDefinition: layout data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values)
[ "def", "_ReadLayoutDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")", ":", "return", "self", ".", "_ReadDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")" ]
Reads a layout data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: LayoutDataTypeDefinition: layout data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "layout", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L605-L628
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadMemberDataTypeDefinitionMember
def _ReadMemberDataTypeDefinitionMember( self, definitions_registry, definition_values, definition_name, supports_conditions=False): """Reads a member data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. supports_conditions (Optional[bool]): True if conditions are supported by the data type definition. Returns: DataTypeDefinition: structure member data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if not definition_values: error_message = 'invalid structure member missing definition values' raise errors.DefinitionReaderError(definition_name, error_message) name = definition_values.get('name', None) type_indicator = definition_values.get('type', None) if not name and type_indicator != definitions.TYPE_INDICATOR_UNION: error_message = 'invalid structure member missing name' raise errors.DefinitionReaderError(definition_name, error_message) # TODO: detect duplicate names. data_type = definition_values.get('data_type', None) type_values = (data_type, type_indicator) type_values = [value for value in type_values if value is not None] if not type_values: error_message = ( 'invalid structure member: {0:s} both data type and type are ' 'missing').format(name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) if len(type_values) > 1: error_message = ( 'invalid structure member: {0:s} data type and type not allowed to ' 'be set at the same time').format(name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) condition = definition_values.get('condition', None) if not supports_conditions and condition: error_message = ( 'invalid structure member: {0:s} unsupported condition').format( name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) value = definition_values.get('value', None) values = definition_values.get('values', None) if None not in (value, values): error_message = ( 'invalid structure member: {0:s} value and values not allowed to ' 'be set at the same time').format(name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) if value: values = [value] supported_values = None if values: supported_values = [] for value in values: if isinstance(value, py2to3.UNICODE_TYPE): value = value.encode('ascii') supported_values.append(value) if type_indicator is not None: data_type_callback = self._DATA_TYPE_CALLBACKS.get(type_indicator, None) if data_type_callback: data_type_callback = getattr(self, data_type_callback, None) if not data_type_callback: error_message = 'unuspported data type definition: {0:s}.'.format( type_indicator) raise errors.DefinitionReaderError(name, error_message) try: data_type_definition = data_type_callback( definitions_registry, definition_values, name, is_member=True) except errors.DefinitionReaderError as exception: error_message = 'in: {0:s} {1:s}'.format( exception.name or '<NAMELESS>', exception.message) raise errors.DefinitionReaderError(definition_name, error_message) if condition or supported_values: definition_object = data_types.MemberDataTypeDefinition( name, data_type_definition, condition=condition, values=supported_values) else: definition_object = data_type_definition elif data_type is not None: data_type_definition = definitions_registry.GetDefinitionByName( data_type) if not data_type_definition: error_message = ( 'invalid structure member: {0:s} undefined data type: ' '{1:s}').format(name or '<NAMELESS>', data_type) raise errors.DefinitionReaderError(definition_name, error_message) unsupported_definition_values = set(definition_values.keys()).difference( self._SUPPORTED_DEFINITION_VALUES_MEMBER_DATA_TYPE) if unsupported_definition_values: error_message = 'unsupported definition values: {0:s}'.format( ', '.join(unsupported_definition_values)) raise errors.DefinitionReaderError(definition_name, error_message) aliases = definition_values.get('aliases', None) description = definition_values.get('description', None) definition_object = data_types.MemberDataTypeDefinition( name, data_type_definition, aliases=aliases, condition=condition, data_type=data_type, description=description, values=supported_values) return definition_object
python
def _ReadMemberDataTypeDefinitionMember( self, definitions_registry, definition_values, definition_name, supports_conditions=False): """Reads a member data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. supports_conditions (Optional[bool]): True if conditions are supported by the data type definition. Returns: DataTypeDefinition: structure member data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if not definition_values: error_message = 'invalid structure member missing definition values' raise errors.DefinitionReaderError(definition_name, error_message) name = definition_values.get('name', None) type_indicator = definition_values.get('type', None) if not name and type_indicator != definitions.TYPE_INDICATOR_UNION: error_message = 'invalid structure member missing name' raise errors.DefinitionReaderError(definition_name, error_message) # TODO: detect duplicate names. data_type = definition_values.get('data_type', None) type_values = (data_type, type_indicator) type_values = [value for value in type_values if value is not None] if not type_values: error_message = ( 'invalid structure member: {0:s} both data type and type are ' 'missing').format(name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) if len(type_values) > 1: error_message = ( 'invalid structure member: {0:s} data type and type not allowed to ' 'be set at the same time').format(name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) condition = definition_values.get('condition', None) if not supports_conditions and condition: error_message = ( 'invalid structure member: {0:s} unsupported condition').format( name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) value = definition_values.get('value', None) values = definition_values.get('values', None) if None not in (value, values): error_message = ( 'invalid structure member: {0:s} value and values not allowed to ' 'be set at the same time').format(name or '<NAMELESS>') raise errors.DefinitionReaderError(definition_name, error_message) if value: values = [value] supported_values = None if values: supported_values = [] for value in values: if isinstance(value, py2to3.UNICODE_TYPE): value = value.encode('ascii') supported_values.append(value) if type_indicator is not None: data_type_callback = self._DATA_TYPE_CALLBACKS.get(type_indicator, None) if data_type_callback: data_type_callback = getattr(self, data_type_callback, None) if not data_type_callback: error_message = 'unuspported data type definition: {0:s}.'.format( type_indicator) raise errors.DefinitionReaderError(name, error_message) try: data_type_definition = data_type_callback( definitions_registry, definition_values, name, is_member=True) except errors.DefinitionReaderError as exception: error_message = 'in: {0:s} {1:s}'.format( exception.name or '<NAMELESS>', exception.message) raise errors.DefinitionReaderError(definition_name, error_message) if condition or supported_values: definition_object = data_types.MemberDataTypeDefinition( name, data_type_definition, condition=condition, values=supported_values) else: definition_object = data_type_definition elif data_type is not None: data_type_definition = definitions_registry.GetDefinitionByName( data_type) if not data_type_definition: error_message = ( 'invalid structure member: {0:s} undefined data type: ' '{1:s}').format(name or '<NAMELESS>', data_type) raise errors.DefinitionReaderError(definition_name, error_message) unsupported_definition_values = set(definition_values.keys()).difference( self._SUPPORTED_DEFINITION_VALUES_MEMBER_DATA_TYPE) if unsupported_definition_values: error_message = 'unsupported definition values: {0:s}'.format( ', '.join(unsupported_definition_values)) raise errors.DefinitionReaderError(definition_name, error_message) aliases = definition_values.get('aliases', None) description = definition_values.get('description', None) definition_object = data_types.MemberDataTypeDefinition( name, data_type_definition, aliases=aliases, condition=condition, data_type=data_type, description=description, values=supported_values) return definition_object
[ "def", "_ReadMemberDataTypeDefinitionMember", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "supports_conditions", "=", "False", ")", ":", "if", "not", "definition_values", ":", "error_message", "=", "'invalid structure member missing definition values'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "name", "=", "definition_values", ".", "get", "(", "'name'", ",", "None", ")", "type_indicator", "=", "definition_values", ".", "get", "(", "'type'", ",", "None", ")", "if", "not", "name", "and", "type_indicator", "!=", "definitions", ".", "TYPE_INDICATOR_UNION", ":", "error_message", "=", "'invalid structure member missing name'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "# TODO: detect duplicate names.", "data_type", "=", "definition_values", ".", "get", "(", "'data_type'", ",", "None", ")", "type_values", "=", "(", "data_type", ",", "type_indicator", ")", "type_values", "=", "[", "value", "for", "value", "in", "type_values", "if", "value", "is", "not", "None", "]", "if", "not", "type_values", ":", "error_message", "=", "(", "'invalid structure member: {0:s} both data type and type are '", "'missing'", ")", ".", "format", "(", "name", "or", "'<NAMELESS>'", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "len", "(", "type_values", ")", ">", "1", ":", "error_message", "=", "(", "'invalid structure member: {0:s} data type and type not allowed to '", "'be set at the same time'", ")", ".", "format", "(", "name", "or", "'<NAMELESS>'", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "condition", "=", "definition_values", ".", "get", "(", "'condition'", ",", "None", ")", "if", "not", "supports_conditions", "and", "condition", ":", "error_message", "=", "(", "'invalid structure member: {0:s} unsupported condition'", ")", ".", "format", "(", "name", "or", "'<NAMELESS>'", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "value", "=", "definition_values", ".", "get", "(", "'value'", ",", "None", ")", "values", "=", "definition_values", ".", "get", "(", "'values'", ",", "None", ")", "if", "None", "not", "in", "(", "value", ",", "values", ")", ":", "error_message", "=", "(", "'invalid structure member: {0:s} value and values not allowed to '", "'be set at the same time'", ")", ".", "format", "(", "name", "or", "'<NAMELESS>'", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "value", ":", "values", "=", "[", "value", "]", "supported_values", "=", "None", "if", "values", ":", "supported_values", "=", "[", "]", "for", "value", "in", "values", ":", "if", "isinstance", "(", "value", ",", "py2to3", ".", "UNICODE_TYPE", ")", ":", "value", "=", "value", ".", "encode", "(", "'ascii'", ")", "supported_values", ".", "append", "(", "value", ")", "if", "type_indicator", "is", "not", "None", ":", "data_type_callback", "=", "self", ".", "_DATA_TYPE_CALLBACKS", ".", "get", "(", "type_indicator", ",", "None", ")", "if", "data_type_callback", ":", "data_type_callback", "=", "getattr", "(", "self", ",", "data_type_callback", ",", "None", ")", "if", "not", "data_type_callback", ":", "error_message", "=", "'unuspported data type definition: {0:s}.'", ".", "format", "(", "type_indicator", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "name", ",", "error_message", ")", "try", ":", "data_type_definition", "=", "data_type_callback", "(", "definitions_registry", ",", "definition_values", ",", "name", ",", "is_member", "=", "True", ")", "except", "errors", ".", "DefinitionReaderError", "as", "exception", ":", "error_message", "=", "'in: {0:s} {1:s}'", ".", "format", "(", "exception", ".", "name", "or", "'<NAMELESS>'", ",", "exception", ".", "message", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "condition", "or", "supported_values", ":", "definition_object", "=", "data_types", ".", "MemberDataTypeDefinition", "(", "name", ",", "data_type_definition", ",", "condition", "=", "condition", ",", "values", "=", "supported_values", ")", "else", ":", "definition_object", "=", "data_type_definition", "elif", "data_type", "is", "not", "None", ":", "data_type_definition", "=", "definitions_registry", ".", "GetDefinitionByName", "(", "data_type", ")", "if", "not", "data_type_definition", ":", "error_message", "=", "(", "'invalid structure member: {0:s} undefined data type: '", "'{1:s}'", ")", ".", "format", "(", "name", "or", "'<NAMELESS>'", ",", "data_type", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "unsupported_definition_values", "=", "set", "(", "definition_values", ".", "keys", "(", ")", ")", ".", "difference", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_MEMBER_DATA_TYPE", ")", "if", "unsupported_definition_values", ":", "error_message", "=", "'unsupported definition values: {0:s}'", ".", "format", "(", "', '", ".", "join", "(", "unsupported_definition_values", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "aliases", "=", "definition_values", ".", "get", "(", "'aliases'", ",", "None", ")", "description", "=", "definition_values", ".", "get", "(", "'description'", ",", "None", ")", "definition_object", "=", "data_types", ".", "MemberDataTypeDefinition", "(", "name", ",", "data_type_definition", ",", "aliases", "=", "aliases", ",", "condition", "=", "condition", ",", "data_type", "=", "data_type", ",", "description", "=", "description", ",", "values", "=", "supported_values", ")", "return", "definition_object" ]
Reads a member data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. supports_conditions (Optional[bool]): True if conditions are supported by the data type definition. Returns: DataTypeDefinition: structure member data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "member", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L630-L755
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadPaddingDataTypeDefinition
def _ReadPaddingDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a padding data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: PaddingtDefinition: padding definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if not is_member: error_message = 'data type only supported as member' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadDataTypeDefinition( definitions_registry, definition_values, data_types.PaddingDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_PADDING) alignment_size = definition_values.get('alignment_size', None) if not alignment_size: error_message = 'missing alignment_size' raise errors.DefinitionReaderError(definition_name, error_message) try: int(alignment_size) except ValueError: error_message = 'unuspported alignment size attribute: {0!s}'.format( alignment_size) raise errors.DefinitionReaderError(definition_name, error_message) if alignment_size not in (2, 4, 8, 16): error_message = 'unuspported alignment size value: {0!s}'.format( alignment_size) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.alignment_size = alignment_size return definition_object
python
def _ReadPaddingDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a padding data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: PaddingtDefinition: padding definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if not is_member: error_message = 'data type only supported as member' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadDataTypeDefinition( definitions_registry, definition_values, data_types.PaddingDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_PADDING) alignment_size = definition_values.get('alignment_size', None) if not alignment_size: error_message = 'missing alignment_size' raise errors.DefinitionReaderError(definition_name, error_message) try: int(alignment_size) except ValueError: error_message = 'unuspported alignment size attribute: {0!s}'.format( alignment_size) raise errors.DefinitionReaderError(definition_name, error_message) if alignment_size not in (2, 4, 8, 16): error_message = 'unuspported alignment size value: {0!s}'.format( alignment_size) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.alignment_size = alignment_size return definition_object
[ "def", "_ReadPaddingDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "not", "is_member", ":", "error_message", "=", "'data type only supported as member'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", "=", "self", ".", "_ReadDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "PaddingDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_DEFINITION_VALUES_PADDING", ")", "alignment_size", "=", "definition_values", ".", "get", "(", "'alignment_size'", ",", "None", ")", "if", "not", "alignment_size", ":", "error_message", "=", "'missing alignment_size'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "try", ":", "int", "(", "alignment_size", ")", "except", "ValueError", ":", "error_message", "=", "'unuspported alignment size attribute: {0!s}'", ".", "format", "(", "alignment_size", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "alignment_size", "not", "in", "(", "2", ",", "4", ",", "8", ",", "16", ")", ":", "error_message", "=", "'unuspported alignment size value: {0!s}'", ".", "format", "(", "alignment_size", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "alignment_size", "=", "alignment_size", "return", "definition_object" ]
Reads a padding data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: PaddingtDefinition: padding definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "padding", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L757-L804
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadSemanticDataTypeDefinition
def _ReadSemanticDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads a semantic data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: SemanticDataTypeDefinition: semantic data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values)
python
def _ReadSemanticDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values): """Reads a semantic data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: SemanticDataTypeDefinition: semantic data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values)
[ "def", "_ReadSemanticDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")", ":", "return", "self", ".", "_ReadDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")" ]
Reads a semantic data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_definition_values (set[str]): names of the supported definition values. Returns: SemanticDataTypeDefinition: semantic data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "semantic", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L806-L829
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadSequenceDataTypeDefinition
def _ReadSequenceDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a sequence data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: SequenceDefinition: sequence data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_MEMBER_DATA_TYPE) else: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_DATA_TYPE) return self._ReadElementSequenceDataTypeDefinition( definitions_registry, definition_values, data_types.SequenceDefinition, definition_name, supported_definition_values)
python
def _ReadSequenceDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a sequence data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: SequenceDefinition: sequence data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_MEMBER_DATA_TYPE) else: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_DATA_TYPE) return self._ReadElementSequenceDataTypeDefinition( definitions_registry, definition_values, data_types.SequenceDefinition, definition_name, supported_definition_values)
[ "def", "_ReadSequenceDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_ELEMENTS_MEMBER_DATA_TYPE", ")", "else", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_ELEMENTS_DATA_TYPE", ")", "return", "self", ".", "_ReadElementSequenceDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "SequenceDefinition", ",", "definition_name", ",", "supported_definition_values", ")" ]
Reads a sequence data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: SequenceDefinition: sequence data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "sequence", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L831-L860
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadStorageDataTypeDefinition
def _ReadStorageDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_attributes, is_member=False): """Reads a storage data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_attributes (set[str]): names of the supported attributes. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StorageDataTypeDefinition: storage data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_MEMBER_DATA_TYPE) else: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_STORAGE_DATA_TYPE) definition_object = self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values) attributes = definition_values.get('attributes', None) if attributes: unsupported_attributes = set(attributes.keys()).difference( supported_attributes) if unsupported_attributes: error_message = 'unsupported attributes: {0:s}'.format( ', '.join(unsupported_attributes)) raise errors.DefinitionReaderError(definition_name, error_message) byte_order = attributes.get('byte_order', definitions.BYTE_ORDER_NATIVE) if byte_order not in definitions.BYTE_ORDERS: error_message = 'unsupported byte-order attribute: {0!s}'.format( byte_order) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.byte_order = byte_order return definition_object
python
def _ReadStorageDataTypeDefinition( self, definitions_registry, definition_values, data_type_definition_class, definition_name, supported_attributes, is_member=False): """Reads a storage data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_attributes (set[str]): names of the supported attributes. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StorageDataTypeDefinition: storage data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_MEMBER_DATA_TYPE) else: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_STORAGE_DATA_TYPE) definition_object = self._ReadDataTypeDefinition( definitions_registry, definition_values, data_type_definition_class, definition_name, supported_definition_values) attributes = definition_values.get('attributes', None) if attributes: unsupported_attributes = set(attributes.keys()).difference( supported_attributes) if unsupported_attributes: error_message = 'unsupported attributes: {0:s}'.format( ', '.join(unsupported_attributes)) raise errors.DefinitionReaderError(definition_name, error_message) byte_order = attributes.get('byte_order', definitions.BYTE_ORDER_NATIVE) if byte_order not in definitions.BYTE_ORDERS: error_message = 'unsupported byte-order attribute: {0!s}'.format( byte_order) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.byte_order = byte_order return definition_object
[ "def", "_ReadStorageDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_attributes", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_MEMBER_DATA_TYPE", ")", "else", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_STORAGE_DATA_TYPE", ")", "definition_object", "=", "self", ".", "_ReadDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_type_definition_class", ",", "definition_name", ",", "supported_definition_values", ")", "attributes", "=", "definition_values", ".", "get", "(", "'attributes'", ",", "None", ")", "if", "attributes", ":", "unsupported_attributes", "=", "set", "(", "attributes", ".", "keys", "(", ")", ")", ".", "difference", "(", "supported_attributes", ")", "if", "unsupported_attributes", ":", "error_message", "=", "'unsupported attributes: {0:s}'", ".", "format", "(", "', '", ".", "join", "(", "unsupported_attributes", ")", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "byte_order", "=", "attributes", ".", "get", "(", "'byte_order'", ",", "definitions", ".", "BYTE_ORDER_NATIVE", ")", "if", "byte_order", "not", "in", "definitions", ".", "BYTE_ORDERS", ":", "error_message", "=", "'unsupported byte-order attribute: {0!s}'", ".", "format", "(", "byte_order", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "byte_order", "=", "byte_order", "return", "definition_object" ]
Reads a storage data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. data_type_definition_class (str): data type definition class. definition_name (str): name of the definition. supported_attributes (set[str]): names of the supported attributes. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StorageDataTypeDefinition: storage data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "storage", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L862-L912
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadStreamDataTypeDefinition
def _ReadStreamDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a stream data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StreamDefinition: stream data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_MEMBER_DATA_TYPE) else: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_DATA_TYPE) return self._ReadElementSequenceDataTypeDefinition( definitions_registry, definition_values, data_types.StreamDefinition, definition_name, supported_definition_values)
python
def _ReadStreamDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a stream data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StreamDefinition: stream data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_MEMBER_DATA_TYPE) else: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_ELEMENTS_DATA_TYPE) return self._ReadElementSequenceDataTypeDefinition( definitions_registry, definition_values, data_types.StreamDefinition, definition_name, supported_definition_values)
[ "def", "_ReadStreamDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_ELEMENTS_MEMBER_DATA_TYPE", ")", "else", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_ELEMENTS_DATA_TYPE", ")", "return", "self", ".", "_ReadElementSequenceDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "StreamDefinition", ",", "definition_name", ",", "supported_definition_values", ")" ]
Reads a stream data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StreamDefinition: stream data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "stream", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L914-L943
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadStringDataTypeDefinition
def _ReadStringDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a string data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StringDefinition: string data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_STRING_MEMBER) else: supported_definition_values = self._SUPPORTED_DEFINITION_VALUES_STRING definition_object = self._ReadElementSequenceDataTypeDefinition( definitions_registry, definition_values, data_types.StringDefinition, definition_name, supported_definition_values) encoding = definition_values.get('encoding', None) if not encoding: error_message = 'missing encoding' raise errors.DefinitionReaderError(definition_name, error_message) definition_object.encoding = encoding return definition_object
python
def _ReadStringDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a string data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StringDefinition: string data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: supported_definition_values = ( self._SUPPORTED_DEFINITION_VALUES_STRING_MEMBER) else: supported_definition_values = self._SUPPORTED_DEFINITION_VALUES_STRING definition_object = self._ReadElementSequenceDataTypeDefinition( definitions_registry, definition_values, data_types.StringDefinition, definition_name, supported_definition_values) encoding = definition_values.get('encoding', None) if not encoding: error_message = 'missing encoding' raise errors.DefinitionReaderError(definition_name, error_message) definition_object.encoding = encoding return definition_object
[ "def", "_ReadStringDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "supported_definition_values", "=", "(", "self", ".", "_SUPPORTED_DEFINITION_VALUES_STRING_MEMBER", ")", "else", ":", "supported_definition_values", "=", "self", ".", "_SUPPORTED_DEFINITION_VALUES_STRING", "definition_object", "=", "self", ".", "_ReadElementSequenceDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "StringDefinition", ",", "definition_name", ",", "supported_definition_values", ")", "encoding", "=", "definition_values", ".", "get", "(", "'encoding'", ",", "None", ")", "if", "not", "encoding", ":", "error_message", "=", "'missing encoding'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "encoding", "=", "encoding", "return", "definition_object" ]
Reads a string data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StringDefinition: string data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "string", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L945-L982
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadStructureDataTypeDefinition
def _ReadStructureDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a structure data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StructureDefinition: structure data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) return self._ReadDataTypeDefinitionWithMembers( definitions_registry, definition_values, data_types.StructureDefinition, definition_name, supports_conditions=True)
python
def _ReadStructureDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a structure data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StructureDefinition: structure data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) return self._ReadDataTypeDefinitionWithMembers( definitions_registry, definition_values, data_types.StructureDefinition, definition_name, supports_conditions=True)
[ "def", "_ReadStructureDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "error_message", "=", "'data type not supported as member'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "return", "self", ".", "_ReadDataTypeDefinitionWithMembers", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "StructureDefinition", ",", "definition_name", ",", "supports_conditions", "=", "True", ")" ]
Reads a structure data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StructureDefinition: structure data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "structure", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L984-L1010
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadStructureFamilyDataTypeDefinition
def _ReadStructureFamilyDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a structure family data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StructureDefinition: structure data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadLayoutDataTypeDefinition( definitions_registry, definition_values, data_types.StructureFamilyDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_STRUCTURE_FAMILY) runtime = definition_values.get('runtime', None) if not runtime: error_message = 'missing runtime' raise errors.DefinitionReaderError(definition_name, error_message) runtime_data_type_definition = definitions_registry.GetDefinitionByName( runtime) if not runtime_data_type_definition: error_message = 'undefined runtime: {0:s}.'.format(runtime) raise errors.DefinitionReaderError(definition_name, error_message) if runtime_data_type_definition.family_definition: error_message = 'runtime: {0:s} already part of a family.'.format(runtime) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.AddRuntimeDefinition(runtime_data_type_definition) members = definition_values.get('members', None) if not members: error_message = 'missing members' raise errors.DefinitionReaderError(definition_name, error_message) for member in members: member_data_type_definition = definitions_registry.GetDefinitionByName( member) if not member_data_type_definition: error_message = 'undefined member: {0:s}.'.format(member) raise errors.DefinitionReaderError(definition_name, error_message) if member_data_type_definition.family_definition: error_message = 'member: {0:s} already part of a family.'.format(member) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.AddMemberDefinition(member_data_type_definition) return definition_object
python
def _ReadStructureFamilyDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads a structure family data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StructureDefinition: structure data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if is_member: error_message = 'data type not supported as member' raise errors.DefinitionReaderError(definition_name, error_message) definition_object = self._ReadLayoutDataTypeDefinition( definitions_registry, definition_values, data_types.StructureFamilyDefinition, definition_name, self._SUPPORTED_DEFINITION_VALUES_STRUCTURE_FAMILY) runtime = definition_values.get('runtime', None) if not runtime: error_message = 'missing runtime' raise errors.DefinitionReaderError(definition_name, error_message) runtime_data_type_definition = definitions_registry.GetDefinitionByName( runtime) if not runtime_data_type_definition: error_message = 'undefined runtime: {0:s}.'.format(runtime) raise errors.DefinitionReaderError(definition_name, error_message) if runtime_data_type_definition.family_definition: error_message = 'runtime: {0:s} already part of a family.'.format(runtime) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.AddRuntimeDefinition(runtime_data_type_definition) members = definition_values.get('members', None) if not members: error_message = 'missing members' raise errors.DefinitionReaderError(definition_name, error_message) for member in members: member_data_type_definition = definitions_registry.GetDefinitionByName( member) if not member_data_type_definition: error_message = 'undefined member: {0:s}.'.format(member) raise errors.DefinitionReaderError(definition_name, error_message) if member_data_type_definition.family_definition: error_message = 'member: {0:s} already part of a family.'.format(member) raise errors.DefinitionReaderError(definition_name, error_message) definition_object.AddMemberDefinition(member_data_type_definition) return definition_object
[ "def", "_ReadStructureFamilyDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "if", "is_member", ":", "error_message", "=", "'data type not supported as member'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", "=", "self", ".", "_ReadLayoutDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "StructureFamilyDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_DEFINITION_VALUES_STRUCTURE_FAMILY", ")", "runtime", "=", "definition_values", ".", "get", "(", "'runtime'", ",", "None", ")", "if", "not", "runtime", ":", "error_message", "=", "'missing runtime'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "runtime_data_type_definition", "=", "definitions_registry", ".", "GetDefinitionByName", "(", "runtime", ")", "if", "not", "runtime_data_type_definition", ":", "error_message", "=", "'undefined runtime: {0:s}.'", ".", "format", "(", "runtime", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "runtime_data_type_definition", ".", "family_definition", ":", "error_message", "=", "'runtime: {0:s} already part of a family.'", ".", "format", "(", "runtime", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "AddRuntimeDefinition", "(", "runtime_data_type_definition", ")", "members", "=", "definition_values", ".", "get", "(", "'members'", ",", "None", ")", "if", "not", "members", ":", "error_message", "=", "'missing members'", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "for", "member", "in", "members", ":", "member_data_type_definition", "=", "definitions_registry", ".", "GetDefinitionByName", "(", "member", ")", "if", "not", "member_data_type_definition", ":", "error_message", "=", "'undefined member: {0:s}.'", ".", "format", "(", "member", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "if", "member_data_type_definition", ".", "family_definition", ":", "error_message", "=", "'member: {0:s} already part of a family.'", ".", "format", "(", "member", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "definition_name", ",", "error_message", ")", "definition_object", ".", "AddMemberDefinition", "(", "member_data_type_definition", ")", "return", "definition_object" ]
Reads a structure family data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: StructureDefinition: structure data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "structure", "family", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1012-L1076
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadUnionDataTypeDefinition
def _ReadUnionDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an union data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: UnionDefinition: union data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadDataTypeDefinitionWithMembers( definitions_registry, definition_values, data_types.UnionDefinition, definition_name, supports_conditions=False)
python
def _ReadUnionDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an union data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: UnionDefinition: union data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadDataTypeDefinitionWithMembers( definitions_registry, definition_values, data_types.UnionDefinition, definition_name, supports_conditions=False)
[ "def", "_ReadUnionDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "return", "self", ".", "_ReadDataTypeDefinitionWithMembers", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "UnionDefinition", ",", "definition_name", ",", "supports_conditions", "=", "False", ")" ]
Reads an union data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: UnionDefinition: union data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "an", "union", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1078-L1100
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsReader._ReadUUIDDataTypeDefinition
def _ReadUUIDDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an UUID data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: UUIDDataTypeDefinition: UUID data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.UUIDDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE, default_size=16, is_member=is_member, supported_size_values=(16, ))
python
def _ReadUUIDDataTypeDefinition( self, definitions_registry, definition_values, definition_name, is_member=False): """Reads an UUID data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: UUIDDataTypeDefinition: UUID data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ return self._ReadFixedSizeDataTypeDefinition( definitions_registry, definition_values, data_types.UUIDDefinition, definition_name, self._SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE, default_size=16, is_member=is_member, supported_size_values=(16, ))
[ "def", "_ReadUUIDDataTypeDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ",", "definition_name", ",", "is_member", "=", "False", ")", ":", "return", "self", ".", "_ReadFixedSizeDataTypeDefinition", "(", "definitions_registry", ",", "definition_values", ",", "data_types", ".", "UUIDDefinition", ",", "definition_name", ",", "self", ".", "_SUPPORTED_ATTRIBUTES_FIXED_SIZE_DATA_TYPE", ",", "default_size", "=", "16", ",", "is_member", "=", "is_member", ",", "supported_size_values", "=", "(", "16", ",", ")", ")" ]
Reads an UUID data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. definition_name (str): name of the definition. is_member (Optional[bool]): True if the data type definition is a member data type definition. Returns: UUIDDataTypeDefinition: UUID data type definition. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "an", "UUID", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1102-L1126
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsFileReader._ReadDefinition
def _ReadDefinition(self, definitions_registry, definition_values): """Reads a data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. Returns: DataTypeDefinition: data type definition or None. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if not definition_values: error_message = 'missing definition values' raise errors.DefinitionReaderError(None, error_message) name = definition_values.get('name', None) if not name: error_message = 'missing name' raise errors.DefinitionReaderError(None, error_message) type_indicator = definition_values.get('type', None) if not type_indicator: error_message = 'invalid definition missing type' raise errors.DefinitionReaderError(name, error_message) data_type_callback = self._DATA_TYPE_CALLBACKS.get(type_indicator, None) if data_type_callback: data_type_callback = getattr(self, data_type_callback, None) if not data_type_callback: error_message = 'unuspported data type definition: {0:s}.'.format( type_indicator) raise errors.DefinitionReaderError(name, error_message) return data_type_callback(definitions_registry, definition_values, name)
python
def _ReadDefinition(self, definitions_registry, definition_values): """Reads a data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. Returns: DataTypeDefinition: data type definition or None. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect. """ if not definition_values: error_message = 'missing definition values' raise errors.DefinitionReaderError(None, error_message) name = definition_values.get('name', None) if not name: error_message = 'missing name' raise errors.DefinitionReaderError(None, error_message) type_indicator = definition_values.get('type', None) if not type_indicator: error_message = 'invalid definition missing type' raise errors.DefinitionReaderError(name, error_message) data_type_callback = self._DATA_TYPE_CALLBACKS.get(type_indicator, None) if data_type_callback: data_type_callback = getattr(self, data_type_callback, None) if not data_type_callback: error_message = 'unuspported data type definition: {0:s}.'.format( type_indicator) raise errors.DefinitionReaderError(name, error_message) return data_type_callback(definitions_registry, definition_values, name)
[ "def", "_ReadDefinition", "(", "self", ",", "definitions_registry", ",", "definition_values", ")", ":", "if", "not", "definition_values", ":", "error_message", "=", "'missing definition values'", "raise", "errors", ".", "DefinitionReaderError", "(", "None", ",", "error_message", ")", "name", "=", "definition_values", ".", "get", "(", "'name'", ",", "None", ")", "if", "not", "name", ":", "error_message", "=", "'missing name'", "raise", "errors", ".", "DefinitionReaderError", "(", "None", ",", "error_message", ")", "type_indicator", "=", "definition_values", ".", "get", "(", "'type'", ",", "None", ")", "if", "not", "type_indicator", ":", "error_message", "=", "'invalid definition missing type'", "raise", "errors", ".", "DefinitionReaderError", "(", "name", ",", "error_message", ")", "data_type_callback", "=", "self", ".", "_DATA_TYPE_CALLBACKS", ".", "get", "(", "type_indicator", ",", "None", ")", "if", "data_type_callback", ":", "data_type_callback", "=", "getattr", "(", "self", ",", "data_type_callback", ",", "None", ")", "if", "not", "data_type_callback", ":", "error_message", "=", "'unuspported data type definition: {0:s}.'", ".", "format", "(", "type_indicator", ")", "raise", "errors", ".", "DefinitionReaderError", "(", "name", ",", "error_message", ")", "return", "data_type_callback", "(", "definitions_registry", ",", "definition_values", ",", "name", ")" ]
Reads a data type definition. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. definition_values (dict[str, object]): definition values. Returns: DataTypeDefinition: data type definition or None. Raises: DefinitionReaderError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "a", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1132-L1169
libyal/dtfabric
dtfabric/reader.py
DataTypeDefinitionsFileReader.ReadFile
def ReadFile(self, definitions_registry, path): """Reads data type definitions from a file into the registry. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. path (str): path of the file to read from. """ with open(path, 'r') as file_object: self.ReadFileObject(definitions_registry, file_object)
python
def ReadFile(self, definitions_registry, path): """Reads data type definitions from a file into the registry. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. path (str): path of the file to read from. """ with open(path, 'r') as file_object: self.ReadFileObject(definitions_registry, file_object)
[ "def", "ReadFile", "(", "self", ",", "definitions_registry", ",", "path", ")", ":", "with", "open", "(", "path", ",", "'r'", ")", "as", "file_object", ":", "self", ".", "ReadFileObject", "(", "definitions_registry", ",", "file_object", ")" ]
Reads data type definitions from a file into the registry. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. path (str): path of the file to read from.
[ "Reads", "data", "type", "definitions", "from", "a", "file", "into", "the", "registry", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1171-L1180
libyal/dtfabric
dtfabric/reader.py
YAMLDataTypeDefinitionsFileReader._GetFormatErrorLocation
def _GetFormatErrorLocation( self, yaml_definition, last_definition_object): """Retrieves a format error location. Args: yaml_definition (dict[str, object]): current YAML definition. last_definition_object (DataTypeDefinition): previous data type definition. Returns: str: format error location. """ name = yaml_definition.get('name', None) if name: error_location = 'in: {0:s}'.format(name or '<NAMELESS>') elif last_definition_object: error_location = 'after: {0:s}'.format(last_definition_object.name) else: error_location = 'at start' return error_location
python
def _GetFormatErrorLocation( self, yaml_definition, last_definition_object): """Retrieves a format error location. Args: yaml_definition (dict[str, object]): current YAML definition. last_definition_object (DataTypeDefinition): previous data type definition. Returns: str: format error location. """ name = yaml_definition.get('name', None) if name: error_location = 'in: {0:s}'.format(name or '<NAMELESS>') elif last_definition_object: error_location = 'after: {0:s}'.format(last_definition_object.name) else: error_location = 'at start' return error_location
[ "def", "_GetFormatErrorLocation", "(", "self", ",", "yaml_definition", ",", "last_definition_object", ")", ":", "name", "=", "yaml_definition", ".", "get", "(", "'name'", ",", "None", ")", "if", "name", ":", "error_location", "=", "'in: {0:s}'", ".", "format", "(", "name", "or", "'<NAMELESS>'", ")", "elif", "last_definition_object", ":", "error_location", "=", "'after: {0:s}'", ".", "format", "(", "last_definition_object", ".", "name", ")", "else", ":", "error_location", "=", "'at start'", "return", "error_location" ]
Retrieves a format error location. Args: yaml_definition (dict[str, object]): current YAML definition. last_definition_object (DataTypeDefinition): previous data type definition. Returns: str: format error location.
[ "Retrieves", "a", "format", "error", "location", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1205-L1225
libyal/dtfabric
dtfabric/reader.py
YAMLDataTypeDefinitionsFileReader.ReadFileObject
def ReadFileObject(self, definitions_registry, file_object): """Reads data type definitions from a file-like object into the registry. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. file_object (file): file-like object to read from. Raises: FormatError: if the definitions values are missing or if the format is incorrect. """ last_definition_object = None error_location = None error_message = None try: yaml_generator = yaml.safe_load_all(file_object) for yaml_definition in yaml_generator: definition_object = self._ReadDefinition( definitions_registry, yaml_definition) if not definition_object: error_location = self._GetFormatErrorLocation( yaml_definition, last_definition_object) error_message = '{0:s} Missing definition object.'.format( error_location) raise errors.FormatError(error_message) definitions_registry.RegisterDefinition(definition_object) last_definition_object = definition_object except errors.DefinitionReaderError as exception: error_message = 'in: {0:s} {1:s}'.format( exception.name or '<NAMELESS>', exception.message) raise errors.FormatError(error_message) except (yaml.reader.ReaderError, yaml.scanner.ScannerError) as exception: error_location = self._GetFormatErrorLocation({}, last_definition_object) error_message = '{0:s} {1!s}'.format(error_location, exception) raise errors.FormatError(error_message)
python
def ReadFileObject(self, definitions_registry, file_object): """Reads data type definitions from a file-like object into the registry. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. file_object (file): file-like object to read from. Raises: FormatError: if the definitions values are missing or if the format is incorrect. """ last_definition_object = None error_location = None error_message = None try: yaml_generator = yaml.safe_load_all(file_object) for yaml_definition in yaml_generator: definition_object = self._ReadDefinition( definitions_registry, yaml_definition) if not definition_object: error_location = self._GetFormatErrorLocation( yaml_definition, last_definition_object) error_message = '{0:s} Missing definition object.'.format( error_location) raise errors.FormatError(error_message) definitions_registry.RegisterDefinition(definition_object) last_definition_object = definition_object except errors.DefinitionReaderError as exception: error_message = 'in: {0:s} {1:s}'.format( exception.name or '<NAMELESS>', exception.message) raise errors.FormatError(error_message) except (yaml.reader.ReaderError, yaml.scanner.ScannerError) as exception: error_location = self._GetFormatErrorLocation({}, last_definition_object) error_message = '{0:s} {1!s}'.format(error_location, exception) raise errors.FormatError(error_message)
[ "def", "ReadFileObject", "(", "self", ",", "definitions_registry", ",", "file_object", ")", ":", "last_definition_object", "=", "None", "error_location", "=", "None", "error_message", "=", "None", "try", ":", "yaml_generator", "=", "yaml", ".", "safe_load_all", "(", "file_object", ")", "for", "yaml_definition", "in", "yaml_generator", ":", "definition_object", "=", "self", ".", "_ReadDefinition", "(", "definitions_registry", ",", "yaml_definition", ")", "if", "not", "definition_object", ":", "error_location", "=", "self", ".", "_GetFormatErrorLocation", "(", "yaml_definition", ",", "last_definition_object", ")", "error_message", "=", "'{0:s} Missing definition object.'", ".", "format", "(", "error_location", ")", "raise", "errors", ".", "FormatError", "(", "error_message", ")", "definitions_registry", ".", "RegisterDefinition", "(", "definition_object", ")", "last_definition_object", "=", "definition_object", "except", "errors", ".", "DefinitionReaderError", "as", "exception", ":", "error_message", "=", "'in: {0:s} {1:s}'", ".", "format", "(", "exception", ".", "name", "or", "'<NAMELESS>'", ",", "exception", ".", "message", ")", "raise", "errors", ".", "FormatError", "(", "error_message", ")", "except", "(", "yaml", ".", "reader", ".", "ReaderError", ",", "yaml", ".", "scanner", ".", "ScannerError", ")", "as", "exception", ":", "error_location", "=", "self", ".", "_GetFormatErrorLocation", "(", "{", "}", ",", "last_definition_object", ")", "error_message", "=", "'{0:s} {1!s}'", ".", "format", "(", "error_location", ",", "exception", ")", "raise", "errors", ".", "FormatError", "(", "error_message", ")" ]
Reads data type definitions from a file-like object into the registry. Args: definitions_registry (DataTypeDefinitionsRegistry): data type definitions registry. file_object (file): file-like object to read from. Raises: FormatError: if the definitions values are missing or if the format is incorrect.
[ "Reads", "data", "type", "definitions", "from", "a", "file", "-", "like", "object", "into", "the", "registry", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/reader.py#L1227-L1267
jorisroovers/pymarkdownlint
pymarkdownlint/filefinder.py
MarkdownFileFinder.find_files
def find_files(path, filter="*.md"): """ Finds files with an (optional) given extension in a given path. """ if os.path.isfile(path): return [path] if os.path.isdir(path): matches = [] for root, dirnames, filenames in os.walk(path): for filename in fnmatch.filter(filenames, filter): matches.append(os.path.join(root, filename)) return matches
python
def find_files(path, filter="*.md"): """ Finds files with an (optional) given extension in a given path. """ if os.path.isfile(path): return [path] if os.path.isdir(path): matches = [] for root, dirnames, filenames in os.walk(path): for filename in fnmatch.filter(filenames, filter): matches.append(os.path.join(root, filename)) return matches
[ "def", "find_files", "(", "path", ",", "filter", "=", "\"*.md\"", ")", ":", "if", "os", ".", "path", ".", "isfile", "(", "path", ")", ":", "return", "[", "path", "]", "if", "os", ".", "path", ".", "isdir", "(", "path", ")", ":", "matches", "=", "[", "]", "for", "root", ",", "dirnames", ",", "filenames", "in", "os", ".", "walk", "(", "path", ")", ":", "for", "filename", "in", "fnmatch", ".", "filter", "(", "filenames", ",", "filter", ")", ":", "matches", ".", "append", "(", "os", ".", "path", ".", "join", "(", "root", ",", "filename", ")", ")", "return", "matches" ]
Finds files with an (optional) given extension in a given path.
[ "Finds", "files", "with", "an", "(", "optional", ")", "given", "extension", "in", "a", "given", "path", "." ]
train
https://github.com/jorisroovers/pymarkdownlint/blob/c1044e25e18afd78b3fda8fd9b00a4f67cfbbc65/pymarkdownlint/filefinder.py#L7-L17
OCHA-DAP/hdx-python-api
src/hdx/data/organization.py
Organization.read_from_hdx
def read_from_hdx(identifier, configuration=None): # type: (str, Optional[Configuration]) -> Optional['Organization'] """Reads the organization given by identifier from HDX and returns Organization object Args: identifier (str): Identifier of organization configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: Optional[Organization]: Organization object if successful read, None if not """ organization = Organization(configuration=configuration) result = organization._load_from_hdx('organization', identifier) if result: return organization return None
python
def read_from_hdx(identifier, configuration=None): # type: (str, Optional[Configuration]) -> Optional['Organization'] """Reads the organization given by identifier from HDX and returns Organization object Args: identifier (str): Identifier of organization configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: Optional[Organization]: Organization object if successful read, None if not """ organization = Organization(configuration=configuration) result = organization._load_from_hdx('organization', identifier) if result: return organization return None
[ "def", "read_from_hdx", "(", "identifier", ",", "configuration", "=", "None", ")", ":", "# type: (str, Optional[Configuration]) -> Optional['Organization']", "organization", "=", "Organization", "(", "configuration", "=", "configuration", ")", "result", "=", "organization", ".", "_load_from_hdx", "(", "'organization'", ",", "identifier", ")", "if", "result", ":", "return", "organization", "return", "None" ]
Reads the organization given by identifier from HDX and returns Organization object Args: identifier (str): Identifier of organization configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: Optional[Organization]: Organization object if successful read, None if not
[ "Reads", "the", "organization", "given", "by", "identifier", "from", "HDX", "and", "returns", "Organization", "object" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/organization.py#L70-L86
OCHA-DAP/hdx-python-api
src/hdx/data/organization.py
Organization.get_users
def get_users(self, capacity=None): # type: (Optional[str]) -> List[User] """Returns the organization's users. Args: capacity (Optional[str]): Filter by capacity eg. member, admin. Defaults to None. Returns: List[User]: Organization's users. """ users = list() usersdicts = self.data.get('users') if usersdicts is not None: for userdata in usersdicts: if capacity is not None and userdata['capacity'] != capacity: continue id = userdata.get('id') if id is None: id = userdata['name'] user = hdx.data.user.User.read_from_hdx(id, configuration=self.configuration) user['capacity'] = userdata['capacity'] users.append(user) return users
python
def get_users(self, capacity=None): # type: (Optional[str]) -> List[User] """Returns the organization's users. Args: capacity (Optional[str]): Filter by capacity eg. member, admin. Defaults to None. Returns: List[User]: Organization's users. """ users = list() usersdicts = self.data.get('users') if usersdicts is not None: for userdata in usersdicts: if capacity is not None and userdata['capacity'] != capacity: continue id = userdata.get('id') if id is None: id = userdata['name'] user = hdx.data.user.User.read_from_hdx(id, configuration=self.configuration) user['capacity'] = userdata['capacity'] users.append(user) return users
[ "def", "get_users", "(", "self", ",", "capacity", "=", "None", ")", ":", "# type: (Optional[str]) -> List[User]", "users", "=", "list", "(", ")", "usersdicts", "=", "self", ".", "data", ".", "get", "(", "'users'", ")", "if", "usersdicts", "is", "not", "None", ":", "for", "userdata", "in", "usersdicts", ":", "if", "capacity", "is", "not", "None", "and", "userdata", "[", "'capacity'", "]", "!=", "capacity", ":", "continue", "id", "=", "userdata", ".", "get", "(", "'id'", ")", "if", "id", "is", "None", ":", "id", "=", "userdata", "[", "'name'", "]", "user", "=", "hdx", ".", "data", ".", "user", ".", "User", ".", "read_from_hdx", "(", "id", ",", "configuration", "=", "self", ".", "configuration", ")", "user", "[", "'capacity'", "]", "=", "userdata", "[", "'capacity'", "]", "users", ".", "append", "(", "user", ")", "return", "users" ]
Returns the organization's users. Args: capacity (Optional[str]): Filter by capacity eg. member, admin. Defaults to None. Returns: List[User]: Organization's users.
[ "Returns", "the", "organization", "s", "users", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/organization.py#L128-L149
OCHA-DAP/hdx-python-api
src/hdx/data/organization.py
Organization.add_update_user
def add_update_user(self, user, capacity=None): # type: (Union[hdx.data.user.User,Dict,str],Optional[str]) -> None """Add new or update existing user in organization with new metadata. Capacity eg. member, admin must be supplied either within the User object or dictionary or using the capacity argument (which takes precedence). Args: user (Union[User,Dict,str]): Either a user id or user metadata either from a User object or a dictionary capacity (Optional[str]): Capacity of user eg. member, admin. Defaults to None. Returns: None """ if isinstance(user, str): user = hdx.data.user.User.read_from_hdx(user, configuration=self.configuration) elif isinstance(user, dict): user = hdx.data.user.User(user, configuration=self.configuration) if isinstance(user, hdx.data.user.User): users = self.data.get('users') if users is None: users = list() self.data['users'] = users if capacity is not None: user['capacity'] = capacity self._addupdate_hdxobject(users, 'name', user) return raise HDXError('Type %s cannot be added as a user!' % type(user).__name__)
python
def add_update_user(self, user, capacity=None): # type: (Union[hdx.data.user.User,Dict,str],Optional[str]) -> None """Add new or update existing user in organization with new metadata. Capacity eg. member, admin must be supplied either within the User object or dictionary or using the capacity argument (which takes precedence). Args: user (Union[User,Dict,str]): Either a user id or user metadata either from a User object or a dictionary capacity (Optional[str]): Capacity of user eg. member, admin. Defaults to None. Returns: None """ if isinstance(user, str): user = hdx.data.user.User.read_from_hdx(user, configuration=self.configuration) elif isinstance(user, dict): user = hdx.data.user.User(user, configuration=self.configuration) if isinstance(user, hdx.data.user.User): users = self.data.get('users') if users is None: users = list() self.data['users'] = users if capacity is not None: user['capacity'] = capacity self._addupdate_hdxobject(users, 'name', user) return raise HDXError('Type %s cannot be added as a user!' % type(user).__name__)
[ "def", "add_update_user", "(", "self", ",", "user", ",", "capacity", "=", "None", ")", ":", "# type: (Union[hdx.data.user.User,Dict,str],Optional[str]) -> None", "if", "isinstance", "(", "user", ",", "str", ")", ":", "user", "=", "hdx", ".", "data", ".", "user", ".", "User", ".", "read_from_hdx", "(", "user", ",", "configuration", "=", "self", ".", "configuration", ")", "elif", "isinstance", "(", "user", ",", "dict", ")", ":", "user", "=", "hdx", ".", "data", ".", "user", ".", "User", "(", "user", ",", "configuration", "=", "self", ".", "configuration", ")", "if", "isinstance", "(", "user", ",", "hdx", ".", "data", ".", "user", ".", "User", ")", ":", "users", "=", "self", ".", "data", ".", "get", "(", "'users'", ")", "if", "users", "is", "None", ":", "users", "=", "list", "(", ")", "self", ".", "data", "[", "'users'", "]", "=", "users", "if", "capacity", "is", "not", "None", ":", "user", "[", "'capacity'", "]", "=", "capacity", "self", ".", "_addupdate_hdxobject", "(", "users", ",", "'name'", ",", "user", ")", "return", "raise", "HDXError", "(", "'Type %s cannot be added as a user!'", "%", "type", "(", "user", ")", ".", "__name__", ")" ]
Add new or update existing user in organization with new metadata. Capacity eg. member, admin must be supplied either within the User object or dictionary or using the capacity argument (which takes precedence). Args: user (Union[User,Dict,str]): Either a user id or user metadata either from a User object or a dictionary capacity (Optional[str]): Capacity of user eg. member, admin. Defaults to None. Returns: None
[ "Add", "new", "or", "update", "existing", "user", "in", "organization", "with", "new", "metadata", ".", "Capacity", "eg", ".", "member", "admin", "must", "be", "supplied", "either", "within", "the", "User", "object", "or", "dictionary", "or", "using", "the", "capacity", "argument", "(", "which", "takes", "precedence", ")", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/organization.py#L151-L178
OCHA-DAP/hdx-python-api
src/hdx/data/organization.py
Organization.add_update_users
def add_update_users(self, users, capacity=None): # type: (List[Union[hdx.data.user.User,Dict,str]],Optional[str]) -> None """Add new or update existing users in organization with new metadata. Capacity eg. member, admin must be supplied either within the User object or dictionary or using the capacity argument (which takes precedence). Args: users (List[Union[User,Dict,str]]): A list of either user ids or users metadata from User objects or dictionaries capacity (Optional[str]): Capacity of users eg. member, admin. Defaults to None. Returns: None """ if not isinstance(users, list): raise HDXError('Users should be a list!') for user in users: self.add_update_user(user, capacity)
python
def add_update_users(self, users, capacity=None): # type: (List[Union[hdx.data.user.User,Dict,str]],Optional[str]) -> None """Add new or update existing users in organization with new metadata. Capacity eg. member, admin must be supplied either within the User object or dictionary or using the capacity argument (which takes precedence). Args: users (List[Union[User,Dict,str]]): A list of either user ids or users metadata from User objects or dictionaries capacity (Optional[str]): Capacity of users eg. member, admin. Defaults to None. Returns: None """ if not isinstance(users, list): raise HDXError('Users should be a list!') for user in users: self.add_update_user(user, capacity)
[ "def", "add_update_users", "(", "self", ",", "users", ",", "capacity", "=", "None", ")", ":", "# type: (List[Union[hdx.data.user.User,Dict,str]],Optional[str]) -> None", "if", "not", "isinstance", "(", "users", ",", "list", ")", ":", "raise", "HDXError", "(", "'Users should be a list!'", ")", "for", "user", "in", "users", ":", "self", ".", "add_update_user", "(", "user", ",", "capacity", ")" ]
Add new or update existing users in organization with new metadata. Capacity eg. member, admin must be supplied either within the User object or dictionary or using the capacity argument (which takes precedence). Args: users (List[Union[User,Dict,str]]): A list of either user ids or users metadata from User objects or dictionaries capacity (Optional[str]): Capacity of users eg. member, admin. Defaults to None. Returns: None
[ "Add", "new", "or", "update", "existing", "users", "in", "organization", "with", "new", "metadata", ".", "Capacity", "eg", ".", "member", "admin", "must", "be", "supplied", "either", "within", "the", "User", "object", "or", "dictionary", "or", "using", "the", "capacity", "argument", "(", "which", "takes", "precedence", ")", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/organization.py#L180-L196
OCHA-DAP/hdx-python-api
src/hdx/data/organization.py
Organization.get_datasets
def get_datasets(self, query='*:*', **kwargs): # type: (str, Any) -> List[hdx.data.dataset.Dataset] """Get list of datasets in organization Args: query (str): Restrict datasets returned to this query (in Solr format). Defaults to '*:*'. **kwargs: See below sort (string): Sorting of the search results. Defaults to 'relevance asc, metadata_modified desc'. rows (int): Number of matching rows to return. Defaults to all datasets (sys.maxsize). start (int): Offset in the complete result for where the set of returned datasets should begin facet (string): Whether to enable faceted results. Default to True. facet.mincount (int): Minimum counts for facet fields should be included in the results facet.limit (int): Maximum number of values the facet fields return (- = unlimited). Defaults to 50. facet.field (List[str]): Fields to facet upon. Default is empty. use_default_schema (bool): Use default package schema instead of custom schema. Defaults to False. Returns: List[Dataset]: List of datasets in organization """ return hdx.data.dataset.Dataset.search_in_hdx(query=query, configuration=self.configuration, fq='organization:%s' % self.data['name'], **kwargs)
python
def get_datasets(self, query='*:*', **kwargs): # type: (str, Any) -> List[hdx.data.dataset.Dataset] """Get list of datasets in organization Args: query (str): Restrict datasets returned to this query (in Solr format). Defaults to '*:*'. **kwargs: See below sort (string): Sorting of the search results. Defaults to 'relevance asc, metadata_modified desc'. rows (int): Number of matching rows to return. Defaults to all datasets (sys.maxsize). start (int): Offset in the complete result for where the set of returned datasets should begin facet (string): Whether to enable faceted results. Default to True. facet.mincount (int): Minimum counts for facet fields should be included in the results facet.limit (int): Maximum number of values the facet fields return (- = unlimited). Defaults to 50. facet.field (List[str]): Fields to facet upon. Default is empty. use_default_schema (bool): Use default package schema instead of custom schema. Defaults to False. Returns: List[Dataset]: List of datasets in organization """ return hdx.data.dataset.Dataset.search_in_hdx(query=query, configuration=self.configuration, fq='organization:%s' % self.data['name'], **kwargs)
[ "def", "get_datasets", "(", "self", ",", "query", "=", "'*:*'", ",", "*", "*", "kwargs", ")", ":", "# type: (str, Any) -> List[hdx.data.dataset.Dataset]", "return", "hdx", ".", "data", ".", "dataset", ".", "Dataset", ".", "search_in_hdx", "(", "query", "=", "query", ",", "configuration", "=", "self", ".", "configuration", ",", "fq", "=", "'organization:%s'", "%", "self", ".", "data", "[", "'name'", "]", ",", "*", "*", "kwargs", ")" ]
Get list of datasets in organization Args: query (str): Restrict datasets returned to this query (in Solr format). Defaults to '*:*'. **kwargs: See below sort (string): Sorting of the search results. Defaults to 'relevance asc, metadata_modified desc'. rows (int): Number of matching rows to return. Defaults to all datasets (sys.maxsize). start (int): Offset in the complete result for where the set of returned datasets should begin facet (string): Whether to enable faceted results. Default to True. facet.mincount (int): Minimum counts for facet fields should be included in the results facet.limit (int): Maximum number of values the facet fields return (- = unlimited). Defaults to 50. facet.field (List[str]): Fields to facet upon. Default is empty. use_default_schema (bool): Use default package schema instead of custom schema. Defaults to False. Returns: List[Dataset]: List of datasets in organization
[ "Get", "list", "of", "datasets", "in", "organization" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/organization.py#L210-L231
OCHA-DAP/hdx-python-api
src/hdx/data/organization.py
Organization.get_all_organization_names
def get_all_organization_names(configuration=None, **kwargs): # type: (Optional[Configuration], Any) -> List[str] """Get all organization names in HDX Args: configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. **kwargs: See below sort (str): Sort the search results according to field name and sort-order. Allowed fields are ‘name’, ‘package_count’ and ‘title’. Defaults to 'name asc'. organizations (List[str]): List of names of the groups to return. all_fields (bool): Return group dictionaries instead of just names. Only core fields are returned - get some more using the include_* options. Defaults to False. include_extras (bool): If all_fields, include the group extra fields. Defaults to False. include_tags (bool): If all_fields, include the group tags. Defaults to False. include_groups: If all_fields, include the groups the groups are in. Defaults to False. Returns: List[str]: List of all organization names in HDX """ organization = Organization(configuration=configuration) organization['id'] = 'all organizations' # only for error message if produced return organization._write_to_hdx('list', kwargs, 'id')
python
def get_all_organization_names(configuration=None, **kwargs): # type: (Optional[Configuration], Any) -> List[str] """Get all organization names in HDX Args: configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. **kwargs: See below sort (str): Sort the search results according to field name and sort-order. Allowed fields are ‘name’, ‘package_count’ and ‘title’. Defaults to 'name asc'. organizations (List[str]): List of names of the groups to return. all_fields (bool): Return group dictionaries instead of just names. Only core fields are returned - get some more using the include_* options. Defaults to False. include_extras (bool): If all_fields, include the group extra fields. Defaults to False. include_tags (bool): If all_fields, include the group tags. Defaults to False. include_groups: If all_fields, include the groups the groups are in. Defaults to False. Returns: List[str]: List of all organization names in HDX """ organization = Organization(configuration=configuration) organization['id'] = 'all organizations' # only for error message if produced return organization._write_to_hdx('list', kwargs, 'id')
[ "def", "get_all_organization_names", "(", "configuration", "=", "None", ",", "*", "*", "kwargs", ")", ":", "# type: (Optional[Configuration], Any) -> List[str]", "organization", "=", "Organization", "(", "configuration", "=", "configuration", ")", "organization", "[", "'id'", "]", "=", "'all organizations'", "# only for error message if produced", "return", "organization", ".", "_write_to_hdx", "(", "'list'", ",", "kwargs", ",", "'id'", ")" ]
Get all organization names in HDX Args: configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. **kwargs: See below sort (str): Sort the search results according to field name and sort-order. Allowed fields are ‘name’, ‘package_count’ and ‘title’. Defaults to 'name asc'. organizations (List[str]): List of names of the groups to return. all_fields (bool): Return group dictionaries instead of just names. Only core fields are returned - get some more using the include_* options. Defaults to False. include_extras (bool): If all_fields, include the group extra fields. Defaults to False. include_tags (bool): If all_fields, include the group tags. Defaults to False. include_groups: If all_fields, include the groups the groups are in. Defaults to False. Returns: List[str]: List of all organization names in HDX
[ "Get", "all", "organization", "names", "in", "HDX" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/organization.py#L234-L253
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._read_from_hdx
def _read_from_hdx(self, object_type, value, fieldname='id', action=None, **kwargs): # type: (str, str, str, Optional[str], Any) -> Tuple[bool, Union[Dict, str]] """Makes a read call to HDX passing in given parameter. Args: object_type (str): Description of HDX object type (for messages) value (str): Value of HDX field fieldname (str): HDX field name. Defaults to id. action (Optional[str]): Replacement CKAN action url to use. Defaults to None. **kwargs: Other fields to pass to CKAN. Returns: Tuple[bool, Union[Dict, str]]: (True/False, HDX object metadata/Error) """ if not fieldname: raise HDXError('Empty %s field name!' % object_type) if action is None: action = self.actions()['show'] data = {fieldname: value} data.update(kwargs) try: result = self.configuration.call_remoteckan(action, data) return True, result except NotFound: return False, '%s=%s: not found!' % (fieldname, value) except Exception as e: raisefrom(HDXError, 'Failed when trying to read: %s=%s! (POST)' % (fieldname, value), e)
python
def _read_from_hdx(self, object_type, value, fieldname='id', action=None, **kwargs): # type: (str, str, str, Optional[str], Any) -> Tuple[bool, Union[Dict, str]] """Makes a read call to HDX passing in given parameter. Args: object_type (str): Description of HDX object type (for messages) value (str): Value of HDX field fieldname (str): HDX field name. Defaults to id. action (Optional[str]): Replacement CKAN action url to use. Defaults to None. **kwargs: Other fields to pass to CKAN. Returns: Tuple[bool, Union[Dict, str]]: (True/False, HDX object metadata/Error) """ if not fieldname: raise HDXError('Empty %s field name!' % object_type) if action is None: action = self.actions()['show'] data = {fieldname: value} data.update(kwargs) try: result = self.configuration.call_remoteckan(action, data) return True, result except NotFound: return False, '%s=%s: not found!' % (fieldname, value) except Exception as e: raisefrom(HDXError, 'Failed when trying to read: %s=%s! (POST)' % (fieldname, value), e)
[ "def", "_read_from_hdx", "(", "self", ",", "object_type", ",", "value", ",", "fieldname", "=", "'id'", ",", "action", "=", "None", ",", "*", "*", "kwargs", ")", ":", "# type: (str, str, str, Optional[str], Any) -> Tuple[bool, Union[Dict, str]]", "if", "not", "fieldname", ":", "raise", "HDXError", "(", "'Empty %s field name!'", "%", "object_type", ")", "if", "action", "is", "None", ":", "action", "=", "self", ".", "actions", "(", ")", "[", "'show'", "]", "data", "=", "{", "fieldname", ":", "value", "}", "data", ".", "update", "(", "kwargs", ")", "try", ":", "result", "=", "self", ".", "configuration", ".", "call_remoteckan", "(", "action", ",", "data", ")", "return", "True", ",", "result", "except", "NotFound", ":", "return", "False", ",", "'%s=%s: not found!'", "%", "(", "fieldname", ",", "value", ")", "except", "Exception", "as", "e", ":", "raisefrom", "(", "HDXError", ",", "'Failed when trying to read: %s=%s! (POST)'", "%", "(", "fieldname", ",", "value", ")", ",", "e", ")" ]
Makes a read call to HDX passing in given parameter. Args: object_type (str): Description of HDX object type (for messages) value (str): Value of HDX field fieldname (str): HDX field name. Defaults to id. action (Optional[str]): Replacement CKAN action url to use. Defaults to None. **kwargs: Other fields to pass to CKAN. Returns: Tuple[bool, Union[Dict, str]]: (True/False, HDX object metadata/Error)
[ "Makes", "a", "read", "call", "to", "HDX", "passing", "in", "given", "parameter", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L95-L122
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._load_from_hdx
def _load_from_hdx(self, object_type, id_field): # type: (str, str) -> bool """Helper method to load the HDX object given by identifier from HDX Args: object_type (str): Description of HDX object type (for messages) id_field (str): HDX object identifier Returns: bool: True if loaded, False if not """ success, result = self._read_from_hdx(object_type, id_field) if success: self.old_data = self.data self.data = result return True logger.debug(result) return False
python
def _load_from_hdx(self, object_type, id_field): # type: (str, str) -> bool """Helper method to load the HDX object given by identifier from HDX Args: object_type (str): Description of HDX object type (for messages) id_field (str): HDX object identifier Returns: bool: True if loaded, False if not """ success, result = self._read_from_hdx(object_type, id_field) if success: self.old_data = self.data self.data = result return True logger.debug(result) return False
[ "def", "_load_from_hdx", "(", "self", ",", "object_type", ",", "id_field", ")", ":", "# type: (str, str) -> bool", "success", ",", "result", "=", "self", ".", "_read_from_hdx", "(", "object_type", ",", "id_field", ")", "if", "success", ":", "self", ".", "old_data", "=", "self", ".", "data", "self", ".", "data", "=", "result", "return", "True", "logger", ".", "debug", "(", "result", ")", "return", "False" ]
Helper method to load the HDX object given by identifier from HDX Args: object_type (str): Description of HDX object type (for messages) id_field (str): HDX object identifier Returns: bool: True if loaded, False if not
[ "Helper", "method", "to", "load", "the", "HDX", "object", "given", "by", "identifier", "from", "HDX" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L124-L141
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._check_load_existing_object
def _check_load_existing_object(self, object_type, id_field_name, operation='update'): # type: (str, str, str) -> None """Check metadata exists and contains HDX object identifier, and if so load HDX object Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier operation (str): Operation to report if error. Defaults to update. Returns: None """ self._check_existing_object(object_type, id_field_name) if not self._load_from_hdx(object_type, self.data[id_field_name]): raise HDXError('No existing %s to %s!' % (object_type, operation))
python
def _check_load_existing_object(self, object_type, id_field_name, operation='update'): # type: (str, str, str) -> None """Check metadata exists and contains HDX object identifier, and if so load HDX object Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier operation (str): Operation to report if error. Defaults to update. Returns: None """ self._check_existing_object(object_type, id_field_name) if not self._load_from_hdx(object_type, self.data[id_field_name]): raise HDXError('No existing %s to %s!' % (object_type, operation))
[ "def", "_check_load_existing_object", "(", "self", ",", "object_type", ",", "id_field_name", ",", "operation", "=", "'update'", ")", ":", "# type: (str, str, str) -> None", "self", ".", "_check_existing_object", "(", "object_type", ",", "id_field_name", ")", "if", "not", "self", ".", "_load_from_hdx", "(", "object_type", ",", "self", ".", "data", "[", "id_field_name", "]", ")", ":", "raise", "HDXError", "(", "'No existing %s to %s!'", "%", "(", "object_type", ",", "operation", ")", ")" ]
Check metadata exists and contains HDX object identifier, and if so load HDX object Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier operation (str): Operation to report if error. Defaults to update. Returns: None
[ "Check", "metadata", "exists", "and", "contains", "HDX", "object", "identifier", "and", "if", "so", "load", "HDX", "object" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L165-L179
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._check_required_fields
def _check_required_fields(self, object_type, ignore_fields): # type: (str, List[str]) -> None """Helper method to check that metadata for HDX object is complete Args: ignore_fields (List[str]): Any fields to ignore in the check Returns: None """ for field in self.configuration[object_type]['required_fields']: if field not in self.data and field not in ignore_fields: raise HDXError('Field %s is missing in %s!' % (field, object_type))
python
def _check_required_fields(self, object_type, ignore_fields): # type: (str, List[str]) -> None """Helper method to check that metadata for HDX object is complete Args: ignore_fields (List[str]): Any fields to ignore in the check Returns: None """ for field in self.configuration[object_type]['required_fields']: if field not in self.data and field not in ignore_fields: raise HDXError('Field %s is missing in %s!' % (field, object_type))
[ "def", "_check_required_fields", "(", "self", ",", "object_type", ",", "ignore_fields", ")", ":", "# type: (str, List[str]) -> None", "for", "field", "in", "self", ".", "configuration", "[", "object_type", "]", "[", "'required_fields'", "]", ":", "if", "field", "not", "in", "self", ".", "data", "and", "field", "not", "in", "ignore_fields", ":", "raise", "HDXError", "(", "'Field %s is missing in %s!'", "%", "(", "field", ",", "object_type", ")", ")" ]
Helper method to check that metadata for HDX object is complete Args: ignore_fields (List[str]): Any fields to ignore in the check Returns: None
[ "Helper", "method", "to", "check", "that", "metadata", "for", "HDX", "object", "is", "complete" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L195-L207
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._merge_hdx_update
def _merge_hdx_update(self, object_type, id_field_name, file_to_upload=None, **kwargs): # type: (str, str, Optional[str], Any) -> None """Helper method to check if HDX object exists and update it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None """ merge_two_dictionaries(self.data, self.old_data) if 'batch_mode' in kwargs: # Whether or not CKAN should change groupings of datasets on /datasets page self.data['batch_mode'] = kwargs['batch_mode'] if 'skip_validation' in kwargs: # Whether or not CKAN should perform validation steps (checking fields present) self.data['skip_validation'] = kwargs['skip_validation'] ignore_field = self.configuration['%s' % object_type].get('ignore_on_update') self.check_required_fields(ignore_fields=[ignore_field]) operation = kwargs.get('operation', 'update') self._save_to_hdx(operation, id_field_name, file_to_upload)
python
def _merge_hdx_update(self, object_type, id_field_name, file_to_upload=None, **kwargs): # type: (str, str, Optional[str], Any) -> None """Helper method to check if HDX object exists and update it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None """ merge_two_dictionaries(self.data, self.old_data) if 'batch_mode' in kwargs: # Whether or not CKAN should change groupings of datasets on /datasets page self.data['batch_mode'] = kwargs['batch_mode'] if 'skip_validation' in kwargs: # Whether or not CKAN should perform validation steps (checking fields present) self.data['skip_validation'] = kwargs['skip_validation'] ignore_field = self.configuration['%s' % object_type].get('ignore_on_update') self.check_required_fields(ignore_fields=[ignore_field]) operation = kwargs.get('operation', 'update') self._save_to_hdx(operation, id_field_name, file_to_upload)
[ "def", "_merge_hdx_update", "(", "self", ",", "object_type", ",", "id_field_name", ",", "file_to_upload", "=", "None", ",", "*", "*", "kwargs", ")", ":", "# type: (str, str, Optional[str], Any) -> None", "merge_two_dictionaries", "(", "self", ".", "data", ",", "self", ".", "old_data", ")", "if", "'batch_mode'", "in", "kwargs", ":", "# Whether or not CKAN should change groupings of datasets on /datasets page", "self", ".", "data", "[", "'batch_mode'", "]", "=", "kwargs", "[", "'batch_mode'", "]", "if", "'skip_validation'", "in", "kwargs", ":", "# Whether or not CKAN should perform validation steps (checking fields present)", "self", ".", "data", "[", "'skip_validation'", "]", "=", "kwargs", "[", "'skip_validation'", "]", "ignore_field", "=", "self", ".", "configuration", "[", "'%s'", "%", "object_type", "]", ".", "get", "(", "'ignore_on_update'", ")", "self", ".", "check_required_fields", "(", "ignore_fields", "=", "[", "ignore_field", "]", ")", "operation", "=", "kwargs", ".", "get", "(", "'operation'", ",", "'update'", ")", "self", ".", "_save_to_hdx", "(", "operation", ",", "id_field_name", ",", "file_to_upload", ")" ]
Helper method to check if HDX object exists and update it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None
[ "Helper", "method", "to", "check", "if", "HDX", "object", "exists", "and", "update", "it" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L209-L231
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._update_in_hdx
def _update_in_hdx(self, object_type, id_field_name, file_to_upload=None, **kwargs): # type: (str, str, Optional[str], Any) -> None """Helper method to check if HDX object exists in HDX and if so, update it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None """ self._check_load_existing_object(object_type, id_field_name) # We load an existing object even thought it may well have been loaded already # to prevent an admittedly unlikely race condition where someone has updated # the object in the intervening time self._merge_hdx_update(object_type, id_field_name, file_to_upload, **kwargs)
python
def _update_in_hdx(self, object_type, id_field_name, file_to_upload=None, **kwargs): # type: (str, str, Optional[str], Any) -> None """Helper method to check if HDX object exists in HDX and if so, update it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None """ self._check_load_existing_object(object_type, id_field_name) # We load an existing object even thought it may well have been loaded already # to prevent an admittedly unlikely race condition where someone has updated # the object in the intervening time self._merge_hdx_update(object_type, id_field_name, file_to_upload, **kwargs)
[ "def", "_update_in_hdx", "(", "self", ",", "object_type", ",", "id_field_name", ",", "file_to_upload", "=", "None", ",", "*", "*", "kwargs", ")", ":", "# type: (str, str, Optional[str], Any) -> None", "self", ".", "_check_load_existing_object", "(", "object_type", ",", "id_field_name", ")", "# We load an existing object even thought it may well have been loaded already", "# to prevent an admittedly unlikely race condition where someone has updated", "# the object in the intervening time", "self", ".", "_merge_hdx_update", "(", "object_type", ",", "id_field_name", ",", "file_to_upload", ",", "*", "*", "kwargs", ")" ]
Helper method to check if HDX object exists in HDX and if so, update it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None
[ "Helper", "method", "to", "check", "if", "HDX", "object", "exists", "in", "HDX", "and", "if", "so", "update", "it" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L243-L262
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._write_to_hdx
def _write_to_hdx(self, action, data, id_field_name, file_to_upload=None): # type: (str, Dict, str, Optional[str]) -> Dict """Creates or updates an HDX object in HDX and return HDX object metadata dict Args: action (str): Action to perform eg. 'create', 'update' data (Dict): Data to write to HDX id_field_name (str): Name of field containing HDX object identifier or None file_to_upload (Optional[str]): File to upload to HDX Returns: Dict: HDX object metadata """ file = None try: if file_to_upload: file = open(file_to_upload, 'rb') files = [('upload', file)] else: files = None return self.configuration.call_remoteckan(self.actions()[action], data, files=files) except Exception as e: raisefrom(HDXError, 'Failed when trying to %s %s! (POST)' % (action, data[id_field_name]), e) finally: if file_to_upload and file: file.close()
python
def _write_to_hdx(self, action, data, id_field_name, file_to_upload=None): # type: (str, Dict, str, Optional[str]) -> Dict """Creates or updates an HDX object in HDX and return HDX object metadata dict Args: action (str): Action to perform eg. 'create', 'update' data (Dict): Data to write to HDX id_field_name (str): Name of field containing HDX object identifier or None file_to_upload (Optional[str]): File to upload to HDX Returns: Dict: HDX object metadata """ file = None try: if file_to_upload: file = open(file_to_upload, 'rb') files = [('upload', file)] else: files = None return self.configuration.call_remoteckan(self.actions()[action], data, files=files) except Exception as e: raisefrom(HDXError, 'Failed when trying to %s %s! (POST)' % (action, data[id_field_name]), e) finally: if file_to_upload and file: file.close()
[ "def", "_write_to_hdx", "(", "self", ",", "action", ",", "data", ",", "id_field_name", ",", "file_to_upload", "=", "None", ")", ":", "# type: (str, Dict, str, Optional[str]) -> Dict", "file", "=", "None", "try", ":", "if", "file_to_upload", ":", "file", "=", "open", "(", "file_to_upload", ",", "'rb'", ")", "files", "=", "[", "(", "'upload'", ",", "file", ")", "]", "else", ":", "files", "=", "None", "return", "self", ".", "configuration", ".", "call_remoteckan", "(", "self", ".", "actions", "(", ")", "[", "action", "]", ",", "data", ",", "files", "=", "files", ")", "except", "Exception", "as", "e", ":", "raisefrom", "(", "HDXError", ",", "'Failed when trying to %s %s! (POST)'", "%", "(", "action", ",", "data", "[", "id_field_name", "]", ")", ",", "e", ")", "finally", ":", "if", "file_to_upload", "and", "file", ":", "file", ".", "close", "(", ")" ]
Creates or updates an HDX object in HDX and return HDX object metadata dict Args: action (str): Action to perform eg. 'create', 'update' data (Dict): Data to write to HDX id_field_name (str): Name of field containing HDX object identifier or None file_to_upload (Optional[str]): File to upload to HDX Returns: Dict: HDX object metadata
[ "Creates", "or", "updates", "an", "HDX", "object", "in", "HDX", "and", "return", "HDX", "object", "metadata", "dict" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L264-L289
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._save_to_hdx
def _save_to_hdx(self, action, id_field_name, file_to_upload=None): # type: (str, str, Optional[str]) -> None """Creates or updates an HDX object in HDX, saving current data and replacing with returned HDX object data from HDX Args: action (str): Action to perform: 'create' or 'update' id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX Returns: None """ result = self._write_to_hdx(action, self.data, id_field_name, file_to_upload) self.old_data = self.data self.data = result
python
def _save_to_hdx(self, action, id_field_name, file_to_upload=None): # type: (str, str, Optional[str]) -> None """Creates or updates an HDX object in HDX, saving current data and replacing with returned HDX object data from HDX Args: action (str): Action to perform: 'create' or 'update' id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX Returns: None """ result = self._write_to_hdx(action, self.data, id_field_name, file_to_upload) self.old_data = self.data self.data = result
[ "def", "_save_to_hdx", "(", "self", ",", "action", ",", "id_field_name", ",", "file_to_upload", "=", "None", ")", ":", "# type: (str, str, Optional[str]) -> None", "result", "=", "self", ".", "_write_to_hdx", "(", "action", ",", "self", ".", "data", ",", "id_field_name", ",", "file_to_upload", ")", "self", ".", "old_data", "=", "self", ".", "data", "self", ".", "data", "=", "result" ]
Creates or updates an HDX object in HDX, saving current data and replacing with returned HDX object data from HDX Args: action (str): Action to perform: 'create' or 'update' id_field_name (str): Name of field containing HDX object identifier file_to_upload (Optional[str]): File to upload to HDX Returns: None
[ "Creates", "or", "updates", "an", "HDX", "object", "in", "HDX", "saving", "current", "data", "and", "replacing", "with", "returned", "HDX", "object", "data", "from", "HDX" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L291-L306
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._create_in_hdx
def _create_in_hdx(self, object_type, id_field_name, name_field_name, file_to_upload=None): # type: (str, str, str, Optional[str]) -> None """Helper method to check if resource exists in HDX and if so, update it, otherwise create it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier name_field_name (str): Name of field containing HDX object name file_to_upload (Optional[str]): File to upload to HDX (if url not supplied) Returns: None """ self.check_required_fields() if id_field_name in self.data and self._load_from_hdx(object_type, self.data[id_field_name]): logger.warning('%s exists. Updating %s' % (object_type, self.data[id_field_name])) self._merge_hdx_update(object_type, id_field_name, file_to_upload) else: self._save_to_hdx('create', name_field_name, file_to_upload)
python
def _create_in_hdx(self, object_type, id_field_name, name_field_name, file_to_upload=None): # type: (str, str, str, Optional[str]) -> None """Helper method to check if resource exists in HDX and if so, update it, otherwise create it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier name_field_name (str): Name of field containing HDX object name file_to_upload (Optional[str]): File to upload to HDX (if url not supplied) Returns: None """ self.check_required_fields() if id_field_name in self.data and self._load_from_hdx(object_type, self.data[id_field_name]): logger.warning('%s exists. Updating %s' % (object_type, self.data[id_field_name])) self._merge_hdx_update(object_type, id_field_name, file_to_upload) else: self._save_to_hdx('create', name_field_name, file_to_upload)
[ "def", "_create_in_hdx", "(", "self", ",", "object_type", ",", "id_field_name", ",", "name_field_name", ",", "file_to_upload", "=", "None", ")", ":", "# type: (str, str, str, Optional[str]) -> None", "self", ".", "check_required_fields", "(", ")", "if", "id_field_name", "in", "self", ".", "data", "and", "self", ".", "_load_from_hdx", "(", "object_type", ",", "self", ".", "data", "[", "id_field_name", "]", ")", ":", "logger", ".", "warning", "(", "'%s exists. Updating %s'", "%", "(", "object_type", ",", "self", ".", "data", "[", "id_field_name", "]", ")", ")", "self", ".", "_merge_hdx_update", "(", "object_type", ",", "id_field_name", ",", "file_to_upload", ")", "else", ":", "self", ".", "_save_to_hdx", "(", "'create'", ",", "name_field_name", ",", "file_to_upload", ")" ]
Helper method to check if resource exists in HDX and if so, update it, otherwise create it Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier name_field_name (str): Name of field containing HDX object name file_to_upload (Optional[str]): File to upload to HDX (if url not supplied) Returns: None
[ "Helper", "method", "to", "check", "if", "resource", "exists", "in", "HDX", "and", "if", "so", "update", "it", "otherwise", "create", "it" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L318-L338
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._delete_from_hdx
def _delete_from_hdx(self, object_type, id_field_name): # type: (str, str) -> None """Helper method to deletes a resource from HDX Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier Returns: None """ if id_field_name not in self.data: raise HDXError('No %s field (mandatory) in %s!' % (id_field_name, object_type)) self._save_to_hdx('delete', id_field_name)
python
def _delete_from_hdx(self, object_type, id_field_name): # type: (str, str) -> None """Helper method to deletes a resource from HDX Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier Returns: None """ if id_field_name not in self.data: raise HDXError('No %s field (mandatory) in %s!' % (id_field_name, object_type)) self._save_to_hdx('delete', id_field_name)
[ "def", "_delete_from_hdx", "(", "self", ",", "object_type", ",", "id_field_name", ")", ":", "# type: (str, str) -> None", "if", "id_field_name", "not", "in", "self", ".", "data", ":", "raise", "HDXError", "(", "'No %s field (mandatory) in %s!'", "%", "(", "id_field_name", ",", "object_type", ")", ")", "self", ".", "_save_to_hdx", "(", "'delete'", ",", "id_field_name", ")" ]
Helper method to deletes a resource from HDX Args: object_type (str): Description of HDX object type (for messages) id_field_name (str): Name of field containing HDX object identifier Returns: None
[ "Helper", "method", "to", "deletes", "a", "resource", "from", "HDX" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L350-L363
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._addupdate_hdxobject
def _addupdate_hdxobject(self, hdxobjects, id_field, new_hdxobject): # type: (List[HDXObjectUpperBound], str, HDXObjectUpperBound) -> HDXObjectUpperBound """Helper function to add a new HDX object to a supplied list of HDX objects or update existing metadata if the object already exists in the list Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to which to add new objects or update existing ones id_field (str): Field on which to match to determine if object already exists in list new_hdxobject (T <= HDXObject): The HDX object to be added/updated Returns: T <= HDXObject: The HDX object which was added or updated """ for hdxobject in hdxobjects: if hdxobject[id_field] == new_hdxobject[id_field]: merge_two_dictionaries(hdxobject, new_hdxobject) return hdxobject hdxobjects.append(new_hdxobject) return new_hdxobject
python
def _addupdate_hdxobject(self, hdxobjects, id_field, new_hdxobject): # type: (List[HDXObjectUpperBound], str, HDXObjectUpperBound) -> HDXObjectUpperBound """Helper function to add a new HDX object to a supplied list of HDX objects or update existing metadata if the object already exists in the list Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to which to add new objects or update existing ones id_field (str): Field on which to match to determine if object already exists in list new_hdxobject (T <= HDXObject): The HDX object to be added/updated Returns: T <= HDXObject: The HDX object which was added or updated """ for hdxobject in hdxobjects: if hdxobject[id_field] == new_hdxobject[id_field]: merge_two_dictionaries(hdxobject, new_hdxobject) return hdxobject hdxobjects.append(new_hdxobject) return new_hdxobject
[ "def", "_addupdate_hdxobject", "(", "self", ",", "hdxobjects", ",", "id_field", ",", "new_hdxobject", ")", ":", "# type: (List[HDXObjectUpperBound], str, HDXObjectUpperBound) -> HDXObjectUpperBound", "for", "hdxobject", "in", "hdxobjects", ":", "if", "hdxobject", "[", "id_field", "]", "==", "new_hdxobject", "[", "id_field", "]", ":", "merge_two_dictionaries", "(", "hdxobject", ",", "new_hdxobject", ")", "return", "hdxobject", "hdxobjects", ".", "append", "(", "new_hdxobject", ")", "return", "new_hdxobject" ]
Helper function to add a new HDX object to a supplied list of HDX objects or update existing metadata if the object already exists in the list Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to which to add new objects or update existing ones id_field (str): Field on which to match to determine if object already exists in list new_hdxobject (T <= HDXObject): The HDX object to be added/updated Returns: T <= HDXObject: The HDX object which was added or updated
[ "Helper", "function", "to", "add", "a", "new", "HDX", "object", "to", "a", "supplied", "list", "of", "HDX", "objects", "or", "update", "existing", "metadata", "if", "the", "object", "already", "exists", "in", "the", "list" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L365-L383
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._remove_hdxobject
def _remove_hdxobject(self, objlist, obj, matchon='id', delete=False): # type: (List[Union[HDXObjectUpperBound,Dict]], Union[HDXObjectUpperBound,Dict,str], str, bool) -> bool """Remove an HDX object from a list within the parent HDX object Args: objlist (List[Union[T <= HDXObject,Dict]]): list of HDX objects obj (Union[T <= HDXObject,Dict,str]): Either an id or hdx object metadata either from an HDX object or a dictionary matchon (str): Field to match on. Defaults to id. delete (bool): Whether to delete HDX object. Defaults to False. Returns: bool: True if object removed, False if not """ if objlist is None: return False if isinstance(obj, six.string_types): obj_id = obj elif isinstance(obj, dict) or isinstance(obj, HDXObject): obj_id = obj.get(matchon) else: raise HDXError('Type of object not a string, dict or T<=HDXObject') if not obj_id: return False for i, objdata in enumerate(objlist): objid = objdata.get(matchon) if objid and objid == obj_id: if delete: objlist[i].delete_from_hdx() del objlist[i] return True return False
python
def _remove_hdxobject(self, objlist, obj, matchon='id', delete=False): # type: (List[Union[HDXObjectUpperBound,Dict]], Union[HDXObjectUpperBound,Dict,str], str, bool) -> bool """Remove an HDX object from a list within the parent HDX object Args: objlist (List[Union[T <= HDXObject,Dict]]): list of HDX objects obj (Union[T <= HDXObject,Dict,str]): Either an id or hdx object metadata either from an HDX object or a dictionary matchon (str): Field to match on. Defaults to id. delete (bool): Whether to delete HDX object. Defaults to False. Returns: bool: True if object removed, False if not """ if objlist is None: return False if isinstance(obj, six.string_types): obj_id = obj elif isinstance(obj, dict) or isinstance(obj, HDXObject): obj_id = obj.get(matchon) else: raise HDXError('Type of object not a string, dict or T<=HDXObject') if not obj_id: return False for i, objdata in enumerate(objlist): objid = objdata.get(matchon) if objid and objid == obj_id: if delete: objlist[i].delete_from_hdx() del objlist[i] return True return False
[ "def", "_remove_hdxobject", "(", "self", ",", "objlist", ",", "obj", ",", "matchon", "=", "'id'", ",", "delete", "=", "False", ")", ":", "# type: (List[Union[HDXObjectUpperBound,Dict]], Union[HDXObjectUpperBound,Dict,str], str, bool) -> bool", "if", "objlist", "is", "None", ":", "return", "False", "if", "isinstance", "(", "obj", ",", "six", ".", "string_types", ")", ":", "obj_id", "=", "obj", "elif", "isinstance", "(", "obj", ",", "dict", ")", "or", "isinstance", "(", "obj", ",", "HDXObject", ")", ":", "obj_id", "=", "obj", ".", "get", "(", "matchon", ")", "else", ":", "raise", "HDXError", "(", "'Type of object not a string, dict or T<=HDXObject'", ")", "if", "not", "obj_id", ":", "return", "False", "for", "i", ",", "objdata", "in", "enumerate", "(", "objlist", ")", ":", "objid", "=", "objdata", ".", "get", "(", "matchon", ")", "if", "objid", "and", "objid", "==", "obj_id", ":", "if", "delete", ":", "objlist", "[", "i", "]", ".", "delete_from_hdx", "(", ")", "del", "objlist", "[", "i", "]", "return", "True", "return", "False" ]
Remove an HDX object from a list within the parent HDX object Args: objlist (List[Union[T <= HDXObject,Dict]]): list of HDX objects obj (Union[T <= HDXObject,Dict,str]): Either an id or hdx object metadata either from an HDX object or a dictionary matchon (str): Field to match on. Defaults to id. delete (bool): Whether to delete HDX object. Defaults to False. Returns: bool: True if object removed, False if not
[ "Remove", "an", "HDX", "object", "from", "a", "list", "within", "the", "parent", "HDX", "object" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L385-L415
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._convert_hdxobjects
def _convert_hdxobjects(self, hdxobjects): # type: (List[HDXObjectUpperBound]) -> List[HDXObjectUpperBound] """Helper function to convert supplied list of HDX objects to a list of dict Args: hdxobjects (List[T <= HDXObject]): List of HDX objects to convert Returns: List[Dict]: List of HDX objects converted to simple dictionaries """ newhdxobjects = list() for hdxobject in hdxobjects: newhdxobjects.append(hdxobject.data) return newhdxobjects
python
def _convert_hdxobjects(self, hdxobjects): # type: (List[HDXObjectUpperBound]) -> List[HDXObjectUpperBound] """Helper function to convert supplied list of HDX objects to a list of dict Args: hdxobjects (List[T <= HDXObject]): List of HDX objects to convert Returns: List[Dict]: List of HDX objects converted to simple dictionaries """ newhdxobjects = list() for hdxobject in hdxobjects: newhdxobjects.append(hdxobject.data) return newhdxobjects
[ "def", "_convert_hdxobjects", "(", "self", ",", "hdxobjects", ")", ":", "# type: (List[HDXObjectUpperBound]) -> List[HDXObjectUpperBound]", "newhdxobjects", "=", "list", "(", ")", "for", "hdxobject", "in", "hdxobjects", ":", "newhdxobjects", ".", "append", "(", "hdxobject", ".", "data", ")", "return", "newhdxobjects" ]
Helper function to convert supplied list of HDX objects to a list of dict Args: hdxobjects (List[T <= HDXObject]): List of HDX objects to convert Returns: List[Dict]: List of HDX objects converted to simple dictionaries
[ "Helper", "function", "to", "convert", "supplied", "list", "of", "HDX", "objects", "to", "a", "list", "of", "dict" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L417-L430
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._copy_hdxobjects
def _copy_hdxobjects(self, hdxobjects, hdxobjectclass, attribute_to_copy=None): # type: (List[HDXObjectUpperBound], type, Optional[str]) -> List[HDXObjectUpperBound] """Helper function to make a deep copy of a supplied list of HDX objects Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to copy hdxobjectclass (type): Type of the HDX Objects to be copied attribute_to_copy (Optional[str]): An attribute to copy over from the HDX object. Defaults to None. Returns: List[T <= HDXObject]: Deep copy of list of HDX objects """ newhdxobjects = list() for hdxobject in hdxobjects: newhdxobjectdata = copy.deepcopy(hdxobject.data) newhdxobject = hdxobjectclass(newhdxobjectdata, configuration=self.configuration) if attribute_to_copy: value = getattr(hdxobject, attribute_to_copy) setattr(newhdxobject, attribute_to_copy, value) newhdxobjects.append(newhdxobject) return newhdxobjects
python
def _copy_hdxobjects(self, hdxobjects, hdxobjectclass, attribute_to_copy=None): # type: (List[HDXObjectUpperBound], type, Optional[str]) -> List[HDXObjectUpperBound] """Helper function to make a deep copy of a supplied list of HDX objects Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to copy hdxobjectclass (type): Type of the HDX Objects to be copied attribute_to_copy (Optional[str]): An attribute to copy over from the HDX object. Defaults to None. Returns: List[T <= HDXObject]: Deep copy of list of HDX objects """ newhdxobjects = list() for hdxobject in hdxobjects: newhdxobjectdata = copy.deepcopy(hdxobject.data) newhdxobject = hdxobjectclass(newhdxobjectdata, configuration=self.configuration) if attribute_to_copy: value = getattr(hdxobject, attribute_to_copy) setattr(newhdxobject, attribute_to_copy, value) newhdxobjects.append(newhdxobject) return newhdxobjects
[ "def", "_copy_hdxobjects", "(", "self", ",", "hdxobjects", ",", "hdxobjectclass", ",", "attribute_to_copy", "=", "None", ")", ":", "# type: (List[HDXObjectUpperBound], type, Optional[str]) -> List[HDXObjectUpperBound]", "newhdxobjects", "=", "list", "(", ")", "for", "hdxobject", "in", "hdxobjects", ":", "newhdxobjectdata", "=", "copy", ".", "deepcopy", "(", "hdxobject", ".", "data", ")", "newhdxobject", "=", "hdxobjectclass", "(", "newhdxobjectdata", ",", "configuration", "=", "self", ".", "configuration", ")", "if", "attribute_to_copy", ":", "value", "=", "getattr", "(", "hdxobject", ",", "attribute_to_copy", ")", "setattr", "(", "newhdxobject", ",", "attribute_to_copy", ",", "value", ")", "newhdxobjects", ".", "append", "(", "newhdxobject", ")", "return", "newhdxobjects" ]
Helper function to make a deep copy of a supplied list of HDX objects Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to copy hdxobjectclass (type): Type of the HDX Objects to be copied attribute_to_copy (Optional[str]): An attribute to copy over from the HDX object. Defaults to None. Returns: List[T <= HDXObject]: Deep copy of list of HDX objects
[ "Helper", "function", "to", "make", "a", "deep", "copy", "of", "a", "supplied", "list", "of", "HDX", "objects" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L432-L452
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._separate_hdxobjects
def _separate_hdxobjects(self, hdxobjects, hdxobjects_name, id_field, hdxobjectclass): # type: (List[HDXObjectUpperBound], str, str, type) -> None """Helper function to take a list of HDX objects contained in the internal dictionary and add them to a supplied list of HDX objects or update existing metadata if any objects already exist in the list. The list in the internal dictionary is then deleted. Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to which to add new objects or update existing ones hdxobjects_name (str): Name of key in internal dictionary from which to obtain list of HDX objects id_field (str): Field on which to match to determine if object already exists in list hdxobjectclass (type): Type of the HDX Object to be added/updated Returns: None """ new_hdxobjects = self.data.get(hdxobjects_name, list()) """:type : List[HDXObjectUpperBound]""" if new_hdxobjects: hdxobject_names = set() for hdxobject in hdxobjects: hdxobject_name = hdxobject[id_field] hdxobject_names.add(hdxobject_name) for new_hdxobject in new_hdxobjects: if hdxobject_name == new_hdxobject[id_field]: merge_two_dictionaries(hdxobject, new_hdxobject) break for new_hdxobject in new_hdxobjects: if not new_hdxobject[id_field] in hdxobject_names: hdxobjects.append(hdxobjectclass(new_hdxobject, configuration=self.configuration)) del self.data[hdxobjects_name]
python
def _separate_hdxobjects(self, hdxobjects, hdxobjects_name, id_field, hdxobjectclass): # type: (List[HDXObjectUpperBound], str, str, type) -> None """Helper function to take a list of HDX objects contained in the internal dictionary and add them to a supplied list of HDX objects or update existing metadata if any objects already exist in the list. The list in the internal dictionary is then deleted. Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to which to add new objects or update existing ones hdxobjects_name (str): Name of key in internal dictionary from which to obtain list of HDX objects id_field (str): Field on which to match to determine if object already exists in list hdxobjectclass (type): Type of the HDX Object to be added/updated Returns: None """ new_hdxobjects = self.data.get(hdxobjects_name, list()) """:type : List[HDXObjectUpperBound]""" if new_hdxobjects: hdxobject_names = set() for hdxobject in hdxobjects: hdxobject_name = hdxobject[id_field] hdxobject_names.add(hdxobject_name) for new_hdxobject in new_hdxobjects: if hdxobject_name == new_hdxobject[id_field]: merge_two_dictionaries(hdxobject, new_hdxobject) break for new_hdxobject in new_hdxobjects: if not new_hdxobject[id_field] in hdxobject_names: hdxobjects.append(hdxobjectclass(new_hdxobject, configuration=self.configuration)) del self.data[hdxobjects_name]
[ "def", "_separate_hdxobjects", "(", "self", ",", "hdxobjects", ",", "hdxobjects_name", ",", "id_field", ",", "hdxobjectclass", ")", ":", "# type: (List[HDXObjectUpperBound], str, str, type) -> None", "new_hdxobjects", "=", "self", ".", "data", ".", "get", "(", "hdxobjects_name", ",", "list", "(", ")", ")", "\"\"\":type : List[HDXObjectUpperBound]\"\"\"", "if", "new_hdxobjects", ":", "hdxobject_names", "=", "set", "(", ")", "for", "hdxobject", "in", "hdxobjects", ":", "hdxobject_name", "=", "hdxobject", "[", "id_field", "]", "hdxobject_names", ".", "add", "(", "hdxobject_name", ")", "for", "new_hdxobject", "in", "new_hdxobjects", ":", "if", "hdxobject_name", "==", "new_hdxobject", "[", "id_field", "]", ":", "merge_two_dictionaries", "(", "hdxobject", ",", "new_hdxobject", ")", "break", "for", "new_hdxobject", "in", "new_hdxobjects", ":", "if", "not", "new_hdxobject", "[", "id_field", "]", "in", "hdxobject_names", ":", "hdxobjects", ".", "append", "(", "hdxobjectclass", "(", "new_hdxobject", ",", "configuration", "=", "self", ".", "configuration", ")", ")", "del", "self", ".", "data", "[", "hdxobjects_name", "]" ]
Helper function to take a list of HDX objects contained in the internal dictionary and add them to a supplied list of HDX objects or update existing metadata if any objects already exist in the list. The list in the internal dictionary is then deleted. Args: hdxobjects (List[T <= HDXObject]): list of HDX objects to which to add new objects or update existing ones hdxobjects_name (str): Name of key in internal dictionary from which to obtain list of HDX objects id_field (str): Field on which to match to determine if object already exists in list hdxobjectclass (type): Type of the HDX Object to be added/updated Returns: None
[ "Helper", "function", "to", "take", "a", "list", "of", "HDX", "objects", "contained", "in", "the", "internal", "dictionary", "and", "add", "them", "to", "a", "supplied", "list", "of", "HDX", "objects", "or", "update", "existing", "metadata", "if", "any", "objects", "already", "exist", "in", "the", "list", ".", "The", "list", "in", "the", "internal", "dictionary", "is", "then", "deleted", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L454-L483
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._get_tags
def _get_tags(self): # type: () -> List[str] """Return the dataset's list of tags Returns: List[str]: list of tags or [] if there are none """ tags = self.data.get('tags', None) if not tags: return list() return [x['name'] for x in tags]
python
def _get_tags(self): # type: () -> List[str] """Return the dataset's list of tags Returns: List[str]: list of tags or [] if there are none """ tags = self.data.get('tags', None) if not tags: return list() return [x['name'] for x in tags]
[ "def", "_get_tags", "(", "self", ")", ":", "# type: () -> List[str]", "tags", "=", "self", ".", "data", ".", "get", "(", "'tags'", ",", "None", ")", "if", "not", "tags", ":", "return", "list", "(", ")", "return", "[", "x", "[", "'name'", "]", "for", "x", "in", "tags", "]" ]
Return the dataset's list of tags Returns: List[str]: list of tags or [] if there are none
[ "Return", "the", "dataset", "s", "list", "of", "tags" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L485-L495
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._add_tag
def _add_tag(self, tag): # type: (str) -> bool """Add a tag Args: tag (str): Tag to add Returns: bool: True if tag added or False if tag already present """ tags = self.data.get('tags', None) if tags: if tag in [x['name'] for x in tags]: return False else: tags = list() tags.append({'name': tag}) self.data['tags'] = tags return True
python
def _add_tag(self, tag): # type: (str) -> bool """Add a tag Args: tag (str): Tag to add Returns: bool: True if tag added or False if tag already present """ tags = self.data.get('tags', None) if tags: if tag in [x['name'] for x in tags]: return False else: tags = list() tags.append({'name': tag}) self.data['tags'] = tags return True
[ "def", "_add_tag", "(", "self", ",", "tag", ")", ":", "# type: (str) -> bool", "tags", "=", "self", ".", "data", ".", "get", "(", "'tags'", ",", "None", ")", "if", "tags", ":", "if", "tag", "in", "[", "x", "[", "'name'", "]", "for", "x", "in", "tags", "]", ":", "return", "False", "else", ":", "tags", "=", "list", "(", ")", "tags", ".", "append", "(", "{", "'name'", ":", "tag", "}", ")", "self", ".", "data", "[", "'tags'", "]", "=", "tags", "return", "True" ]
Add a tag Args: tag (str): Tag to add Returns: bool: True if tag added or False if tag already present
[ "Add", "a", "tag" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L497-L515
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._add_tags
def _add_tags(self, tags): # type: (List[str]) -> bool """Add a list of tag Args: tags (List[str]): list of tags to add Returns: bool: True if all tags added or False if any already present. """ alltagsadded = True for tag in tags: if not self._add_tag(tag): alltagsadded = False return alltagsadded
python
def _add_tags(self, tags): # type: (List[str]) -> bool """Add a list of tag Args: tags (List[str]): list of tags to add Returns: bool: True if all tags added or False if any already present. """ alltagsadded = True for tag in tags: if not self._add_tag(tag): alltagsadded = False return alltagsadded
[ "def", "_add_tags", "(", "self", ",", "tags", ")", ":", "# type: (List[str]) -> bool", "alltagsadded", "=", "True", "for", "tag", "in", "tags", ":", "if", "not", "self", ".", "_add_tag", "(", "tag", ")", ":", "alltagsadded", "=", "False", "return", "alltagsadded" ]
Add a list of tag Args: tags (List[str]): list of tags to add Returns: bool: True if all tags added or False if any already present.
[ "Add", "a", "list", "of", "tag" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L517-L531
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._get_stringlist_from_commastring
def _get_stringlist_from_commastring(self, field): # type: (str) -> List[str] """Return list of strings from comma separated list Args: field (str): Field containing comma separated list Returns: List[str]: List of strings """ strings = self.data.get(field) if strings: return strings.split(',') else: return list()
python
def _get_stringlist_from_commastring(self, field): # type: (str) -> List[str] """Return list of strings from comma separated list Args: field (str): Field containing comma separated list Returns: List[str]: List of strings """ strings = self.data.get(field) if strings: return strings.split(',') else: return list()
[ "def", "_get_stringlist_from_commastring", "(", "self", ",", "field", ")", ":", "# type: (str) -> List[str]", "strings", "=", "self", ".", "data", ".", "get", "(", "field", ")", "if", "strings", ":", "return", "strings", ".", "split", "(", "','", ")", "else", ":", "return", "list", "(", ")" ]
Return list of strings from comma separated list Args: field (str): Field containing comma separated list Returns: List[str]: List of strings
[ "Return", "list", "of", "strings", "from", "comma", "separated", "list" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L533-L547
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._add_string_to_commastring
def _add_string_to_commastring(self, field, string): # type: (str, str) -> bool """Add a string to a comma separated list of strings Args: field (str): Field containing comma separated list string (str): String to add Returns: bool: True if string added or False if string already present """ if string in self._get_stringlist_from_commastring(field): return False strings = '%s,%s' % (self.data.get(field, ''), string) if strings[0] == ',': strings = strings[1:] self.data[field] = strings return True
python
def _add_string_to_commastring(self, field, string): # type: (str, str) -> bool """Add a string to a comma separated list of strings Args: field (str): Field containing comma separated list string (str): String to add Returns: bool: True if string added or False if string already present """ if string in self._get_stringlist_from_commastring(field): return False strings = '%s,%s' % (self.data.get(field, ''), string) if strings[0] == ',': strings = strings[1:] self.data[field] = strings return True
[ "def", "_add_string_to_commastring", "(", "self", ",", "field", ",", "string", ")", ":", "# type: (str, str) -> bool", "if", "string", "in", "self", ".", "_get_stringlist_from_commastring", "(", "field", ")", ":", "return", "False", "strings", "=", "'%s,%s'", "%", "(", "self", ".", "data", ".", "get", "(", "field", ",", "''", ")", ",", "string", ")", "if", "strings", "[", "0", "]", "==", "','", ":", "strings", "=", "strings", "[", "1", ":", "]", "self", ".", "data", "[", "field", "]", "=", "strings", "return", "True" ]
Add a string to a comma separated list of strings Args: field (str): Field containing comma separated list string (str): String to add Returns: bool: True if string added or False if string already present
[ "Add", "a", "string", "to", "a", "comma", "separated", "list", "of", "strings" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L549-L566
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._add_strings_to_commastring
def _add_strings_to_commastring(self, field, strings): # type: (str, List[str]) -> bool """Add a list of strings to a comma separated list of strings Args: field (str): Field containing comma separated list strings (List[str]): list of strings to add Returns: bool: True if all strings added or False if any already present. """ allstringsadded = True for string in strings: if not self._add_string_to_commastring(field, string): allstringsadded = False return allstringsadded
python
def _add_strings_to_commastring(self, field, strings): # type: (str, List[str]) -> bool """Add a list of strings to a comma separated list of strings Args: field (str): Field containing comma separated list strings (List[str]): list of strings to add Returns: bool: True if all strings added or False if any already present. """ allstringsadded = True for string in strings: if not self._add_string_to_commastring(field, string): allstringsadded = False return allstringsadded
[ "def", "_add_strings_to_commastring", "(", "self", ",", "field", ",", "strings", ")", ":", "# type: (str, List[str]) -> bool", "allstringsadded", "=", "True", "for", "string", "in", "strings", ":", "if", "not", "self", ".", "_add_string_to_commastring", "(", "field", ",", "string", ")", ":", "allstringsadded", "=", "False", "return", "allstringsadded" ]
Add a list of strings to a comma separated list of strings Args: field (str): Field containing comma separated list strings (List[str]): list of strings to add Returns: bool: True if all strings added or False if any already present.
[ "Add", "a", "list", "of", "strings", "to", "a", "comma", "separated", "list", "of", "strings" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L568-L583
OCHA-DAP/hdx-python-api
src/hdx/data/hdxobject.py
HDXObject._remove_string_from_commastring
def _remove_string_from_commastring(self, field, string): # type: (str, str) -> bool """Remove a string from a comma separated list of strings Args: field (str): Field containing comma separated list string (str): String to remove Returns: bool: True if string removed or False if not """ commastring = self.data.get(field, '') if string in commastring: self.data[field] = commastring.replace(string, '') return True return False
python
def _remove_string_from_commastring(self, field, string): # type: (str, str) -> bool """Remove a string from a comma separated list of strings Args: field (str): Field containing comma separated list string (str): String to remove Returns: bool: True if string removed or False if not """ commastring = self.data.get(field, '') if string in commastring: self.data[field] = commastring.replace(string, '') return True return False
[ "def", "_remove_string_from_commastring", "(", "self", ",", "field", ",", "string", ")", ":", "# type: (str, str) -> bool", "commastring", "=", "self", ".", "data", ".", "get", "(", "field", ",", "''", ")", "if", "string", "in", "commastring", ":", "self", ".", "data", "[", "field", "]", "=", "commastring", ".", "replace", "(", "string", ",", "''", ")", "return", "True", "return", "False" ]
Remove a string from a comma separated list of strings Args: field (str): Field containing comma separated list string (str): String to remove Returns: bool: True if string removed or False if not
[ "Remove", "a", "string", "from", "a", "comma", "separated", "list", "of", "strings" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/hdxobject.py#L585-L600
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.read_from_hdx
def read_from_hdx(identifier, configuration=None): # type: (str, Optional[Configuration]) -> Optional['Resource'] """Reads the resource given by identifier from HDX and returns Resource object Args: identifier (str): Identifier of resource configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: Optional[Resource]: Resource object if successful read, None if not """ if is_valid_uuid(identifier) is False: raise HDXError('%s is not a valid resource id!' % identifier) resource = Resource(configuration=configuration) result = resource._load_from_hdx('resource', identifier) if result: return resource return None
python
def read_from_hdx(identifier, configuration=None): # type: (str, Optional[Configuration]) -> Optional['Resource'] """Reads the resource given by identifier from HDX and returns Resource object Args: identifier (str): Identifier of resource configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: Optional[Resource]: Resource object if successful read, None if not """ if is_valid_uuid(identifier) is False: raise HDXError('%s is not a valid resource id!' % identifier) resource = Resource(configuration=configuration) result = resource._load_from_hdx('resource', identifier) if result: return resource return None
[ "def", "read_from_hdx", "(", "identifier", ",", "configuration", "=", "None", ")", ":", "# type: (str, Optional[Configuration]) -> Optional['Resource']", "if", "is_valid_uuid", "(", "identifier", ")", "is", "False", ":", "raise", "HDXError", "(", "'%s is not a valid resource id!'", "%", "identifier", ")", "resource", "=", "Resource", "(", "configuration", "=", "configuration", ")", "result", "=", "resource", ".", "_load_from_hdx", "(", "'resource'", ",", "identifier", ")", "if", "result", ":", "return", "resource", "return", "None" ]
Reads the resource given by identifier from HDX and returns Resource object Args: identifier (str): Identifier of resource configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: Optional[Resource]: Resource object if successful read, None if not
[ "Reads", "the", "resource", "given", "by", "identifier", "from", "HDX", "and", "returns", "Resource", "object" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L83-L101
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.set_file_to_upload
def set_file_to_upload(self, file_to_upload): # type: (str) -> None """Delete any existing url and set the file uploaded to the local path provided Args: file_to_upload (str): Local path to file to upload Returns: None """ if 'url' in self.data: del self.data['url'] self.file_to_upload = file_to_upload
python
def set_file_to_upload(self, file_to_upload): # type: (str) -> None """Delete any existing url and set the file uploaded to the local path provided Args: file_to_upload (str): Local path to file to upload Returns: None """ if 'url' in self.data: del self.data['url'] self.file_to_upload = file_to_upload
[ "def", "set_file_to_upload", "(", "self", ",", "file_to_upload", ")", ":", "# type: (str) -> None", "if", "'url'", "in", "self", ".", "data", ":", "del", "self", ".", "data", "[", "'url'", "]", "self", ".", "file_to_upload", "=", "file_to_upload" ]
Delete any existing url and set the file uploaded to the local path provided Args: file_to_upload (str): Local path to file to upload Returns: None
[ "Delete", "any", "existing", "url", "and", "set", "the", "file", "uploaded", "to", "the", "local", "path", "provided" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L133-L145
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.check_url_filetoupload
def check_url_filetoupload(self): # type: () -> None """Check if url or file to upload provided for resource and add resource_type and url_type if not supplied Returns: None """ if self.file_to_upload is None: if 'url' in self.data: if 'resource_type' not in self.data: self.data['resource_type'] = 'api' if 'url_type' not in self.data: self.data['url_type'] = 'api' else: raise HDXError('Either a url or a file to upload must be supplied!') else: if 'url' in self.data: if self.data['url'] != hdx.data.dataset.Dataset.temporary_url: raise HDXError('Either a url or a file to upload must be supplied not both!') if 'resource_type' not in self.data: self.data['resource_type'] = 'file.upload' if 'url_type' not in self.data: self.data['url_type'] = 'upload' if 'tracking_summary' in self.data: del self.data['tracking_summary']
python
def check_url_filetoupload(self): # type: () -> None """Check if url or file to upload provided for resource and add resource_type and url_type if not supplied Returns: None """ if self.file_to_upload is None: if 'url' in self.data: if 'resource_type' not in self.data: self.data['resource_type'] = 'api' if 'url_type' not in self.data: self.data['url_type'] = 'api' else: raise HDXError('Either a url or a file to upload must be supplied!') else: if 'url' in self.data: if self.data['url'] != hdx.data.dataset.Dataset.temporary_url: raise HDXError('Either a url or a file to upload must be supplied not both!') if 'resource_type' not in self.data: self.data['resource_type'] = 'file.upload' if 'url_type' not in self.data: self.data['url_type'] = 'upload' if 'tracking_summary' in self.data: del self.data['tracking_summary']
[ "def", "check_url_filetoupload", "(", "self", ")", ":", "# type: () -> None", "if", "self", ".", "file_to_upload", "is", "None", ":", "if", "'url'", "in", "self", ".", "data", ":", "if", "'resource_type'", "not", "in", "self", ".", "data", ":", "self", ".", "data", "[", "'resource_type'", "]", "=", "'api'", "if", "'url_type'", "not", "in", "self", ".", "data", ":", "self", ".", "data", "[", "'url_type'", "]", "=", "'api'", "else", ":", "raise", "HDXError", "(", "'Either a url or a file to upload must be supplied!'", ")", "else", ":", "if", "'url'", "in", "self", ".", "data", ":", "if", "self", ".", "data", "[", "'url'", "]", "!=", "hdx", ".", "data", ".", "dataset", ".", "Dataset", ".", "temporary_url", ":", "raise", "HDXError", "(", "'Either a url or a file to upload must be supplied not both!'", ")", "if", "'resource_type'", "not", "in", "self", ".", "data", ":", "self", ".", "data", "[", "'resource_type'", "]", "=", "'file.upload'", "if", "'url_type'", "not", "in", "self", ".", "data", ":", "self", ".", "data", "[", "'url_type'", "]", "=", "'upload'", "if", "'tracking_summary'", "in", "self", ".", "data", ":", "del", "self", ".", "data", "[", "'tracking_summary'", "]" ]
Check if url or file to upload provided for resource and add resource_type and url_type if not supplied Returns: None
[ "Check", "if", "url", "or", "file", "to", "upload", "provided", "for", "resource", "and", "add", "resource_type", "and", "url_type", "if", "not", "supplied" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L147-L171
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.update_in_hdx
def update_in_hdx(self, **kwargs): # type: (Any) -> None """Check if resource exists in HDX and if so, update it Args: **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None """ self._check_load_existing_object('resource', 'id') if self.file_to_upload and 'url' in self.data: del self.data['url'] self._merge_hdx_update('resource', 'id', self.file_to_upload, **kwargs)
python
def update_in_hdx(self, **kwargs): # type: (Any) -> None """Check if resource exists in HDX and if so, update it Args: **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None """ self._check_load_existing_object('resource', 'id') if self.file_to_upload and 'url' in self.data: del self.data['url'] self._merge_hdx_update('resource', 'id', self.file_to_upload, **kwargs)
[ "def", "update_in_hdx", "(", "self", ",", "*", "*", "kwargs", ")", ":", "# type: (Any) -> None", "self", ".", "_check_load_existing_object", "(", "'resource'", ",", "'id'", ")", "if", "self", ".", "file_to_upload", "and", "'url'", "in", "self", ".", "data", ":", "del", "self", ".", "data", "[", "'url'", "]", "self", ".", "_merge_hdx_update", "(", "'resource'", ",", "'id'", ",", "self", ".", "file_to_upload", ",", "*", "*", "kwargs", ")" ]
Check if resource exists in HDX and if so, update it Args: **kwargs: See below operation (string): Operation to perform eg. patch. Defaults to update. Returns: None
[ "Check", "if", "resource", "exists", "in", "HDX", "and", "if", "so", "update", "it" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L187-L201
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.create_in_hdx
def create_in_hdx(self): # type: () -> None """Check if resource exists in HDX and if so, update it, otherwise create it Returns: None """ self.check_required_fields() id = self.data.get('id') if id and self._load_from_hdx('resource', id): logger.warning('%s exists. Updating %s' % ('resource', id)) if self.file_to_upload and 'url' in self.data: del self.data['url'] self._merge_hdx_update('resource', 'id', self.file_to_upload) else: self._save_to_hdx('create', 'name', self.file_to_upload)
python
def create_in_hdx(self): # type: () -> None """Check if resource exists in HDX and if so, update it, otherwise create it Returns: None """ self.check_required_fields() id = self.data.get('id') if id and self._load_from_hdx('resource', id): logger.warning('%s exists. Updating %s' % ('resource', id)) if self.file_to_upload and 'url' in self.data: del self.data['url'] self._merge_hdx_update('resource', 'id', self.file_to_upload) else: self._save_to_hdx('create', 'name', self.file_to_upload)
[ "def", "create_in_hdx", "(", "self", ")", ":", "# type: () -> None", "self", ".", "check_required_fields", "(", ")", "id", "=", "self", ".", "data", ".", "get", "(", "'id'", ")", "if", "id", "and", "self", ".", "_load_from_hdx", "(", "'resource'", ",", "id", ")", ":", "logger", ".", "warning", "(", "'%s exists. Updating %s'", "%", "(", "'resource'", ",", "id", ")", ")", "if", "self", ".", "file_to_upload", "and", "'url'", "in", "self", ".", "data", ":", "del", "self", ".", "data", "[", "'url'", "]", "self", ".", "_merge_hdx_update", "(", "'resource'", ",", "'id'", ",", "self", ".", "file_to_upload", ")", "else", ":", "self", ".", "_save_to_hdx", "(", "'create'", ",", "'name'", ",", "self", ".", "file_to_upload", ")" ]
Check if resource exists in HDX and if so, update it, otherwise create it Returns: None
[ "Check", "if", "resource", "exists", "in", "HDX", "and", "if", "so", "update", "it", "otherwise", "create", "it" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L203-L218
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.get_dataset
def get_dataset(self): # type: () -> hdx.data.dataset.Dataset """Return dataset containing this resource Returns: hdx.data.dataset.Dataset: Dataset containing this resource """ package_id = self.data.get('package_id') if package_id is None: raise HDXError('Resource has no package id!') return hdx.data.dataset.Dataset.read_from_hdx(package_id)
python
def get_dataset(self): # type: () -> hdx.data.dataset.Dataset """Return dataset containing this resource Returns: hdx.data.dataset.Dataset: Dataset containing this resource """ package_id = self.data.get('package_id') if package_id is None: raise HDXError('Resource has no package id!') return hdx.data.dataset.Dataset.read_from_hdx(package_id)
[ "def", "get_dataset", "(", "self", ")", ":", "# type: () -> hdx.data.dataset.Dataset", "package_id", "=", "self", ".", "data", ".", "get", "(", "'package_id'", ")", "if", "package_id", "is", "None", ":", "raise", "HDXError", "(", "'Resource has no package id!'", ")", "return", "hdx", ".", "data", ".", "dataset", ".", "Dataset", ".", "read_from_hdx", "(", "package_id", ")" ]
Return dataset containing this resource Returns: hdx.data.dataset.Dataset: Dataset containing this resource
[ "Return", "dataset", "containing", "this", "resource" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L229-L239
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.search_in_hdx
def search_in_hdx(query, configuration=None, **kwargs): # type: (str, Optional[Configuration], Any) -> List['Resource'] """Searches for resources in HDX. NOTE: Does not search dataset metadata! Args: query (str): Query configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. **kwargs: See below order_by (str): A field on the Resource model that orders the results offset (int): Apply an offset to the query limit (int): Apply a limit to the query Returns: List[Resource]: List of resources resulting from query """ resources = [] resource = Resource(configuration=configuration) success, result = resource._read_from_hdx('resource', query, 'query', Resource.actions()['search']) if result: count = result.get('count', None) if count: for resourcedict in result['results']: resource = Resource(resourcedict, configuration=configuration) resources.append(resource) else: logger.debug(result) return resources
python
def search_in_hdx(query, configuration=None, **kwargs): # type: (str, Optional[Configuration], Any) -> List['Resource'] """Searches for resources in HDX. NOTE: Does not search dataset metadata! Args: query (str): Query configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. **kwargs: See below order_by (str): A field on the Resource model that orders the results offset (int): Apply an offset to the query limit (int): Apply a limit to the query Returns: List[Resource]: List of resources resulting from query """ resources = [] resource = Resource(configuration=configuration) success, result = resource._read_from_hdx('resource', query, 'query', Resource.actions()['search']) if result: count = result.get('count', None) if count: for resourcedict in result['results']: resource = Resource(resourcedict, configuration=configuration) resources.append(resource) else: logger.debug(result) return resources
[ "def", "search_in_hdx", "(", "query", ",", "configuration", "=", "None", ",", "*", "*", "kwargs", ")", ":", "# type: (str, Optional[Configuration], Any) -> List['Resource']", "resources", "=", "[", "]", "resource", "=", "Resource", "(", "configuration", "=", "configuration", ")", "success", ",", "result", "=", "resource", ".", "_read_from_hdx", "(", "'resource'", ",", "query", ",", "'query'", ",", "Resource", ".", "actions", "(", ")", "[", "'search'", "]", ")", "if", "result", ":", "count", "=", "result", ".", "get", "(", "'count'", ",", "None", ")", "if", "count", ":", "for", "resourcedict", "in", "result", "[", "'results'", "]", ":", "resource", "=", "Resource", "(", "resourcedict", ",", "configuration", "=", "configuration", ")", "resources", ".", "append", "(", "resource", ")", "else", ":", "logger", ".", "debug", "(", "result", ")", "return", "resources" ]
Searches for resources in HDX. NOTE: Does not search dataset metadata! Args: query (str): Query configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. **kwargs: See below order_by (str): A field on the Resource model that orders the results offset (int): Apply an offset to the query limit (int): Apply a limit to the query Returns: List[Resource]: List of resources resulting from query
[ "Searches", "for", "resources", "in", "HDX", ".", "NOTE", ":", "Does", "not", "search", "dataset", "metadata!" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L242-L268
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.download
def download(self, folder=None): # type: (Optional[str]) -> Tuple[str, str] """Download resource store to provided folder or temporary folder if no folder supplied Args: folder (Optional[str]): Folder to download resource to. Defaults to None. Returns: Tuple[str, str]: (URL downloaded, Path to downloaded file) """ # Download the resource url = self.data.get('url', None) if not url: raise HDXError('No URL to download!') logger.debug('Downloading %s' % url) filename = self.data['name'] format = '.%s' % self.data['format'] if format not in filename: filename = '%s%s' % (filename, format) with Download(full_agent=self.configuration.get_user_agent()) as downloader: path = downloader.download_file(url, folder, filename) return url, path
python
def download(self, folder=None): # type: (Optional[str]) -> Tuple[str, str] """Download resource store to provided folder or temporary folder if no folder supplied Args: folder (Optional[str]): Folder to download resource to. Defaults to None. Returns: Tuple[str, str]: (URL downloaded, Path to downloaded file) """ # Download the resource url = self.data.get('url', None) if not url: raise HDXError('No URL to download!') logger.debug('Downloading %s' % url) filename = self.data['name'] format = '.%s' % self.data['format'] if format not in filename: filename = '%s%s' % (filename, format) with Download(full_agent=self.configuration.get_user_agent()) as downloader: path = downloader.download_file(url, folder, filename) return url, path
[ "def", "download", "(", "self", ",", "folder", "=", "None", ")", ":", "# type: (Optional[str]) -> Tuple[str, str]", "# Download the resource", "url", "=", "self", ".", "data", ".", "get", "(", "'url'", ",", "None", ")", "if", "not", "url", ":", "raise", "HDXError", "(", "'No URL to download!'", ")", "logger", ".", "debug", "(", "'Downloading %s'", "%", "url", ")", "filename", "=", "self", ".", "data", "[", "'name'", "]", "format", "=", "'.%s'", "%", "self", ".", "data", "[", "'format'", "]", "if", "format", "not", "in", "filename", ":", "filename", "=", "'%s%s'", "%", "(", "filename", ",", "format", ")", "with", "Download", "(", "full_agent", "=", "self", ".", "configuration", ".", "get_user_agent", "(", ")", ")", "as", "downloader", ":", "path", "=", "downloader", ".", "download_file", "(", "url", ",", "folder", ",", "filename", ")", "return", "url", ",", "path" ]
Download resource store to provided folder or temporary folder if no folder supplied Args: folder (Optional[str]): Folder to download resource to. Defaults to None. Returns: Tuple[str, str]: (URL downloaded, Path to downloaded file)
[ "Download", "resource", "store", "to", "provided", "folder", "or", "temporary", "folder", "if", "no", "folder", "supplied" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L270-L292
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.get_all_resource_ids_in_datastore
def get_all_resource_ids_in_datastore(configuration=None): # type: (Optional[Configuration]) -> List[str] """Get list of resources that have a datastore returning their ids. Args: configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: List[str]: List of resource ids that are in the datastore """ resource = Resource(configuration=configuration) success, result = resource._read_from_hdx('datastore', '_table_metadata', 'resource_id', Resource.actions()['datastore_search'], limit=10000) resource_ids = list() if not success: logger.debug(result) else: for record in result['records']: resource_ids.append(record['name']) return resource_ids
python
def get_all_resource_ids_in_datastore(configuration=None): # type: (Optional[Configuration]) -> List[str] """Get list of resources that have a datastore returning their ids. Args: configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: List[str]: List of resource ids that are in the datastore """ resource = Resource(configuration=configuration) success, result = resource._read_from_hdx('datastore', '_table_metadata', 'resource_id', Resource.actions()['datastore_search'], limit=10000) resource_ids = list() if not success: logger.debug(result) else: for record in result['records']: resource_ids.append(record['name']) return resource_ids
[ "def", "get_all_resource_ids_in_datastore", "(", "configuration", "=", "None", ")", ":", "# type: (Optional[Configuration]) -> List[str]", "resource", "=", "Resource", "(", "configuration", "=", "configuration", ")", "success", ",", "result", "=", "resource", ".", "_read_from_hdx", "(", "'datastore'", ",", "'_table_metadata'", ",", "'resource_id'", ",", "Resource", ".", "actions", "(", ")", "[", "'datastore_search'", "]", ",", "limit", "=", "10000", ")", "resource_ids", "=", "list", "(", ")", "if", "not", "success", ":", "logger", ".", "debug", "(", "result", ")", "else", ":", "for", "record", "in", "result", "[", "'records'", "]", ":", "resource_ids", ".", "append", "(", "record", "[", "'name'", "]", ")", "return", "resource_ids" ]
Get list of resources that have a datastore returning their ids. Args: configuration (Optional[Configuration]): HDX configuration. Defaults to global configuration. Returns: List[str]: List of resource ids that are in the datastore
[ "Get", "list", "of", "resources", "that", "have", "a", "datastore", "returning", "their", "ids", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L295-L314
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.has_datastore
def has_datastore(self): # type: () -> bool """Check if the resource has a datastore. Returns: bool: Whether the resource has a datastore or not """ success, result = self._read_from_hdx('datastore', self.data['id'], 'resource_id', self.actions()['datastore_search']) if not success: logger.debug(result) else: if result: return True return False
python
def has_datastore(self): # type: () -> bool """Check if the resource has a datastore. Returns: bool: Whether the resource has a datastore or not """ success, result = self._read_from_hdx('datastore', self.data['id'], 'resource_id', self.actions()['datastore_search']) if not success: logger.debug(result) else: if result: return True return False
[ "def", "has_datastore", "(", "self", ")", ":", "# type: () -> bool", "success", ",", "result", "=", "self", ".", "_read_from_hdx", "(", "'datastore'", ",", "self", ".", "data", "[", "'id'", "]", ",", "'resource_id'", ",", "self", ".", "actions", "(", ")", "[", "'datastore_search'", "]", ")", "if", "not", "success", ":", "logger", ".", "debug", "(", "result", ")", "else", ":", "if", "result", ":", "return", "True", "return", "False" ]
Check if the resource has a datastore. Returns: bool: Whether the resource has a datastore or not
[ "Check", "if", "the", "resource", "has", "a", "datastore", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L316-L330
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.delete_datastore
def delete_datastore(self): # type: () -> None """Delete a resource from the HDX datastore Returns: None """ success, result = self._read_from_hdx('datastore', self.data['id'], 'resource_id', self.actions()['datastore_delete'], force=True) if not success: logger.debug(result)
python
def delete_datastore(self): # type: () -> None """Delete a resource from the HDX datastore Returns: None """ success, result = self._read_from_hdx('datastore', self.data['id'], 'resource_id', self.actions()['datastore_delete'], force=True) if not success: logger.debug(result)
[ "def", "delete_datastore", "(", "self", ")", ":", "# type: () -> None", "success", ",", "result", "=", "self", ".", "_read_from_hdx", "(", "'datastore'", ",", "self", ".", "data", "[", "'id'", "]", ",", "'resource_id'", ",", "self", ".", "actions", "(", ")", "[", "'datastore_delete'", "]", ",", "force", "=", "True", ")", "if", "not", "success", ":", "logger", ".", "debug", "(", "result", ")" ]
Delete a resource from the HDX datastore Returns: None
[ "Delete", "a", "resource", "from", "the", "HDX", "datastore" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L332-L343
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.create_datastore
def create_datastore(self, schema=None, primary_key=None, delete_first=0, path=None): # type: (Optional[List[Dict]], Optional[str], int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX. If no schema is provided all fields are assumed to be text. If path is not supplied, the file is first downloaded from HDX. Args: schema (List[Dict]): List of fields and types of form {'id': 'FIELD', 'type': 'TYPE'}. Defaults to None. primary_key (Optional[str]): Primary key of schema. Defaults to None. delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ if delete_first == 0: pass elif delete_first == 1: self.delete_datastore() elif delete_first == 2: if primary_key is None: self.delete_datastore() else: raise HDXError('delete_first must be 0, 1 or 2! (0 = No, 1 = Yes, 2 = Delete if no primary key)') if path is None: # Download the resource url, path = self.download() delete_after_download = True else: url = path delete_after_download = False def convert_to_text(extended_rows): for number, headers, row in extended_rows: for i, val in enumerate(row): row[i] = str(val) yield (number, headers, row) with Download(full_agent=self.configuration.get_user_agent()) as downloader: try: stream = downloader.get_tabular_stream(path, headers=1, post_parse=[convert_to_text], bytes_sample_size=1000000) nonefieldname = False if schema is None: schema = list() for fieldname in stream.headers: if fieldname is not None: schema.append({'id': fieldname, 'type': 'text'}) else: nonefieldname = True data = {'resource_id': self.data['id'], 'force': True, 'fields': schema, 'primary_key': primary_key} self._write_to_hdx('datastore_create', data, 'resource_id') if primary_key is None: method = 'insert' else: method = 'upsert' logger.debug('Uploading data from %s to datastore' % url) offset = 0 chunksize = 100 rowset = stream.read(keyed=True, limit=chunksize) while len(rowset) != 0: if nonefieldname: for row in rowset: del row[None] data = {'resource_id': self.data['id'], 'force': True, 'method': method, 'records': rowset} self._write_to_hdx('datastore_upsert', data, 'resource_id') rowset = stream.read(keyed=True, limit=chunksize) logger.debug('Uploading: %s' % offset) offset += chunksize except Exception as e: raisefrom(HDXError, 'Upload to datastore of %s failed!' % url, e) finally: if delete_after_download: remove(path)
python
def create_datastore(self, schema=None, primary_key=None, delete_first=0, path=None): # type: (Optional[List[Dict]], Optional[str], int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX. If no schema is provided all fields are assumed to be text. If path is not supplied, the file is first downloaded from HDX. Args: schema (List[Dict]): List of fields and types of form {'id': 'FIELD', 'type': 'TYPE'}. Defaults to None. primary_key (Optional[str]): Primary key of schema. Defaults to None. delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ if delete_first == 0: pass elif delete_first == 1: self.delete_datastore() elif delete_first == 2: if primary_key is None: self.delete_datastore() else: raise HDXError('delete_first must be 0, 1 or 2! (0 = No, 1 = Yes, 2 = Delete if no primary key)') if path is None: # Download the resource url, path = self.download() delete_after_download = True else: url = path delete_after_download = False def convert_to_text(extended_rows): for number, headers, row in extended_rows: for i, val in enumerate(row): row[i] = str(val) yield (number, headers, row) with Download(full_agent=self.configuration.get_user_agent()) as downloader: try: stream = downloader.get_tabular_stream(path, headers=1, post_parse=[convert_to_text], bytes_sample_size=1000000) nonefieldname = False if schema is None: schema = list() for fieldname in stream.headers: if fieldname is not None: schema.append({'id': fieldname, 'type': 'text'}) else: nonefieldname = True data = {'resource_id': self.data['id'], 'force': True, 'fields': schema, 'primary_key': primary_key} self._write_to_hdx('datastore_create', data, 'resource_id') if primary_key is None: method = 'insert' else: method = 'upsert' logger.debug('Uploading data from %s to datastore' % url) offset = 0 chunksize = 100 rowset = stream.read(keyed=True, limit=chunksize) while len(rowset) != 0: if nonefieldname: for row in rowset: del row[None] data = {'resource_id': self.data['id'], 'force': True, 'method': method, 'records': rowset} self._write_to_hdx('datastore_upsert', data, 'resource_id') rowset = stream.read(keyed=True, limit=chunksize) logger.debug('Uploading: %s' % offset) offset += chunksize except Exception as e: raisefrom(HDXError, 'Upload to datastore of %s failed!' % url, e) finally: if delete_after_download: remove(path)
[ "def", "create_datastore", "(", "self", ",", "schema", "=", "None", ",", "primary_key", "=", "None", ",", "delete_first", "=", "0", ",", "path", "=", "None", ")", ":", "# type: (Optional[List[Dict]], Optional[str], int, Optional[str]) -> None", "if", "delete_first", "==", "0", ":", "pass", "elif", "delete_first", "==", "1", ":", "self", ".", "delete_datastore", "(", ")", "elif", "delete_first", "==", "2", ":", "if", "primary_key", "is", "None", ":", "self", ".", "delete_datastore", "(", ")", "else", ":", "raise", "HDXError", "(", "'delete_first must be 0, 1 or 2! (0 = No, 1 = Yes, 2 = Delete if no primary key)'", ")", "if", "path", "is", "None", ":", "# Download the resource", "url", ",", "path", "=", "self", ".", "download", "(", ")", "delete_after_download", "=", "True", "else", ":", "url", "=", "path", "delete_after_download", "=", "False", "def", "convert_to_text", "(", "extended_rows", ")", ":", "for", "number", ",", "headers", ",", "row", "in", "extended_rows", ":", "for", "i", ",", "val", "in", "enumerate", "(", "row", ")", ":", "row", "[", "i", "]", "=", "str", "(", "val", ")", "yield", "(", "number", ",", "headers", ",", "row", ")", "with", "Download", "(", "full_agent", "=", "self", ".", "configuration", ".", "get_user_agent", "(", ")", ")", "as", "downloader", ":", "try", ":", "stream", "=", "downloader", ".", "get_tabular_stream", "(", "path", ",", "headers", "=", "1", ",", "post_parse", "=", "[", "convert_to_text", "]", ",", "bytes_sample_size", "=", "1000000", ")", "nonefieldname", "=", "False", "if", "schema", "is", "None", ":", "schema", "=", "list", "(", ")", "for", "fieldname", "in", "stream", ".", "headers", ":", "if", "fieldname", "is", "not", "None", ":", "schema", ".", "append", "(", "{", "'id'", ":", "fieldname", ",", "'type'", ":", "'text'", "}", ")", "else", ":", "nonefieldname", "=", "True", "data", "=", "{", "'resource_id'", ":", "self", ".", "data", "[", "'id'", "]", ",", "'force'", ":", "True", ",", "'fields'", ":", "schema", ",", "'primary_key'", ":", "primary_key", "}", "self", ".", "_write_to_hdx", "(", "'datastore_create'", ",", "data", ",", "'resource_id'", ")", "if", "primary_key", "is", "None", ":", "method", "=", "'insert'", "else", ":", "method", "=", "'upsert'", "logger", ".", "debug", "(", "'Uploading data from %s to datastore'", "%", "url", ")", "offset", "=", "0", "chunksize", "=", "100", "rowset", "=", "stream", ".", "read", "(", "keyed", "=", "True", ",", "limit", "=", "chunksize", ")", "while", "len", "(", "rowset", ")", "!=", "0", ":", "if", "nonefieldname", ":", "for", "row", "in", "rowset", ":", "del", "row", "[", "None", "]", "data", "=", "{", "'resource_id'", ":", "self", ".", "data", "[", "'id'", "]", ",", "'force'", ":", "True", ",", "'method'", ":", "method", ",", "'records'", ":", "rowset", "}", "self", ".", "_write_to_hdx", "(", "'datastore_upsert'", ",", "data", ",", "'resource_id'", ")", "rowset", "=", "stream", ".", "read", "(", "keyed", "=", "True", ",", "limit", "=", "chunksize", ")", "logger", ".", "debug", "(", "'Uploading: %s'", "%", "offset", ")", "offset", "+=", "chunksize", "except", "Exception", "as", "e", ":", "raisefrom", "(", "HDXError", ",", "'Upload to datastore of %s failed!'", "%", "url", ",", "e", ")", "finally", ":", "if", "delete_after_download", ":", "remove", "(", "path", ")" ]
For tabular data, create a resource in the HDX datastore which enables data preview in HDX. If no schema is provided all fields are assumed to be text. If path is not supplied, the file is first downloaded from HDX. Args: schema (List[Dict]): List of fields and types of form {'id': 'FIELD', 'type': 'TYPE'}. Defaults to None. primary_key (Optional[str]): Primary key of schema. Defaults to None. delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None
[ "For", "tabular", "data", "create", "a", "resource", "in", "the", "HDX", "datastore", "which", "enables", "data", "preview", "in", "HDX", ".", "If", "no", "schema", "is", "provided", "all", "fields", "are", "assumed", "to", "be", "text", ".", "If", "path", "is", "not", "supplied", "the", "file", "is", "first", "downloaded", "from", "HDX", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L345-L418
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.create_datastore_from_dict_schema
def create_datastore_from_dict_schema(self, data, delete_first=0, path=None): # type: (dict, int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a dictionary containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: data (dict): Dictionary containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ schema = data['schema'] primary_key = data.get('primary_key') self.create_datastore(schema, primary_key, delete_first, path=path)
python
def create_datastore_from_dict_schema(self, data, delete_first=0, path=None): # type: (dict, int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a dictionary containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: data (dict): Dictionary containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ schema = data['schema'] primary_key = data.get('primary_key') self.create_datastore(schema, primary_key, delete_first, path=path)
[ "def", "create_datastore_from_dict_schema", "(", "self", ",", "data", ",", "delete_first", "=", "0", ",", "path", "=", "None", ")", ":", "# type: (dict, int, Optional[str]) -> None", "schema", "=", "data", "[", "'schema'", "]", "primary_key", "=", "data", ".", "get", "(", "'primary_key'", ")", "self", ".", "create_datastore", "(", "schema", ",", "primary_key", ",", "delete_first", ",", "path", "=", "path", ")" ]
For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a dictionary containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: data (dict): Dictionary containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None
[ "For", "tabular", "data", "create", "a", "resource", "in", "the", "HDX", "datastore", "which", "enables", "data", "preview", "in", "HDX", "from", "a", "dictionary", "containing", "a", "list", "of", "fields", "and", "types", "of", "form", "{", "id", ":", "FIELD", "type", ":", "TYPE", "}", "and", "optionally", "a", "primary", "key", ".", "If", "path", "is", "not", "supplied", "the", "file", "is", "first", "downloaded", "from", "HDX", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L420-L436
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.create_datastore_from_yaml_schema
def create_datastore_from_yaml_schema(self, yaml_path, delete_first=0, path=None): # type: (str, Optional[int], Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a YAML file containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: yaml_path (str): Path to YAML file containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ data = load_yaml(yaml_path) self.create_datastore_from_dict_schema(data, delete_first, path=path)
python
def create_datastore_from_yaml_schema(self, yaml_path, delete_first=0, path=None): # type: (str, Optional[int], Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a YAML file containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: yaml_path (str): Path to YAML file containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ data = load_yaml(yaml_path) self.create_datastore_from_dict_schema(data, delete_first, path=path)
[ "def", "create_datastore_from_yaml_schema", "(", "self", ",", "yaml_path", ",", "delete_first", "=", "0", ",", "path", "=", "None", ")", ":", "# type: (str, Optional[int], Optional[str]) -> None", "data", "=", "load_yaml", "(", "yaml_path", ")", "self", ".", "create_datastore_from_dict_schema", "(", "data", ",", "delete_first", ",", "path", "=", "path", ")" ]
For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a YAML file containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: yaml_path (str): Path to YAML file containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None
[ "For", "tabular", "data", "create", "a", "resource", "in", "the", "HDX", "datastore", "which", "enables", "data", "preview", "in", "HDX", "from", "a", "YAML", "file", "containing", "a", "list", "of", "fields", "and", "types", "of", "form", "{", "id", ":", "FIELD", "type", ":", "TYPE", "}", "and", "optionally", "a", "primary", "key", ".", "If", "path", "is", "not", "supplied", "the", "file", "is", "first", "downloaded", "from", "HDX", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L438-L454
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.create_datastore_from_json_schema
def create_datastore_from_json_schema(self, json_path, delete_first=0, path=None): # type: (str, int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a JSON file containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: json_path (str): Path to JSON file containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ data = load_json(json_path) self.create_datastore_from_dict_schema(data, delete_first, path=path)
python
def create_datastore_from_json_schema(self, json_path, delete_first=0, path=None): # type: (str, int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a JSON file containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: json_path (str): Path to JSON file containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ data = load_json(json_path) self.create_datastore_from_dict_schema(data, delete_first, path=path)
[ "def", "create_datastore_from_json_schema", "(", "self", ",", "json_path", ",", "delete_first", "=", "0", ",", "path", "=", "None", ")", ":", "# type: (str, int, Optional[str]) -> None", "data", "=", "load_json", "(", "json_path", ")", "self", ".", "create_datastore_from_dict_schema", "(", "data", ",", "delete_first", ",", "path", "=", "path", ")" ]
For tabular data, create a resource in the HDX datastore which enables data preview in HDX from a JSON file containing a list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} and optionally a primary key. If path is not supplied, the file is first downloaded from HDX. Args: json_path (str): Path to JSON file containing list of fields and types of form {'id': 'FIELD', 'type': 'TYPE'} delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None
[ "For", "tabular", "data", "create", "a", "resource", "in", "the", "HDX", "datastore", "which", "enables", "data", "preview", "in", "HDX", "from", "a", "JSON", "file", "containing", "a", "list", "of", "fields", "and", "types", "of", "form", "{", "id", ":", "FIELD", "type", ":", "TYPE", "}", "and", "optionally", "a", "primary", "key", ".", "If", "path", "is", "not", "supplied", "the", "file", "is", "first", "downloaded", "from", "HDX", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L456-L471
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.create_datastore_for_topline
def create_datastore_for_topline(self, delete_first=0, path=None): # type: (int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX using the built in YAML definition for a topline. If path is not supplied, the file is first downloaded from HDX. Args: delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ data = load_yaml(script_dir_plus_file(join('..', 'hdx_datasource_topline.yml'), Resource)) self.create_datastore_from_dict_schema(data, delete_first, path=path)
python
def create_datastore_for_topline(self, delete_first=0, path=None): # type: (int, Optional[str]) -> None """For tabular data, create a resource in the HDX datastore which enables data preview in HDX using the built in YAML definition for a topline. If path is not supplied, the file is first downloaded from HDX. Args: delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ data = load_yaml(script_dir_plus_file(join('..', 'hdx_datasource_topline.yml'), Resource)) self.create_datastore_from_dict_schema(data, delete_first, path=path)
[ "def", "create_datastore_for_topline", "(", "self", ",", "delete_first", "=", "0", ",", "path", "=", "None", ")", ":", "# type: (int, Optional[str]) -> None", "data", "=", "load_yaml", "(", "script_dir_plus_file", "(", "join", "(", "'..'", ",", "'hdx_datasource_topline.yml'", ")", ",", "Resource", ")", ")", "self", ".", "create_datastore_from_dict_schema", "(", "data", ",", "delete_first", ",", "path", "=", "path", ")" ]
For tabular data, create a resource in the HDX datastore which enables data preview in HDX using the built in YAML definition for a topline. If path is not supplied, the file is first downloaded from HDX. Args: delete_first (int): Delete datastore before creation. 0 = No, 1 = Yes, 2 = If no primary key. Defaults to 0. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None
[ "For", "tabular", "data", "create", "a", "resource", "in", "the", "HDX", "datastore", "which", "enables", "data", "preview", "in", "HDX", "using", "the", "built", "in", "YAML", "definition", "for", "a", "topline", ".", "If", "path", "is", "not", "supplied", "the", "file", "is", "first", "downloaded", "from", "HDX", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L473-L486
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.update_datastore
def update_datastore(self, schema=None, primary_key=None, path=None): # type: (Optional[List[Dict]], Optional[str], Optional[str]) -> None """For tabular data, update a resource in the HDX datastore which enables data preview in HDX. If no schema is provided all fields are assumed to be text. If path is not supplied, the file is first downloaded from HDX. Args: schema (List[Dict]): List of fields and types of form {'id': 'FIELD', 'type': 'TYPE'}. Defaults to None. primary_key (Optional[str]): Primary key of schema. Defaults to None. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ self.create_datastore(schema, primary_key, 2, path=path)
python
def update_datastore(self, schema=None, primary_key=None, path=None): # type: (Optional[List[Dict]], Optional[str], Optional[str]) -> None """For tabular data, update a resource in the HDX datastore which enables data preview in HDX. If no schema is provided all fields are assumed to be text. If path is not supplied, the file is first downloaded from HDX. Args: schema (List[Dict]): List of fields and types of form {'id': 'FIELD', 'type': 'TYPE'}. Defaults to None. primary_key (Optional[str]): Primary key of schema. Defaults to None. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None """ self.create_datastore(schema, primary_key, 2, path=path)
[ "def", "update_datastore", "(", "self", ",", "schema", "=", "None", ",", "primary_key", "=", "None", ",", "path", "=", "None", ")", ":", "# type: (Optional[List[Dict]], Optional[str], Optional[str]) -> None", "self", ".", "create_datastore", "(", "schema", ",", "primary_key", ",", "2", ",", "path", "=", "path", ")" ]
For tabular data, update a resource in the HDX datastore which enables data preview in HDX. If no schema is provided all fields are assumed to be text. If path is not supplied, the file is first downloaded from HDX. Args: schema (List[Dict]): List of fields and types of form {'id': 'FIELD', 'type': 'TYPE'}. Defaults to None. primary_key (Optional[str]): Primary key of schema. Defaults to None. path (Optional[str]): Local path to file that was uploaded. Defaults to None. Returns: None
[ "For", "tabular", "data", "update", "a", "resource", "in", "the", "HDX", "datastore", "which", "enables", "data", "preview", "in", "HDX", ".", "If", "no", "schema", "is", "provided", "all", "fields", "are", "assumed", "to", "be", "text", ".", "If", "path", "is", "not", "supplied", "the", "file", "is", "first", "downloaded", "from", "HDX", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L488-L502
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource._get_resource_view
def _get_resource_view(self, resource_view): # type: (Union[ResourceView,Dict]) -> ResourceView """Get resource view id Args: resource_view (Union[ResourceView,Dict]): ResourceView metadata from a ResourceView object or dictionary Returns: ResourceView: ResourceView object """ if isinstance(resource_view, dict): resource_view = ResourceView(resource_view, configuration=self.configuration) if isinstance(resource_view, ResourceView): return resource_view raise HDXError('Type %s is not a valid resource view!' % type(resource_view).__name__)
python
def _get_resource_view(self, resource_view): # type: (Union[ResourceView,Dict]) -> ResourceView """Get resource view id Args: resource_view (Union[ResourceView,Dict]): ResourceView metadata from a ResourceView object or dictionary Returns: ResourceView: ResourceView object """ if isinstance(resource_view, dict): resource_view = ResourceView(resource_view, configuration=self.configuration) if isinstance(resource_view, ResourceView): return resource_view raise HDXError('Type %s is not a valid resource view!' % type(resource_view).__name__)
[ "def", "_get_resource_view", "(", "self", ",", "resource_view", ")", ":", "# type: (Union[ResourceView,Dict]) -> ResourceView", "if", "isinstance", "(", "resource_view", ",", "dict", ")", ":", "resource_view", "=", "ResourceView", "(", "resource_view", ",", "configuration", "=", "self", ".", "configuration", ")", "if", "isinstance", "(", "resource_view", ",", "ResourceView", ")", ":", "return", "resource_view", "raise", "HDXError", "(", "'Type %s is not a valid resource view!'", "%", "type", "(", "resource_view", ")", ".", "__name__", ")" ]
Get resource view id Args: resource_view (Union[ResourceView,Dict]): ResourceView metadata from a ResourceView object or dictionary Returns: ResourceView: ResourceView object
[ "Get", "resource", "view", "id" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L571-L585
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.add_update_resource_views
def add_update_resource_views(self, resource_views): # type: (List[Union[ResourceView,Dict]]) -> None """Add new or update existing resource views in resource with new metadata. Args: resource_views (List[Union[ResourceView,Dict]]): A list of resource views metadata from ResourceView objects or dictionaries Returns: None """ if not isinstance(resource_views, list): raise HDXError('ResourceViews should be a list!') for resource_view in resource_views: self.add_update_resource_view(resource_view)
python
def add_update_resource_views(self, resource_views): # type: (List[Union[ResourceView,Dict]]) -> None """Add new or update existing resource views in resource with new metadata. Args: resource_views (List[Union[ResourceView,Dict]]): A list of resource views metadata from ResourceView objects or dictionaries Returns: None """ if not isinstance(resource_views, list): raise HDXError('ResourceViews should be a list!') for resource_view in resource_views: self.add_update_resource_view(resource_view)
[ "def", "add_update_resource_views", "(", "self", ",", "resource_views", ")", ":", "# type: (List[Union[ResourceView,Dict]]) -> None", "if", "not", "isinstance", "(", "resource_views", ",", "list", ")", ":", "raise", "HDXError", "(", "'ResourceViews should be a list!'", ")", "for", "resource_view", "in", "resource_views", ":", "self", ".", "add_update_resource_view", "(", "resource_view", ")" ]
Add new or update existing resource views in resource with new metadata. Args: resource_views (List[Union[ResourceView,Dict]]): A list of resource views metadata from ResourceView objects or dictionaries Returns: None
[ "Add", "new", "or", "update", "existing", "resource", "views", "in", "resource", "with", "new", "metadata", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L601-L614
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.reorder_resource_views
def reorder_resource_views(self, resource_views): # type: (List[Union[ResourceView,Dict,str]]) -> None """Order resource views in resource. Args: resource_views (List[Union[ResourceView,Dict,str]]): A list of either resource view ids or resource views metadata from ResourceView objects or dictionaries Returns: None """ if not isinstance(resource_views, list): raise HDXError('ResourceViews should be a list!') ids = list() for resource_view in resource_views: if isinstance(resource_view, str): resource_view_id = resource_view else: resource_view_id = resource_view['id'] if is_valid_uuid(resource_view_id) is False: raise HDXError('%s is not a valid resource view id!' % resource_view) ids.append(resource_view_id) _, result = self._read_from_hdx('resource view', self.data['id'], 'id', ResourceView.actions()['reorder'], order=ids)
python
def reorder_resource_views(self, resource_views): # type: (List[Union[ResourceView,Dict,str]]) -> None """Order resource views in resource. Args: resource_views (List[Union[ResourceView,Dict,str]]): A list of either resource view ids or resource views metadata from ResourceView objects or dictionaries Returns: None """ if not isinstance(resource_views, list): raise HDXError('ResourceViews should be a list!') ids = list() for resource_view in resource_views: if isinstance(resource_view, str): resource_view_id = resource_view else: resource_view_id = resource_view['id'] if is_valid_uuid(resource_view_id) is False: raise HDXError('%s is not a valid resource view id!' % resource_view) ids.append(resource_view_id) _, result = self._read_from_hdx('resource view', self.data['id'], 'id', ResourceView.actions()['reorder'], order=ids)
[ "def", "reorder_resource_views", "(", "self", ",", "resource_views", ")", ":", "# type: (List[Union[ResourceView,Dict,str]]) -> None", "if", "not", "isinstance", "(", "resource_views", ",", "list", ")", ":", "raise", "HDXError", "(", "'ResourceViews should be a list!'", ")", "ids", "=", "list", "(", ")", "for", "resource_view", "in", "resource_views", ":", "if", "isinstance", "(", "resource_view", ",", "str", ")", ":", "resource_view_id", "=", "resource_view", "else", ":", "resource_view_id", "=", "resource_view", "[", "'id'", "]", "if", "is_valid_uuid", "(", "resource_view_id", ")", "is", "False", ":", "raise", "HDXError", "(", "'%s is not a valid resource view id!'", "%", "resource_view", ")", "ids", ".", "append", "(", "resource_view_id", ")", "_", ",", "result", "=", "self", ".", "_read_from_hdx", "(", "'resource view'", ",", "self", ".", "data", "[", "'id'", "]", ",", "'id'", ",", "ResourceView", ".", "actions", "(", ")", "[", "'reorder'", "]", ",", "order", "=", "ids", ")" ]
Order resource views in resource. Args: resource_views (List[Union[ResourceView,Dict,str]]): A list of either resource view ids or resource views metadata from ResourceView objects or dictionaries Returns: None
[ "Order", "resource", "views", "in", "resource", "." ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L616-L638
OCHA-DAP/hdx-python-api
src/hdx/data/resource.py
Resource.delete_resource_view
def delete_resource_view(self, resource_view): # type: (Union[ResourceView,Dict,str]) -> None """Delete a resource view from the resource and HDX Args: resource_view (Union[ResourceView,Dict,str]): Either a resource view id or resource view metadata either from a ResourceView object or a dictionary Returns: None """ if isinstance(resource_view, str): if is_valid_uuid(resource_view) is False: raise HDXError('%s is not a valid resource view id!' % resource_view) resource_view = ResourceView({'id': resource_view}, configuration=self.configuration) else: resource_view = self._get_resource_view(resource_view) if 'id' not in resource_view: found = False title = resource_view.get('title') for rv in self.get_resource_views(): if resource_view['title'] == rv['title']: resource_view = rv found = True break if not found: raise HDXError('No resource views have title %s in this resource!' % title) resource_view.delete_from_hdx()
python
def delete_resource_view(self, resource_view): # type: (Union[ResourceView,Dict,str]) -> None """Delete a resource view from the resource and HDX Args: resource_view (Union[ResourceView,Dict,str]): Either a resource view id or resource view metadata either from a ResourceView object or a dictionary Returns: None """ if isinstance(resource_view, str): if is_valid_uuid(resource_view) is False: raise HDXError('%s is not a valid resource view id!' % resource_view) resource_view = ResourceView({'id': resource_view}, configuration=self.configuration) else: resource_view = self._get_resource_view(resource_view) if 'id' not in resource_view: found = False title = resource_view.get('title') for rv in self.get_resource_views(): if resource_view['title'] == rv['title']: resource_view = rv found = True break if not found: raise HDXError('No resource views have title %s in this resource!' % title) resource_view.delete_from_hdx()
[ "def", "delete_resource_view", "(", "self", ",", "resource_view", ")", ":", "# type: (Union[ResourceView,Dict,str]) -> None", "if", "isinstance", "(", "resource_view", ",", "str", ")", ":", "if", "is_valid_uuid", "(", "resource_view", ")", "is", "False", ":", "raise", "HDXError", "(", "'%s is not a valid resource view id!'", "%", "resource_view", ")", "resource_view", "=", "ResourceView", "(", "{", "'id'", ":", "resource_view", "}", ",", "configuration", "=", "self", ".", "configuration", ")", "else", ":", "resource_view", "=", "self", ".", "_get_resource_view", "(", "resource_view", ")", "if", "'id'", "not", "in", "resource_view", ":", "found", "=", "False", "title", "=", "resource_view", ".", "get", "(", "'title'", ")", "for", "rv", "in", "self", ".", "get_resource_views", "(", ")", ":", "if", "resource_view", "[", "'title'", "]", "==", "rv", "[", "'title'", "]", ":", "resource_view", "=", "rv", "found", "=", "True", "break", "if", "not", "found", ":", "raise", "HDXError", "(", "'No resource views have title %s in this resource!'", "%", "title", ")", "resource_view", ".", "delete_from_hdx", "(", ")" ]
Delete a resource view from the resource and HDX Args: resource_view (Union[ResourceView,Dict,str]): Either a resource view id or resource view metadata either from a ResourceView object or a dictionary Returns: None
[ "Delete", "a", "resource", "view", "from", "the", "resource", "and", "HDX" ]
train
https://github.com/OCHA-DAP/hdx-python-api/blob/212440f54f73805826a16db77dbcb6033b18a313/src/hdx/data/resource.py#L640-L666
dmirecki/pyMorfologik
pymorfologik/output_parser.py
parse_for_simple_stems
def parse_for_simple_stems(output, skip_empty=False, skip_same_stems=True): """ Parses the output stem lines to produce a list with possible stems for each word in the output. :param skip_empty: set True to skip lines without stems (default is False) :returns: a list of tuples, each containing an original text word and a list of stems for the given word """ lines_with_stems = _get_lines_with_stems(output) stems = list() last_word = None for line in lines_with_stems: word, stem, _ = line.split("\t") stem = stem if stem != '-' else None if skip_empty and (stem is None): continue if last_word != word: stems.append((word, [])) ## append new stem only if not on list already stem = None if skip_same_stems and stem in stems[-1][1] else stem if stem is not None: stems[-1][1].append(stem) last_word = word return stems
python
def parse_for_simple_stems(output, skip_empty=False, skip_same_stems=True): """ Parses the output stem lines to produce a list with possible stems for each word in the output. :param skip_empty: set True to skip lines without stems (default is False) :returns: a list of tuples, each containing an original text word and a list of stems for the given word """ lines_with_stems = _get_lines_with_stems(output) stems = list() last_word = None for line in lines_with_stems: word, stem, _ = line.split("\t") stem = stem if stem != '-' else None if skip_empty and (stem is None): continue if last_word != word: stems.append((word, [])) ## append new stem only if not on list already stem = None if skip_same_stems and stem in stems[-1][1] else stem if stem is not None: stems[-1][1].append(stem) last_word = word return stems
[ "def", "parse_for_simple_stems", "(", "output", ",", "skip_empty", "=", "False", ",", "skip_same_stems", "=", "True", ")", ":", "lines_with_stems", "=", "_get_lines_with_stems", "(", "output", ")", "stems", "=", "list", "(", ")", "last_word", "=", "None", "for", "line", "in", "lines_with_stems", ":", "word", ",", "stem", ",", "_", "=", "line", ".", "split", "(", "\"\\t\"", ")", "stem", "=", "stem", "if", "stem", "!=", "'-'", "else", "None", "if", "skip_empty", "and", "(", "stem", "is", "None", ")", ":", "continue", "if", "last_word", "!=", "word", ":", "stems", ".", "append", "(", "(", "word", ",", "[", "]", ")", ")", "## append new stem only if not on list already", "stem", "=", "None", "if", "skip_same_stems", "and", "stem", "in", "stems", "[", "-", "1", "]", "[", "1", "]", "else", "stem", "if", "stem", "is", "not", "None", ":", "stems", "[", "-", "1", "]", "[", "1", "]", ".", "append", "(", "stem", ")", "last_word", "=", "word", "return", "stems" ]
Parses the output stem lines to produce a list with possible stems for each word in the output. :param skip_empty: set True to skip lines without stems (default is False) :returns: a list of tuples, each containing an original text word and a list of stems for the given word
[ "Parses", "the", "output", "stem", "lines", "to", "produce", "a", "list", "with", "possible", "stems", "for", "each", "word", "in", "the", "output", "." ]
train
https://github.com/dmirecki/pyMorfologik/blob/e4d93a82e8b4c7a108f01e0456fbeb8024df0259/pymorfologik/output_parser.py#L19-L49
libyal/dtfabric
dtfabric/runtime/runtime.py
StructureValuesClassFactory._CreateClassTemplate
def _CreateClassTemplate(cls, data_type_definition): """Creates the class template. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: str: class template. """ type_name = data_type_definition.name type_description = data_type_definition.description or type_name while type_description.endswith('.'): type_description = type_description[:-1] class_attributes_description = [] init_arguments = [] instance_attributes = [] for member_definition in data_type_definition.members: attribute_name = member_definition.name description = member_definition.description or attribute_name while description.endswith('.'): description = description[:-1] member_data_type = getattr(member_definition, 'member_data_type', '') if isinstance(member_definition, data_types.MemberDataTypeDefinition): member_definition = member_definition.member_data_type_definition member_type_indicator = member_definition.TYPE_INDICATOR if member_type_indicator == definitions.TYPE_INDICATOR_SEQUENCE: element_type_indicator = member_definition.element_data_type member_type_indicator = 'tuple[{0:s}]'.format(element_type_indicator) else: member_type_indicator = cls._PYTHON_NATIVE_TYPES.get( member_type_indicator, member_data_type) argument = '{0:s}=None'.format(attribute_name) definition = ' self.{0:s} = {0:s}'.format(attribute_name) description = ' {0:s} ({1:s}): {2:s}.'.format( attribute_name, member_type_indicator, description) class_attributes_description.append(description) init_arguments.append(argument) instance_attributes.append(definition) class_attributes_description = '\n'.join( sorted(class_attributes_description)) init_arguments = ', '.join(init_arguments) instance_attributes = '\n'.join(sorted(instance_attributes)) template_values = { 'class_attributes_description': class_attributes_description, 'init_arguments': init_arguments, 'instance_attributes': instance_attributes, 'type_description': type_description, 'type_name': type_name} return cls._CLASS_TEMPLATE.format(**template_values)
python
def _CreateClassTemplate(cls, data_type_definition): """Creates the class template. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: str: class template. """ type_name = data_type_definition.name type_description = data_type_definition.description or type_name while type_description.endswith('.'): type_description = type_description[:-1] class_attributes_description = [] init_arguments = [] instance_attributes = [] for member_definition in data_type_definition.members: attribute_name = member_definition.name description = member_definition.description or attribute_name while description.endswith('.'): description = description[:-1] member_data_type = getattr(member_definition, 'member_data_type', '') if isinstance(member_definition, data_types.MemberDataTypeDefinition): member_definition = member_definition.member_data_type_definition member_type_indicator = member_definition.TYPE_INDICATOR if member_type_indicator == definitions.TYPE_INDICATOR_SEQUENCE: element_type_indicator = member_definition.element_data_type member_type_indicator = 'tuple[{0:s}]'.format(element_type_indicator) else: member_type_indicator = cls._PYTHON_NATIVE_TYPES.get( member_type_indicator, member_data_type) argument = '{0:s}=None'.format(attribute_name) definition = ' self.{0:s} = {0:s}'.format(attribute_name) description = ' {0:s} ({1:s}): {2:s}.'.format( attribute_name, member_type_indicator, description) class_attributes_description.append(description) init_arguments.append(argument) instance_attributes.append(definition) class_attributes_description = '\n'.join( sorted(class_attributes_description)) init_arguments = ', '.join(init_arguments) instance_attributes = '\n'.join(sorted(instance_attributes)) template_values = { 'class_attributes_description': class_attributes_description, 'init_arguments': init_arguments, 'instance_attributes': instance_attributes, 'type_description': type_description, 'type_name': type_name} return cls._CLASS_TEMPLATE.format(**template_values)
[ "def", "_CreateClassTemplate", "(", "cls", ",", "data_type_definition", ")", ":", "type_name", "=", "data_type_definition", ".", "name", "type_description", "=", "data_type_definition", ".", "description", "or", "type_name", "while", "type_description", ".", "endswith", "(", "'.'", ")", ":", "type_description", "=", "type_description", "[", ":", "-", "1", "]", "class_attributes_description", "=", "[", "]", "init_arguments", "=", "[", "]", "instance_attributes", "=", "[", "]", "for", "member_definition", "in", "data_type_definition", ".", "members", ":", "attribute_name", "=", "member_definition", ".", "name", "description", "=", "member_definition", ".", "description", "or", "attribute_name", "while", "description", ".", "endswith", "(", "'.'", ")", ":", "description", "=", "description", "[", ":", "-", "1", "]", "member_data_type", "=", "getattr", "(", "member_definition", ",", "'member_data_type'", ",", "''", ")", "if", "isinstance", "(", "member_definition", ",", "data_types", ".", "MemberDataTypeDefinition", ")", ":", "member_definition", "=", "member_definition", ".", "member_data_type_definition", "member_type_indicator", "=", "member_definition", ".", "TYPE_INDICATOR", "if", "member_type_indicator", "==", "definitions", ".", "TYPE_INDICATOR_SEQUENCE", ":", "element_type_indicator", "=", "member_definition", ".", "element_data_type", "member_type_indicator", "=", "'tuple[{0:s}]'", ".", "format", "(", "element_type_indicator", ")", "else", ":", "member_type_indicator", "=", "cls", ".", "_PYTHON_NATIVE_TYPES", ".", "get", "(", "member_type_indicator", ",", "member_data_type", ")", "argument", "=", "'{0:s}=None'", ".", "format", "(", "attribute_name", ")", "definition", "=", "' self.{0:s} = {0:s}'", ".", "format", "(", "attribute_name", ")", "description", "=", "' {0:s} ({1:s}): {2:s}.'", ".", "format", "(", "attribute_name", ",", "member_type_indicator", ",", "description", ")", "class_attributes_description", ".", "append", "(", "description", ")", "init_arguments", ".", "append", "(", "argument", ")", "instance_attributes", ".", "append", "(", "definition", ")", "class_attributes_description", "=", "'\\n'", ".", "join", "(", "sorted", "(", "class_attributes_description", ")", ")", "init_arguments", "=", "', '", ".", "join", "(", "init_arguments", ")", "instance_attributes", "=", "'\\n'", ".", "join", "(", "sorted", "(", "instance_attributes", ")", ")", "template_values", "=", "{", "'class_attributes_description'", ":", "class_attributes_description", ",", "'init_arguments'", ":", "init_arguments", ",", "'instance_attributes'", ":", "instance_attributes", ",", "'type_description'", ":", "type_description", ",", "'type_name'", ":", "type_name", "}", "return", "cls", ".", "_CLASS_TEMPLATE", ".", "format", "(", "*", "*", "template_values", ")" ]
Creates the class template. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: str: class template.
[ "Creates", "the", "class", "template", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/runtime.py#L45-L106
libyal/dtfabric
dtfabric/runtime/runtime.py
StructureValuesClassFactory._IsIdentifier
def _IsIdentifier(cls, string): """Checks if a string contains an identifier. Args: string (str): string to check. Returns: bool: True if the string contains an identifier, False otherwise. """ return ( string and not string[0].isdigit() and all(character.isalnum() or character == '_' for character in string))
python
def _IsIdentifier(cls, string): """Checks if a string contains an identifier. Args: string (str): string to check. Returns: bool: True if the string contains an identifier, False otherwise. """ return ( string and not string[0].isdigit() and all(character.isalnum() or character == '_' for character in string))
[ "def", "_IsIdentifier", "(", "cls", ",", "string", ")", ":", "return", "(", "string", "and", "not", "string", "[", "0", "]", ".", "isdigit", "(", ")", "and", "all", "(", "character", ".", "isalnum", "(", ")", "or", "character", "==", "'_'", "for", "character", "in", "string", ")", ")" ]
Checks if a string contains an identifier. Args: string (str): string to check. Returns: bool: True if the string contains an identifier, False otherwise.
[ "Checks", "if", "a", "string", "contains", "an", "identifier", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/runtime.py#L109-L120
libyal/dtfabric
dtfabric/runtime/runtime.py
StructureValuesClassFactory._ValidateDataTypeDefinition
def _ValidateDataTypeDefinition(cls, data_type_definition): """Validates the data type definition. Args: data_type_definition (DataTypeDefinition): data type definition. Raises: ValueError: if the data type definition is not considered valid. """ if not cls._IsIdentifier(data_type_definition.name): raise ValueError( 'Data type definition name: {0!s} not a valid identifier'.format( data_type_definition.name)) if keyword.iskeyword(data_type_definition.name): raise ValueError( 'Data type definition name: {0!s} matches keyword'.format( data_type_definition.name)) members = getattr(data_type_definition, 'members', None) if not members: raise ValueError( 'Data type definition name: {0!s} missing members'.format( data_type_definition.name)) defined_attribute_names = set() for member_definition in members: attribute_name = member_definition.name if not cls._IsIdentifier(attribute_name): raise ValueError('Attribute name: {0!s} not a valid identifier'.format( attribute_name)) if attribute_name.startswith('_'): raise ValueError('Attribute name: {0!s} starts with underscore'.format( attribute_name)) if keyword.iskeyword(attribute_name): raise ValueError('Attribute name: {0!s} matches keyword'.format( attribute_name)) if attribute_name in defined_attribute_names: raise ValueError('Attribute name: {0!s} already defined'.format( attribute_name)) defined_attribute_names.add(attribute_name)
python
def _ValidateDataTypeDefinition(cls, data_type_definition): """Validates the data type definition. Args: data_type_definition (DataTypeDefinition): data type definition. Raises: ValueError: if the data type definition is not considered valid. """ if not cls._IsIdentifier(data_type_definition.name): raise ValueError( 'Data type definition name: {0!s} not a valid identifier'.format( data_type_definition.name)) if keyword.iskeyword(data_type_definition.name): raise ValueError( 'Data type definition name: {0!s} matches keyword'.format( data_type_definition.name)) members = getattr(data_type_definition, 'members', None) if not members: raise ValueError( 'Data type definition name: {0!s} missing members'.format( data_type_definition.name)) defined_attribute_names = set() for member_definition in members: attribute_name = member_definition.name if not cls._IsIdentifier(attribute_name): raise ValueError('Attribute name: {0!s} not a valid identifier'.format( attribute_name)) if attribute_name.startswith('_'): raise ValueError('Attribute name: {0!s} starts with underscore'.format( attribute_name)) if keyword.iskeyword(attribute_name): raise ValueError('Attribute name: {0!s} matches keyword'.format( attribute_name)) if attribute_name in defined_attribute_names: raise ValueError('Attribute name: {0!s} already defined'.format( attribute_name)) defined_attribute_names.add(attribute_name)
[ "def", "_ValidateDataTypeDefinition", "(", "cls", ",", "data_type_definition", ")", ":", "if", "not", "cls", ".", "_IsIdentifier", "(", "data_type_definition", ".", "name", ")", ":", "raise", "ValueError", "(", "'Data type definition name: {0!s} not a valid identifier'", ".", "format", "(", "data_type_definition", ".", "name", ")", ")", "if", "keyword", ".", "iskeyword", "(", "data_type_definition", ".", "name", ")", ":", "raise", "ValueError", "(", "'Data type definition name: {0!s} matches keyword'", ".", "format", "(", "data_type_definition", ".", "name", ")", ")", "members", "=", "getattr", "(", "data_type_definition", ",", "'members'", ",", "None", ")", "if", "not", "members", ":", "raise", "ValueError", "(", "'Data type definition name: {0!s} missing members'", ".", "format", "(", "data_type_definition", ".", "name", ")", ")", "defined_attribute_names", "=", "set", "(", ")", "for", "member_definition", "in", "members", ":", "attribute_name", "=", "member_definition", ".", "name", "if", "not", "cls", ".", "_IsIdentifier", "(", "attribute_name", ")", ":", "raise", "ValueError", "(", "'Attribute name: {0!s} not a valid identifier'", ".", "format", "(", "attribute_name", ")", ")", "if", "attribute_name", ".", "startswith", "(", "'_'", ")", ":", "raise", "ValueError", "(", "'Attribute name: {0!s} starts with underscore'", ".", "format", "(", "attribute_name", ")", ")", "if", "keyword", ".", "iskeyword", "(", "attribute_name", ")", ":", "raise", "ValueError", "(", "'Attribute name: {0!s} matches keyword'", ".", "format", "(", "attribute_name", ")", ")", "if", "attribute_name", "in", "defined_attribute_names", ":", "raise", "ValueError", "(", "'Attribute name: {0!s} already defined'", ".", "format", "(", "attribute_name", ")", ")", "defined_attribute_names", ".", "add", "(", "attribute_name", ")" ]
Validates the data type definition. Args: data_type_definition (DataTypeDefinition): data type definition. Raises: ValueError: if the data type definition is not considered valid.
[ "Validates", "the", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/runtime.py#L123-L169
libyal/dtfabric
dtfabric/runtime/runtime.py
StructureValuesClassFactory.CreateClass
def CreateClass(cls, data_type_definition): """Creates a new structure values class. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: class: structure values class. """ cls._ValidateDataTypeDefinition(data_type_definition) class_definition = cls._CreateClassTemplate(data_type_definition) namespace = { '__builtins__' : { 'object': builtins.object, 'super': builtins.super}, '__name__': '{0:s}'.format(data_type_definition.name)} if sys.version_info[0] >= 3: # pylint: disable=no-member namespace['__builtins__']['__build_class__'] = builtins.__build_class__ exec(class_definition, namespace) # pylint: disable=exec-used return namespace[data_type_definition.name]
python
def CreateClass(cls, data_type_definition): """Creates a new structure values class. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: class: structure values class. """ cls._ValidateDataTypeDefinition(data_type_definition) class_definition = cls._CreateClassTemplate(data_type_definition) namespace = { '__builtins__' : { 'object': builtins.object, 'super': builtins.super}, '__name__': '{0:s}'.format(data_type_definition.name)} if sys.version_info[0] >= 3: # pylint: disable=no-member namespace['__builtins__']['__build_class__'] = builtins.__build_class__ exec(class_definition, namespace) # pylint: disable=exec-used return namespace[data_type_definition.name]
[ "def", "CreateClass", "(", "cls", ",", "data_type_definition", ")", ":", "cls", ".", "_ValidateDataTypeDefinition", "(", "data_type_definition", ")", "class_definition", "=", "cls", ".", "_CreateClassTemplate", "(", "data_type_definition", ")", "namespace", "=", "{", "'__builtins__'", ":", "{", "'object'", ":", "builtins", ".", "object", ",", "'super'", ":", "builtins", ".", "super", "}", ",", "'__name__'", ":", "'{0:s}'", ".", "format", "(", "data_type_definition", ".", "name", ")", "}", "if", "sys", ".", "version_info", "[", "0", "]", ">=", "3", ":", "# pylint: disable=no-member", "namespace", "[", "'__builtins__'", "]", "[", "'__build_class__'", "]", "=", "builtins", ".", "__build_class__", "exec", "(", "class_definition", ",", "namespace", ")", "# pylint: disable=exec-used", "return", "namespace", "[", "data_type_definition", ".", "name", "]" ]
Creates a new structure values class. Args: data_type_definition (DataTypeDefinition): data type definition. Returns: class: structure values class.
[ "Creates", "a", "new", "structure", "values", "class", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/runtime/runtime.py#L172-L197
libyal/dtfabric
dtfabric/registry.py
DataTypeDefinitionsRegistry.DeregisterDefinition
def DeregisterDefinition(self, data_type_definition): """Deregisters a data type definition. The data type definitions are identified based on their lower case name. Args: data_type_definition (DataTypeDefinition): data type definition. Raises: KeyError: if a data type definition is not set for the corresponding name. """ name = data_type_definition.name.lower() if name not in self._definitions: raise KeyError('Definition not set for name: {0:s}.'.format( data_type_definition.name)) del self._definitions[name]
python
def DeregisterDefinition(self, data_type_definition): """Deregisters a data type definition. The data type definitions are identified based on their lower case name. Args: data_type_definition (DataTypeDefinition): data type definition. Raises: KeyError: if a data type definition is not set for the corresponding name. """ name = data_type_definition.name.lower() if name not in self._definitions: raise KeyError('Definition not set for name: {0:s}.'.format( data_type_definition.name)) del self._definitions[name]
[ "def", "DeregisterDefinition", "(", "self", ",", "data_type_definition", ")", ":", "name", "=", "data_type_definition", ".", "name", ".", "lower", "(", ")", "if", "name", "not", "in", "self", ".", "_definitions", ":", "raise", "KeyError", "(", "'Definition not set for name: {0:s}.'", ".", "format", "(", "data_type_definition", ".", "name", ")", ")", "del", "self", ".", "_definitions", "[", "name", "]" ]
Deregisters a data type definition. The data type definitions are identified based on their lower case name. Args: data_type_definition (DataTypeDefinition): data type definition. Raises: KeyError: if a data type definition is not set for the corresponding name.
[ "Deregisters", "a", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/registry.py#L19-L36
libyal/dtfabric
dtfabric/registry.py
DataTypeDefinitionsRegistry.GetDefinitionByName
def GetDefinitionByName(self, name): """Retrieves a specific data type definition by name. Args: name (str): name of the data type definition. Returns: DataTypeDefinition: data type definition or None if not available. """ lookup_name = name.lower() if lookup_name not in self._definitions: lookup_name = self._aliases.get(name, None) return self._definitions.get(lookup_name, None)
python
def GetDefinitionByName(self, name): """Retrieves a specific data type definition by name. Args: name (str): name of the data type definition. Returns: DataTypeDefinition: data type definition or None if not available. """ lookup_name = name.lower() if lookup_name not in self._definitions: lookup_name = self._aliases.get(name, None) return self._definitions.get(lookup_name, None)
[ "def", "GetDefinitionByName", "(", "self", ",", "name", ")", ":", "lookup_name", "=", "name", ".", "lower", "(", ")", "if", "lookup_name", "not", "in", "self", ".", "_definitions", ":", "lookup_name", "=", "self", ".", "_aliases", ".", "get", "(", "name", ",", "None", ")", "return", "self", ".", "_definitions", ".", "get", "(", "lookup_name", ",", "None", ")" ]
Retrieves a specific data type definition by name. Args: name (str): name of the data type definition. Returns: DataTypeDefinition: data type definition or None if not available.
[ "Retrieves", "a", "specific", "data", "type", "definition", "by", "name", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/registry.py#L38-L51
libyal/dtfabric
dtfabric/registry.py
DataTypeDefinitionsRegistry.RegisterDefinition
def RegisterDefinition(self, data_type_definition): """Registers a data type definition. The data type definitions are identified based on their lower case name. Args: data_type_definition (DataTypeDefinition): data type definitions. Raises: KeyError: if data type definition is already set for the corresponding name. """ name_lower = data_type_definition.name.lower() if name_lower in self._definitions: raise KeyError('Definition already set for name: {0:s}.'.format( data_type_definition.name)) if data_type_definition.name in self._aliases: raise KeyError('Alias already set for name: {0:s}.'.format( data_type_definition.name)) for alias in data_type_definition.aliases: if alias in self._aliases: raise KeyError('Alias already set for name: {0:s}.'.format(alias)) self._definitions[name_lower] = data_type_definition for alias in data_type_definition.aliases: self._aliases[alias] = name_lower if data_type_definition.TYPE_INDICATOR == definitions.TYPE_INDICATOR_FORMAT: self._format_definitions.append(name_lower)
python
def RegisterDefinition(self, data_type_definition): """Registers a data type definition. The data type definitions are identified based on their lower case name. Args: data_type_definition (DataTypeDefinition): data type definitions. Raises: KeyError: if data type definition is already set for the corresponding name. """ name_lower = data_type_definition.name.lower() if name_lower in self._definitions: raise KeyError('Definition already set for name: {0:s}.'.format( data_type_definition.name)) if data_type_definition.name in self._aliases: raise KeyError('Alias already set for name: {0:s}.'.format( data_type_definition.name)) for alias in data_type_definition.aliases: if alias in self._aliases: raise KeyError('Alias already set for name: {0:s}.'.format(alias)) self._definitions[name_lower] = data_type_definition for alias in data_type_definition.aliases: self._aliases[alias] = name_lower if data_type_definition.TYPE_INDICATOR == definitions.TYPE_INDICATOR_FORMAT: self._format_definitions.append(name_lower)
[ "def", "RegisterDefinition", "(", "self", ",", "data_type_definition", ")", ":", "name_lower", "=", "data_type_definition", ".", "name", ".", "lower", "(", ")", "if", "name_lower", "in", "self", ".", "_definitions", ":", "raise", "KeyError", "(", "'Definition already set for name: {0:s}.'", ".", "format", "(", "data_type_definition", ".", "name", ")", ")", "if", "data_type_definition", ".", "name", "in", "self", ".", "_aliases", ":", "raise", "KeyError", "(", "'Alias already set for name: {0:s}.'", ".", "format", "(", "data_type_definition", ".", "name", ")", ")", "for", "alias", "in", "data_type_definition", ".", "aliases", ":", "if", "alias", "in", "self", ".", "_aliases", ":", "raise", "KeyError", "(", "'Alias already set for name: {0:s}.'", ".", "format", "(", "alias", ")", ")", "self", ".", "_definitions", "[", "name_lower", "]", "=", "data_type_definition", "for", "alias", "in", "data_type_definition", ".", "aliases", ":", "self", ".", "_aliases", "[", "alias", "]", "=", "name_lower", "if", "data_type_definition", ".", "TYPE_INDICATOR", "==", "definitions", ".", "TYPE_INDICATOR_FORMAT", ":", "self", ".", "_format_definitions", ".", "append", "(", "name_lower", ")" ]
Registers a data type definition. The data type definitions are identified based on their lower case name. Args: data_type_definition (DataTypeDefinition): data type definitions. Raises: KeyError: if data type definition is already set for the corresponding name.
[ "Registers", "a", "data", "type", "definition", "." ]
train
https://github.com/libyal/dtfabric/blob/0d2b5719fa257f6e5c661a406737ebcf8c8db266/dtfabric/registry.py#L61-L92
jorisroovers/pymarkdownlint
pymarkdownlint/config.py
LintConfig.apply_on_csv_string
def apply_on_csv_string(rules_str, func): """ Splits a given string by comma, trims whitespace on the resulting strings and applies a given ```func``` to each item. """ splitted = rules_str.split(",") for str in splitted: func(str.strip())
python
def apply_on_csv_string(rules_str, func): """ Splits a given string by comma, trims whitespace on the resulting strings and applies a given ```func``` to each item. """ splitted = rules_str.split(",") for str in splitted: func(str.strip())
[ "def", "apply_on_csv_string", "(", "rules_str", ",", "func", ")", ":", "splitted", "=", "rules_str", ".", "split", "(", "\",\"", ")", "for", "str", "in", "splitted", ":", "func", "(", "str", ".", "strip", "(", ")", ")" ]
Splits a given string by comma, trims whitespace on the resulting strings and applies a given ```func``` to each item.
[ "Splits", "a", "given", "string", "by", "comma", "trims", "whitespace", "on", "the", "resulting", "strings", "and", "applies", "a", "given", "func", "to", "each", "item", "." ]
train
https://github.com/jorisroovers/pymarkdownlint/blob/c1044e25e18afd78b3fda8fd9b00a4f67cfbbc65/pymarkdownlint/config.py#L41-L46