Search is not available for this dataset
identifier
stringlengths
1
155
parameters
stringlengths
2
6.09k
docstring
stringlengths
11
63.4k
docstring_summary
stringlengths
0
63.4k
function
stringlengths
29
99.8k
function_tokens
list
start_point
list
end_point
list
language
stringclasses
1 value
docstring_language
stringlengths
2
7
docstring_language_predictions
stringlengths
18
23
is_langid_reliable
stringclasses
2 values
setup_platform
(hass, config, add_entities, discovery_info=None)
Set up the Nest Sensor. No longer used.
Set up the Nest Sensor.
def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the Nest Sensor. No longer used. """
[ "def", "setup_platform", "(", "hass", ",", "config", ",", "add_entities", ",", "discovery_info", "=", "None", ")", ":" ]
[ 69, 0 ]
[ 73, 7 ]
python
en
['en', 'pt', 'en']
True
async_setup_legacy_entry
(hass, entry, async_add_entities)
Set up a Nest sensor based on a config entry.
Set up a Nest sensor based on a config entry.
async def async_setup_legacy_entry(hass, entry, async_add_entities): """Set up a Nest sensor based on a config entry.""" nest = hass.data[DATA_NEST] discovery_info = hass.data.get(DATA_NEST_CONFIG, {}).get(CONF_SENSORS, {}) # Add all available sensors if no Nest sensor config is set if discovery_info == {}: conditions = _VALID_SENSOR_TYPES else: conditions = discovery_info.get(CONF_MONITORED_CONDITIONS, {}) for variable in conditions: if variable in _SENSOR_TYPES_DEPRECATED: if variable in DEPRECATED_WEATHER_VARS: wstr = ( "Nest no longer provides weather data like %s. See " "https://www.home-assistant.io/integrations/#weather " "for a list of other weather integrations to use." % variable ) else: wstr = ( f"{variable} is no a longer supported " "monitored_conditions. See " "https://www.home-assistant.io/integrations/" "binary_sensor.nest/ for valid options." ) _LOGGER.error(wstr) def get_sensors(): """Get the Nest sensors.""" all_sensors = [] for structure in nest.structures(): all_sensors += [ NestBasicSensor(structure, None, variable) for variable in conditions if variable in STRUCTURE_SENSOR_TYPES ] for structure, device in nest.thermostats(): all_sensors += [ NestBasicSensor(structure, device, variable) for variable in conditions if variable in SENSOR_TYPES ] all_sensors += [ NestTempSensor(structure, device, variable) for variable in conditions if variable in TEMP_SENSOR_TYPES ] for structure, device in nest.smoke_co_alarms(): all_sensors += [ NestBasicSensor(structure, device, variable) for variable in conditions if variable in PROTECT_SENSOR_TYPES ] structures_has_camera = {} for structure, device in nest.cameras(): structures_has_camera[structure] = True for structure in structures_has_camera: all_sensors += [ NestBasicSensor(structure, None, variable) for variable in conditions if variable in STRUCTURE_CAMERA_SENSOR_TYPES ] return all_sensors async_add_entities(await hass.async_add_executor_job(get_sensors), True)
[ "async", "def", "async_setup_legacy_entry", "(", "hass", ",", "entry", ",", "async_add_entities", ")", ":", "nest", "=", "hass", ".", "data", "[", "DATA_NEST", "]", "discovery_info", "=", "hass", ".", "data", ".", "get", "(", "DATA_NEST_CONFIG", ",", "{", ...
[ 76, 0 ]
[ 146, 76 ]
python
en
['en', 'da', 'en']
True
NestBasicSensor.state
(self)
Return the state of the sensor.
Return the state of the sensor.
def state(self): """Return the state of the sensor.""" return self._state
[ "def", "state", "(", "self", ")", ":", "return", "self", ".", "_state" ]
[ 153, 4 ]
[ 155, 26 ]
python
en
['en', 'en', 'en']
True
NestBasicSensor.device_class
(self)
Return the device class of the sensor.
Return the device class of the sensor.
def device_class(self): """Return the device class of the sensor.""" return SENSOR_DEVICE_CLASSES.get(self.variable)
[ "def", "device_class", "(", "self", ")", ":", "return", "SENSOR_DEVICE_CLASSES", ".", "get", "(", "self", ".", "variable", ")" ]
[ 158, 4 ]
[ 160, 55 ]
python
en
['en', 'en', 'en']
True
NestBasicSensor.update
(self)
Retrieve latest state.
Retrieve latest state.
def update(self): """Retrieve latest state.""" self._unit = SENSOR_UNITS.get(self.variable) if self.variable in VARIABLE_NAME_MAPPING: self._state = getattr(self.device, VARIABLE_NAME_MAPPING[self.variable]) elif self.variable in VALUE_MAPPING: state = getattr(self.device, self.variable) self._state = VALUE_MAPPING[self.variable].get(state, state) elif self.variable in PROTECT_SENSOR_TYPES and self.variable != "color_status": # keep backward compatibility state = getattr(self.device, self.variable) self._state = state.capitalize() if state is not None else None else: self._state = getattr(self.device, self.variable)
[ "def", "update", "(", "self", ")", ":", "self", ".", "_unit", "=", "SENSOR_UNITS", ".", "get", "(", "self", ".", "variable", ")", "if", "self", ".", "variable", "in", "VARIABLE_NAME_MAPPING", ":", "self", ".", "_state", "=", "getattr", "(", "self", "."...
[ 162, 4 ]
[ 176, 61 ]
python
en
['es', 'sk', 'en']
False
NestTempSensor.state
(self)
Return the state of the sensor.
Return the state of the sensor.
def state(self): """Return the state of the sensor.""" return self._state
[ "def", "state", "(", "self", ")", ":", "return", "self", ".", "_state" ]
[ 183, 4 ]
[ 185, 26 ]
python
en
['en', 'en', 'en']
True
NestTempSensor.device_class
(self)
Return the device class of the sensor.
Return the device class of the sensor.
def device_class(self): """Return the device class of the sensor.""" return DEVICE_CLASS_TEMPERATURE
[ "def", "device_class", "(", "self", ")", ":", "return", "DEVICE_CLASS_TEMPERATURE" ]
[ 188, 4 ]
[ 190, 39 ]
python
en
['en', 'en', 'en']
True
NestTempSensor.update
(self)
Retrieve latest state.
Retrieve latest state.
def update(self): """Retrieve latest state.""" if self.device.temperature_scale == "C": self._unit = TEMP_CELSIUS else: self._unit = TEMP_FAHRENHEIT temp = getattr(self.device, self.variable) if temp is None: self._state = None if isinstance(temp, tuple): low, high = temp self._state = f"{int(low)}-{int(high)}" else: self._state = round(temp, 1)
[ "def", "update", "(", "self", ")", ":", "if", "self", ".", "device", ".", "temperature_scale", "==", "\"C\"", ":", "self", ".", "_unit", "=", "TEMP_CELSIUS", "else", ":", "self", ".", "_unit", "=", "TEMP_FAHRENHEIT", "temp", "=", "getattr", "(", "self", ...
[ 192, 4 ]
[ 207, 40 ]
python
en
['es', 'sk', 'en']
False
setup_platform
(hass, config, add_entities, discovery_info=None)
Set up the platform in Home Assistant and Case Information.
Set up the platform in Home Assistant and Case Information.
def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the platform in Home Assistant and Case Information.""" uscis = UscisSensor(config["case_id"], config[CONF_NAME]) uscis.update() if uscis.valid_case_id: add_entities([uscis]) else: _LOGGER.error("Setup USCIS Sensor Fail check if your Case ID is Valid")
[ "def", "setup_platform", "(", "hass", ",", "config", ",", "add_entities", ",", "discovery_info", "=", "None", ")", ":", "uscis", "=", "UscisSensor", "(", "config", "[", "\"case_id\"", "]", ",", "config", "[", "CONF_NAME", "]", ")", "uscis", ".", "update", ...
[ 25, 0 ]
[ 32, 79 ]
python
en
['en', 'en', 'en']
True
UscisSensor.__init__
(self, case, name)
Initialize the sensor.
Initialize the sensor.
def __init__(self, case, name): """Initialize the sensor.""" self._state = None self._case_id = case self._attributes = None self.valid_case_id = None self._name = name
[ "def", "__init__", "(", "self", ",", "case", ",", "name", ")", ":", "self", ".", "_state", "=", "None", "self", ".", "_case_id", "=", "case", "self", ".", "_attributes", "=", "None", "self", ".", "valid_case_id", "=", "None", "self", ".", "_name", "=...
[ 43, 4 ]
[ 49, 25 ]
python
en
['en', 'en', 'en']
True
UscisSensor.name
(self)
Return the name.
Return the name.
def name(self): """Return the name.""" return self._name
[ "def", "name", "(", "self", ")", ":", "return", "self", ".", "_name" ]
[ 52, 4 ]
[ 54, 25 ]
python
en
['en', 'ig', 'en']
True
UscisSensor.state
(self)
Return the state.
Return the state.
def state(self): """Return the state.""" return self._state
[ "def", "state", "(", "self", ")", ":", "return", "self", ".", "_state" ]
[ 57, 4 ]
[ 59, 26 ]
python
en
['en', 'en', 'en']
True
UscisSensor.device_state_attributes
(self)
Return the state attributes.
Return the state attributes.
def device_state_attributes(self): """Return the state attributes.""" return self._attributes
[ "def", "device_state_attributes", "(", "self", ")", ":", "return", "self", ".", "_attributes" ]
[ 62, 4 ]
[ 64, 31 ]
python
en
['en', 'en', 'en']
True
UscisSensor.update
(self)
Fetch data from the USCIS website and update state attributes.
Fetch data from the USCIS website and update state attributes.
def update(self): """Fetch data from the USCIS website and update state attributes.""" try: status = uscisstatus.get_case_status(self._case_id) self._attributes = {self.CURRENT_STATUS: status["status"]} self._state = status["date"] self.valid_case_id = True except ValueError: _LOGGER("Please Check that you have valid USCIS case id") self.valid_case_id = False
[ "def", "update", "(", "self", ")", ":", "try", ":", "status", "=", "uscisstatus", ".", "get_case_status", "(", "self", ".", "_case_id", ")", "self", ".", "_attributes", "=", "{", "self", ".", "CURRENT_STATUS", ":", "status", "[", "\"status\"", "]", "}", ...
[ 67, 4 ]
[ 77, 38 ]
python
en
['en', 'en', 'en']
True
load_tf_weights_in_albert
(model, config, tf_checkpoint_path)
Load tf checkpoints in a pytorch model.
Load tf checkpoints in a pytorch model.
def load_tf_weights_in_albert(model, config, tf_checkpoint_path): """ Load tf checkpoints in a pytorch model.""" try: import re import numpy as np import tensorflow as tf except ImportError: logger.error( "Loading a TensorFlow model in PyTorch, requires TensorFlow to be installed. Please see " "https://www.tensorflow.org/install/ for installation instructions." ) raise tf_path = os.path.abspath(tf_checkpoint_path) logger.info("Converting TensorFlow checkpoint from {}".format(tf_path)) # Load weights from TF model init_vars = tf.train.list_variables(tf_path) names = [] arrays = [] for name, shape in init_vars: logger.info("Loading TF weight {} with shape {}".format(name, shape)) array = tf.train.load_variable(tf_path, name) names.append(name) arrays.append(array) for name, array in zip(names, arrays): print(name) for name, array in zip(names, arrays): original_name = name # If saved from the TF HUB module name = name.replace("module/", "") # Renaming and simplifying name = name.replace("ffn_1", "ffn") name = name.replace("bert/", "albert/") name = name.replace("attention_1", "attention") name = name.replace("transform/", "") name = name.replace("LayerNorm_1", "full_layer_layer_norm") name = name.replace("LayerNorm", "attention/LayerNorm") name = name.replace("transformer/", "") # The feed forward layer had an 'intermediate' step which has been abstracted away name = name.replace("intermediate/dense/", "") name = name.replace("ffn/intermediate/output/dense/", "ffn_output/") # ALBERT attention was split between self and output which have been abstracted away name = name.replace("/output/", "/") name = name.replace("/self/", "/") # The pooler is a linear layer name = name.replace("pooler/dense", "pooler") # The classifier was simplified to predictions from cls/predictions name = name.replace("cls/predictions", "predictions") name = name.replace("predictions/attention", "predictions") # Naming was changed to be more explicit name = name.replace("embeddings/attention", "embeddings") name = name.replace("inner_group_", "albert_layers/") name = name.replace("group_", "albert_layer_groups/") # Classifier if len(name.split("/")) == 1 and ("output_bias" in name or "output_weights" in name): name = "classifier/" + name # No ALBERT model currently handles the next sentence prediction task if "seq_relationship" in name: name = name.replace("seq_relationship/output_", "sop_classifier/classifier/") name = name.replace("weights", "weight") name = name.split("/") # Ignore the gradients applied by the LAMB/ADAM optimizers. if ( "adam_m" in name or "adam_v" in name or "AdamWeightDecayOptimizer" in name or "AdamWeightDecayOptimizer_1" in name or "global_step" in name ): logger.info("Skipping {}".format("/".join(name))) continue pointer = model for m_name in name: if re.fullmatch(r"[A-Za-z]+_\d+", m_name): scope_names = re.split(r"_(\d+)", m_name) else: scope_names = [m_name] if scope_names[0] == "kernel" or scope_names[0] == "gamma": pointer = getattr(pointer, "weight") elif scope_names[0] == "output_bias" or scope_names[0] == "beta": pointer = getattr(pointer, "bias") elif scope_names[0] == "output_weights": pointer = getattr(pointer, "weight") elif scope_names[0] == "squad": pointer = getattr(pointer, "classifier") else: try: pointer = getattr(pointer, scope_names[0]) except AttributeError: logger.info("Skipping {}".format("/".join(name))) continue if len(scope_names) >= 2: num = int(scope_names[1]) pointer = pointer[num] if m_name[-11:] == "_embeddings": pointer = getattr(pointer, "weight") elif m_name == "kernel": array = np.transpose(array) try: assert ( pointer.shape == array.shape ), f"Pointer shape {pointer.shape} and array shape {array.shape} mismatched" except AssertionError as e: e.args += (pointer.shape, array.shape) raise print("Initialize PyTorch weight {} from {}".format(name, original_name)) pointer.data = torch.from_numpy(array) return model
[ "def", "load_tf_weights_in_albert", "(", "model", ",", "config", ",", "tf_checkpoint_path", ")", ":", "try", ":", "import", "re", "import", "numpy", "as", "np", "import", "tensorflow", "as", "tf", "except", "ImportError", ":", "logger", ".", "error", "(", "\...
[ 72, 0 ]
[ 196, 16 ]
python
en
['en', 'en', 'en']
True
AlbertPreTrainedModel._init_weights
(self, module)
Initialize the weights.
Initialize the weights.
def _init_weights(self, module): """Initialize the weights.""" if isinstance(module, nn.Linear): # Slightly different from the TF version which uses truncated_normal for initialization # cf https://github.com/pytorch/pytorch/pull/5617 module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) if module.bias is not None: module.bias.data.zero_() elif isinstance(module, nn.Embedding): module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) if module.padding_idx is not None: module.weight.data[module.padding_idx].zero_() elif isinstance(module, nn.LayerNorm): module.bias.data.zero_() module.weight.data.fill_(1.0)
[ "def", "_init_weights", "(", "self", ",", "module", ")", ":", "if", "isinstance", "(", "module", ",", "nn", ".", "Linear", ")", ":", "# Slightly different from the TF version which uses truncated_normal for initialization", "# cf https://github.com/pytorch/pytorch/pull/5617", ...
[ 491, 4 ]
[ 505, 41 ]
python
en
['en', 'en', 'en']
True
HerbertTokenizerFast.build_inputs_with_special_tokens
( self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None )
Build model inputs from a sequence or a pair of sequence for sequence classification tasks by concatenating and adding special tokens. An HerBERT, like BERT sequence has the following format: - single sequence: ``<s> X </s>`` - pair of sequences: ``<s> A </s> B </s>`` Args: token_ids_0 (:obj:`List[int]`): List of IDs to which the special tokens will be added. token_ids_1 (:obj:`List[int]`, `optional`): Optional second list of IDs for sequence pairs. Returns: :obj:`List[int]`: List of `input IDs <../glossary.html#input-ids>`__ with the appropriate special tokens.
Build model inputs from a sequence or a pair of sequence for sequence classification tasks by concatenating and adding special tokens. An HerBERT, like BERT sequence has the following format:
def build_inputs_with_special_tokens( self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None ) -> List[int]: """ Build model inputs from a sequence or a pair of sequence for sequence classification tasks by concatenating and adding special tokens. An HerBERT, like BERT sequence has the following format: - single sequence: ``<s> X </s>`` - pair of sequences: ``<s> A </s> B </s>`` Args: token_ids_0 (:obj:`List[int]`): List of IDs to which the special tokens will be added. token_ids_1 (:obj:`List[int]`, `optional`): Optional second list of IDs for sequence pairs. Returns: :obj:`List[int]`: List of `input IDs <../glossary.html#input-ids>`__ with the appropriate special tokens. """ cls = [self.cls_token_id] sep = [self.sep_token_id] if token_ids_1 is None: return cls + token_ids_0 + sep return cls + token_ids_0 + sep + token_ids_1 + sep
[ "def", "build_inputs_with_special_tokens", "(", "self", ",", "token_ids_0", ":", "List", "[", "int", "]", ",", "token_ids_1", ":", "Optional", "[", "List", "[", "int", "]", "]", "=", "None", ")", "->", "List", "[", "int", "]", ":", "cls", "=", "[", "...
[ 82, 4 ]
[ 107, 58 ]
python
en
['en', 'error', 'th']
False
HerbertTokenizerFast.get_special_tokens_mask
( self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None, already_has_special_tokens: bool = False )
Retrieve sequence ids from a token list that has no special tokens added. This method is called when adding special tokens using the tokenizer ``prepare_for_model`` method. Args: token_ids_0 (:obj:`List[int]`): List of IDs. token_ids_1 (:obj:`List[int]`, `optional`): Optional second list of IDs for sequence pairs. already_has_special_tokens (:obj:`bool`, `optional`, defaults to :obj:`False`): Whether or not the token list is already formatted with special tokens for the model. Returns: :obj:`List[int]`: A list of integers in the range [0, 1]: 1 for a special token, 0 for a sequence token.
Retrieve sequence ids from a token list that has no special tokens added. This method is called when adding special tokens using the tokenizer ``prepare_for_model`` method.
def get_special_tokens_mask( self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None, already_has_special_tokens: bool = False ) -> List[int]: """ Retrieve sequence ids from a token list that has no special tokens added. This method is called when adding special tokens using the tokenizer ``prepare_for_model`` method. Args: token_ids_0 (:obj:`List[int]`): List of IDs. token_ids_1 (:obj:`List[int]`, `optional`): Optional second list of IDs for sequence pairs. already_has_special_tokens (:obj:`bool`, `optional`, defaults to :obj:`False`): Whether or not the token list is already formatted with special tokens for the model. Returns: :obj:`List[int]`: A list of integers in the range [0, 1]: 1 for a special token, 0 for a sequence token. """ if already_has_special_tokens: if token_ids_1 is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return list(map(lambda x: 1 if x in [self.sep_token_id, self.cls_token_id] else 0, token_ids_0)) if token_ids_1 is None: return [1] + ([0] * len(token_ids_0)) + [1] return [1] + ([0] * len(token_ids_0)) + [1] + ([0] * len(token_ids_1)) + [1]
[ "def", "get_special_tokens_mask", "(", "self", ",", "token_ids_0", ":", "List", "[", "int", "]", ",", "token_ids_1", ":", "Optional", "[", "List", "[", "int", "]", "]", "=", "None", ",", "already_has_special_tokens", ":", "bool", "=", "False", ")", "->", ...
[ 109, 4 ]
[ 137, 84 ]
python
en
['en', 'error', 'th']
False
HerbertTokenizerFast.create_token_type_ids_from_sequences
( self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None )
Create a mask from the two sequences passed to be used in a sequence-pair classification task. HerBERT, like BERT sequence pair mask has the following format: :: 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 | first sequence | second sequence | Args: token_ids_0 (:obj:`List[int]`): List of IDs. token_ids_1 (:obj:`List[int]`, `optional`): Optional second list of IDs for sequence pairs. Returns: :obj:`List[int]`: List of `token type IDs <../glossary.html#token-type-ids>`_ according to the given sequence(s).
Create a mask from the two sequences passed to be used in a sequence-pair classification task. HerBERT, like BERT sequence pair mask has the following format:
def create_token_type_ids_from_sequences( self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None ) -> List[int]: """ Create a mask from the two sequences passed to be used in a sequence-pair classification task. HerBERT, like BERT sequence pair mask has the following format: :: 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 | first sequence | second sequence | Args: token_ids_0 (:obj:`List[int]`): List of IDs. token_ids_1 (:obj:`List[int]`, `optional`): Optional second list of IDs for sequence pairs. Returns: :obj:`List[int]`: List of `token type IDs <../glossary.html#token-type-ids>`_ according to the given sequence(s). """ sep = [self.sep_token_id] cls = [self.cls_token_id] if token_ids_1 is None: return len(cls + token_ids_0 + sep) * [0] return len(cls + token_ids_0 + sep) * [0] + len(token_ids_1 + sep) * [1]
[ "def", "create_token_type_ids_from_sequences", "(", "self", ",", "token_ids_0", ":", "List", "[", "int", "]", ",", "token_ids_1", ":", "Optional", "[", "List", "[", "int", "]", "]", "=", "None", ")", "->", "List", "[", "int", "]", ":", "sep", "=", "[",...
[ 139, 4 ]
[ 166, 80 ]
python
en
['en', 'error', 'th']
False
setup_platform
(hass, config, add_entities, discovery_info=None)
Set up the available PM sensors.
Set up the available PM sensors.
def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the available PM sensors.""" try: coll = pm.PMDataCollector( config.get(CONF_SERIAL_DEVICE), pm.SUPPORTED_SENSORS[config.get(CONF_BRAND)] ) except KeyError: _LOGGER.error( "Brand %s not supported\n supported brands: %s", config.get(CONF_BRAND), pm.SUPPORTED_SENSORS.keys(), ) return except OSError as err: _LOGGER.error( "Could not open serial connection to %s (%s)", config.get(CONF_SERIAL_DEVICE), err, ) return dev = [] for pmname in coll.supported_values(): if config.get(CONF_NAME) is not None: name = "{} PM{}".format(config.get(CONF_NAME), pmname) else: name = f"PM{pmname}" dev.append(ParticulateMatterSensor(coll, name, pmname)) add_entities(dev)
[ "def", "setup_platform", "(", "hass", ",", "config", ",", "add_entities", ",", "discovery_info", "=", "None", ")", ":", "try", ":", "coll", "=", "pm", ".", "PMDataCollector", "(", "config", ".", "get", "(", "CONF_SERIAL_DEVICE", ")", ",", "pm", ".", "SUP...
[ 25, 0 ]
[ 55, 21 ]
python
en
['en', 'en', 'en']
True
ParticulateMatterSensor.__init__
(self, pmDataCollector, name, pmname)
Initialize a new PM sensor.
Initialize a new PM sensor.
def __init__(self, pmDataCollector, name, pmname): """Initialize a new PM sensor.""" self._name = name self._pmname = pmname self._state = None self._collector = pmDataCollector
[ "def", "__init__", "(", "self", ",", "pmDataCollector", ",", "name", ",", "pmname", ")", ":", "self", ".", "_name", "=", "name", "self", ".", "_pmname", "=", "pmname", "self", ".", "_state", "=", "None", "self", ".", "_collector", "=", "pmDataCollector" ...
[ 61, 4 ]
[ 66, 41 ]
python
en
['en', 'co', 'en']
True
ParticulateMatterSensor.name
(self)
Return the name of the sensor.
Return the name of the sensor.
def name(self): """Return the name of the sensor.""" return self._name
[ "def", "name", "(", "self", ")", ":", "return", "self", ".", "_name" ]
[ 69, 4 ]
[ 71, 25 ]
python
en
['en', 'mi', 'en']
True
ParticulateMatterSensor.state
(self)
Return the state of the sensor.
Return the state of the sensor.
def state(self): """Return the state of the sensor.""" return self._state
[ "def", "state", "(", "self", ")", ":", "return", "self", ".", "_state" ]
[ 74, 4 ]
[ 76, 26 ]
python
en
['en', 'en', 'en']
True
ParticulateMatterSensor.unit_of_measurement
(self)
Return the unit of measurement of this entity, if any.
Return the unit of measurement of this entity, if any.
def unit_of_measurement(self): """Return the unit of measurement of this entity, if any.""" return CONCENTRATION_MICROGRAMS_PER_CUBIC_METER
[ "def", "unit_of_measurement", "(", "self", ")", ":", "return", "CONCENTRATION_MICROGRAMS_PER_CUBIC_METER" ]
[ 79, 4 ]
[ 81, 55 ]
python
en
['en', 'en', 'en']
True
ParticulateMatterSensor.update
(self)
Read from sensor and update the state.
Read from sensor and update the state.
def update(self): """Read from sensor and update the state.""" _LOGGER.debug("Reading data from PM sensor") try: self._state = self._collector.read_data()[self._pmname] except KeyError: _LOGGER.error("Could not read PM%s value", self._pmname)
[ "def", "update", "(", "self", ")", ":", "_LOGGER", ".", "debug", "(", "\"Reading data from PM sensor\"", ")", "try", ":", "self", ".", "_state", "=", "self", ".", "_collector", ".", "read_data", "(", ")", "[", "self", ".", "_pmname", "]", "except", "KeyE...
[ 83, 4 ]
[ 89, 68 ]
python
en
['en', 'en', 'en']
True
async_setup_platform
(hass, config, async_add_entities, discovery_info=None)
Set up a generic IP Camera.
Set up a generic IP Camera.
async def async_setup_platform(hass, config, async_add_entities, discovery_info=None): """Set up a generic IP Camera.""" await async_setup_reload_service(hass, DOMAIN, PLATFORMS) async_add_entities([GenericCamera(hass, config)])
[ "async", "def", "async_setup_platform", "(", "hass", ",", "config", ",", "async_add_entities", ",", "discovery_info", "=", "None", ")", ":", "await", "async_setup_reload_service", "(", "hass", ",", "DOMAIN", ",", "PLATFORMS", ")", "async_add_entities", "(", "[", ...
[ 62, 0 ]
[ 67, 53 ]
python
en
['en', 'su', 'en']
True
GenericCamera.__init__
(self, hass, device_info)
Initialize a generic camera.
Initialize a generic camera.
def __init__(self, hass, device_info): """Initialize a generic camera.""" super().__init__() self.hass = hass self._authentication = device_info.get(CONF_AUTHENTICATION) self._name = device_info.get(CONF_NAME) self._still_image_url = device_info[CONF_STILL_IMAGE_URL] self._stream_source = device_info.get(CONF_STREAM_SOURCE) self._still_image_url.hass = hass if self._stream_source is not None: self._stream_source.hass = hass self._limit_refetch = device_info[CONF_LIMIT_REFETCH_TO_URL_CHANGE] self._frame_interval = 1 / device_info[CONF_FRAMERATE] self._supported_features = SUPPORT_STREAM if self._stream_source else 0 self.content_type = device_info[CONF_CONTENT_TYPE] self.verify_ssl = device_info[CONF_VERIFY_SSL] username = device_info.get(CONF_USERNAME) password = device_info.get(CONF_PASSWORD) if username and password: if self._authentication == HTTP_DIGEST_AUTHENTICATION: self._auth = HTTPDigestAuth(username, password) else: self._auth = aiohttp.BasicAuth(username, password=password) else: self._auth = None self._last_url = None self._last_image = None
[ "def", "__init__", "(", "self", ",", "hass", ",", "device_info", ")", ":", "super", "(", ")", ".", "__init__", "(", ")", "self", ".", "hass", "=", "hass", "self", ".", "_authentication", "=", "device_info", ".", "get", "(", "CONF_AUTHENTICATION", ")", ...
[ 73, 4 ]
[ 102, 31 ]
python
co
['en', 'co', 'it']
False
GenericCamera.supported_features
(self)
Return supported features for this camera.
Return supported features for this camera.
def supported_features(self): """Return supported features for this camera.""" return self._supported_features
[ "def", "supported_features", "(", "self", ")", ":", "return", "self", ".", "_supported_features" ]
[ 105, 4 ]
[ 107, 39 ]
python
en
['en', 'en', 'en']
True
GenericCamera.frame_interval
(self)
Return the interval between frames of the mjpeg stream.
Return the interval between frames of the mjpeg stream.
def frame_interval(self): """Return the interval between frames of the mjpeg stream.""" return self._frame_interval
[ "def", "frame_interval", "(", "self", ")", ":", "return", "self", ".", "_frame_interval" ]
[ 110, 4 ]
[ 112, 35 ]
python
en
['en', 'en', 'en']
True
GenericCamera.camera_image
(self)
Return bytes of camera image.
Return bytes of camera image.
def camera_image(self): """Return bytes of camera image.""" return asyncio.run_coroutine_threadsafe( self.async_camera_image(), self.hass.loop ).result()
[ "def", "camera_image", "(", "self", ")", ":", "return", "asyncio", ".", "run_coroutine_threadsafe", "(", "self", ".", "async_camera_image", "(", ")", ",", "self", ".", "hass", ".", "loop", ")", ".", "result", "(", ")" ]
[ 114, 4 ]
[ 118, 18 ]
python
en
['en', 'zu', 'en']
True
GenericCamera.async_camera_image
(self)
Return a still image response from the camera.
Return a still image response from the camera.
async def async_camera_image(self): """Return a still image response from the camera.""" try: url = self._still_image_url.async_render(parse_result=False) except TemplateError as err: _LOGGER.error("Error parsing template %s: %s", self._still_image_url, err) return self._last_image if url == self._last_url and self._limit_refetch: return self._last_image # aiohttp don't support DigestAuth yet if self._authentication == HTTP_DIGEST_AUTHENTICATION: def fetch(): """Read image from a URL.""" try: response = requests.get( url, timeout=10, auth=self._auth, verify=self.verify_ssl ) return response.content except requests.exceptions.RequestException as error: _LOGGER.error( "Error getting new camera image from %s: %s", self._name, error ) return self._last_image self._last_image = await self.hass.async_add_executor_job(fetch) # async else: try: websession = async_get_clientsession( self.hass, verify_ssl=self.verify_ssl ) with async_timeout.timeout(10): response = await websession.get(url, auth=self._auth) self._last_image = await response.read() except asyncio.TimeoutError: _LOGGER.error("Timeout getting camera image from %s", self._name) return self._last_image except aiohttp.ClientError as err: _LOGGER.error( "Error getting new camera image from %s: %s", self._name, err ) return self._last_image self._last_url = url return self._last_image
[ "async", "def", "async_camera_image", "(", "self", ")", ":", "try", ":", "url", "=", "self", ".", "_still_image_url", ".", "async_render", "(", "parse_result", "=", "False", ")", "except", "TemplateError", "as", "err", ":", "_LOGGER", ".", "error", "(", "\...
[ 120, 4 ]
[ 167, 31 ]
python
en
['en', 'en', 'en']
True
GenericCamera.name
(self)
Return the name of this device.
Return the name of this device.
def name(self): """Return the name of this device.""" return self._name
[ "def", "name", "(", "self", ")", ":", "return", "self", ".", "_name" ]
[ 170, 4 ]
[ 172, 25 ]
python
en
['en', 'en', 'en']
True
GenericCamera.stream_source
(self)
Return the source of the stream.
Return the source of the stream.
async def stream_source(self): """Return the source of the stream.""" if self._stream_source is None: return None try: return self._stream_source.async_render(parse_result=False) except TemplateError as err: _LOGGER.error("Error parsing template %s: %s", self._stream_source, err) return None
[ "async", "def", "stream_source", "(", "self", ")", ":", "if", "self", ".", "_stream_source", "is", "None", ":", "return", "None", "try", ":", "return", "self", ".", "_stream_source", ".", "async_render", "(", "parse_result", "=", "False", ")", "except", "T...
[ 174, 4 ]
[ 183, 23 ]
python
en
['en', 'en', 'en']
True
BaseTrainer.train
(self)
Override the method to train.
Override the method to train.
def train(self): """ Override the method to train. """ raise NotImplementedError
[ "def", "train", "(", "self", ")", ":", "raise", "NotImplementedError" ]
[ 9, 4 ]
[ 13, 33 ]
python
en
['en', 'error', 'th']
False
BaseTrainer.validate
(self)
Override the method to validate.
Override the method to validate.
def validate(self): """ Override the method to validate. """ raise NotImplementedError
[ "def", "validate", "(", "self", ")", ":", "raise", "NotImplementedError" ]
[ 16, 4 ]
[ 20, 33 ]
python
en
['en', 'error', 'th']
False
BaseTrainer.export
(self, file)
Override the method to export to file. Parameters ---------- file : str File path to export to.
Override the method to export to file.
def export(self, file): """ Override the method to export to file. Parameters ---------- file : str File path to export to. """ raise NotImplementedError
[ "def", "export", "(", "self", ",", "file", ")", ":", "raise", "NotImplementedError" ]
[ 23, 4 ]
[ 32, 33 ]
python
en
['en', 'error', 'th']
False
BaseTrainer.checkpoint
(self)
Override to dump a checkpoint.
Override to dump a checkpoint.
def checkpoint(self): """ Override to dump a checkpoint. """ raise NotImplementedError
[ "def", "checkpoint", "(", "self", ")", ":", "raise", "NotImplementedError" ]
[ 35, 4 ]
[ 39, 33 ]
python
en
['en', 'error', 'th']
False
mock_credstash
()
Mock credstash so it doesn't connect to the internet.
Mock credstash so it doesn't connect to the internet.
def mock_credstash(): """Mock credstash so it doesn't connect to the internet.""" with patch.object(yaml_loader, "credstash") as mock_credstash: mock_credstash.getSecret.return_value = None yield mock_credstash
[ "def", "mock_credstash", "(", ")", ":", "with", "patch", ".", "object", "(", "yaml_loader", ",", "\"credstash\"", ")", "as", "mock_credstash", ":", "mock_credstash", ".", "getSecret", ".", "return_value", "=", "None", "yield", "mock_credstash" ]
[ 18, 0 ]
[ 22, 28 ]
python
en
['en', 'en', 'en']
True
test_simple_list
()
Test simple list.
Test simple list.
def test_simple_list(): """Test simple list.""" conf = "config:\n - simple\n - list" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["config"] == ["simple", "list"]
[ "def", "test_simple_list", "(", ")", ":", "conf", "=", "\"config:\\n - simple\\n - list\"", "with", "io", ".", "StringIO", "(", "conf", ")", "as", "file", ":", "doc", "=", "yaml_loader", ".", "yaml", ".", "safe_load", "(", "file", ")", "assert", "doc", "...
[ 25, 0 ]
[ 30, 46 ]
python
en
['en', 'et', 'en']
True
test_simple_dict
()
Test simple dict.
Test simple dict.
def test_simple_dict(): """Test simple dict.""" conf = "key: value" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["key"] == "value"
[ "def", "test_simple_dict", "(", ")", ":", "conf", "=", "\"key: value\"", "with", "io", ".", "StringIO", "(", "conf", ")", "as", "file", ":", "doc", "=", "yaml_loader", ".", "yaml", ".", "safe_load", "(", "file", ")", "assert", "doc", "[", "\"key\"", "]...
[ 33, 0 ]
[ 38, 32 ]
python
en
['it', 'la', 'en']
False
test_unhashable_key
()
Test an unhashable key.
Test an unhashable key.
def test_unhashable_key(): """Test an unhashable key.""" files = {YAML_CONFIG_FILE: "message:\n {{ states.state }}"} with pytest.raises(HomeAssistantError), patch_yaml_files(files): load_yaml_config_file(YAML_CONFIG_FILE)
[ "def", "test_unhashable_key", "(", ")", ":", "files", "=", "{", "YAML_CONFIG_FILE", ":", "\"message:\\n {{ states.state }}\"", "}", "with", "pytest", ".", "raises", "(", "HomeAssistantError", ")", ",", "patch_yaml_files", "(", "files", ")", ":", "load_yaml_config_f...
[ 41, 0 ]
[ 45, 47 ]
python
en
['en', 'ga', 'es']
False
test_no_key
()
Test item without a key.
Test item without a key.
def test_no_key(): """Test item without a key.""" files = {YAML_CONFIG_FILE: "a: a\nnokeyhere"} with pytest.raises(HomeAssistantError), patch_yaml_files(files): yaml.load_yaml(YAML_CONFIG_FILE)
[ "def", "test_no_key", "(", ")", ":", "files", "=", "{", "YAML_CONFIG_FILE", ":", "\"a: a\\nnokeyhere\"", "}", "with", "pytest", ".", "raises", "(", "HomeAssistantError", ")", ",", "patch_yaml_files", "(", "files", ")", ":", "yaml", ".", "load_yaml", "(", "YA...
[ 48, 0 ]
[ 52, 40 ]
python
en
['en', 'en', 'en']
True
test_environment_variable
()
Test config file with environment variable.
Test config file with environment variable.
def test_environment_variable(): """Test config file with environment variable.""" os.environ["PASSWORD"] = "secret_password" conf = "password: !env_var PASSWORD" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["password"] == "secret_password" del os.environ["PASSWORD"]
[ "def", "test_environment_variable", "(", ")", ":", "os", ".", "environ", "[", "\"PASSWORD\"", "]", "=", "\"secret_password\"", "conf", "=", "\"password: !env_var PASSWORD\"", "with", "io", ".", "StringIO", "(", "conf", ")", "as", "file", ":", "doc", "=", "yaml...
[ 55, 0 ]
[ 62, 30 ]
python
en
['en', 'en', 'en']
True
test_environment_variable_default
()
Test config file with default value for environment variable.
Test config file with default value for environment variable.
def test_environment_variable_default(): """Test config file with default value for environment variable.""" conf = "password: !env_var PASSWORD secret_password" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["password"] == "secret_password"
[ "def", "test_environment_variable_default", "(", ")", ":", "conf", "=", "\"password: !env_var PASSWORD secret_password\"", "with", "io", ".", "StringIO", "(", "conf", ")", "as", "file", ":", "doc", "=", "yaml_loader", ".", "yaml", ".", "safe_load", "(", "file", ...
[ 65, 0 ]
[ 70, 47 ]
python
en
['en', 'en', 'en']
True
test_invalid_environment_variable
()
Test config file with no environment variable sat.
Test config file with no environment variable sat.
def test_invalid_environment_variable(): """Test config file with no environment variable sat.""" conf = "password: !env_var PASSWORD" with pytest.raises(HomeAssistantError): with io.StringIO(conf) as file: yaml_loader.yaml.safe_load(file)
[ "def", "test_invalid_environment_variable", "(", ")", ":", "conf", "=", "\"password: !env_var PASSWORD\"", "with", "pytest", ".", "raises", "(", "HomeAssistantError", ")", ":", "with", "io", ".", "StringIO", "(", "conf", ")", "as", "file", ":", "yaml_loader", "....
[ 73, 0 ]
[ 78, 44 ]
python
en
['en', 'en', 'en']
True
test_include_yaml
()
Test include yaml.
Test include yaml.
def test_include_yaml(): """Test include yaml.""" with patch_yaml_files({"test.yaml": "value"}): conf = "key: !include test.yaml" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["key"] == "value" with patch_yaml_files({"test.yaml": None}): conf = "key: !include test.yaml" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["key"] == {}
[ "def", "test_include_yaml", "(", ")", ":", "with", "patch_yaml_files", "(", "{", "\"test.yaml\"", ":", "\"value\"", "}", ")", ":", "conf", "=", "\"key: !include test.yaml\"", "with", "io", ".", "StringIO", "(", "conf", ")", "as", "file", ":", "doc", "=", "...
[ 81, 0 ]
[ 93, 35 ]
python
en
['tr', 'en', 'en']
True
test_include_dir_list
(mock_walk)
Test include dir list yaml.
Test include dir list yaml.
def test_include_dir_list(mock_walk): """Test include dir list yaml.""" mock_walk.return_value = [["/test", [], ["two.yaml", "one.yaml"]]] with patch_yaml_files({"/test/one.yaml": "one", "/test/two.yaml": "two"}): conf = "key: !include_dir_list /test" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["key"] == sorted(["one", "two"])
[ "def", "test_include_dir_list", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "]", ",", "[", "\"two.yaml\"", ",", "\"one.yaml\"", "]", "]", "]", "with", "patch_yaml_files", "(", "{", "\"/test/one.yaml\"", ...
[ 97, 0 ]
[ 105, 55 ]
python
az
['oc', 'az', 'sw']
False
test_include_dir_list_recursive
(mock_walk)
Test include dir recursive list yaml.
Test include dir recursive list yaml.
def test_include_dir_list_recursive(mock_walk): """Test include dir recursive list yaml.""" mock_walk.return_value = [ ["/test", ["tmp2", ".ignore", "ignore"], ["zero.yaml"]], ["/test/tmp2", [], ["one.yaml", "two.yaml"]], ["/test/ignore", [], [".ignore.yaml"]], ] with patch_yaml_files( { "/test/zero.yaml": "zero", "/test/tmp2/one.yaml": "one", "/test/tmp2/two.yaml": "two", } ): conf = "key: !include_dir_list /test" with io.StringIO(conf) as file: assert ( ".ignore" in mock_walk.return_value[0][1] ), "Expecting .ignore in here" doc = yaml_loader.yaml.safe_load(file) assert "tmp2" in mock_walk.return_value[0][1] assert ".ignore" not in mock_walk.return_value[0][1] assert sorted(doc["key"]) == sorted(["zero", "one", "two"])
[ "def", "test_include_dir_list_recursive", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "\"tmp2\"", ",", "\".ignore\"", ",", "\"ignore\"", "]", ",", "[", "\"zero.yaml\"", "]", "]", ",", "[", "\"/test/tmp2\"...
[ 109, 0 ]
[ 132, 71 ]
python
ro
['tr', 'ro', 'sw']
False
test_include_dir_named
(mock_walk)
Test include dir named yaml.
Test include dir named yaml.
def test_include_dir_named(mock_walk): """Test include dir named yaml.""" mock_walk.return_value = [ ["/test", [], ["first.yaml", "second.yaml", "secrets.yaml"]] ] with patch_yaml_files({"/test/first.yaml": "one", "/test/second.yaml": "two"}): conf = "key: !include_dir_named /test" correct = {"first": "one", "second": "two"} with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["key"] == correct
[ "def", "test_include_dir_named", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "]", ",", "[", "\"first.yaml\"", ",", "\"second.yaml\"", ",", "\"secrets.yaml\"", "]", "]", "]", "with", "patch_yaml_files", "(...
[ 136, 0 ]
[ 147, 40 ]
python
en
['tr', 'en', 'sw']
False
test_include_dir_named_recursive
(mock_walk)
Test include dir named yaml.
Test include dir named yaml.
def test_include_dir_named_recursive(mock_walk): """Test include dir named yaml.""" mock_walk.return_value = [ ["/test", ["tmp2", ".ignore", "ignore"], ["first.yaml"]], ["/test/tmp2", [], ["second.yaml", "third.yaml"]], ["/test/ignore", [], [".ignore.yaml"]], ] with patch_yaml_files( { "/test/first.yaml": "one", "/test/tmp2/second.yaml": "two", "/test/tmp2/third.yaml": "three", } ): conf = "key: !include_dir_named /test" correct = {"first": "one", "second": "two", "third": "three"} with io.StringIO(conf) as file: assert ( ".ignore" in mock_walk.return_value[0][1] ), "Expecting .ignore in here" doc = yaml_loader.yaml.safe_load(file) assert "tmp2" in mock_walk.return_value[0][1] assert ".ignore" not in mock_walk.return_value[0][1] assert doc["key"] == correct
[ "def", "test_include_dir_named_recursive", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "\"tmp2\"", ",", "\".ignore\"", ",", "\"ignore\"", "]", ",", "[", "\"first.yaml\"", "]", "]", ",", "[", "\"/test/tmp2...
[ 151, 0 ]
[ 175, 40 ]
python
en
['tr', 'en', 'sw']
False
test_include_dir_merge_list
(mock_walk)
Test include dir merge list yaml.
Test include dir merge list yaml.
def test_include_dir_merge_list(mock_walk): """Test include dir merge list yaml.""" mock_walk.return_value = [["/test", [], ["first.yaml", "second.yaml"]]] with patch_yaml_files( {"/test/first.yaml": "- one", "/test/second.yaml": "- two\n- three"} ): conf = "key: !include_dir_merge_list /test" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert sorted(doc["key"]) == sorted(["one", "two", "three"])
[ "def", "test_include_dir_merge_list", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "]", ",", "[", "\"first.yaml\"", ",", "\"second.yaml\"", "]", "]", "]", "with", "patch_yaml_files", "(", "{", "\"/test/fir...
[ 179, 0 ]
[ 189, 72 ]
python
tr
['tr', 'cy', 'tr']
True
test_include_dir_merge_list_recursive
(mock_walk)
Test include dir merge list yaml.
Test include dir merge list yaml.
def test_include_dir_merge_list_recursive(mock_walk): """Test include dir merge list yaml.""" mock_walk.return_value = [ ["/test", ["tmp2", ".ignore", "ignore"], ["first.yaml"]], ["/test/tmp2", [], ["second.yaml", "third.yaml"]], ["/test/ignore", [], [".ignore.yaml"]], ] with patch_yaml_files( { "/test/first.yaml": "- one", "/test/tmp2/second.yaml": "- two", "/test/tmp2/third.yaml": "- three\n- four", } ): conf = "key: !include_dir_merge_list /test" with io.StringIO(conf) as file: assert ( ".ignore" in mock_walk.return_value[0][1] ), "Expecting .ignore in here" doc = yaml_loader.yaml.safe_load(file) assert "tmp2" in mock_walk.return_value[0][1] assert ".ignore" not in mock_walk.return_value[0][1] assert sorted(doc["key"]) == sorted(["one", "two", "three", "four"])
[ "def", "test_include_dir_merge_list_recursive", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "\"tmp2\"", ",", "\".ignore\"", ",", "\"ignore\"", "]", ",", "[", "\"first.yaml\"", "]", "]", ",", "[", "\"/test...
[ 193, 0 ]
[ 216, 80 ]
python
tr
['tr', 'cy', 'tr']
True
test_include_dir_merge_named
(mock_walk)
Test include dir merge named yaml.
Test include dir merge named yaml.
def test_include_dir_merge_named(mock_walk): """Test include dir merge named yaml.""" mock_walk.return_value = [["/test", [], ["first.yaml", "second.yaml"]]] files = { "/test/first.yaml": "key1: one", "/test/second.yaml": "key2: two\nkey3: three", } with patch_yaml_files(files): conf = "key: !include_dir_merge_named /test" with io.StringIO(conf) as file: doc = yaml_loader.yaml.safe_load(file) assert doc["key"] == {"key1": "one", "key2": "two", "key3": "three"}
[ "def", "test_include_dir_merge_named", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "]", ",", "[", "\"first.yaml\"", ",", "\"second.yaml\"", "]", "]", "]", "files", "=", "{", "\"/test/first.yaml\"", ":", ...
[ 220, 0 ]
[ 233, 80 ]
python
en
['tr', 'en', 'sw']
False
test_include_dir_merge_named_recursive
(mock_walk)
Test include dir merge named yaml.
Test include dir merge named yaml.
def test_include_dir_merge_named_recursive(mock_walk): """Test include dir merge named yaml.""" mock_walk.return_value = [ ["/test", ["tmp2", ".ignore", "ignore"], ["first.yaml"]], ["/test/tmp2", [], ["second.yaml", "third.yaml"]], ["/test/ignore", [], [".ignore.yaml"]], ] with patch_yaml_files( { "/test/first.yaml": "key1: one", "/test/tmp2/second.yaml": "key2: two", "/test/tmp2/third.yaml": "key3: three\nkey4: four", } ): conf = "key: !include_dir_merge_named /test" with io.StringIO(conf) as file: assert ( ".ignore" in mock_walk.return_value[0][1] ), "Expecting .ignore in here" doc = yaml_loader.yaml.safe_load(file) assert "tmp2" in mock_walk.return_value[0][1] assert ".ignore" not in mock_walk.return_value[0][1] assert doc["key"] == { "key1": "one", "key2": "two", "key3": "three", "key4": "four", }
[ "def", "test_include_dir_merge_named_recursive", "(", "mock_walk", ")", ":", "mock_walk", ".", "return_value", "=", "[", "[", "\"/test\"", ",", "[", "\"tmp2\"", ",", "\".ignore\"", ",", "\"ignore\"", "]", ",", "[", "\"first.yaml\"", "]", "]", ",", "[", "\"/tes...
[ 237, 0 ]
[ 265, 13 ]
python
en
['tr', 'en', 'sw']
False
test_load_yaml_encoding_error
(mock_open)
Test raising a UnicodeDecodeError.
Test raising a UnicodeDecodeError.
def test_load_yaml_encoding_error(mock_open): """Test raising a UnicodeDecodeError.""" mock_open.side_effect = UnicodeDecodeError("", b"", 1, 0, "") with pytest.raises(HomeAssistantError): yaml_loader.load_yaml("test")
[ "def", "test_load_yaml_encoding_error", "(", "mock_open", ")", ":", "mock_open", ".", "side_effect", "=", "UnicodeDecodeError", "(", "\"\"", ",", "b\"\"", ",", "1", ",", "0", ",", "\"\"", ")", "with", "pytest", ".", "raises", "(", "HomeAssistantError", ")", ...
[ 269, 0 ]
[ 273, 37 ]
python
de
['en', 'de', 'pt']
False
test_dump
()
The that the dump method returns empty None values.
The that the dump method returns empty None values.
def test_dump(): """The that the dump method returns empty None values.""" assert yaml.dump({"a": None, "b": "b"}) == "a:\nb: b\n"
[ "def", "test_dump", "(", ")", ":", "assert", "yaml", ".", "dump", "(", "{", "\"a\"", ":", "None", ",", "\"b\"", ":", "\"b\"", "}", ")", "==", "\"a:\\nb: b\\n\"" ]
[ 276, 0 ]
[ 278, 59 ]
python
en
['en', 'en', 'en']
True
test_dump_unicode
()
The that the dump method returns empty None values.
The that the dump method returns empty None values.
def test_dump_unicode(): """The that the dump method returns empty None values.""" assert yaml.dump({"a": None, "b": "привет"}) == "a:\nb: привет\n"
[ "def", "test_dump_unicode", "(", ")", ":", "assert", "yaml", ".", "dump", "(", "{", "\"a\"", ":", "None", ",", "\"b\"", ":", "\"привет\"}) == ", "\"", "a", "\\n", ": привет\\n\"" ]
[ 281, 0 ]
[ 283, 81 ]
python
en
['en', 'en', 'en']
True
load_yaml
(fname, string)
Write a string to file and return the parsed yaml.
Write a string to file and return the parsed yaml.
def load_yaml(fname, string): """Write a string to file and return the parsed yaml.""" FILES[fname] = string with patch_yaml_files(FILES): return load_yaml_config_file(fname)
[ "def", "load_yaml", "(", "fname", ",", "string", ")", ":", "FILES", "[", "fname", "]", "=", "string", "with", "patch_yaml_files", "(", "FILES", ")", ":", "return", "load_yaml_config_file", "(", "fname", ")" ]
[ 289, 0 ]
[ 293, 43 ]
python
en
['en', 'en', 'en']
True
test_representing_yaml_loaded_data
()
Test we can represent YAML loaded data.
Test we can represent YAML loaded data.
def test_representing_yaml_loaded_data(): """Test we can represent YAML loaded data.""" files = {YAML_CONFIG_FILE: 'key: [1, "2", 3]'} with patch_yaml_files(files): data = load_yaml_config_file(YAML_CONFIG_FILE) assert yaml.dump(data) == "key:\n- 1\n- '2'\n- 3\n"
[ "def", "test_representing_yaml_loaded_data", "(", ")", ":", "files", "=", "{", "YAML_CONFIG_FILE", ":", "'key: [1, \"2\", 3]'", "}", "with", "patch_yaml_files", "(", "files", ")", ":", "data", "=", "load_yaml_config_file", "(", "YAML_CONFIG_FILE", ")", "assert", "ya...
[ 449, 0 ]
[ 454, 55 ]
python
en
['en', 'en', 'en']
True
test_duplicate_key
(caplog)
Test duplicate dict keys.
Test duplicate dict keys.
def test_duplicate_key(caplog): """Test duplicate dict keys.""" files = {YAML_CONFIG_FILE: "key: thing1\nkey: thing2"} with patch_yaml_files(files): load_yaml_config_file(YAML_CONFIG_FILE) assert "contains duplicate key" in caplog.text
[ "def", "test_duplicate_key", "(", "caplog", ")", ":", "files", "=", "{", "YAML_CONFIG_FILE", ":", "\"key: thing1\\nkey: thing2\"", "}", "with", "patch_yaml_files", "(", "files", ")", ":", "load_yaml_config_file", "(", "YAML_CONFIG_FILE", ")", "assert", "\"contains dup...
[ 457, 0 ]
[ 462, 50 ]
python
en
['ro', 'id', 'en']
False
test_placeholder_class
()
Test placeholder class.
Test placeholder class.
def test_placeholder_class(): """Test placeholder class.""" placeholder = yaml_loader.Placeholder("hello") placeholder2 = yaml_loader.Placeholder("hello") assert placeholder.name == "hello" assert placeholder == placeholder2 assert len({placeholder, placeholder2}) == 1
[ "def", "test_placeholder_class", "(", ")", ":", "placeholder", "=", "yaml_loader", ".", "Placeholder", "(", "\"hello\"", ")", "placeholder2", "=", "yaml_loader", ".", "Placeholder", "(", "\"hello\"", ")", "assert", "placeholder", ".", "name", "==", "\"hello\"", ...
[ 465, 0 ]
[ 473, 48 ]
python
en
['en', 'en', 'en']
True
test_placeholder
()
Test loading placeholders.
Test loading placeholders.
def test_placeholder(): """Test loading placeholders.""" data = {"hello": yaml.Placeholder("test_name")} assert yaml.parse_yaml(yaml.dump(data)) == data
[ "def", "test_placeholder", "(", ")", ":", "data", "=", "{", "\"hello\"", ":", "yaml", ".", "Placeholder", "(", "\"test_name\"", ")", "}", "assert", "yaml", ".", "parse_yaml", "(", "yaml", ".", "dump", "(", "data", ")", ")", "==", "data" ]
[ 476, 0 ]
[ 479, 51 ]
python
en
['en', 'en', 'en']
True
FakeKeyring.__init__
(self, secrets_dict)
Store keyring dictionary.
Store keyring dictionary.
def __init__(self, secrets_dict): """Store keyring dictionary.""" self._secrets = secrets_dict
[ "def", "__init__", "(", "self", ",", "secrets_dict", ")", ":", "self", ".", "_secrets", "=", "secrets_dict" ]
[ 299, 4 ]
[ 301, 36 ]
python
en
['id', 'en', 'en']
True
FakeKeyring.get_password
(self, domain, name)
Retrieve password.
Retrieve password.
def get_password(self, domain, name): """Retrieve password.""" assert domain == yaml._SECRET_NAMESPACE return self._secrets.get(name)
[ "def", "get_password", "(", "self", ",", "domain", ",", "name", ")", ":", "assert", "domain", "==", "yaml", ".", "_SECRET_NAMESPACE", "return", "self", ".", "_secrets", ".", "get", "(", "name", ")" ]
[ 304, 4 ]
[ 307, 38 ]
python
en
['fr', 'sr', 'en']
False
TestSecrets.setUp
(self)
Create & load secrets file.
Create & load secrets file.
def setUp(self): """Create & load secrets file.""" config_dir = get_test_config_dir() yaml.clear_secret_cache() self._yaml_path = os.path.join(config_dir, YAML_CONFIG_FILE) self._secret_path = os.path.join(config_dir, yaml.SECRET_YAML) self._sub_folder_path = os.path.join(config_dir, "subFolder") self._unrelated_path = os.path.join(config_dir, "unrelated") load_yaml( self._secret_path, "http_pw: pwhttp\n" "comp1_un: un1\n" "comp1_pw: pw1\n" "stale_pw: not_used\n" "logger: debug\n", ) self._yaml = load_yaml( self._yaml_path, "http:\n" " api_password: !secret http_pw\n" "component:\n" " username: !secret comp1_un\n" " password: !secret comp1_pw\n" "", )
[ "def", "setUp", "(", "self", ")", ":", "config_dir", "=", "get_test_config_dir", "(", ")", "yaml", ".", "clear_secret_cache", "(", ")", "self", ".", "_yaml_path", "=", "os", ".", "path", ".", "join", "(", "config_dir", ",", "YAML_CONFIG_FILE", ")", "self",...
[ 315, 4 ]
[ 340, 9 ]
python
en
['en', 'sm', 'en']
True
TestSecrets.tearDown
(self)
Clean up secrets.
Clean up secrets.
def tearDown(self): """Clean up secrets.""" yaml.clear_secret_cache() FILES.clear()
[ "def", "tearDown", "(", "self", ")", ":", "yaml", ".", "clear_secret_cache", "(", ")", "FILES", ".", "clear", "(", ")" ]
[ 342, 4 ]
[ 345, 21 ]
python
en
['en', 'fr', 'en']
True
TestSecrets.test_secrets_from_yaml
(self)
Did secrets load ok.
Did secrets load ok.
def test_secrets_from_yaml(self): """Did secrets load ok.""" expected = {"api_password": "pwhttp"} assert expected == self._yaml["http"] expected = {"username": "un1", "password": "pw1"} assert expected == self._yaml["component"]
[ "def", "test_secrets_from_yaml", "(", "self", ")", ":", "expected", "=", "{", "\"api_password\"", ":", "\"pwhttp\"", "}", "assert", "expected", "==", "self", ".", "_yaml", "[", "\"http\"", "]", "expected", "=", "{", "\"username\"", ":", "\"un1\"", ",", "\"pa...
[ 347, 4 ]
[ 353, 50 ]
python
en
['en', 'bg', 'en']
True
TestSecrets.test_secrets_from_parent_folder
(self)
Test loading secrets from parent folder.
Test loading secrets from parent folder.
def test_secrets_from_parent_folder(self): """Test loading secrets from parent folder.""" expected = {"api_password": "pwhttp"} self._yaml = load_yaml( os.path.join(self._sub_folder_path, "sub.yaml"), "http:\n" " api_password: !secret http_pw\n" "component:\n" " username: !secret comp1_un\n" " password: !secret comp1_pw\n" "", ) assert expected == self._yaml["http"]
[ "def", "test_secrets_from_parent_folder", "(", "self", ")", ":", "expected", "=", "{", "\"api_password\"", ":", "\"pwhttp\"", "}", "self", ".", "_yaml", "=", "load_yaml", "(", "os", ".", "path", ".", "join", "(", "self", ".", "_sub_folder_path", ",", "\"sub....
[ 355, 4 ]
[ 368, 45 ]
python
en
['en', 'da', 'en']
True
TestSecrets.test_secret_overrides_parent
(self)
Test loading current directory secret overrides the parent.
Test loading current directory secret overrides the parent.
def test_secret_overrides_parent(self): """Test loading current directory secret overrides the parent.""" expected = {"api_password": "override"} load_yaml( os.path.join(self._sub_folder_path, yaml.SECRET_YAML), "http_pw: override" ) self._yaml = load_yaml( os.path.join(self._sub_folder_path, "sub.yaml"), "http:\n" " api_password: !secret http_pw\n" "component:\n" " username: !secret comp1_un\n" " password: !secret comp1_pw\n" "", ) assert expected == self._yaml["http"]
[ "def", "test_secret_overrides_parent", "(", "self", ")", ":", "expected", "=", "{", "\"api_password\"", ":", "\"override\"", "}", "load_yaml", "(", "os", ".", "path", ".", "join", "(", "self", ".", "_sub_folder_path", ",", "yaml", ".", "SECRET_YAML", ")", ",...
[ 370, 4 ]
[ 386, 45 ]
python
en
['en', 'da', 'en']
True
TestSecrets.test_secrets_from_unrelated_fails
(self)
Test loading secrets from unrelated folder fails.
Test loading secrets from unrelated folder fails.
def test_secrets_from_unrelated_fails(self): """Test loading secrets from unrelated folder fails.""" load_yaml(os.path.join(self._unrelated_path, yaml.SECRET_YAML), "test: failure") with pytest.raises(HomeAssistantError): load_yaml( os.path.join(self._sub_folder_path, "sub.yaml"), "http:\n api_password: !secret test", )
[ "def", "test_secrets_from_unrelated_fails", "(", "self", ")", ":", "load_yaml", "(", "os", ".", "path", ".", "join", "(", "self", ".", "_unrelated_path", ",", "yaml", ".", "SECRET_YAML", ")", ",", "\"test: failure\"", ")", "with", "pytest", ".", "raises", "(...
[ 388, 4 ]
[ 395, 13 ]
python
en
['en', 'en', 'en']
True
TestSecrets.test_secrets_keyring
(self)
Test keyring fallback & get_password.
Test keyring fallback & get_password.
def test_secrets_keyring(self): """Test keyring fallback & get_password.""" yaml_loader.keyring = None # Ensure its not there yaml_str = "http:\n api_password: !secret http_pw_keyring" with pytest.raises(HomeAssistantError): load_yaml(self._yaml_path, yaml_str) yaml_loader.keyring = FakeKeyring({"http_pw_keyring": "yeah"}) _yaml = load_yaml(self._yaml_path, yaml_str) assert {"http": {"api_password": "yeah"}} == _yaml
[ "def", "test_secrets_keyring", "(", "self", ")", ":", "yaml_loader", ".", "keyring", "=", "None", "# Ensure its not there", "yaml_str", "=", "\"http:\\n api_password: !secret http_pw_keyring\"", "with", "pytest", ".", "raises", "(", "HomeAssistantError", ")", ":", "loa...
[ 397, 4 ]
[ 406, 58 ]
python
de
['de', 'jv', 'en']
False
TestSecrets.test_secrets_credstash
(self, mock_credstash)
Test credstash fallback & get_password.
Test credstash fallback & get_password.
def test_secrets_credstash(self, mock_credstash): """Test credstash fallback & get_password.""" mock_credstash.getSecret.return_value = "yeah" yaml_str = "http:\n api_password: !secret http_pw_credstash" _yaml = load_yaml(self._yaml_path, yaml_str) log = logging.getLogger() log.error(_yaml["http"]) assert {"api_password": "yeah"} == _yaml["http"]
[ "def", "test_secrets_credstash", "(", "self", ",", "mock_credstash", ")", ":", "mock_credstash", ".", "getSecret", ".", "return_value", "=", "\"yeah\"", "yaml_str", "=", "\"http:\\n api_password: !secret http_pw_credstash\"", "_yaml", "=", "load_yaml", "(", "self", "."...
[ 409, 4 ]
[ 416, 56 ]
python
da
['de', 'da', 'en']
False
TestSecrets.test_secrets_logger_removed
(self)
Ensure logger: debug was removed.
Ensure logger: debug was removed.
def test_secrets_logger_removed(self): """Ensure logger: debug was removed.""" with pytest.raises(HomeAssistantError): load_yaml(self._yaml_path, "api_password: !secret logger")
[ "def", "test_secrets_logger_removed", "(", "self", ")", ":", "with", "pytest", ".", "raises", "(", "HomeAssistantError", ")", ":", "load_yaml", "(", "self", ".", "_yaml_path", ",", "\"api_password: !secret logger\"", ")" ]
[ 418, 4 ]
[ 421, 70 ]
python
en
['en', 'ceb', 'en']
True
TestSecrets.test_bad_logger_value
(self, mock_error)
Ensure logger: debug was removed.
Ensure logger: debug was removed.
def test_bad_logger_value(self, mock_error): """Ensure logger: debug was removed.""" yaml.clear_secret_cache() load_yaml(self._secret_path, "logger: info\npw: abc") load_yaml(self._yaml_path, "api_password: !secret pw") assert mock_error.call_count == 1, "Expected an error about logger: value"
[ "def", "test_bad_logger_value", "(", "self", ",", "mock_error", ")", ":", "yaml", ".", "clear_secret_cache", "(", ")", "load_yaml", "(", "self", ".", "_secret_path", ",", "\"logger: info\\npw: abc\"", ")", "load_yaml", "(", "self", ".", "_yaml_path", ",", "\"api...
[ 424, 4 ]
[ 429, 82 ]
python
en
['en', 'ceb', 'en']
True
TestSecrets.test_secrets_are_not_dict
(self)
Did secrets handle non-dict file.
Did secrets handle non-dict file.
def test_secrets_are_not_dict(self): """Did secrets handle non-dict file.""" FILES[ self._secret_path ] = "- http_pw: pwhttp\n comp1_un: un1\n comp1_pw: pw1\n" yaml.clear_secret_cache() with pytest.raises(HomeAssistantError): load_yaml( self._yaml_path, "http:\n" " api_password: !secret http_pw\n" "component:\n" " username: !secret comp1_un\n" " password: !secret comp1_pw\n" "", )
[ "def", "test_secrets_are_not_dict", "(", "self", ")", ":", "FILES", "[", "self", ".", "_secret_path", "]", "=", "\"- http_pw: pwhttp\\n comp1_un: un1\\n comp1_pw: pw1\\n\"", "yaml", ".", "clear_secret_cache", "(", ")", "with", "pytest", ".", "raises", "(", "HomeAssi...
[ 431, 4 ]
[ 446, 13 ]
python
it
['en', 'it', 'it']
True
test_air_con_create_sensors
(hass)
Test creation of aircon sensors.
Test creation of aircon sensors.
async def test_air_con_create_sensors(hass): """Test creation of aircon sensors.""" await async_init_integration(hass) state = hass.states.get("sensor.air_conditioning_power") assert state.state == "ON" state = hass.states.get("sensor.air_conditioning_link") assert state.state == "ONLINE" state = hass.states.get("sensor.air_conditioning_link") assert state.state == "ONLINE" state = hass.states.get("sensor.air_conditioning_tado_mode") assert state.state == "HOME" state = hass.states.get("sensor.air_conditioning_temperature") assert state.state == "24.76" state = hass.states.get("sensor.air_conditioning_ac") assert state.state == "ON" state = hass.states.get("sensor.air_conditioning_overlay") assert state.state == "True" state = hass.states.get("sensor.air_conditioning_humidity") assert state.state == "60.9" state = hass.states.get("sensor.air_conditioning_open_window") assert state.state == "False"
[ "async", "def", "test_air_con_create_sensors", "(", "hass", ")", ":", "await", "async_init_integration", "(", "hass", ")", "state", "=", "hass", ".", "states", ".", "get", "(", "\"sensor.air_conditioning_power\"", ")", "assert", "state", ".", "state", "==", "\"O...
[ 5, 0 ]
[ 35, 33 ]
python
en
['en', 'en', 'en']
True
test_heater_create_sensors
(hass)
Test creation of heater sensors.
Test creation of heater sensors.
async def test_heater_create_sensors(hass): """Test creation of heater sensors.""" await async_init_integration(hass) state = hass.states.get("sensor.baseboard_heater_power") assert state.state == "ON" state = hass.states.get("sensor.baseboard_heater_link") assert state.state == "ONLINE" state = hass.states.get("sensor.baseboard_heater_link") assert state.state == "ONLINE" state = hass.states.get("sensor.baseboard_heater_tado_mode") assert state.state == "HOME" state = hass.states.get("sensor.baseboard_heater_temperature") assert state.state == "20.65" state = hass.states.get("sensor.baseboard_heater_early_start") assert state.state == "False" state = hass.states.get("sensor.baseboard_heater_overlay") assert state.state == "True" state = hass.states.get("sensor.baseboard_heater_humidity") assert state.state == "45.2" state = hass.states.get("sensor.baseboard_heater_open_window") assert state.state == "False"
[ "async", "def", "test_heater_create_sensors", "(", "hass", ")", ":", "await", "async_init_integration", "(", "hass", ")", "state", "=", "hass", ".", "states", ".", "get", "(", "\"sensor.baseboard_heater_power\"", ")", "assert", "state", ".", "state", "==", "\"ON...
[ 38, 0 ]
[ 68, 33 ]
python
en
['en', 'fi', 'en']
True
test_water_heater_create_sensors
(hass)
Test creation of water heater sensors.
Test creation of water heater sensors.
async def test_water_heater_create_sensors(hass): """Test creation of water heater sensors.""" await async_init_integration(hass) state = hass.states.get("sensor.water_heater_tado_mode") assert state.state == "HOME" state = hass.states.get("sensor.water_heater_link") assert state.state == "ONLINE" state = hass.states.get("sensor.water_heater_overlay") assert state.state == "False" state = hass.states.get("sensor.water_heater_power") assert state.state == "ON"
[ "async", "def", "test_water_heater_create_sensors", "(", "hass", ")", ":", "await", "async_init_integration", "(", "hass", ")", "state", "=", "hass", ".", "states", ".", "get", "(", "\"sensor.water_heater_tado_mode\"", ")", "assert", "state", ".", "state", "==", ...
[ 71, 0 ]
[ 86, 30 ]
python
en
['en', 'en', 'en']
True
test_home_create_sensors
(hass)
Test creation of home sensors.
Test creation of home sensors.
async def test_home_create_sensors(hass): """Test creation of home sensors.""" await async_init_integration(hass) state = hass.states.get("sensor.home_name_tado_bridge_status") assert state.state == "True"
[ "async", "def", "test_home_create_sensors", "(", "hass", ")", ":", "await", "async_init_integration", "(", "hass", ")", "state", "=", "hass", ".", "states", ".", "get", "(", "\"sensor.home_name_tado_bridge_status\"", ")", "assert", "state", ".", "state", "==", "...
[ 89, 0 ]
[ 95, 32 ]
python
en
['en', 'en', 'en']
True
async_setup_platform
(hass, config, async_add_entities, discovery_info=None)
Set up the media player demo platform.
Set up the media player demo platform.
async def async_setup_platform(hass, config, async_add_entities, discovery_info=None): """Set up the media player demo platform.""" async_add_entities( [ DemoYoutubePlayer( "Living Room", "eyU3bRy2x44", "♥♥ The Best Fireplace Video (3 hours)", 300, ), DemoYoutubePlayer( "Bedroom", "kxopViU98Xo", "Epic sax guy 10 hours", 360000 ), DemoMusicPlayer(), DemoTVShowPlayer(), ] )
[ "async", "def", "async_setup_platform", "(", "hass", ",", "config", ",", "async_add_entities", ",", "discovery_info", "=", "None", ")", ":", "async_add_entities", "(", "[", "DemoYoutubePlayer", "(", "\"Living Room\"", ",", "\"eyU3bRy2x44\"", ",", "\"♥♥ The Best Firepl...
[ 28, 0 ]
[ 44, 5 ]
python
en
['en', 'da', 'en']
True
async_setup_entry
(hass, config_entry, async_add_entities)
Set up the Demo config entry.
Set up the Demo config entry.
async def async_setup_entry(hass, config_entry, async_add_entities): """Set up the Demo config entry.""" await async_setup_platform(hass, {}, async_add_entities)
[ "async", "def", "async_setup_entry", "(", "hass", ",", "config_entry", ",", "async_add_entities", ")", ":", "await", "async_setup_platform", "(", "hass", ",", "{", "}", ",", "async_add_entities", ")" ]
[ 47, 0 ]
[ 49, 60 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.__init__
(self, name, device_class=None)
Initialize the demo device.
Initialize the demo device.
def __init__(self, name, device_class=None): """Initialize the demo device.""" self._name = name self._player_state = STATE_PLAYING self._volume_level = 1.0 self._volume_muted = False self._shuffle = False self._sound_mode_list = SOUND_MODE_LIST self._sound_mode = DEFAULT_SOUND_MODE self._device_class = device_class
[ "def", "__init__", "(", "self", ",", "name", ",", "device_class", "=", "None", ")", ":", "self", ".", "_name", "=", "name", "self", ".", "_player_state", "=", "STATE_PLAYING", "self", ".", "_volume_level", "=", "1.0", "self", ".", "_volume_muted", "=", "...
[ 102, 4 ]
[ 111, 41 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.should_poll
(self)
Push an update after each command.
Push an update after each command.
def should_poll(self): """Push an update after each command.""" return False
[ "def", "should_poll", "(", "self", ")", ":", "return", "False" ]
[ 114, 4 ]
[ 116, 20 ]
python
en
['en', 'lb', 'en']
True
AbstractDemoPlayer.name
(self)
Return the name of the media player.
Return the name of the media player.
def name(self): """Return the name of the media player.""" return self._name
[ "def", "name", "(", "self", ")", ":", "return", "self", ".", "_name" ]
[ 119, 4 ]
[ 121, 25 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.state
(self)
Return the state of the player.
Return the state of the player.
def state(self): """Return the state of the player.""" return self._player_state
[ "def", "state", "(", "self", ")", ":", "return", "self", ".", "_player_state" ]
[ 124, 4 ]
[ 126, 33 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.volume_level
(self)
Return the volume level of the media player (0..1).
Return the volume level of the media player (0..1).
def volume_level(self): """Return the volume level of the media player (0..1).""" return self._volume_level
[ "def", "volume_level", "(", "self", ")", ":", "return", "self", ".", "_volume_level" ]
[ 129, 4 ]
[ 131, 33 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.is_volume_muted
(self)
Return boolean if volume is currently muted.
Return boolean if volume is currently muted.
def is_volume_muted(self): """Return boolean if volume is currently muted.""" return self._volume_muted
[ "def", "is_volume_muted", "(", "self", ")", ":", "return", "self", ".", "_volume_muted" ]
[ 134, 4 ]
[ 136, 33 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.shuffle
(self)
Boolean if shuffling is enabled.
Boolean if shuffling is enabled.
def shuffle(self): """Boolean if shuffling is enabled.""" return self._shuffle
[ "def", "shuffle", "(", "self", ")", ":", "return", "self", ".", "_shuffle" ]
[ 139, 4 ]
[ 141, 28 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.sound_mode
(self)
Return the current sound mode.
Return the current sound mode.
def sound_mode(self): """Return the current sound mode.""" return self._sound_mode
[ "def", "sound_mode", "(", "self", ")", ":", "return", "self", ".", "_sound_mode" ]
[ 144, 4 ]
[ 146, 31 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.sound_mode_list
(self)
Return a list of available sound modes.
Return a list of available sound modes.
def sound_mode_list(self): """Return a list of available sound modes.""" return self._sound_mode_list
[ "def", "sound_mode_list", "(", "self", ")", ":", "return", "self", ".", "_sound_mode_list" ]
[ 149, 4 ]
[ 151, 36 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.device_class
(self)
Return the device class of the media player.
Return the device class of the media player.
def device_class(self): """Return the device class of the media player.""" return self._device_class
[ "def", "device_class", "(", "self", ")", ":", "return", "self", ".", "_device_class" ]
[ 154, 4 ]
[ 156, 33 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.turn_on
(self)
Turn the media player on.
Turn the media player on.
def turn_on(self): """Turn the media player on.""" self._player_state = STATE_PLAYING self.schedule_update_ha_state()
[ "def", "turn_on", "(", "self", ")", ":", "self", ".", "_player_state", "=", "STATE_PLAYING", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 158, 4 ]
[ 161, 39 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.turn_off
(self)
Turn the media player off.
Turn the media player off.
def turn_off(self): """Turn the media player off.""" self._player_state = STATE_OFF self.schedule_update_ha_state()
[ "def", "turn_off", "(", "self", ")", ":", "self", ".", "_player_state", "=", "STATE_OFF", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 163, 4 ]
[ 166, 39 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.mute_volume
(self, mute)
Mute the volume.
Mute the volume.
def mute_volume(self, mute): """Mute the volume.""" self._volume_muted = mute self.schedule_update_ha_state()
[ "def", "mute_volume", "(", "self", ",", "mute", ")", ":", "self", ".", "_volume_muted", "=", "mute", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 168, 4 ]
[ 171, 39 ]
python
en
['en', 'sn', 'en']
True
AbstractDemoPlayer.volume_up
(self)
Increase volume.
Increase volume.
def volume_up(self): """Increase volume.""" self._volume_level = min(1.0, self._volume_level + 0.1) self.schedule_update_ha_state()
[ "def", "volume_up", "(", "self", ")", ":", "self", ".", "_volume_level", "=", "min", "(", "1.0", ",", "self", ".", "_volume_level", "+", "0.1", ")", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 173, 4 ]
[ 176, 39 ]
python
en
['en', 'af', 'en']
False
AbstractDemoPlayer.volume_down
(self)
Decrease volume.
Decrease volume.
def volume_down(self): """Decrease volume.""" self._volume_level = max(0.0, self._volume_level - 0.1) self.schedule_update_ha_state()
[ "def", "volume_down", "(", "self", ")", ":", "self", ".", "_volume_level", "=", "max", "(", "0.0", ",", "self", ".", "_volume_level", "-", "0.1", ")", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 178, 4 ]
[ 181, 39 ]
python
en
['en', 'de', 'en']
False
AbstractDemoPlayer.set_volume_level
(self, volume)
Set the volume level, range 0..1.
Set the volume level, range 0..1.
def set_volume_level(self, volume): """Set the volume level, range 0..1.""" self._volume_level = volume self.schedule_update_ha_state()
[ "def", "set_volume_level", "(", "self", ",", "volume", ")", ":", "self", ".", "_volume_level", "=", "volume", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 183, 4 ]
[ 186, 39 ]
python
en
['en', 'zu', 'en']
True
AbstractDemoPlayer.media_play
(self)
Send play command.
Send play command.
def media_play(self): """Send play command.""" self._player_state = STATE_PLAYING self.schedule_update_ha_state()
[ "def", "media_play", "(", "self", ")", ":", "self", ".", "_player_state", "=", "STATE_PLAYING", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 188, 4 ]
[ 191, 39 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.media_pause
(self)
Send pause command.
Send pause command.
def media_pause(self): """Send pause command.""" self._player_state = STATE_PAUSED self.schedule_update_ha_state()
[ "def", "media_pause", "(", "self", ")", ":", "self", ".", "_player_state", "=", "STATE_PAUSED", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 193, 4 ]
[ 196, 39 ]
python
en
['en', 'en', 'en']
True
AbstractDemoPlayer.set_shuffle
(self, shuffle)
Enable/disable shuffle mode.
Enable/disable shuffle mode.
def set_shuffle(self, shuffle): """Enable/disable shuffle mode.""" self._shuffle = shuffle self.schedule_update_ha_state()
[ "def", "set_shuffle", "(", "self", ",", "shuffle", ")", ":", "self", ".", "_shuffle", "=", "shuffle", "self", ".", "schedule_update_ha_state", "(", ")" ]
[ 198, 4 ]
[ 201, 39 ]
python
en
['en', 'st', 'en']
True