id int32 0 252k | repo stringlengths 7 55 | path stringlengths 4 127 | func_name stringlengths 1 88 | original_string stringlengths 75 19.8k | language stringclasses 1 value | code stringlengths 75 19.8k | code_tokens list | docstring stringlengths 3 17.3k | docstring_tokens list | sha stringlengths 40 40 | url stringlengths 87 242 |
|---|---|---|---|---|---|---|---|---|---|---|---|
226,900 | google/openhtf | openhtf/util/multicast.py | send | def send(query,
address=DEFAULT_ADDRESS,
port=DEFAULT_PORT,
ttl=DEFAULT_TTL,
local_only=False,
timeout_s=2):
"""Sends a query to the given multicast socket and returns responses.
Args:
query: The string query to send.
address: Multicast IP address component of the socket to send to.
port: Multicast UDP port component of the socket to send to.
ttl: TTL for multicast messages. 1 to keep traffic in-network.
timeout_s: Seconds to wait for responses.
Returns: A set of all responses that arrived before the timeout expired.
Responses are tuples of (sender_address, message).
"""
# Set up the socket as a UDP Multicast socket with the given timeout.
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
sock.setsockopt(socket.IPPROTO_IP, socket.IP_MULTICAST_TTL, ttl)
if local_only:
# Set outgoing interface to localhost to ensure no packets leave this host.
sock.setsockopt(
socket.IPPROTO_IP,
socket.IP_MULTICAST_IF,
struct.pack('!L', LOCALHOST_ADDRESS))
sock.settimeout(timeout_s)
sock.sendto(query.encode('utf-8'), (address, port))
# Set up our thread-safe Queue for handling responses.
recv_queue = queue.Queue()
def _handle_responses():
while True:
try:
data, address = sock.recvfrom(MAX_MESSAGE_BYTES)
data = data.decode('utf-8')
except socket.timeout:
recv_queue.put(None)
break
else:
_LOG.debug('Multicast response to query "%s": %s:%s',
query, address[0], data)
recv_queue.put((address[0], str(data)))
# Yield responses as they come in, giving up once timeout expires.
response_thread = threading.Thread(target=_handle_responses)
response_thread.start()
while response_thread.is_alive():
recv_tuple = recv_queue.get()
if not recv_tuple:
break
yield recv_tuple
response_thread.join() | python | def send(query,
address=DEFAULT_ADDRESS,
port=DEFAULT_PORT,
ttl=DEFAULT_TTL,
local_only=False,
timeout_s=2):
"""Sends a query to the given multicast socket and returns responses.
Args:
query: The string query to send.
address: Multicast IP address component of the socket to send to.
port: Multicast UDP port component of the socket to send to.
ttl: TTL for multicast messages. 1 to keep traffic in-network.
timeout_s: Seconds to wait for responses.
Returns: A set of all responses that arrived before the timeout expired.
Responses are tuples of (sender_address, message).
"""
# Set up the socket as a UDP Multicast socket with the given timeout.
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
sock.setsockopt(socket.IPPROTO_IP, socket.IP_MULTICAST_TTL, ttl)
if local_only:
# Set outgoing interface to localhost to ensure no packets leave this host.
sock.setsockopt(
socket.IPPROTO_IP,
socket.IP_MULTICAST_IF,
struct.pack('!L', LOCALHOST_ADDRESS))
sock.settimeout(timeout_s)
sock.sendto(query.encode('utf-8'), (address, port))
# Set up our thread-safe Queue for handling responses.
recv_queue = queue.Queue()
def _handle_responses():
while True:
try:
data, address = sock.recvfrom(MAX_MESSAGE_BYTES)
data = data.decode('utf-8')
except socket.timeout:
recv_queue.put(None)
break
else:
_LOG.debug('Multicast response to query "%s": %s:%s',
query, address[0], data)
recv_queue.put((address[0], str(data)))
# Yield responses as they come in, giving up once timeout expires.
response_thread = threading.Thread(target=_handle_responses)
response_thread.start()
while response_thread.is_alive():
recv_tuple = recv_queue.get()
if not recv_tuple:
break
yield recv_tuple
response_thread.join() | [
"def",
"send",
"(",
"query",
",",
"address",
"=",
"DEFAULT_ADDRESS",
",",
"port",
"=",
"DEFAULT_PORT",
",",
"ttl",
"=",
"DEFAULT_TTL",
",",
"local_only",
"=",
"False",
",",
"timeout_s",
"=",
"2",
")",
":",
"# Set up the socket as a UDP Multicast socket with the gi... | Sends a query to the given multicast socket and returns responses.
Args:
query: The string query to send.
address: Multicast IP address component of the socket to send to.
port: Multicast UDP port component of the socket to send to.
ttl: TTL for multicast messages. 1 to keep traffic in-network.
timeout_s: Seconds to wait for responses.
Returns: A set of all responses that arrived before the timeout expired.
Responses are tuples of (sender_address, message). | [
"Sends",
"a",
"query",
"to",
"the",
"given",
"multicast",
"socket",
"and",
"returns",
"responses",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/multicast.py#L135-L188 |
226,901 | google/openhtf | openhtf/util/multicast.py | MulticastListener.run | def run(self):
"""Listen for pings until stopped."""
self._live = True
self._sock.settimeout(self.LISTEN_TIMEOUT_S)
# Passing in INADDR_ANY means the kernel will choose the default interface.
# The localhost address is used to receive messages sent in "local_only"
# mode and the default address is used to receive all other messages.
for interface_ip in (socket.INADDR_ANY, LOCALHOST_ADDRESS):
self._sock.setsockopt(
socket.IPPROTO_IP,
socket.IP_ADD_MEMBERSHIP,
# IP_ADD_MEMBERSHIP takes the 8-byte group address followed by the IP
# assigned to the interface on which to listen.
struct.pack('!4sL', socket.inet_aton(self.address), interface_ip))
if sys.platform == 'darwin':
self._sock.setsockopt(socket.SOL_SOCKET,
socket.SO_REUSEPORT,
1) # Allow multiple listeners to bind.
else:
self._sock.setsockopt(socket.SOL_SOCKET,
socket.SO_REUSEADDR,
1) # Allow multiple listeners to bind.
self._sock.bind((self.address, self.port))
while self._live:
try:
data, address = self._sock.recvfrom(MAX_MESSAGE_BYTES)
data = data.decode('utf-8')
log_line = 'Received multicast message from %s: %s' % (address, data)
response = self._callback(data)
if response is not None:
log_line += ', responding with %s bytes' % len(response)
# Send replies out-of-band instead of with the same multicast socket
# so that multiple processes on the same host can listen for
# requests and reply (if they all try to use the multicast socket
# to reply, they conflict and this sendto fails).
response = response.encode('utf-8')
socket.socket(socket.AF_INET, socket.SOCK_DGRAM).sendto(
response, address)
_LOG.debug(log_line)
except socket.timeout:
pass
except socket.error:
_LOG.debug('Error receiving multicast message', exc_info=True) | python | def run(self):
"""Listen for pings until stopped."""
self._live = True
self._sock.settimeout(self.LISTEN_TIMEOUT_S)
# Passing in INADDR_ANY means the kernel will choose the default interface.
# The localhost address is used to receive messages sent in "local_only"
# mode and the default address is used to receive all other messages.
for interface_ip in (socket.INADDR_ANY, LOCALHOST_ADDRESS):
self._sock.setsockopt(
socket.IPPROTO_IP,
socket.IP_ADD_MEMBERSHIP,
# IP_ADD_MEMBERSHIP takes the 8-byte group address followed by the IP
# assigned to the interface on which to listen.
struct.pack('!4sL', socket.inet_aton(self.address), interface_ip))
if sys.platform == 'darwin':
self._sock.setsockopt(socket.SOL_SOCKET,
socket.SO_REUSEPORT,
1) # Allow multiple listeners to bind.
else:
self._sock.setsockopt(socket.SOL_SOCKET,
socket.SO_REUSEADDR,
1) # Allow multiple listeners to bind.
self._sock.bind((self.address, self.port))
while self._live:
try:
data, address = self._sock.recvfrom(MAX_MESSAGE_BYTES)
data = data.decode('utf-8')
log_line = 'Received multicast message from %s: %s' % (address, data)
response = self._callback(data)
if response is not None:
log_line += ', responding with %s bytes' % len(response)
# Send replies out-of-band instead of with the same multicast socket
# so that multiple processes on the same host can listen for
# requests and reply (if they all try to use the multicast socket
# to reply, they conflict and this sendto fails).
response = response.encode('utf-8')
socket.socket(socket.AF_INET, socket.SOCK_DGRAM).sendto(
response, address)
_LOG.debug(log_line)
except socket.timeout:
pass
except socket.error:
_LOG.debug('Error receiving multicast message', exc_info=True) | [
"def",
"run",
"(",
"self",
")",
":",
"self",
".",
"_live",
"=",
"True",
"self",
".",
"_sock",
".",
"settimeout",
"(",
"self",
".",
"LISTEN_TIMEOUT_S",
")",
"# Passing in INADDR_ANY means the kernel will choose the default interface.",
"# The localhost address is used to r... | Listen for pings until stopped. | [
"Listen",
"for",
"pings",
"until",
"stopped",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/multicast.py#L87-L132 |
226,902 | google/openhtf | openhtf/plugs/usb/fastboot_protocol.py | FastbootProtocol._handle_progress | def _handle_progress(self, total, progress_callback): # pylint: disable=no-self-use
"""Calls the callback with the current progress and total ."""
current = 0
while True:
current += yield
try:
progress_callback(current, total)
except Exception: # pylint: disable=broad-except
_LOG.exception('Progress callback raised an exception. %s',
progress_callback)
continue | python | def _handle_progress(self, total, progress_callback): # pylint: disable=no-self-use
"""Calls the callback with the current progress and total ."""
current = 0
while True:
current += yield
try:
progress_callback(current, total)
except Exception: # pylint: disable=broad-except
_LOG.exception('Progress callback raised an exception. %s',
progress_callback)
continue | [
"def",
"_handle_progress",
"(",
"self",
",",
"total",
",",
"progress_callback",
")",
":",
"# pylint: disable=no-self-use",
"current",
"=",
"0",
"while",
"True",
":",
"current",
"+=",
"yield",
"try",
":",
"progress_callback",
"(",
"current",
",",
"total",
")",
... | Calls the callback with the current progress and total . | [
"Calls",
"the",
"callback",
"with",
"the",
"current",
"progress",
"and",
"total",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/fastboot_protocol.py#L162-L172 |
226,903 | google/openhtf | openhtf/plugs/usb/fastboot_protocol.py | FastbootCommands._simple_command | def _simple_command(self, command, arg=None, **kwargs):
"""Send a simple command."""
self._protocol.send_command(command, arg)
return self._protocol.handle_simple_responses(**kwargs) | python | def _simple_command(self, command, arg=None, **kwargs):
"""Send a simple command."""
self._protocol.send_command(command, arg)
return self._protocol.handle_simple_responses(**kwargs) | [
"def",
"_simple_command",
"(",
"self",
",",
"command",
",",
"arg",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"self",
".",
"_protocol",
".",
"send_command",
"(",
"command",
",",
"arg",
")",
"return",
"self",
".",
"_protocol",
".",
"handle_simple_res... | Send a simple command. | [
"Send",
"a",
"simple",
"command",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/fastboot_protocol.py#L210-L213 |
226,904 | google/openhtf | openhtf/output/servers/dashboard_server.py | _discover | def _discover(**kwargs):
"""Yields info about station servers announcing themselves via multicast."""
query = station_server.MULTICAST_QUERY
for host, response in multicast.send(query, **kwargs):
try:
result = json.loads(response)
except ValueError:
_LOG.warn('Received bad JSON over multicast from %s: %s', host, response)
try:
yield StationInfo(result['cell'], host, result['port'],
result['station_id'], 'ONLINE',
result.get('test_description'),
result['test_name'])
except KeyError:
if 'last_activity_time_millis' in result:
_LOG.debug('Received old station API response on multicast. Ignoring.')
else:
_LOG.warn('Received bad multicast response from %s: %s', host, response) | python | def _discover(**kwargs):
"""Yields info about station servers announcing themselves via multicast."""
query = station_server.MULTICAST_QUERY
for host, response in multicast.send(query, **kwargs):
try:
result = json.loads(response)
except ValueError:
_LOG.warn('Received bad JSON over multicast from %s: %s', host, response)
try:
yield StationInfo(result['cell'], host, result['port'],
result['station_id'], 'ONLINE',
result.get('test_description'),
result['test_name'])
except KeyError:
if 'last_activity_time_millis' in result:
_LOG.debug('Received old station API response on multicast. Ignoring.')
else:
_LOG.warn('Received bad multicast response from %s: %s', host, response) | [
"def",
"_discover",
"(",
"*",
"*",
"kwargs",
")",
":",
"query",
"=",
"station_server",
".",
"MULTICAST_QUERY",
"for",
"host",
",",
"response",
"in",
"multicast",
".",
"send",
"(",
"query",
",",
"*",
"*",
"kwargs",
")",
":",
"try",
":",
"result",
"=",
... | Yields info about station servers announcing themselves via multicast. | [
"Yields",
"info",
"about",
"station",
"servers",
"announcing",
"themselves",
"via",
"multicast",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/servers/dashboard_server.py#L35-L52 |
226,905 | google/openhtf | openhtf/output/servers/dashboard_server.py | DashboardPubSub.update_stations | def update_stations(cls, station_info_list):
"""Called by the station discovery loop to update the station map."""
with cls.station_map_lock:
# By default, assume old stations are unreachable.
for host_port, station_info in six.iteritems(cls.station_map):
cls.station_map[host_port] = station_info._replace(status='UNREACHABLE')
for station_info in station_info_list:
host_port = '%s:%s' % (station_info.host, station_info.port)
cls.station_map[host_port] = station_info | python | def update_stations(cls, station_info_list):
"""Called by the station discovery loop to update the station map."""
with cls.station_map_lock:
# By default, assume old stations are unreachable.
for host_port, station_info in six.iteritems(cls.station_map):
cls.station_map[host_port] = station_info._replace(status='UNREACHABLE')
for station_info in station_info_list:
host_port = '%s:%s' % (station_info.host, station_info.port)
cls.station_map[host_port] = station_info | [
"def",
"update_stations",
"(",
"cls",
",",
"station_info_list",
")",
":",
"with",
"cls",
".",
"station_map_lock",
":",
"# By default, assume old stations are unreachable.",
"for",
"host_port",
",",
"station_info",
"in",
"six",
".",
"iteritems",
"(",
"cls",
".",
"sta... | Called by the station discovery loop to update the station map. | [
"Called",
"by",
"the",
"station",
"discovery",
"loop",
"to",
"update",
"the",
"station",
"map",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/servers/dashboard_server.py#L79-L89 |
226,906 | google/openhtf | openhtf/output/servers/dashboard_server.py | DashboardPubSub.publish_if_new | def publish_if_new(cls):
"""If the station map has changed, publish the new information."""
message = cls.make_message()
if message != cls.last_message:
super(DashboardPubSub, cls).publish(message)
cls.last_message = message | python | def publish_if_new(cls):
"""If the station map has changed, publish the new information."""
message = cls.make_message()
if message != cls.last_message:
super(DashboardPubSub, cls).publish(message)
cls.last_message = message | [
"def",
"publish_if_new",
"(",
"cls",
")",
":",
"message",
"=",
"cls",
".",
"make_message",
"(",
")",
"if",
"message",
"!=",
"cls",
".",
"last_message",
":",
"super",
"(",
"DashboardPubSub",
",",
"cls",
")",
".",
"publish",
"(",
"message",
")",
"cls",
"... | If the station map has changed, publish the new information. | [
"If",
"the",
"station",
"map",
"has",
"changed",
"publish",
"the",
"new",
"information",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/servers/dashboard_server.py#L92-L97 |
226,907 | google/openhtf | examples/phase_groups.py | run_basic_group | def run_basic_group():
"""Run the basic phase group example.
In this example, there are no terminal phases; all phases are run.
"""
test = htf.Test(htf.PhaseGroup(
setup=[setup_phase],
main=[main_phase],
teardown=[teardown_phase],
))
test.execute() | python | def run_basic_group():
"""Run the basic phase group example.
In this example, there are no terminal phases; all phases are run.
"""
test = htf.Test(htf.PhaseGroup(
setup=[setup_phase],
main=[main_phase],
teardown=[teardown_phase],
))
test.execute() | [
"def",
"run_basic_group",
"(",
")",
":",
"test",
"=",
"htf",
".",
"Test",
"(",
"htf",
".",
"PhaseGroup",
"(",
"setup",
"=",
"[",
"setup_phase",
"]",
",",
"main",
"=",
"[",
"main_phase",
"]",
",",
"teardown",
"=",
"[",
"teardown_phase",
"]",
",",
")",... | Run the basic phase group example.
In this example, there are no terminal phases; all phases are run. | [
"Run",
"the",
"basic",
"phase",
"group",
"example",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/examples/phase_groups.py#L54-L64 |
226,908 | google/openhtf | examples/phase_groups.py | run_setup_error_group | def run_setup_error_group():
"""Run the phase group example where an error occurs in a setup phase.
The terminal setup phase shortcuts the test. The main phases are
skipped. The PhaseGroup is not entered, so the teardown phases are also
skipped.
"""
test = htf.Test(htf.PhaseGroup(
setup=[error_setup_phase],
main=[main_phase],
teardown=[teardown_phase],
))
test.execute() | python | def run_setup_error_group():
"""Run the phase group example where an error occurs in a setup phase.
The terminal setup phase shortcuts the test. The main phases are
skipped. The PhaseGroup is not entered, so the teardown phases are also
skipped.
"""
test = htf.Test(htf.PhaseGroup(
setup=[error_setup_phase],
main=[main_phase],
teardown=[teardown_phase],
))
test.execute() | [
"def",
"run_setup_error_group",
"(",
")",
":",
"test",
"=",
"htf",
".",
"Test",
"(",
"htf",
".",
"PhaseGroup",
"(",
"setup",
"=",
"[",
"error_setup_phase",
"]",
",",
"main",
"=",
"[",
"main_phase",
"]",
",",
"teardown",
"=",
"[",
"teardown_phase",
"]",
... | Run the phase group example where an error occurs in a setup phase.
The terminal setup phase shortcuts the test. The main phases are
skipped. The PhaseGroup is not entered, so the teardown phases are also
skipped. | [
"Run",
"the",
"phase",
"group",
"example",
"where",
"an",
"error",
"occurs",
"in",
"a",
"setup",
"phase",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/examples/phase_groups.py#L67-L79 |
226,909 | google/openhtf | examples/phase_groups.py | run_main_error_group | def run_main_error_group():
"""Run the phase group example where an error occurs in a main phase.
The main phase in this example is terminal. The PhaseGroup was entered
because the setup phases ran without error, so the teardown phases are run.
The other main phase is skipped.
"""
test = htf.Test(htf.PhaseGroup(
setup=[setup_phase],
main=[error_main_phase, main_phase],
teardown=[teardown_phase],
))
test.execute() | python | def run_main_error_group():
"""Run the phase group example where an error occurs in a main phase.
The main phase in this example is terminal. The PhaseGroup was entered
because the setup phases ran without error, so the teardown phases are run.
The other main phase is skipped.
"""
test = htf.Test(htf.PhaseGroup(
setup=[setup_phase],
main=[error_main_phase, main_phase],
teardown=[teardown_phase],
))
test.execute() | [
"def",
"run_main_error_group",
"(",
")",
":",
"test",
"=",
"htf",
".",
"Test",
"(",
"htf",
".",
"PhaseGroup",
"(",
"setup",
"=",
"[",
"setup_phase",
"]",
",",
"main",
"=",
"[",
"error_main_phase",
",",
"main_phase",
"]",
",",
"teardown",
"=",
"[",
"tea... | Run the phase group example where an error occurs in a main phase.
The main phase in this example is terminal. The PhaseGroup was entered
because the setup phases ran without error, so the teardown phases are run.
The other main phase is skipped. | [
"Run",
"the",
"phase",
"group",
"example",
"where",
"an",
"error",
"occurs",
"in",
"a",
"main",
"phase",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/examples/phase_groups.py#L82-L94 |
226,910 | google/openhtf | examples/phase_groups.py | run_nested_groups | def run_nested_groups():
"""Run the nested groups example.
This example shows a PhaseGroup in a PhaseGroup. No phase is terminal, so all
are run in the order;
main_phase
inner_main_phase
inner_teardown_phase
teardown_phase
"""
test = htf.Test(
htf.PhaseGroup(
main=[
main_phase,
htf.PhaseGroup.with_teardown(inner_teardown_phase)(
inner_main_phase),
],
teardown=[teardown_phase]
)
)
test.execute() | python | def run_nested_groups():
"""Run the nested groups example.
This example shows a PhaseGroup in a PhaseGroup. No phase is terminal, so all
are run in the order;
main_phase
inner_main_phase
inner_teardown_phase
teardown_phase
"""
test = htf.Test(
htf.PhaseGroup(
main=[
main_phase,
htf.PhaseGroup.with_teardown(inner_teardown_phase)(
inner_main_phase),
],
teardown=[teardown_phase]
)
)
test.execute() | [
"def",
"run_nested_groups",
"(",
")",
":",
"test",
"=",
"htf",
".",
"Test",
"(",
"htf",
".",
"PhaseGroup",
"(",
"main",
"=",
"[",
"main_phase",
",",
"htf",
".",
"PhaseGroup",
".",
"with_teardown",
"(",
"inner_teardown_phase",
")",
"(",
"inner_main_phase",
... | Run the nested groups example.
This example shows a PhaseGroup in a PhaseGroup. No phase is terminal, so all
are run in the order;
main_phase
inner_main_phase
inner_teardown_phase
teardown_phase | [
"Run",
"the",
"nested",
"groups",
"example",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/examples/phase_groups.py#L97-L117 |
226,911 | google/openhtf | examples/phase_groups.py | run_nested_error_groups | def run_nested_error_groups():
"""Run nested groups example where an error occurs in nested main phase.
In this example, the first main phase in the nested PhaseGroup errors out.
The other inner main phase is skipped, as is the outer main phase. Both
PhaseGroups were entered, so both teardown phases are run.
"""
test = htf.Test(
htf.PhaseGroup(
main=[
htf.PhaseGroup.with_teardown(inner_teardown_phase)(
error_main_phase, main_phase),
main_phase,
],
teardown=[teardown_phase],
)
)
test.execute() | python | def run_nested_error_groups():
"""Run nested groups example where an error occurs in nested main phase.
In this example, the first main phase in the nested PhaseGroup errors out.
The other inner main phase is skipped, as is the outer main phase. Both
PhaseGroups were entered, so both teardown phases are run.
"""
test = htf.Test(
htf.PhaseGroup(
main=[
htf.PhaseGroup.with_teardown(inner_teardown_phase)(
error_main_phase, main_phase),
main_phase,
],
teardown=[teardown_phase],
)
)
test.execute() | [
"def",
"run_nested_error_groups",
"(",
")",
":",
"test",
"=",
"htf",
".",
"Test",
"(",
"htf",
".",
"PhaseGroup",
"(",
"main",
"=",
"[",
"htf",
".",
"PhaseGroup",
".",
"with_teardown",
"(",
"inner_teardown_phase",
")",
"(",
"error_main_phase",
",",
"main_phas... | Run nested groups example where an error occurs in nested main phase.
In this example, the first main phase in the nested PhaseGroup errors out.
The other inner main phase is skipped, as is the outer main phase. Both
PhaseGroups were entered, so both teardown phases are run. | [
"Run",
"nested",
"groups",
"example",
"where",
"an",
"error",
"occurs",
"in",
"nested",
"main",
"phase",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/examples/phase_groups.py#L120-L137 |
226,912 | google/openhtf | bin/units_from_xls.py | main | def main():
"""Main entry point for UNECE code .xls parsing."""
parser = argparse.ArgumentParser(
description='Reads in a .xls file and generates a units module for '
'OpenHTF.',
prog='python units_from_xls.py')
parser.add_argument('xlsfile', type=str,
help='the .xls file to parse')
parser.add_argument(
'--outfile',
type=str,
default=os.path.join(os.path.dirname(__file__), os.path.pardir,
'openhtf','util', 'units.py'),
help='where to put the generated .py file.')
args = parser.parse_args()
if not os.path.exists(args.xlsfile):
print('Unable to locate the file "%s".' % args.xlsfile)
parser.print_help()
sys.exit()
unit_defs = unit_defs_from_sheet(
xlrd.open_workbook(args.xlsfile).sheet_by_name(SHEET_NAME),
COLUMN_NAMES)
_, tmp_path = mkstemp()
with open(tmp_path, 'w') as new_file:
new_file.write(PRE)
new_file.writelines(
[line.encode('utf8', 'replace') for line in unit_defs])
new_file.write(POST)
new_file.flush()
os.remove(args.outfile)
shutil.move(tmp_path, args.outfile) | python | def main():
"""Main entry point for UNECE code .xls parsing."""
parser = argparse.ArgumentParser(
description='Reads in a .xls file and generates a units module for '
'OpenHTF.',
prog='python units_from_xls.py')
parser.add_argument('xlsfile', type=str,
help='the .xls file to parse')
parser.add_argument(
'--outfile',
type=str,
default=os.path.join(os.path.dirname(__file__), os.path.pardir,
'openhtf','util', 'units.py'),
help='where to put the generated .py file.')
args = parser.parse_args()
if not os.path.exists(args.xlsfile):
print('Unable to locate the file "%s".' % args.xlsfile)
parser.print_help()
sys.exit()
unit_defs = unit_defs_from_sheet(
xlrd.open_workbook(args.xlsfile).sheet_by_name(SHEET_NAME),
COLUMN_NAMES)
_, tmp_path = mkstemp()
with open(tmp_path, 'w') as new_file:
new_file.write(PRE)
new_file.writelines(
[line.encode('utf8', 'replace') for line in unit_defs])
new_file.write(POST)
new_file.flush()
os.remove(args.outfile)
shutil.move(tmp_path, args.outfile) | [
"def",
"main",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"'Reads in a .xls file and generates a units module for '",
"'OpenHTF.'",
",",
"prog",
"=",
"'python units_from_xls.py'",
")",
"parser",
".",
"add_argument",
"(",
... | Main entry point for UNECE code .xls parsing. | [
"Main",
"entry",
"point",
"for",
"UNECE",
"code",
".",
"xls",
"parsing",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/bin/units_from_xls.py#L169-L203 |
226,913 | google/openhtf | bin/units_from_xls.py | unit_defs_from_sheet | def unit_defs_from_sheet(sheet, column_names):
"""A generator that parses a worksheet containing UNECE code definitions.
Args:
sheet: An xldr.sheet object representing a UNECE code worksheet.
column_names: A list/tuple with the expected column names corresponding to
the unit name, code and suffix in that order.
Yields: Lines of Python source code that define OpenHTF Unit objects.
"""
seen = set()
try:
col_indices = {}
rows = sheet.get_rows()
# Find the indices for the columns we care about.
for idx, cell in enumerate(six.next(rows)):
if cell.value in column_names:
col_indices[cell.value] = idx
# loop over all remaining rows and pull out units.
for row in rows:
name = row[col_indices[column_names[0]]].value.replace("'", r'\'')
code = row[col_indices[column_names[1]]].value
suffix = row[col_indices[column_names[2]]].value.replace("'", r'\'')
key = unit_key_from_name(name)
if key in seen:
continue
seen.add(key)
# Split on ' or ' to support the units like '% or pct'
for suffix in suffix.split(' or '):
yield "%s = UnitDescriptor('%s', '%s', '''%s''')\n" % (
key, name, code, suffix)
yield "ALL_UNITS.append(%s)\n" % key
except xlrd.XLRDError:
sys.stdout.write('Unable to process the .xls file.') | python | def unit_defs_from_sheet(sheet, column_names):
"""A generator that parses a worksheet containing UNECE code definitions.
Args:
sheet: An xldr.sheet object representing a UNECE code worksheet.
column_names: A list/tuple with the expected column names corresponding to
the unit name, code and suffix in that order.
Yields: Lines of Python source code that define OpenHTF Unit objects.
"""
seen = set()
try:
col_indices = {}
rows = sheet.get_rows()
# Find the indices for the columns we care about.
for idx, cell in enumerate(six.next(rows)):
if cell.value in column_names:
col_indices[cell.value] = idx
# loop over all remaining rows and pull out units.
for row in rows:
name = row[col_indices[column_names[0]]].value.replace("'", r'\'')
code = row[col_indices[column_names[1]]].value
suffix = row[col_indices[column_names[2]]].value.replace("'", r'\'')
key = unit_key_from_name(name)
if key in seen:
continue
seen.add(key)
# Split on ' or ' to support the units like '% or pct'
for suffix in suffix.split(' or '):
yield "%s = UnitDescriptor('%s', '%s', '''%s''')\n" % (
key, name, code, suffix)
yield "ALL_UNITS.append(%s)\n" % key
except xlrd.XLRDError:
sys.stdout.write('Unable to process the .xls file.') | [
"def",
"unit_defs_from_sheet",
"(",
"sheet",
",",
"column_names",
")",
":",
"seen",
"=",
"set",
"(",
")",
"try",
":",
"col_indices",
"=",
"{",
"}",
"rows",
"=",
"sheet",
".",
"get_rows",
"(",
")",
"# Find the indices for the columns we care about.",
"for",
"id... | A generator that parses a worksheet containing UNECE code definitions.
Args:
sheet: An xldr.sheet object representing a UNECE code worksheet.
column_names: A list/tuple with the expected column names corresponding to
the unit name, code and suffix in that order.
Yields: Lines of Python source code that define OpenHTF Unit objects. | [
"A",
"generator",
"that",
"parses",
"a",
"worksheet",
"containing",
"UNECE",
"code",
"definitions",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/bin/units_from_xls.py#L206-L242 |
226,914 | google/openhtf | bin/units_from_xls.py | unit_key_from_name | def unit_key_from_name(name):
"""Return a legal python name for the given name for use as a unit key."""
result = name
for old, new in six.iteritems(UNIT_KEY_REPLACEMENTS):
result = result.replace(old, new)
# Collapse redundant underscores and convert to uppercase.
result = re.sub(r'_+', '_', result.upper())
return result | python | def unit_key_from_name(name):
"""Return a legal python name for the given name for use as a unit key."""
result = name
for old, new in six.iteritems(UNIT_KEY_REPLACEMENTS):
result = result.replace(old, new)
# Collapse redundant underscores and convert to uppercase.
result = re.sub(r'_+', '_', result.upper())
return result | [
"def",
"unit_key_from_name",
"(",
"name",
")",
":",
"result",
"=",
"name",
"for",
"old",
",",
"new",
"in",
"six",
".",
"iteritems",
"(",
"UNIT_KEY_REPLACEMENTS",
")",
":",
"result",
"=",
"result",
".",
"replace",
"(",
"old",
",",
"new",
")",
"# Collapse ... | Return a legal python name for the given name for use as a unit key. | [
"Return",
"a",
"legal",
"python",
"name",
"for",
"the",
"given",
"name",
"for",
"use",
"as",
"a",
"unit",
"key",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/bin/units_from_xls.py#L245-L255 |
226,915 | google/openhtf | openhtf/plugs/usb/adb_message.py | make_wire_commands | def make_wire_commands(*ids):
"""Assemble the commands."""
cmd_to_wire = {
cmd: sum(ord(c) << (i * 8) for i, c in enumerate(cmd)) for cmd in ids
}
wire_to_cmd = {wire: cmd for cmd, wire in six.iteritems(cmd_to_wire)}
return cmd_to_wire, wire_to_cmd | python | def make_wire_commands(*ids):
"""Assemble the commands."""
cmd_to_wire = {
cmd: sum(ord(c) << (i * 8) for i, c in enumerate(cmd)) for cmd in ids
}
wire_to_cmd = {wire: cmd for cmd, wire in six.iteritems(cmd_to_wire)}
return cmd_to_wire, wire_to_cmd | [
"def",
"make_wire_commands",
"(",
"*",
"ids",
")",
":",
"cmd_to_wire",
"=",
"{",
"cmd",
":",
"sum",
"(",
"ord",
"(",
"c",
")",
"<<",
"(",
"i",
"*",
"8",
")",
"for",
"i",
",",
"c",
"in",
"enumerate",
"(",
"cmd",
")",
")",
"for",
"cmd",
"in",
"... | Assemble the commands. | [
"Assemble",
"the",
"commands",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_message.py#L55-L61 |
226,916 | google/openhtf | openhtf/plugs/usb/adb_message.py | RawAdbMessage.to_adb_message | def to_adb_message(self, data):
"""Turn the data into an ADB message."""
message = AdbMessage(AdbMessage.WIRE_TO_CMD.get(self.cmd),
self.arg0, self.arg1, data)
if (len(data) != self.data_length or
message.data_crc32 != self.data_checksum):
raise usb_exceptions.AdbDataIntegrityError(
'%s (%s) received invalid data: %s', message, self, repr(data))
return message | python | def to_adb_message(self, data):
"""Turn the data into an ADB message."""
message = AdbMessage(AdbMessage.WIRE_TO_CMD.get(self.cmd),
self.arg0, self.arg1, data)
if (len(data) != self.data_length or
message.data_crc32 != self.data_checksum):
raise usb_exceptions.AdbDataIntegrityError(
'%s (%s) received invalid data: %s', message, self, repr(data))
return message | [
"def",
"to_adb_message",
"(",
"self",
",",
"data",
")",
":",
"message",
"=",
"AdbMessage",
"(",
"AdbMessage",
".",
"WIRE_TO_CMD",
".",
"get",
"(",
"self",
".",
"cmd",
")",
",",
"self",
".",
"arg0",
",",
"self",
".",
"arg1",
",",
"data",
")",
"if",
... | Turn the data into an ADB message. | [
"Turn",
"the",
"data",
"into",
"an",
"ADB",
"message",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_message.py#L70-L78 |
226,917 | google/openhtf | openhtf/plugs/usb/adb_message.py | AdbTransportAdapter.write_message | def write_message(self, message, timeout):
"""Send the given message over this transport.
Args:
message: The AdbMessage to send.
timeout: Use this timeout for the entire write operation, it should be an
instance of timeouts.PolledTimeout.
"""
with self._writer_lock:
self._transport.write(message.header, timeout.remaining_ms)
# Use any remaining time to send the data. Note that if we get this far,
# we always at least try to send the data (with a minimum of 10ms timeout)
# because we don't want the remote end to get out of sync because we sent
# a header but no data.
if timeout.has_expired():
_LOG.warning('Timed out between AdbMessage header and data, sending '
'data anyway with 10ms timeout')
timeout = timeouts.PolledTimeout.from_millis(10)
self._transport.write(message.data, timeout.remaining_ms) | python | def write_message(self, message, timeout):
"""Send the given message over this transport.
Args:
message: The AdbMessage to send.
timeout: Use this timeout for the entire write operation, it should be an
instance of timeouts.PolledTimeout.
"""
with self._writer_lock:
self._transport.write(message.header, timeout.remaining_ms)
# Use any remaining time to send the data. Note that if we get this far,
# we always at least try to send the data (with a minimum of 10ms timeout)
# because we don't want the remote end to get out of sync because we sent
# a header but no data.
if timeout.has_expired():
_LOG.warning('Timed out between AdbMessage header and data, sending '
'data anyway with 10ms timeout')
timeout = timeouts.PolledTimeout.from_millis(10)
self._transport.write(message.data, timeout.remaining_ms) | [
"def",
"write_message",
"(",
"self",
",",
"message",
",",
"timeout",
")",
":",
"with",
"self",
".",
"_writer_lock",
":",
"self",
".",
"_transport",
".",
"write",
"(",
"message",
".",
"header",
",",
"timeout",
".",
"remaining_ms",
")",
"# Use any remaining ti... | Send the given message over this transport.
Args:
message: The AdbMessage to send.
timeout: Use this timeout for the entire write operation, it should be an
instance of timeouts.PolledTimeout. | [
"Send",
"the",
"given",
"message",
"over",
"this",
"transport",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_message.py#L109-L128 |
226,918 | google/openhtf | openhtf/plugs/usb/adb_message.py | AdbTransportAdapter.read_message | def read_message(self, timeout):
"""Read an AdbMessage from this transport.
Args:
timeout: Timeout for the entire read operation, in the form of a
timeouts.PolledTimeout instance. Note that for packets with a data
payload, two USB reads are performed.
Returns:
The ADB message read from the device.
Raises:
UsbReadFailedError: There's an error during read, including timeout.
AdbProtocolError: A message is incorrectly formatted.
AdbTimeoutError: timeout is already expired, or expires before we read the
entire message, specifically between reading header and data packets.
"""
with self._reader_lock:
raw_header = self._transport.read(
struct.calcsize(AdbMessage.HEADER_STRUCT_FORMAT),
timeout.remaining_ms)
if not raw_header:
raise usb_exceptions.AdbProtocolError('Adb connection lost')
try:
raw_message = RawAdbMessage(*struct.unpack(
AdbMessage.HEADER_STRUCT_FORMAT, raw_header))
except struct.error as exception:
raise usb_exceptions.AdbProtocolError(
'Unable to unpack ADB command (%s): %s (%s)',
AdbMessage.HEADER_STRUCT_FORMAT, raw_header, exception)
if raw_message.data_length > 0:
if timeout.has_expired():
_LOG.warning('Timed out between AdbMessage header and data, reading '
'data anyway with 10ms timeout')
timeout = timeouts.PolledTimeout.from_millis(10)
data = self._transport.read(raw_message.data_length,
timeout.remaining_ms)
else:
data = ''
return raw_message.to_adb_message(data) | python | def read_message(self, timeout):
"""Read an AdbMessage from this transport.
Args:
timeout: Timeout for the entire read operation, in the form of a
timeouts.PolledTimeout instance. Note that for packets with a data
payload, two USB reads are performed.
Returns:
The ADB message read from the device.
Raises:
UsbReadFailedError: There's an error during read, including timeout.
AdbProtocolError: A message is incorrectly formatted.
AdbTimeoutError: timeout is already expired, or expires before we read the
entire message, specifically between reading header and data packets.
"""
with self._reader_lock:
raw_header = self._transport.read(
struct.calcsize(AdbMessage.HEADER_STRUCT_FORMAT),
timeout.remaining_ms)
if not raw_header:
raise usb_exceptions.AdbProtocolError('Adb connection lost')
try:
raw_message = RawAdbMessage(*struct.unpack(
AdbMessage.HEADER_STRUCT_FORMAT, raw_header))
except struct.error as exception:
raise usb_exceptions.AdbProtocolError(
'Unable to unpack ADB command (%s): %s (%s)',
AdbMessage.HEADER_STRUCT_FORMAT, raw_header, exception)
if raw_message.data_length > 0:
if timeout.has_expired():
_LOG.warning('Timed out between AdbMessage header and data, reading '
'data anyway with 10ms timeout')
timeout = timeouts.PolledTimeout.from_millis(10)
data = self._transport.read(raw_message.data_length,
timeout.remaining_ms)
else:
data = ''
return raw_message.to_adb_message(data) | [
"def",
"read_message",
"(",
"self",
",",
"timeout",
")",
":",
"with",
"self",
".",
"_reader_lock",
":",
"raw_header",
"=",
"self",
".",
"_transport",
".",
"read",
"(",
"struct",
".",
"calcsize",
"(",
"AdbMessage",
".",
"HEADER_STRUCT_FORMAT",
")",
",",
"ti... | Read an AdbMessage from this transport.
Args:
timeout: Timeout for the entire read operation, in the form of a
timeouts.PolledTimeout instance. Note that for packets with a data
payload, two USB reads are performed.
Returns:
The ADB message read from the device.
Raises:
UsbReadFailedError: There's an error during read, including timeout.
AdbProtocolError: A message is incorrectly formatted.
AdbTimeoutError: timeout is already expired, or expires before we read the
entire message, specifically between reading header and data packets. | [
"Read",
"an",
"AdbMessage",
"from",
"this",
"transport",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_message.py#L130-L172 |
226,919 | google/openhtf | openhtf/plugs/usb/adb_message.py | AdbTransportAdapter.read_until | def read_until(self, expected_commands, timeout):
"""Read AdbMessages from this transport until we get an expected command.
The ADB protocol specifies that before a successful CNXN handshake, any
other packets must be ignored, so this method provides the ability to
ignore unwanted commands. It's primarily used during the initial
connection to the device. See Read() for more details, including more
exceptions that may be raised.
Args:
expected_commands: Iterable of expected command responses, like
('CNXN', 'AUTH').
timeout: timeouts.PolledTimeout object to use for timeout.
Returns:
The ADB message received that matched one of expected_commands.
Raises:
AdbProtocolError: If timeout expires between reads, this can happen
if we are getting spammed with unexpected commands.
"""
msg = timeouts.loop_until_timeout_or_valid(
timeout, lambda: self.read_message(timeout),
lambda m: m.command in expected_commands, 0)
if msg.command not in expected_commands:
raise usb_exceptions.AdbTimeoutError(
'Timed out establishing connection, waiting for: %s',
expected_commands)
return msg | python | def read_until(self, expected_commands, timeout):
"""Read AdbMessages from this transport until we get an expected command.
The ADB protocol specifies that before a successful CNXN handshake, any
other packets must be ignored, so this method provides the ability to
ignore unwanted commands. It's primarily used during the initial
connection to the device. See Read() for more details, including more
exceptions that may be raised.
Args:
expected_commands: Iterable of expected command responses, like
('CNXN', 'AUTH').
timeout: timeouts.PolledTimeout object to use for timeout.
Returns:
The ADB message received that matched one of expected_commands.
Raises:
AdbProtocolError: If timeout expires between reads, this can happen
if we are getting spammed with unexpected commands.
"""
msg = timeouts.loop_until_timeout_or_valid(
timeout, lambda: self.read_message(timeout),
lambda m: m.command in expected_commands, 0)
if msg.command not in expected_commands:
raise usb_exceptions.AdbTimeoutError(
'Timed out establishing connection, waiting for: %s',
expected_commands)
return msg | [
"def",
"read_until",
"(",
"self",
",",
"expected_commands",
",",
"timeout",
")",
":",
"msg",
"=",
"timeouts",
".",
"loop_until_timeout_or_valid",
"(",
"timeout",
",",
"lambda",
":",
"self",
".",
"read_message",
"(",
"timeout",
")",
",",
"lambda",
"m",
":",
... | Read AdbMessages from this transport until we get an expected command.
The ADB protocol specifies that before a successful CNXN handshake, any
other packets must be ignored, so this method provides the ability to
ignore unwanted commands. It's primarily used during the initial
connection to the device. See Read() for more details, including more
exceptions that may be raised.
Args:
expected_commands: Iterable of expected command responses, like
('CNXN', 'AUTH').
timeout: timeouts.PolledTimeout object to use for timeout.
Returns:
The ADB message received that matched one of expected_commands.
Raises:
AdbProtocolError: If timeout expires between reads, this can happen
if we are getting spammed with unexpected commands. | [
"Read",
"AdbMessages",
"from",
"this",
"transport",
"until",
"we",
"get",
"an",
"expected",
"command",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_message.py#L174-L202 |
226,920 | google/openhtf | openhtf/plugs/usb/adb_message.py | AdbMessage.header | def header(self):
"""The message header."""
return struct.pack(
self.HEADER_STRUCT_FORMAT, self._command, self.arg0, self.arg1,
len(self.data), self.data_crc32, self.magic) | python | def header(self):
"""The message header."""
return struct.pack(
self.HEADER_STRUCT_FORMAT, self._command, self.arg0, self.arg1,
len(self.data), self.data_crc32, self.magic) | [
"def",
"header",
"(",
"self",
")",
":",
"return",
"struct",
".",
"pack",
"(",
"self",
".",
"HEADER_STRUCT_FORMAT",
",",
"self",
".",
"_command",
",",
"self",
".",
"arg0",
",",
"self",
".",
"arg1",
",",
"len",
"(",
"self",
".",
"data",
")",
",",
"se... | The message header. | [
"The",
"message",
"header",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_message.py#L276-L280 |
226,921 | google/openhtf | openhtf/plugs/usb/filesync_service.py | _make_message_type | def _make_message_type(name, attributes, has_data=True):
"""Make a message type for the AdbTransport subclasses."""
def assert_command_is(self, command): # pylint: disable=invalid-name
"""Assert that a message's command matches the given command."""
if self.command != command:
raise usb_exceptions.AdbProtocolError(
'Expected %s command, received %s', command, self)
return type(name, (collections.namedtuple(name, attributes),),
{
'assert_command_is': assert_command_is,
'has_data': has_data,
# Struct format on the wire has an unsigned int for each attr.
'struct_format': '<%sI' % len(attributes.split()),
}) | python | def _make_message_type(name, attributes, has_data=True):
"""Make a message type for the AdbTransport subclasses."""
def assert_command_is(self, command): # pylint: disable=invalid-name
"""Assert that a message's command matches the given command."""
if self.command != command:
raise usb_exceptions.AdbProtocolError(
'Expected %s command, received %s', command, self)
return type(name, (collections.namedtuple(name, attributes),),
{
'assert_command_is': assert_command_is,
'has_data': has_data,
# Struct format on the wire has an unsigned int for each attr.
'struct_format': '<%sI' % len(attributes.split()),
}) | [
"def",
"_make_message_type",
"(",
"name",
",",
"attributes",
",",
"has_data",
"=",
"True",
")",
":",
"def",
"assert_command_is",
"(",
"self",
",",
"command",
")",
":",
"# pylint: disable=invalid-name",
"\"\"\"Assert that a message's command matches the given command.\"\"\""... | Make a message type for the AdbTransport subclasses. | [
"Make",
"a",
"message",
"type",
"for",
"the",
"AdbTransport",
"subclasses",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L130-L145 |
226,922 | google/openhtf | openhtf/plugs/usb/filesync_service.py | FilesyncService.stat | def stat(self, filename, timeout=None):
"""Return device file stat."""
transport = StatFilesyncTransport(self.stream)
transport.write_data('STAT', filename, timeout)
stat_msg = transport.read_message(timeout)
stat_msg.assert_command_is('STAT')
return DeviceFileStat(filename, stat_msg.mode, stat_msg.size, stat_msg.time) | python | def stat(self, filename, timeout=None):
"""Return device file stat."""
transport = StatFilesyncTransport(self.stream)
transport.write_data('STAT', filename, timeout)
stat_msg = transport.read_message(timeout)
stat_msg.assert_command_is('STAT')
return DeviceFileStat(filename, stat_msg.mode, stat_msg.size, stat_msg.time) | [
"def",
"stat",
"(",
"self",
",",
"filename",
",",
"timeout",
"=",
"None",
")",
":",
"transport",
"=",
"StatFilesyncTransport",
"(",
"self",
".",
"stream",
")",
"transport",
".",
"write_data",
"(",
"'STAT'",
",",
"filename",
",",
"timeout",
")",
"stat_msg",... | Return device file stat. | [
"Return",
"device",
"file",
"stat",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L173-L179 |
226,923 | google/openhtf | openhtf/plugs/usb/filesync_service.py | FilesyncService.list | def list(self, path, timeout=None):
"""List directory contents on the device.
Args:
path: List the contents of this directory.
timeout: Timeout to use for this operation.
Returns:
Generator yielding DeviceFileStat tuples representing the contents of
the requested path.
"""
transport = DentFilesyncTransport(self.stream)
transport.write_data('LIST', path, timeout)
return (DeviceFileStat(dent_msg.name, dent_msg.mode,
dent_msg.size, dent_msg.time) for dent_msg in
transport.read_until_done('DENT', timeout)) | python | def list(self, path, timeout=None):
"""List directory contents on the device.
Args:
path: List the contents of this directory.
timeout: Timeout to use for this operation.
Returns:
Generator yielding DeviceFileStat tuples representing the contents of
the requested path.
"""
transport = DentFilesyncTransport(self.stream)
transport.write_data('LIST', path, timeout)
return (DeviceFileStat(dent_msg.name, dent_msg.mode,
dent_msg.size, dent_msg.time) for dent_msg in
transport.read_until_done('DENT', timeout)) | [
"def",
"list",
"(",
"self",
",",
"path",
",",
"timeout",
"=",
"None",
")",
":",
"transport",
"=",
"DentFilesyncTransport",
"(",
"self",
".",
"stream",
")",
"transport",
".",
"write_data",
"(",
"'LIST'",
",",
"path",
",",
"timeout",
")",
"return",
"(",
... | List directory contents on the device.
Args:
path: List the contents of this directory.
timeout: Timeout to use for this operation.
Returns:
Generator yielding DeviceFileStat tuples representing the contents of
the requested path. | [
"List",
"directory",
"contents",
"on",
"the",
"device",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L181-L196 |
226,924 | google/openhtf | openhtf/plugs/usb/filesync_service.py | FilesyncService.recv | def recv(self, filename, dest_file, timeout=None):
"""Retrieve a file from the device into the file-like dest_file."""
transport = DataFilesyncTransport(self.stream)
transport.write_data('RECV', filename, timeout)
for data_msg in transport.read_until_done('DATA', timeout):
dest_file.write(data_msg.data) | python | def recv(self, filename, dest_file, timeout=None):
"""Retrieve a file from the device into the file-like dest_file."""
transport = DataFilesyncTransport(self.stream)
transport.write_data('RECV', filename, timeout)
for data_msg in transport.read_until_done('DATA', timeout):
dest_file.write(data_msg.data) | [
"def",
"recv",
"(",
"self",
",",
"filename",
",",
"dest_file",
",",
"timeout",
"=",
"None",
")",
":",
"transport",
"=",
"DataFilesyncTransport",
"(",
"self",
".",
"stream",
")",
"transport",
".",
"write_data",
"(",
"'RECV'",
",",
"filename",
",",
"timeout"... | Retrieve a file from the device into the file-like dest_file. | [
"Retrieve",
"a",
"file",
"from",
"the",
"device",
"into",
"the",
"file",
"-",
"like",
"dest_file",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L198-L203 |
226,925 | google/openhtf | openhtf/plugs/usb/filesync_service.py | FilesyncService._check_for_fail_message | def _check_for_fail_message(self, transport, exc_info, timeout): # pylint: disable=no-self-use
"""Check for a 'FAIL' message from transport.
This method always raises, if 'FAIL' was read, it will raise an
AdbRemoteError with the message, otherwise it will raise based on
exc_info, which should be a tuple as per sys.exc_info().
Args:
transport: Transport from which to read for a 'FAIL' message.
exc_info: Exception info to raise if no 'FAIL' is read.
timeout: Timeout to use for the read operation.
Raises:
AdbRemoteError: If a 'FAIL' is read, otherwise raises exc_info.
"""
try:
transport.read_message(timeout)
except usb_exceptions.CommonUsbError:
# If we got a remote error, raise that exception.
if sys.exc_info()[0] is usb_exceptions.AdbRemoteError:
raise
# Otherwise reraise the original exception.
raise_with_traceback(exc_info[0](exc_info[1]), traceback=exc_info[2]) | python | def _check_for_fail_message(self, transport, exc_info, timeout): # pylint: disable=no-self-use
"""Check for a 'FAIL' message from transport.
This method always raises, if 'FAIL' was read, it will raise an
AdbRemoteError with the message, otherwise it will raise based on
exc_info, which should be a tuple as per sys.exc_info().
Args:
transport: Transport from which to read for a 'FAIL' message.
exc_info: Exception info to raise if no 'FAIL' is read.
timeout: Timeout to use for the read operation.
Raises:
AdbRemoteError: If a 'FAIL' is read, otherwise raises exc_info.
"""
try:
transport.read_message(timeout)
except usb_exceptions.CommonUsbError:
# If we got a remote error, raise that exception.
if sys.exc_info()[0] is usb_exceptions.AdbRemoteError:
raise
# Otherwise reraise the original exception.
raise_with_traceback(exc_info[0](exc_info[1]), traceback=exc_info[2]) | [
"def",
"_check_for_fail_message",
"(",
"self",
",",
"transport",
",",
"exc_info",
",",
"timeout",
")",
":",
"# pylint: disable=no-self-use",
"try",
":",
"transport",
".",
"read_message",
"(",
"timeout",
")",
"except",
"usb_exceptions",
".",
"CommonUsbError",
":",
... | Check for a 'FAIL' message from transport.
This method always raises, if 'FAIL' was read, it will raise an
AdbRemoteError with the message, otherwise it will raise based on
exc_info, which should be a tuple as per sys.exc_info().
Args:
transport: Transport from which to read for a 'FAIL' message.
exc_info: Exception info to raise if no 'FAIL' is read.
timeout: Timeout to use for the read operation.
Raises:
AdbRemoteError: If a 'FAIL' is read, otherwise raises exc_info. | [
"Check",
"for",
"a",
"FAIL",
"message",
"from",
"transport",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L205-L227 |
226,926 | google/openhtf | openhtf/plugs/usb/filesync_service.py | AbstractFilesyncTransport.write_data | def write_data(self, command, data, timeout=None):
"""Shortcut for writing specifically a DataMessage."""
self.write_message(FilesyncMessageTypes.DataMessage(command, data), timeout) | python | def write_data(self, command, data, timeout=None):
"""Shortcut for writing specifically a DataMessage."""
self.write_message(FilesyncMessageTypes.DataMessage(command, data), timeout) | [
"def",
"write_data",
"(",
"self",
",",
"command",
",",
"data",
",",
"timeout",
"=",
"None",
")",
":",
"self",
".",
"write_message",
"(",
"FilesyncMessageTypes",
".",
"DataMessage",
"(",
"command",
",",
"data",
")",
",",
"timeout",
")"
] | Shortcut for writing specifically a DataMessage. | [
"Shortcut",
"for",
"writing",
"specifically",
"a",
"DataMessage",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L362-L364 |
226,927 | google/openhtf | openhtf/plugs/usb/filesync_service.py | AbstractFilesyncTransport.read_until_done | def read_until_done(self, command, timeout=None):
"""Yield messages read until we receive a 'DONE' command.
Read messages of the given command until we receive a 'DONE' command. If a
command different than the requested one is received, an AdbProtocolError
is raised.
Args:
command: The command to expect, like 'DENT' or 'DATA'.
timeout: The timeouts.PolledTimeout to use for this operation.
Yields:
Messages read, of type self.RECV_MSG_TYPE, see read_message().
Raises:
AdbProtocolError: If an unexpected command is read.
AdbRemoteError: If a 'FAIL' message is read.
"""
message = self.read_message(timeout)
while message.command != 'DONE':
message.assert_command_is(command)
yield message
message = self.read_message(timeout) | python | def read_until_done(self, command, timeout=None):
"""Yield messages read until we receive a 'DONE' command.
Read messages of the given command until we receive a 'DONE' command. If a
command different than the requested one is received, an AdbProtocolError
is raised.
Args:
command: The command to expect, like 'DENT' or 'DATA'.
timeout: The timeouts.PolledTimeout to use for this operation.
Yields:
Messages read, of type self.RECV_MSG_TYPE, see read_message().
Raises:
AdbProtocolError: If an unexpected command is read.
AdbRemoteError: If a 'FAIL' message is read.
"""
message = self.read_message(timeout)
while message.command != 'DONE':
message.assert_command_is(command)
yield message
message = self.read_message(timeout) | [
"def",
"read_until_done",
"(",
"self",
",",
"command",
",",
"timeout",
"=",
"None",
")",
":",
"message",
"=",
"self",
".",
"read_message",
"(",
"timeout",
")",
"while",
"message",
".",
"command",
"!=",
"'DONE'",
":",
"message",
".",
"assert_command_is",
"(... | Yield messages read until we receive a 'DONE' command.
Read messages of the given command until we receive a 'DONE' command. If a
command different than the requested one is received, an AdbProtocolError
is raised.
Args:
command: The command to expect, like 'DENT' or 'DATA'.
timeout: The timeouts.PolledTimeout to use for this operation.
Yields:
Messages read, of type self.RECV_MSG_TYPE, see read_message().
Raises:
AdbProtocolError: If an unexpected command is read.
AdbRemoteError: If a 'FAIL' message is read. | [
"Yield",
"messages",
"read",
"until",
"we",
"receive",
"a",
"DONE",
"command",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L391-L413 |
226,928 | google/openhtf | openhtf/plugs/usb/filesync_service.py | AbstractFilesyncTransport.read_message | def read_message(self, timeout=None):
"""Read a message from this transport and return it.
Reads a message of RECV_MSG_TYPE and returns it. Note that this method
abstracts the data length and data read so that the caller simply gets the
data along with the header in the returned message.
Args:
timeout: timeouts.PolledTimeout to use for the operation.
Returns:
An instance of self.RECV_MSG_TYPE that was read from self.stream.
Raises:
AdbProtocolError: If an invalid response is received.
AdbRemoteError: If a FAIL response is received.
"""
raw_data = self.stream.read(
struct.calcsize(self.RECV_MSG_TYPE.struct_format), timeout)
try:
raw_message = struct.unpack(self.RECV_MSG_TYPE.struct_format, raw_data)
except struct.error:
raise usb_exceptions.AdbProtocolError(
'%s expected format "%s", got data %s', self,
self.RECV_MSG_TYPE.struct_format, raw_data)
if raw_message[0] not in self.WIRE_TO_CMD:
raise usb_exceptions.AdbProtocolError(
'Unrecognized command id: %s', raw_message)
# Swap out the wire command with the string equivalent.
raw_message = (self.WIRE_TO_CMD[raw_message[0]],) + raw_message[1:]
if self.RECV_MSG_TYPE.has_data and raw_message[-1]:
# For messages that have data, the length of the data is the last field
# in the struct. We do another read and swap out that length for the
# actual data read before we create the namedtuple to return.
data_len = raw_message[-1]
raw_message = raw_message[:-1] + (self.stream.read(data_len, timeout),)
if raw_message[0] not in self.VALID_RESPONSES:
raise usb_exceptions.AdbProtocolError(
'%s not a valid response for %s', raw_message[0], self)
if raw_message[0] == 'FAIL':
raise usb_exceptions.AdbRemoteError(
'Remote ADB failure: %s', raw_message)
return self.RECV_MSG_TYPE(*raw_message) | python | def read_message(self, timeout=None):
"""Read a message from this transport and return it.
Reads a message of RECV_MSG_TYPE and returns it. Note that this method
abstracts the data length and data read so that the caller simply gets the
data along with the header in the returned message.
Args:
timeout: timeouts.PolledTimeout to use for the operation.
Returns:
An instance of self.RECV_MSG_TYPE that was read from self.stream.
Raises:
AdbProtocolError: If an invalid response is received.
AdbRemoteError: If a FAIL response is received.
"""
raw_data = self.stream.read(
struct.calcsize(self.RECV_MSG_TYPE.struct_format), timeout)
try:
raw_message = struct.unpack(self.RECV_MSG_TYPE.struct_format, raw_data)
except struct.error:
raise usb_exceptions.AdbProtocolError(
'%s expected format "%s", got data %s', self,
self.RECV_MSG_TYPE.struct_format, raw_data)
if raw_message[0] not in self.WIRE_TO_CMD:
raise usb_exceptions.AdbProtocolError(
'Unrecognized command id: %s', raw_message)
# Swap out the wire command with the string equivalent.
raw_message = (self.WIRE_TO_CMD[raw_message[0]],) + raw_message[1:]
if self.RECV_MSG_TYPE.has_data and raw_message[-1]:
# For messages that have data, the length of the data is the last field
# in the struct. We do another read and swap out that length for the
# actual data read before we create the namedtuple to return.
data_len = raw_message[-1]
raw_message = raw_message[:-1] + (self.stream.read(data_len, timeout),)
if raw_message[0] not in self.VALID_RESPONSES:
raise usb_exceptions.AdbProtocolError(
'%s not a valid response for %s', raw_message[0], self)
if raw_message[0] == 'FAIL':
raise usb_exceptions.AdbRemoteError(
'Remote ADB failure: %s', raw_message)
return self.RECV_MSG_TYPE(*raw_message) | [
"def",
"read_message",
"(",
"self",
",",
"timeout",
"=",
"None",
")",
":",
"raw_data",
"=",
"self",
".",
"stream",
".",
"read",
"(",
"struct",
".",
"calcsize",
"(",
"self",
".",
"RECV_MSG_TYPE",
".",
"struct_format",
")",
",",
"timeout",
")",
"try",
":... | Read a message from this transport and return it.
Reads a message of RECV_MSG_TYPE and returns it. Note that this method
abstracts the data length and data read so that the caller simply gets the
data along with the header in the returned message.
Args:
timeout: timeouts.PolledTimeout to use for the operation.
Returns:
An instance of self.RECV_MSG_TYPE that was read from self.stream.
Raises:
AdbProtocolError: If an invalid response is received.
AdbRemoteError: If a FAIL response is received. | [
"Read",
"a",
"message",
"from",
"this",
"transport",
"and",
"return",
"it",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/filesync_service.py#L415-L461 |
226,929 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | _lazy_load_units_by_code | def _lazy_load_units_by_code():
"""Populate dict of units by code iff UNITS_BY_CODE is empty."""
if UNITS_BY_CODE:
# already populated
return
for unit in units.UNITS_BY_NAME.values():
UNITS_BY_CODE[unit.code] = unit | python | def _lazy_load_units_by_code():
"""Populate dict of units by code iff UNITS_BY_CODE is empty."""
if UNITS_BY_CODE:
# already populated
return
for unit in units.UNITS_BY_NAME.values():
UNITS_BY_CODE[unit.code] = unit | [
"def",
"_lazy_load_units_by_code",
"(",
")",
":",
"if",
"UNITS_BY_CODE",
":",
"# already populated",
"return",
"for",
"unit",
"in",
"units",
".",
"UNITS_BY_NAME",
".",
"values",
"(",
")",
":",
"UNITS_BY_CODE",
"[",
"unit",
".",
"code",
"]",
"=",
"unit"
] | Populate dict of units by code iff UNITS_BY_CODE is empty. | [
"Populate",
"dict",
"of",
"units",
"by",
"code",
"iff",
"UNITS_BY_CODE",
"is",
"empty",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L52-L59 |
226,930 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | _populate_basic_data | def _populate_basic_data(mfg_event, record):
"""Copies data from the OpenHTF TestRecord to the MfgEvent proto."""
# TODO:
# * Missing in proto: set run name from metadata.
# * `part_tags` field on proto is unused
# * `timings` field on proto is unused.
# * Handle arbitrary units as uom_code/uom_suffix.
# Populate non-repeated fields.
mfg_event.dut_serial = record.dut_id
mfg_event.start_time_ms = record.start_time_millis
mfg_event.end_time_ms = record.end_time_millis
mfg_event.tester_name = record.station_id
mfg_event.test_name = record.metadata.get('test_name') or record.station_id
mfg_event.test_status = test_runs_converter.OUTCOME_MAP[record.outcome]
mfg_event.operator_name = record.metadata.get('operator_name', '')
mfg_event.test_version = str(record.metadata.get('test_version', ''))
mfg_event.test_description = record.metadata.get('test_description', '')
# Populate part_tags.
mfg_event.part_tags.extend(record.metadata.get('part_tags', []))
# Populate phases.
for phase in record.phases:
mfg_phase = mfg_event.phases.add()
mfg_phase.name = phase.name
mfg_phase.description = phase.codeinfo.sourcecode
mfg_phase.timing.start_time_millis = phase.start_time_millis
mfg_phase.timing.end_time_millis = phase.end_time_millis
# Populate failure codes.
for details in record.outcome_details:
failure_code = mfg_event.failure_codes.add()
failure_code.code = details.code
failure_code.details = details.description
# Populate test logs.
for log_record in record.log_records:
test_log = mfg_event.test_logs.add()
test_log.timestamp_millis = log_record.timestamp_millis
test_log.log_message = log_record.message
test_log.logger_name = log_record.logger_name
test_log.levelno = log_record.level
if log_record.level <= logging.DEBUG:
test_log.level = test_runs_pb2.TestRunLogMessage.DEBUG
elif log_record.level <= logging.INFO:
test_log.level = test_runs_pb2.TestRunLogMessage.INFO
elif log_record.level <= logging.WARNING:
test_log.level = test_runs_pb2.TestRunLogMessage.WARNING
elif log_record.level <= logging.ERROR:
test_log.level = test_runs_pb2.TestRunLogMessage.ERROR
elif log_record.level <= logging.CRITICAL:
test_log.level = test_runs_pb2.TestRunLogMessage.CRITICAL
test_log.log_source = log_record.source
test_log.lineno = log_record.lineno | python | def _populate_basic_data(mfg_event, record):
"""Copies data from the OpenHTF TestRecord to the MfgEvent proto."""
# TODO:
# * Missing in proto: set run name from metadata.
# * `part_tags` field on proto is unused
# * `timings` field on proto is unused.
# * Handle arbitrary units as uom_code/uom_suffix.
# Populate non-repeated fields.
mfg_event.dut_serial = record.dut_id
mfg_event.start_time_ms = record.start_time_millis
mfg_event.end_time_ms = record.end_time_millis
mfg_event.tester_name = record.station_id
mfg_event.test_name = record.metadata.get('test_name') or record.station_id
mfg_event.test_status = test_runs_converter.OUTCOME_MAP[record.outcome]
mfg_event.operator_name = record.metadata.get('operator_name', '')
mfg_event.test_version = str(record.metadata.get('test_version', ''))
mfg_event.test_description = record.metadata.get('test_description', '')
# Populate part_tags.
mfg_event.part_tags.extend(record.metadata.get('part_tags', []))
# Populate phases.
for phase in record.phases:
mfg_phase = mfg_event.phases.add()
mfg_phase.name = phase.name
mfg_phase.description = phase.codeinfo.sourcecode
mfg_phase.timing.start_time_millis = phase.start_time_millis
mfg_phase.timing.end_time_millis = phase.end_time_millis
# Populate failure codes.
for details in record.outcome_details:
failure_code = mfg_event.failure_codes.add()
failure_code.code = details.code
failure_code.details = details.description
# Populate test logs.
for log_record in record.log_records:
test_log = mfg_event.test_logs.add()
test_log.timestamp_millis = log_record.timestamp_millis
test_log.log_message = log_record.message
test_log.logger_name = log_record.logger_name
test_log.levelno = log_record.level
if log_record.level <= logging.DEBUG:
test_log.level = test_runs_pb2.TestRunLogMessage.DEBUG
elif log_record.level <= logging.INFO:
test_log.level = test_runs_pb2.TestRunLogMessage.INFO
elif log_record.level <= logging.WARNING:
test_log.level = test_runs_pb2.TestRunLogMessage.WARNING
elif log_record.level <= logging.ERROR:
test_log.level = test_runs_pb2.TestRunLogMessage.ERROR
elif log_record.level <= logging.CRITICAL:
test_log.level = test_runs_pb2.TestRunLogMessage.CRITICAL
test_log.log_source = log_record.source
test_log.lineno = log_record.lineno | [
"def",
"_populate_basic_data",
"(",
"mfg_event",
",",
"record",
")",
":",
"# TODO:",
"# * Missing in proto: set run name from metadata.",
"# * `part_tags` field on proto is unused",
"# * `timings` field on proto is unused.",
"# * Handle arbitrary units as uom_code/uom_suffix.",
"# ... | Copies data from the OpenHTF TestRecord to the MfgEvent proto. | [
"Copies",
"data",
"from",
"the",
"OpenHTF",
"TestRecord",
"to",
"the",
"MfgEvent",
"proto",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L105-L159 |
226,931 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | _attach_record_as_json | def _attach_record_as_json(mfg_event, record):
"""Attach a copy of the record as JSON so we have an un-mangled copy."""
attachment = mfg_event.attachment.add()
attachment.name = TEST_RECORD_ATTACHMENT_NAME
test_record_dict = htf_data.convert_to_base_types(record)
attachment.value_binary = _convert_object_to_json(test_record_dict)
attachment.type = test_runs_pb2.TEXT_UTF8 | python | def _attach_record_as_json(mfg_event, record):
"""Attach a copy of the record as JSON so we have an un-mangled copy."""
attachment = mfg_event.attachment.add()
attachment.name = TEST_RECORD_ATTACHMENT_NAME
test_record_dict = htf_data.convert_to_base_types(record)
attachment.value_binary = _convert_object_to_json(test_record_dict)
attachment.type = test_runs_pb2.TEXT_UTF8 | [
"def",
"_attach_record_as_json",
"(",
"mfg_event",
",",
"record",
")",
":",
"attachment",
"=",
"mfg_event",
".",
"attachment",
".",
"add",
"(",
")",
"attachment",
".",
"name",
"=",
"TEST_RECORD_ATTACHMENT_NAME",
"test_record_dict",
"=",
"htf_data",
".",
"convert_t... | Attach a copy of the record as JSON so we have an un-mangled copy. | [
"Attach",
"a",
"copy",
"of",
"the",
"record",
"as",
"JSON",
"so",
"we",
"have",
"an",
"un",
"-",
"mangled",
"copy",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L162-L168 |
226,932 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | _attach_config | def _attach_config(mfg_event, record):
"""Attaches the OpenHTF config file as JSON."""
if 'config' not in record.metadata:
return
attachment = mfg_event.attachment.add()
attachment.name = 'config'
attachment.value_binary = _convert_object_to_json(record.metadata['config'])
attachment.type = test_runs_pb2.TEXT_UTF8 | python | def _attach_config(mfg_event, record):
"""Attaches the OpenHTF config file as JSON."""
if 'config' not in record.metadata:
return
attachment = mfg_event.attachment.add()
attachment.name = 'config'
attachment.value_binary = _convert_object_to_json(record.metadata['config'])
attachment.type = test_runs_pb2.TEXT_UTF8 | [
"def",
"_attach_config",
"(",
"mfg_event",
",",
"record",
")",
":",
"if",
"'config'",
"not",
"in",
"record",
".",
"metadata",
":",
"return",
"attachment",
"=",
"mfg_event",
".",
"attachment",
".",
"add",
"(",
")",
"attachment",
".",
"name",
"=",
"'config'"... | Attaches the OpenHTF config file as JSON. | [
"Attaches",
"the",
"OpenHTF",
"config",
"file",
"as",
"JSON",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L186-L193 |
226,933 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | phase_uniquizer | def phase_uniquizer(all_phases):
"""Makes the names of phase measurement and attachments unique.
This function will make the names of measurements and attachments unique.
It modifies the input all_phases.
Args:
all_phases: the phases to make unique
Returns:
the phases now modified.
"""
measurement_name_maker = UniqueNameMaker(
itertools.chain.from_iterable(
phase.measurements.keys() for phase in all_phases
if phase.measurements))
attachment_names = list(itertools.chain.from_iterable(
phase.attachments.keys() for phase in all_phases))
attachment_names.extend(itertools.chain.from_iterable([
'multidim_' + name for name, meas in phase.measurements.items()
if meas.dimensions is not None
] for phase in all_phases if phase.measurements))
attachment_name_maker = UniqueNameMaker(attachment_names)
for phase in all_phases:
# Make measurements unique.
for name, _ in sorted(phase.measurements.items()):
old_name = name
name = measurement_name_maker.make_unique(name)
phase.measurements[old_name].name = name
phase.measurements[name] = phase.measurements.pop(old_name)
# Make attachments unique.
for name, _ in sorted(phase.attachments.items()):
old_name = name
name = attachment_name_maker.make_unique(name)
phase.attachments[old_name].name = name
phase.attachments[name] = phase.attachments.pop(old_name)
return all_phases | python | def phase_uniquizer(all_phases):
"""Makes the names of phase measurement and attachments unique.
This function will make the names of measurements and attachments unique.
It modifies the input all_phases.
Args:
all_phases: the phases to make unique
Returns:
the phases now modified.
"""
measurement_name_maker = UniqueNameMaker(
itertools.chain.from_iterable(
phase.measurements.keys() for phase in all_phases
if phase.measurements))
attachment_names = list(itertools.chain.from_iterable(
phase.attachments.keys() for phase in all_phases))
attachment_names.extend(itertools.chain.from_iterable([
'multidim_' + name for name, meas in phase.measurements.items()
if meas.dimensions is not None
] for phase in all_phases if phase.measurements))
attachment_name_maker = UniqueNameMaker(attachment_names)
for phase in all_phases:
# Make measurements unique.
for name, _ in sorted(phase.measurements.items()):
old_name = name
name = measurement_name_maker.make_unique(name)
phase.measurements[old_name].name = name
phase.measurements[name] = phase.measurements.pop(old_name)
# Make attachments unique.
for name, _ in sorted(phase.attachments.items()):
old_name = name
name = attachment_name_maker.make_unique(name)
phase.attachments[old_name].name = name
phase.attachments[name] = phase.attachments.pop(old_name)
return all_phases | [
"def",
"phase_uniquizer",
"(",
"all_phases",
")",
":",
"measurement_name_maker",
"=",
"UniqueNameMaker",
"(",
"itertools",
".",
"chain",
".",
"from_iterable",
"(",
"phase",
".",
"measurements",
".",
"keys",
"(",
")",
"for",
"phase",
"in",
"all_phases",
"if",
"... | Makes the names of phase measurement and attachments unique.
This function will make the names of measurements and attachments unique.
It modifies the input all_phases.
Args:
all_phases: the phases to make unique
Returns:
the phases now modified. | [
"Makes",
"the",
"names",
"of",
"phase",
"measurement",
"and",
"attachments",
"unique",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L224-L261 |
226,934 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | multidim_measurement_to_attachment | def multidim_measurement_to_attachment(name, measurement):
"""Convert a multi-dim measurement to an `openhtf.test_record.Attachment`."""
dimensions = list(measurement.dimensions)
if measurement.units:
dimensions.append(
measurements.Dimension.from_unit_descriptor(measurement.units))
dims = []
for d in dimensions:
if d.suffix is None:
suffix = u''
# Ensure that the suffix is unicode. It's typically str/bytes because
# units.py looks them up against str/bytes.
elif isinstance(d.suffix, unicode):
suffix = d.suffix
else:
suffix = d.suffix.decode('utf8')
dims.append({
'uom_suffix': suffix,
'uom_code': d.code,
'name': d.name,
})
# Refer to the module docstring for the expected schema.
dimensioned_measured_value = measurement.measured_value
value = (sorted(dimensioned_measured_value.value, key=lambda x: x[0])
if dimensioned_measured_value.is_value_set else None)
outcome_str = MEASUREMENT_OUTCOME_TO_TEST_RUN_STATUS_NAME[measurement.outcome]
data = _convert_object_to_json({
'outcome': outcome_str,
'name': name,
'dimensions': dims,
'value': value,
})
attachment = htf_test_record.Attachment(data, test_runs_pb2.MULTIDIM_JSON)
return attachment | python | def multidim_measurement_to_attachment(name, measurement):
"""Convert a multi-dim measurement to an `openhtf.test_record.Attachment`."""
dimensions = list(measurement.dimensions)
if measurement.units:
dimensions.append(
measurements.Dimension.from_unit_descriptor(measurement.units))
dims = []
for d in dimensions:
if d.suffix is None:
suffix = u''
# Ensure that the suffix is unicode. It's typically str/bytes because
# units.py looks them up against str/bytes.
elif isinstance(d.suffix, unicode):
suffix = d.suffix
else:
suffix = d.suffix.decode('utf8')
dims.append({
'uom_suffix': suffix,
'uom_code': d.code,
'name': d.name,
})
# Refer to the module docstring for the expected schema.
dimensioned_measured_value = measurement.measured_value
value = (sorted(dimensioned_measured_value.value, key=lambda x: x[0])
if dimensioned_measured_value.is_value_set else None)
outcome_str = MEASUREMENT_OUTCOME_TO_TEST_RUN_STATUS_NAME[measurement.outcome]
data = _convert_object_to_json({
'outcome': outcome_str,
'name': name,
'dimensions': dims,
'value': value,
})
attachment = htf_test_record.Attachment(data, test_runs_pb2.MULTIDIM_JSON)
return attachment | [
"def",
"multidim_measurement_to_attachment",
"(",
"name",
",",
"measurement",
")",
":",
"dimensions",
"=",
"list",
"(",
"measurement",
".",
"dimensions",
")",
"if",
"measurement",
".",
"units",
":",
"dimensions",
".",
"append",
"(",
"measurements",
".",
"Dimensi... | Convert a multi-dim measurement to an `openhtf.test_record.Attachment`. | [
"Convert",
"a",
"multi",
"-",
"dim",
"measurement",
"to",
"an",
"openhtf",
".",
"test_record",
".",
"Attachment",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L264-L300 |
226,935 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | convert_multidim_measurements | def convert_multidim_measurements(all_phases):
"""Converts each multidim measurements into attachments for all phases.."""
# Combine actual attachments with attachments we make from multi-dim
# measurements.
attachment_names = list(itertools.chain.from_iterable(
phase.attachments.keys() for phase in all_phases))
attachment_names.extend(itertools.chain.from_iterable([
'multidim_' + name for name, meas in phase.measurements.items()
if meas.dimensions is not None
] for phase in all_phases if phase.measurements))
attachment_name_maker = UniqueNameMaker(attachment_names)
for phase in all_phases:
# Process multi-dim measurements into unique attachments.
for name, measurement in sorted(phase.measurements.items()):
if measurement.dimensions:
old_name = name
name = attachment_name_maker.make_unique('multidim_%s' % name)
attachment = multidim_measurement_to_attachment(name, measurement)
phase.attachments[name] = attachment
phase.measurements.pop(old_name)
return all_phases | python | def convert_multidim_measurements(all_phases):
"""Converts each multidim measurements into attachments for all phases.."""
# Combine actual attachments with attachments we make from multi-dim
# measurements.
attachment_names = list(itertools.chain.from_iterable(
phase.attachments.keys() for phase in all_phases))
attachment_names.extend(itertools.chain.from_iterable([
'multidim_' + name for name, meas in phase.measurements.items()
if meas.dimensions is not None
] for phase in all_phases if phase.measurements))
attachment_name_maker = UniqueNameMaker(attachment_names)
for phase in all_phases:
# Process multi-dim measurements into unique attachments.
for name, measurement in sorted(phase.measurements.items()):
if measurement.dimensions:
old_name = name
name = attachment_name_maker.make_unique('multidim_%s' % name)
attachment = multidim_measurement_to_attachment(name, measurement)
phase.attachments[name] = attachment
phase.measurements.pop(old_name)
return all_phases | [
"def",
"convert_multidim_measurements",
"(",
"all_phases",
")",
":",
"# Combine actual attachments with attachments we make from multi-dim",
"# measurements.",
"attachment_names",
"=",
"list",
"(",
"itertools",
".",
"chain",
".",
"from_iterable",
"(",
"phase",
".",
"attachmen... | Converts each multidim measurements into attachments for all phases.. | [
"Converts",
"each",
"multidim",
"measurements",
"into",
"attachments",
"for",
"all",
"phases",
".."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L303-L324 |
226,936 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | attachment_to_multidim_measurement | def attachment_to_multidim_measurement(attachment, name=None):
"""Convert an OpenHTF test record attachment to a multi-dim measurement.
This is a best effort attempt to reverse, as some data is lost in converting
from a multidim to an attachment.
Args:
attachment: an `openhtf.test_record.Attachment` from a multi-dim.
name: an optional name for the measurement. If not provided will use the
name included in the attachment.
Returns:
An multi-dim `openhtf.Measurement`.
"""
data = json.loads(attachment.data)
name = name or data.get('name')
# attachment_dimn are a list of dicts with keys 'uom_suffix' and 'uom_code'
attachment_dims = data.get('dimensions', [])
# attachment_value is a list of lists [[t1, x1, y1, f1], [t2, x2, y2, f2]]
attachment_values = data.get('value')
attachment_outcome_str = data.get('outcome')
if attachment_outcome_str not in TEST_RUN_STATUS_NAME_TO_MEASUREMENT_OUTCOME:
# Fpr backward compatibility with saved data we'll convert integers to str
try:
attachment_outcome_str = test_runs_pb2.Status.Name(
int(attachment_outcome_str))
except ValueError:
attachment_outcome_str = None
# Convert test status outcome str to measurement outcome
outcome = TEST_RUN_STATUS_NAME_TO_MEASUREMENT_OUTCOME.get(
attachment_outcome_str)
# convert dimensions into htf.Dimensions
_lazy_load_units_by_code()
dims = []
for d in attachment_dims:
# Try to convert into htf.Dimension including backwards compatibility.
unit = UNITS_BY_CODE.get(d.get('uom_code'), units.NONE)
description = d.get('name', '')
dims.append(measurements.Dimension(description=description, unit=unit))
# Attempt to determine if units are included.
if attachment_values and len(dims) == len(attachment_values[0]):
# units provided
units_ = dims[-1].unit
dimensions = dims[:-1]
else:
units_ = None
dimensions = dims
# created dimensioned_measured_value and populate with values.
measured_value = measurements.DimensionedMeasuredValue(
name=name,
num_dimensions=len(dimensions)
)
for row in attachment_values:
coordinates = tuple(row[:-1])
val = row[-1]
measured_value[coordinates] = val
measurement = measurements.Measurement(
name=name,
units=units_,
dimensions=tuple(dimensions),
measured_value=measured_value,
outcome=outcome
)
return measurement | python | def attachment_to_multidim_measurement(attachment, name=None):
"""Convert an OpenHTF test record attachment to a multi-dim measurement.
This is a best effort attempt to reverse, as some data is lost in converting
from a multidim to an attachment.
Args:
attachment: an `openhtf.test_record.Attachment` from a multi-dim.
name: an optional name for the measurement. If not provided will use the
name included in the attachment.
Returns:
An multi-dim `openhtf.Measurement`.
"""
data = json.loads(attachment.data)
name = name or data.get('name')
# attachment_dimn are a list of dicts with keys 'uom_suffix' and 'uom_code'
attachment_dims = data.get('dimensions', [])
# attachment_value is a list of lists [[t1, x1, y1, f1], [t2, x2, y2, f2]]
attachment_values = data.get('value')
attachment_outcome_str = data.get('outcome')
if attachment_outcome_str not in TEST_RUN_STATUS_NAME_TO_MEASUREMENT_OUTCOME:
# Fpr backward compatibility with saved data we'll convert integers to str
try:
attachment_outcome_str = test_runs_pb2.Status.Name(
int(attachment_outcome_str))
except ValueError:
attachment_outcome_str = None
# Convert test status outcome str to measurement outcome
outcome = TEST_RUN_STATUS_NAME_TO_MEASUREMENT_OUTCOME.get(
attachment_outcome_str)
# convert dimensions into htf.Dimensions
_lazy_load_units_by_code()
dims = []
for d in attachment_dims:
# Try to convert into htf.Dimension including backwards compatibility.
unit = UNITS_BY_CODE.get(d.get('uom_code'), units.NONE)
description = d.get('name', '')
dims.append(measurements.Dimension(description=description, unit=unit))
# Attempt to determine if units are included.
if attachment_values and len(dims) == len(attachment_values[0]):
# units provided
units_ = dims[-1].unit
dimensions = dims[:-1]
else:
units_ = None
dimensions = dims
# created dimensioned_measured_value and populate with values.
measured_value = measurements.DimensionedMeasuredValue(
name=name,
num_dimensions=len(dimensions)
)
for row in attachment_values:
coordinates = tuple(row[:-1])
val = row[-1]
measured_value[coordinates] = val
measurement = measurements.Measurement(
name=name,
units=units_,
dimensions=tuple(dimensions),
measured_value=measured_value,
outcome=outcome
)
return measurement | [
"def",
"attachment_to_multidim_measurement",
"(",
"attachment",
",",
"name",
"=",
"None",
")",
":",
"data",
"=",
"json",
".",
"loads",
"(",
"attachment",
".",
"data",
")",
"name",
"=",
"name",
"or",
"data",
".",
"get",
"(",
"'name'",
")",
"# attachment_dim... | Convert an OpenHTF test record attachment to a multi-dim measurement.
This is a best effort attempt to reverse, as some data is lost in converting
from a multidim to an attachment.
Args:
attachment: an `openhtf.test_record.Attachment` from a multi-dim.
name: an optional name for the measurement. If not provided will use the
name included in the attachment.
Returns:
An multi-dim `openhtf.Measurement`. | [
"Convert",
"an",
"OpenHTF",
"test",
"record",
"attachment",
"to",
"a",
"multi",
"-",
"dim",
"measurement",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L427-L497 |
226,937 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | PhaseCopier._copy_unidimensional_measurement | def _copy_unidimensional_measurement(
self, phase, name, measurement, mfg_event):
"""Copy uni-dimensional measurements to the MfgEvent."""
mfg_measurement = mfg_event.measurement.add()
# Copy basic measurement fields.
mfg_measurement.name = name
if measurement.docstring:
mfg_measurement.description = measurement.docstring
mfg_measurement.parameter_tag.append(phase.name)
if (measurement.units and
measurement.units.code in test_runs_converter.UOM_CODE_MAP):
mfg_measurement.unit_code = (
test_runs_converter.UOM_CODE_MAP[measurement.units.code])
# Copy failed measurements as failure_codes. This happens early to include
# unset measurements.
if (measurement.outcome != measurements.Outcome.PASS and
phase.outcome != htf_test_record.PhaseOutcome.SKIP):
failure_code = mfg_event.failure_codes.add()
failure_code.code = name
failure_code.details = '\n'.join(str(v) for v in measurement.validators)
# Copy measurement value.
measured_value = measurement.measured_value
status_str = MEASUREMENT_OUTCOME_TO_TEST_RUN_STATUS_NAME[
measurement.outcome]
mfg_measurement.status = test_runs_pb2.Status.Value(status_str)
if not measured_value.is_value_set:
return
value = measured_value.value
if isinstance(value, numbers.Number):
mfg_measurement.numeric_value = float(value)
elif isinstance(value, bytes):
# text_value expects unicode or ascii-compatible strings, so we must
# 'decode' it, even if it's actually just garbage bytestring data.
mfg_measurement.text_value = unicode(value, errors='replace')
elif isinstance(value, unicode):
# Don't waste time and potential errors decoding unicode.
mfg_measurement.text_value = value
else:
# Coercing to string.
mfg_measurement.text_value = str(value)
# Copy measurement validators.
for validator in measurement.validators:
if isinstance(validator, validators.RangeValidatorBase):
if validator.minimum is not None:
mfg_measurement.numeric_minimum = float(validator.minimum)
if validator.maximum is not None:
mfg_measurement.numeric_maximum = float(validator.maximum)
elif isinstance(validator, validators.RegexMatcher):
mfg_measurement.expected_text = validator.regex
else:
mfg_measurement.description += '\nValidator: ' + str(validator) | python | def _copy_unidimensional_measurement(
self, phase, name, measurement, mfg_event):
"""Copy uni-dimensional measurements to the MfgEvent."""
mfg_measurement = mfg_event.measurement.add()
# Copy basic measurement fields.
mfg_measurement.name = name
if measurement.docstring:
mfg_measurement.description = measurement.docstring
mfg_measurement.parameter_tag.append(phase.name)
if (measurement.units and
measurement.units.code in test_runs_converter.UOM_CODE_MAP):
mfg_measurement.unit_code = (
test_runs_converter.UOM_CODE_MAP[measurement.units.code])
# Copy failed measurements as failure_codes. This happens early to include
# unset measurements.
if (measurement.outcome != measurements.Outcome.PASS and
phase.outcome != htf_test_record.PhaseOutcome.SKIP):
failure_code = mfg_event.failure_codes.add()
failure_code.code = name
failure_code.details = '\n'.join(str(v) for v in measurement.validators)
# Copy measurement value.
measured_value = measurement.measured_value
status_str = MEASUREMENT_OUTCOME_TO_TEST_RUN_STATUS_NAME[
measurement.outcome]
mfg_measurement.status = test_runs_pb2.Status.Value(status_str)
if not measured_value.is_value_set:
return
value = measured_value.value
if isinstance(value, numbers.Number):
mfg_measurement.numeric_value = float(value)
elif isinstance(value, bytes):
# text_value expects unicode or ascii-compatible strings, so we must
# 'decode' it, even if it's actually just garbage bytestring data.
mfg_measurement.text_value = unicode(value, errors='replace')
elif isinstance(value, unicode):
# Don't waste time and potential errors decoding unicode.
mfg_measurement.text_value = value
else:
# Coercing to string.
mfg_measurement.text_value = str(value)
# Copy measurement validators.
for validator in measurement.validators:
if isinstance(validator, validators.RangeValidatorBase):
if validator.minimum is not None:
mfg_measurement.numeric_minimum = float(validator.minimum)
if validator.maximum is not None:
mfg_measurement.numeric_maximum = float(validator.maximum)
elif isinstance(validator, validators.RegexMatcher):
mfg_measurement.expected_text = validator.regex
else:
mfg_measurement.description += '\nValidator: ' + str(validator) | [
"def",
"_copy_unidimensional_measurement",
"(",
"self",
",",
"phase",
",",
"name",
",",
"measurement",
",",
"mfg_event",
")",
":",
"mfg_measurement",
"=",
"mfg_event",
".",
"measurement",
".",
"add",
"(",
")",
"# Copy basic measurement fields.",
"mfg_measurement",
"... | Copy uni-dimensional measurements to the MfgEvent. | [
"Copy",
"uni",
"-",
"dimensional",
"measurements",
"to",
"the",
"MfgEvent",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L341-L396 |
226,938 | google/openhtf | openhtf/output/proto/mfg_event_converter.py | PhaseCopier._copy_attachment | def _copy_attachment(self, name, data, mimetype, mfg_event):
"""Copies an attachment to mfg_event."""
attachment = mfg_event.attachment.add()
attachment.name = name
if isinstance(data, unicode):
data = data.encode('utf8')
attachment.value_binary = data
if mimetype in test_runs_converter.MIMETYPE_MAP:
attachment.type = test_runs_converter.MIMETYPE_MAP[mimetype]
elif mimetype == test_runs_pb2.MULTIDIM_JSON:
attachment.type = mimetype
else:
attachment.type = test_runs_pb2.BINARY | python | def _copy_attachment(self, name, data, mimetype, mfg_event):
"""Copies an attachment to mfg_event."""
attachment = mfg_event.attachment.add()
attachment.name = name
if isinstance(data, unicode):
data = data.encode('utf8')
attachment.value_binary = data
if mimetype in test_runs_converter.MIMETYPE_MAP:
attachment.type = test_runs_converter.MIMETYPE_MAP[mimetype]
elif mimetype == test_runs_pb2.MULTIDIM_JSON:
attachment.type = mimetype
else:
attachment.type = test_runs_pb2.BINARY | [
"def",
"_copy_attachment",
"(",
"self",
",",
"name",
",",
"data",
",",
"mimetype",
",",
"mfg_event",
")",
":",
"attachment",
"=",
"mfg_event",
".",
"attachment",
".",
"add",
"(",
")",
"attachment",
".",
"name",
"=",
"name",
"if",
"isinstance",
"(",
"data... | Copies an attachment to mfg_event. | [
"Copies",
"an",
"attachment",
"to",
"mfg_event",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/proto/mfg_event_converter.py#L403-L415 |
226,939 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStream.write | def write(self, data, timeout_ms=None):
"""Write data to this stream.
Args:
data: Data to write.
timeout_ms: Timeout to use for the write/Ack transaction, in
milliseconds (or as a PolledTimeout object).
Raises:
AdbProtocolError: If an ACK is not received.
AdbStreamClosedError: If the stream is already closed, or gets closed
before the write completes.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
# Break the data up into our transport's maxdata sized WRTE messages.
while data:
self._transport.write(
data[:self._transport.adb_connection.maxdata], timeout)
data = data[self._transport.adb_connection.maxdata:] | python | def write(self, data, timeout_ms=None):
"""Write data to this stream.
Args:
data: Data to write.
timeout_ms: Timeout to use for the write/Ack transaction, in
milliseconds (or as a PolledTimeout object).
Raises:
AdbProtocolError: If an ACK is not received.
AdbStreamClosedError: If the stream is already closed, or gets closed
before the write completes.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
# Break the data up into our transport's maxdata sized WRTE messages.
while data:
self._transport.write(
data[:self._transport.adb_connection.maxdata], timeout)
data = data[self._transport.adb_connection.maxdata:] | [
"def",
"write",
"(",
"self",
",",
"data",
",",
"timeout_ms",
"=",
"None",
")",
":",
"timeout",
"=",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
"# Break the data up into our transport's maxdata sized WRTE messages.",
"while",
"data",
... | Write data to this stream.
Args:
data: Data to write.
timeout_ms: Timeout to use for the write/Ack transaction, in
milliseconds (or as a PolledTimeout object).
Raises:
AdbProtocolError: If an ACK is not received.
AdbStreamClosedError: If the stream is already closed, or gets closed
before the write completes. | [
"Write",
"data",
"to",
"this",
"stream",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L167-L185 |
226,940 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStream.read | def read(self, length=0, timeout_ms=None):
"""Reads data from the remote end of this stream.
Internally, this data will have been contained in AdbMessages, but
users of streams shouldn't need to care about the transport mechanism.
Args:
length: If provided, the number of bytes to read, otherwise all available
data will be returned (at least one byte).
timeout_ms: Time to wait for a message to come in for this stream, in
milliseconds (or as a PolledTimeout object).
Returns:
Data that was read, or None if the end of the stream was reached.
Raises:
AdbProtocolError: Received an unexpected wonky non-stream packet (like a
CNXN ADB message).
AdbStreamClosedError: The stream is already closed.
AdbTimeoutError: Timed out waiting for a message.
"""
return self._transport.read(
length, timeouts.PolledTimeout.from_millis(timeout_ms)) | python | def read(self, length=0, timeout_ms=None):
"""Reads data from the remote end of this stream.
Internally, this data will have been contained in AdbMessages, but
users of streams shouldn't need to care about the transport mechanism.
Args:
length: If provided, the number of bytes to read, otherwise all available
data will be returned (at least one byte).
timeout_ms: Time to wait for a message to come in for this stream, in
milliseconds (or as a PolledTimeout object).
Returns:
Data that was read, or None if the end of the stream was reached.
Raises:
AdbProtocolError: Received an unexpected wonky non-stream packet (like a
CNXN ADB message).
AdbStreamClosedError: The stream is already closed.
AdbTimeoutError: Timed out waiting for a message.
"""
return self._transport.read(
length, timeouts.PolledTimeout.from_millis(timeout_ms)) | [
"def",
"read",
"(",
"self",
",",
"length",
"=",
"0",
",",
"timeout_ms",
"=",
"None",
")",
":",
"return",
"self",
".",
"_transport",
".",
"read",
"(",
"length",
",",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
")"
] | Reads data from the remote end of this stream.
Internally, this data will have been contained in AdbMessages, but
users of streams shouldn't need to care about the transport mechanism.
Args:
length: If provided, the number of bytes to read, otherwise all available
data will be returned (at least one byte).
timeout_ms: Time to wait for a message to come in for this stream, in
milliseconds (or as a PolledTimeout object).
Returns:
Data that was read, or None if the end of the stream was reached.
Raises:
AdbProtocolError: Received an unexpected wonky non-stream packet (like a
CNXN ADB message).
AdbStreamClosedError: The stream is already closed.
AdbTimeoutError: Timed out waiting for a message. | [
"Reads",
"data",
"from",
"the",
"remote",
"end",
"of",
"this",
"stream",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L187-L209 |
226,941 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStream.read_until_close | def read_until_close(self, timeout_ms=None):
"""Yield data until this stream is closed.
Args:
timeout_ms: Timeout in milliseconds to keep reading (or a PolledTimeout
object).
Yields:
Data read from a single call to self.read(), until the stream is closed
or timeout is reached.
Raises:
AdbTimeoutError: On timeout.
"""
while True:
try:
yield self.read(timeout_ms=timeout_ms)
except usb_exceptions.AdbStreamClosedError:
break | python | def read_until_close(self, timeout_ms=None):
"""Yield data until this stream is closed.
Args:
timeout_ms: Timeout in milliseconds to keep reading (or a PolledTimeout
object).
Yields:
Data read from a single call to self.read(), until the stream is closed
or timeout is reached.
Raises:
AdbTimeoutError: On timeout.
"""
while True:
try:
yield self.read(timeout_ms=timeout_ms)
except usb_exceptions.AdbStreamClosedError:
break | [
"def",
"read_until_close",
"(",
"self",
",",
"timeout_ms",
"=",
"None",
")",
":",
"while",
"True",
":",
"try",
":",
"yield",
"self",
".",
"read",
"(",
"timeout_ms",
"=",
"timeout_ms",
")",
"except",
"usb_exceptions",
".",
"AdbStreamClosedError",
":",
"break"... | Yield data until this stream is closed.
Args:
timeout_ms: Timeout in milliseconds to keep reading (or a PolledTimeout
object).
Yields:
Data read from a single call to self.read(), until the stream is closed
or timeout is reached.
Raises:
AdbTimeoutError: On timeout. | [
"Yield",
"data",
"until",
"this",
"stream",
"is",
"closed",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L211-L229 |
226,942 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport._set_or_check_remote_id | def _set_or_check_remote_id(self, remote_id):
"""Set or check the remote id."""
if not self.remote_id:
assert self.closed_state == self.ClosedState.PENDING, 'Bad ClosedState!'
self.remote_id = remote_id
self.closed_state = self.ClosedState.OPEN
elif self.remote_id != remote_id:
raise usb_exceptions.AdbProtocolError(
'%s remote-id change to %s', self, remote_id) | python | def _set_or_check_remote_id(self, remote_id):
"""Set or check the remote id."""
if not self.remote_id:
assert self.closed_state == self.ClosedState.PENDING, 'Bad ClosedState!'
self.remote_id = remote_id
self.closed_state = self.ClosedState.OPEN
elif self.remote_id != remote_id:
raise usb_exceptions.AdbProtocolError(
'%s remote-id change to %s', self, remote_id) | [
"def",
"_set_or_check_remote_id",
"(",
"self",
",",
"remote_id",
")",
":",
"if",
"not",
"self",
".",
"remote_id",
":",
"assert",
"self",
".",
"closed_state",
"==",
"self",
".",
"ClosedState",
".",
"PENDING",
",",
"'Bad ClosedState!'",
"self",
".",
"remote_id",... | Set or check the remote id. | [
"Set",
"or",
"check",
"the",
"remote",
"id",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L277-L285 |
226,943 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport._handle_message | def _handle_message(self, message, handle_wrte=True):
"""Handle a message that was read for this stream.
For each message type, this means:
OKAY: Check id's and make sure we are expecting an OKAY. Clear the
self._expecting_okay flag so any pending write()'s know.
CLSE: Set our internal state to closed.
WRTE: Add the data read to our internal read buffer. Note we don't
return the actual data because it may not be this thread that needs it.
Args:
message: Message that was read.
handle_wrte: If True, we can handle WRTE messages, otherwise raise.
Raises:
AdbProtocolError: If we get a WRTE message but handle_wrte is False.
"""
if message.command == 'OKAY':
self._set_or_check_remote_id(message.arg0)
if not self._expecting_okay:
raise usb_exceptions.AdbProtocolError(
'%s received unexpected OKAY: %s', self, message)
self._expecting_okay = False
elif message.command == 'CLSE':
self.closed_state = self.ClosedState.CLOSED
elif not handle_wrte:
raise usb_exceptions.AdbProtocolError(
'%s received WRTE before OKAY/CLSE: %s', self, message)
else:
with self._read_buffer_lock:
self._read_buffer.append(message.data)
self._buffer_size += len(message.data) | python | def _handle_message(self, message, handle_wrte=True):
"""Handle a message that was read for this stream.
For each message type, this means:
OKAY: Check id's and make sure we are expecting an OKAY. Clear the
self._expecting_okay flag so any pending write()'s know.
CLSE: Set our internal state to closed.
WRTE: Add the data read to our internal read buffer. Note we don't
return the actual data because it may not be this thread that needs it.
Args:
message: Message that was read.
handle_wrte: If True, we can handle WRTE messages, otherwise raise.
Raises:
AdbProtocolError: If we get a WRTE message but handle_wrte is False.
"""
if message.command == 'OKAY':
self._set_or_check_remote_id(message.arg0)
if not self._expecting_okay:
raise usb_exceptions.AdbProtocolError(
'%s received unexpected OKAY: %s', self, message)
self._expecting_okay = False
elif message.command == 'CLSE':
self.closed_state = self.ClosedState.CLOSED
elif not handle_wrte:
raise usb_exceptions.AdbProtocolError(
'%s received WRTE before OKAY/CLSE: %s', self, message)
else:
with self._read_buffer_lock:
self._read_buffer.append(message.data)
self._buffer_size += len(message.data) | [
"def",
"_handle_message",
"(",
"self",
",",
"message",
",",
"handle_wrte",
"=",
"True",
")",
":",
"if",
"message",
".",
"command",
"==",
"'OKAY'",
":",
"self",
".",
"_set_or_check_remote_id",
"(",
"message",
".",
"arg0",
")",
"if",
"not",
"self",
".",
"_... | Handle a message that was read for this stream.
For each message type, this means:
OKAY: Check id's and make sure we are expecting an OKAY. Clear the
self._expecting_okay flag so any pending write()'s know.
CLSE: Set our internal state to closed.
WRTE: Add the data read to our internal read buffer. Note we don't
return the actual data because it may not be this thread that needs it.
Args:
message: Message that was read.
handle_wrte: If True, we can handle WRTE messages, otherwise raise.
Raises:
AdbProtocolError: If we get a WRTE message but handle_wrte is False. | [
"Handle",
"a",
"message",
"that",
"was",
"read",
"for",
"this",
"stream",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L314-L345 |
226,944 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport._read_messages_until_true | def _read_messages_until_true(self, predicate, timeout):
"""Read a message from this stream and handle it.
This method tries to read a message from this stream, blocking until a
message is read. Once read, it will handle it accordingly by calling
self._handle_message().
This is repeated as long as predicate() returns False. There is some
locking used internally here so that we don't end up with multiple threads
blocked on a call to read_for_stream when another thread has read the
message that caused predicate() to become True.
Args:
predicate: Callable, keep reading messages until it returns true. Note
that predicate() should not block, as doing so may cause this method to
hang beyond its timeout.
timeout: Timeout to use for this call.
Raises:
AdbStreamClosedError: If this stream is already closed.
"""
while not predicate():
# Hold the message_received Lock while we try to acquire the reader_lock
# and waiting on the message_received condition, to prevent another reader
# thread from notifying the condition between us failing to acquire the
# reader_lock and waiting on the condition.
self._message_received.acquire()
if self._reader_lock.acquire(False):
try:
# Release the message_received Lock while we do the read so other
# threads can wait() on the condition without having to block on
# acquiring the message_received Lock (we may have a longer timeout
# than them, so that would be bad).
self._message_received.release()
# We are now the thread responsible for reading a message. Check
# predicate() to make sure nobody else read a message between our last
# check and acquiring the reader Lock.
if predicate():
return
# Read and handle a message, using our timeout.
self._handle_message(
self.adb_connection.read_for_stream(self, timeout))
# Notify anyone interested that we handled a message, causing them to
# check their predicate again.
with self._message_received:
self._message_received.notify_all()
finally:
self._reader_lock.release()
else:
# There is some other thread reading a message. Since we are already
# holding the message_received Lock, we can immediately do the wait.
try:
self._message_received.wait(timeout.remaining)
if timeout.has_expired():
raise usb_exceptions.AdbTimeoutError(
'%s timed out reading messages.', self)
finally:
# Make sure we release this even if an exception occurred.
self._message_received.release() | python | def _read_messages_until_true(self, predicate, timeout):
"""Read a message from this stream and handle it.
This method tries to read a message from this stream, blocking until a
message is read. Once read, it will handle it accordingly by calling
self._handle_message().
This is repeated as long as predicate() returns False. There is some
locking used internally here so that we don't end up with multiple threads
blocked on a call to read_for_stream when another thread has read the
message that caused predicate() to become True.
Args:
predicate: Callable, keep reading messages until it returns true. Note
that predicate() should not block, as doing so may cause this method to
hang beyond its timeout.
timeout: Timeout to use for this call.
Raises:
AdbStreamClosedError: If this stream is already closed.
"""
while not predicate():
# Hold the message_received Lock while we try to acquire the reader_lock
# and waiting on the message_received condition, to prevent another reader
# thread from notifying the condition between us failing to acquire the
# reader_lock and waiting on the condition.
self._message_received.acquire()
if self._reader_lock.acquire(False):
try:
# Release the message_received Lock while we do the read so other
# threads can wait() on the condition without having to block on
# acquiring the message_received Lock (we may have a longer timeout
# than them, so that would be bad).
self._message_received.release()
# We are now the thread responsible for reading a message. Check
# predicate() to make sure nobody else read a message between our last
# check and acquiring the reader Lock.
if predicate():
return
# Read and handle a message, using our timeout.
self._handle_message(
self.adb_connection.read_for_stream(self, timeout))
# Notify anyone interested that we handled a message, causing them to
# check their predicate again.
with self._message_received:
self._message_received.notify_all()
finally:
self._reader_lock.release()
else:
# There is some other thread reading a message. Since we are already
# holding the message_received Lock, we can immediately do the wait.
try:
self._message_received.wait(timeout.remaining)
if timeout.has_expired():
raise usb_exceptions.AdbTimeoutError(
'%s timed out reading messages.', self)
finally:
# Make sure we release this even if an exception occurred.
self._message_received.release() | [
"def",
"_read_messages_until_true",
"(",
"self",
",",
"predicate",
",",
"timeout",
")",
":",
"while",
"not",
"predicate",
"(",
")",
":",
"# Hold the message_received Lock while we try to acquire the reader_lock",
"# and waiting on the message_received condition, to prevent another ... | Read a message from this stream and handle it.
This method tries to read a message from this stream, blocking until a
message is read. Once read, it will handle it accordingly by calling
self._handle_message().
This is repeated as long as predicate() returns False. There is some
locking used internally here so that we don't end up with multiple threads
blocked on a call to read_for_stream when another thread has read the
message that caused predicate() to become True.
Args:
predicate: Callable, keep reading messages until it returns true. Note
that predicate() should not block, as doing so may cause this method to
hang beyond its timeout.
timeout: Timeout to use for this call.
Raises:
AdbStreamClosedError: If this stream is already closed. | [
"Read",
"a",
"message",
"from",
"this",
"stream",
"and",
"handle",
"it",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L347-L408 |
226,945 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport.ensure_opened | def ensure_opened(self, timeout):
"""Ensure this stream transport was successfully opened.
Checks to make sure we receive our initial OKAY message. This must be
called after creating this AdbStreamTransport and before calling read() or
write().
Args:
timeout: timeouts.PolledTimeout to use for this operation.
Returns:
True if this stream was successfully opened, False if the service was
not recognized by the remote endpoint. If False is returned, then this
AdbStreamTransport will be already closed.
Raises:
AdbProtocolError: If we receive a WRTE message instead of OKAY/CLSE.
"""
self._handle_message(self.adb_connection.read_for_stream(self, timeout),
handle_wrte=False)
return self.is_open() | python | def ensure_opened(self, timeout):
"""Ensure this stream transport was successfully opened.
Checks to make sure we receive our initial OKAY message. This must be
called after creating this AdbStreamTransport and before calling read() or
write().
Args:
timeout: timeouts.PolledTimeout to use for this operation.
Returns:
True if this stream was successfully opened, False if the service was
not recognized by the remote endpoint. If False is returned, then this
AdbStreamTransport will be already closed.
Raises:
AdbProtocolError: If we receive a WRTE message instead of OKAY/CLSE.
"""
self._handle_message(self.adb_connection.read_for_stream(self, timeout),
handle_wrte=False)
return self.is_open() | [
"def",
"ensure_opened",
"(",
"self",
",",
"timeout",
")",
":",
"self",
".",
"_handle_message",
"(",
"self",
".",
"adb_connection",
".",
"read_for_stream",
"(",
"self",
",",
"timeout",
")",
",",
"handle_wrte",
"=",
"False",
")",
"return",
"self",
".",
"is_o... | Ensure this stream transport was successfully opened.
Checks to make sure we receive our initial OKAY message. This must be
called after creating this AdbStreamTransport and before calling read() or
write().
Args:
timeout: timeouts.PolledTimeout to use for this operation.
Returns:
True if this stream was successfully opened, False if the service was
not recognized by the remote endpoint. If False is returned, then this
AdbStreamTransport will be already closed.
Raises:
AdbProtocolError: If we receive a WRTE message instead of OKAY/CLSE. | [
"Ensure",
"this",
"stream",
"transport",
"was",
"successfully",
"opened",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L410-L430 |
226,946 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport.enqueue_message | def enqueue_message(self, message, timeout):
"""Add the given message to this transport's queue.
This method also handles ACKing any WRTE messages.
Args:
message: The AdbMessage to enqueue.
timeout: The timeout to use for the operation. Specifically, WRTE
messages cause an OKAY to be sent; timeout is used for that send.
"""
# Ack WRTE messages immediately, handle our OPEN ack if it gets enqueued.
if message.command == 'WRTE':
self._send_command('OKAY', timeout=timeout)
elif message.command == 'OKAY':
self._set_or_check_remote_id(message.arg0)
self.message_queue.put(message) | python | def enqueue_message(self, message, timeout):
"""Add the given message to this transport's queue.
This method also handles ACKing any WRTE messages.
Args:
message: The AdbMessage to enqueue.
timeout: The timeout to use for the operation. Specifically, WRTE
messages cause an OKAY to be sent; timeout is used for that send.
"""
# Ack WRTE messages immediately, handle our OPEN ack if it gets enqueued.
if message.command == 'WRTE':
self._send_command('OKAY', timeout=timeout)
elif message.command == 'OKAY':
self._set_or_check_remote_id(message.arg0)
self.message_queue.put(message) | [
"def",
"enqueue_message",
"(",
"self",
",",
"message",
",",
"timeout",
")",
":",
"# Ack WRTE messages immediately, handle our OPEN ack if it gets enqueued.",
"if",
"message",
".",
"command",
"==",
"'WRTE'",
":",
"self",
".",
"_send_command",
"(",
"'OKAY'",
",",
"timeo... | Add the given message to this transport's queue.
This method also handles ACKing any WRTE messages.
Args:
message: The AdbMessage to enqueue.
timeout: The timeout to use for the operation. Specifically, WRTE
messages cause an OKAY to be sent; timeout is used for that send. | [
"Add",
"the",
"given",
"message",
"to",
"this",
"transport",
"s",
"queue",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L440-L455 |
226,947 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport.write | def write(self, data, timeout):
"""Write data to this stream, using the given timeouts.PolledTimeout."""
if not self.remote_id:
raise usb_exceptions.AdbStreamClosedError(
'Cannot write() to half-opened %s', self)
if self.closed_state != self.ClosedState.OPEN:
raise usb_exceptions.AdbStreamClosedError(
'Cannot write() to closed %s', self)
elif self._expecting_okay:
raise usb_exceptions.AdbProtocolError(
'Previous WRTE failed, %s in unknown state', self)
# Make sure we only have one WRTE in flight at a time, because ADB doesn't
# identify which WRTE it is ACK'ing when it sends the OKAY message back.
with self._write_lock:
self._expecting_okay = True
self._send_command('WRTE', timeout, data)
self._read_messages_until_true(lambda: not self._expecting_okay, timeout) | python | def write(self, data, timeout):
"""Write data to this stream, using the given timeouts.PolledTimeout."""
if not self.remote_id:
raise usb_exceptions.AdbStreamClosedError(
'Cannot write() to half-opened %s', self)
if self.closed_state != self.ClosedState.OPEN:
raise usb_exceptions.AdbStreamClosedError(
'Cannot write() to closed %s', self)
elif self._expecting_okay:
raise usb_exceptions.AdbProtocolError(
'Previous WRTE failed, %s in unknown state', self)
# Make sure we only have one WRTE in flight at a time, because ADB doesn't
# identify which WRTE it is ACK'ing when it sends the OKAY message back.
with self._write_lock:
self._expecting_okay = True
self._send_command('WRTE', timeout, data)
self._read_messages_until_true(lambda: not self._expecting_okay, timeout) | [
"def",
"write",
"(",
"self",
",",
"data",
",",
"timeout",
")",
":",
"if",
"not",
"self",
".",
"remote_id",
":",
"raise",
"usb_exceptions",
".",
"AdbStreamClosedError",
"(",
"'Cannot write() to half-opened %s'",
",",
"self",
")",
"if",
"self",
".",
"closed_stat... | Write data to this stream, using the given timeouts.PolledTimeout. | [
"Write",
"data",
"to",
"this",
"stream",
"using",
"the",
"given",
"timeouts",
".",
"PolledTimeout",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L457-L474 |
226,948 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbStreamTransport.read | def read(self, length, timeout):
"""Read 'length' bytes from this stream transport.
Args:
length: If not 0, read this many bytes from the stream, otherwise read all
available data (at least one byte).
timeout: timeouts.PolledTimeout to use for this read operation.
Returns:
The bytes read from this stream.
"""
self._read_messages_until_true(
lambda: self._buffer_size and self._buffer_size >= length, timeout)
with self._read_buffer_lock:
data, push_back = ''.join(self._read_buffer), ''
if length:
data, push_back = data[:length], data[length:]
self._read_buffer.clear()
self._buffer_size = len(push_back)
if push_back:
self._read_buffer.appendleft(push_back)
return data | python | def read(self, length, timeout):
"""Read 'length' bytes from this stream transport.
Args:
length: If not 0, read this many bytes from the stream, otherwise read all
available data (at least one byte).
timeout: timeouts.PolledTimeout to use for this read operation.
Returns:
The bytes read from this stream.
"""
self._read_messages_until_true(
lambda: self._buffer_size and self._buffer_size >= length, timeout)
with self._read_buffer_lock:
data, push_back = ''.join(self._read_buffer), ''
if length:
data, push_back = data[:length], data[length:]
self._read_buffer.clear()
self._buffer_size = len(push_back)
if push_back:
self._read_buffer.appendleft(push_back)
return data | [
"def",
"read",
"(",
"self",
",",
"length",
",",
"timeout",
")",
":",
"self",
".",
"_read_messages_until_true",
"(",
"lambda",
":",
"self",
".",
"_buffer_size",
"and",
"self",
".",
"_buffer_size",
">=",
"length",
",",
"timeout",
")",
"with",
"self",
".",
... | Read 'length' bytes from this stream transport.
Args:
length: If not 0, read this many bytes from the stream, otherwise read all
available data (at least one byte).
timeout: timeouts.PolledTimeout to use for this read operation.
Returns:
The bytes read from this stream. | [
"Read",
"length",
"bytes",
"from",
"this",
"stream",
"transport",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L476-L498 |
226,949 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection._make_stream_transport | def _make_stream_transport(self):
"""Create an AdbStreamTransport with a newly allocated local_id."""
msg_queue = queue.Queue()
with self._stream_transport_map_lock:
# Start one past the last id we used, and grab the first available one.
# This mimics the ADB behavior of 'increment an unsigned and let it
# overflow', but with a check to ensure we don't reuse an id in use,
# even though that's unlikely with 2^32 - 1 of them available. We try
# at most 64 id's, if we've wrapped around and there isn't one available
# in the first 64, there's a problem, better to fail fast than hang for
# a potentially very long time (STREAM_ID_LIMIT can be very large).
self._last_id_used = (self._last_id_used % STREAM_ID_LIMIT) + 1
for local_id in itertools.islice(
itertools.chain(
range(self._last_id_used, STREAM_ID_LIMIT),
range(1, self._last_id_used)), 64):
if local_id not in list(self._stream_transport_map.keys()):
self._last_id_used = local_id
break
else:
raise usb_exceptions.AdbStreamUnavailableError('Ran out of local ids!')
# Ignore this warning - the for loop will always have at least one
# iteration, so local_id will always be set.
# pylint: disable=undefined-loop-variable
stream_transport = AdbStreamTransport(self, local_id, msg_queue)
self._stream_transport_map[local_id] = stream_transport
return stream_transport | python | def _make_stream_transport(self):
"""Create an AdbStreamTransport with a newly allocated local_id."""
msg_queue = queue.Queue()
with self._stream_transport_map_lock:
# Start one past the last id we used, and grab the first available one.
# This mimics the ADB behavior of 'increment an unsigned and let it
# overflow', but with a check to ensure we don't reuse an id in use,
# even though that's unlikely with 2^32 - 1 of them available. We try
# at most 64 id's, if we've wrapped around and there isn't one available
# in the first 64, there's a problem, better to fail fast than hang for
# a potentially very long time (STREAM_ID_LIMIT can be very large).
self._last_id_used = (self._last_id_used % STREAM_ID_LIMIT) + 1
for local_id in itertools.islice(
itertools.chain(
range(self._last_id_used, STREAM_ID_LIMIT),
range(1, self._last_id_used)), 64):
if local_id not in list(self._stream_transport_map.keys()):
self._last_id_used = local_id
break
else:
raise usb_exceptions.AdbStreamUnavailableError('Ran out of local ids!')
# Ignore this warning - the for loop will always have at least one
# iteration, so local_id will always be set.
# pylint: disable=undefined-loop-variable
stream_transport = AdbStreamTransport(self, local_id, msg_queue)
self._stream_transport_map[local_id] = stream_transport
return stream_transport | [
"def",
"_make_stream_transport",
"(",
"self",
")",
":",
"msg_queue",
"=",
"queue",
".",
"Queue",
"(",
")",
"with",
"self",
".",
"_stream_transport_map_lock",
":",
"# Start one past the last id we used, and grab the first available one.",
"# This mimics the ADB behavior of 'incr... | Create an AdbStreamTransport with a newly allocated local_id. | [
"Create",
"an",
"AdbStreamTransport",
"with",
"a",
"newly",
"allocated",
"local_id",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L568-L594 |
226,950 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection._handle_message_for_stream | def _handle_message_for_stream(self, stream_transport, message, timeout):
"""Handle an incoming message, check if it's for the given stream.
If the message is not for the stream, then add it to the appropriate
message queue.
Args:
stream_transport: AdbStreamTransport currently waiting on a message.
message: Message to check and handle.
timeout: Timeout to use for the operation, should be an instance of
timeouts.PolledTimeout.
Returns:
The message read if it was for this stream, None otherwise.
Raises:
AdbProtocolError: If we receive an unexpected message type.
"""
if message.command not in ('OKAY', 'CLSE', 'WRTE'):
raise usb_exceptions.AdbProtocolError(
'%s received unexpected message: %s', self, message)
if message.arg1 == stream_transport.local_id:
# Ack writes immediately.
if message.command == 'WRTE':
# Make sure we don't get a WRTE before an OKAY/CLSE message.
if not stream_transport.remote_id:
raise usb_exceptions.AdbProtocolError(
'%s received WRTE before OKAY/CLSE: %s',
stream_transport, message)
self.transport.write_message(adb_message.AdbMessage(
'OKAY', stream_transport.local_id, stream_transport.remote_id),
timeout)
elif message.command == 'CLSE':
self.close_stream_transport(stream_transport, timeout)
return message
else:
# Message was not for this stream, add it to the right stream's queue.
with self._stream_transport_map_lock:
dest_transport = self._stream_transport_map.get(message.arg1)
if dest_transport:
if message.command == 'CLSE':
self.close_stream_transport(dest_transport, timeout)
dest_transport.enqueue_message(message, timeout)
else:
_LOG.warning('Received message for unknown local-id: %s', message) | python | def _handle_message_for_stream(self, stream_transport, message, timeout):
"""Handle an incoming message, check if it's for the given stream.
If the message is not for the stream, then add it to the appropriate
message queue.
Args:
stream_transport: AdbStreamTransport currently waiting on a message.
message: Message to check and handle.
timeout: Timeout to use for the operation, should be an instance of
timeouts.PolledTimeout.
Returns:
The message read if it was for this stream, None otherwise.
Raises:
AdbProtocolError: If we receive an unexpected message type.
"""
if message.command not in ('OKAY', 'CLSE', 'WRTE'):
raise usb_exceptions.AdbProtocolError(
'%s received unexpected message: %s', self, message)
if message.arg1 == stream_transport.local_id:
# Ack writes immediately.
if message.command == 'WRTE':
# Make sure we don't get a WRTE before an OKAY/CLSE message.
if not stream_transport.remote_id:
raise usb_exceptions.AdbProtocolError(
'%s received WRTE before OKAY/CLSE: %s',
stream_transport, message)
self.transport.write_message(adb_message.AdbMessage(
'OKAY', stream_transport.local_id, stream_transport.remote_id),
timeout)
elif message.command == 'CLSE':
self.close_stream_transport(stream_transport, timeout)
return message
else:
# Message was not for this stream, add it to the right stream's queue.
with self._stream_transport_map_lock:
dest_transport = self._stream_transport_map.get(message.arg1)
if dest_transport:
if message.command == 'CLSE':
self.close_stream_transport(dest_transport, timeout)
dest_transport.enqueue_message(message, timeout)
else:
_LOG.warning('Received message for unknown local-id: %s', message) | [
"def",
"_handle_message_for_stream",
"(",
"self",
",",
"stream_transport",
",",
"message",
",",
"timeout",
")",
":",
"if",
"message",
".",
"command",
"not",
"in",
"(",
"'OKAY'",
",",
"'CLSE'",
",",
"'WRTE'",
")",
":",
"raise",
"usb_exceptions",
".",
"AdbProt... | Handle an incoming message, check if it's for the given stream.
If the message is not for the stream, then add it to the appropriate
message queue.
Args:
stream_transport: AdbStreamTransport currently waiting on a message.
message: Message to check and handle.
timeout: Timeout to use for the operation, should be an instance of
timeouts.PolledTimeout.
Returns:
The message read if it was for this stream, None otherwise.
Raises:
AdbProtocolError: If we receive an unexpected message type. | [
"Handle",
"an",
"incoming",
"message",
"check",
"if",
"it",
"s",
"for",
"the",
"given",
"stream",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L596-L642 |
226,951 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection.open_stream | def open_stream(self, destination, timeout_ms=None):
"""Opens a new stream to a destination service on the device.
Not the same as the posix 'open' or any other Open methods, this
corresponds to the OPEN message described in the ADB protocol
documentation mentioned above. It creates a stream (uniquely identified
by remote/local ids) that connects to a particular service endpoint.
Args:
destination: The service:command string, see ADB documentation.
timeout_ms: Timeout in milliseconds for the Open to succeed (or as a
PolledTimeout object).
Raises:
AdbProtocolError: Wrong local_id sent to us, or we didn't get a ready
response.
Returns:
An AdbStream object that can be used to read/write data to the specified
service endpoint, or None if the requested service couldn't be opened.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
stream_transport = self._make_stream_transport()
self.transport.write_message(
adb_message.AdbMessage(
command='OPEN',
arg0=stream_transport.local_id, arg1=0,
data=destination + '\0'),
timeout)
if not stream_transport.ensure_opened(timeout):
return None
return AdbStream(destination, stream_transport) | python | def open_stream(self, destination, timeout_ms=None):
"""Opens a new stream to a destination service on the device.
Not the same as the posix 'open' or any other Open methods, this
corresponds to the OPEN message described in the ADB protocol
documentation mentioned above. It creates a stream (uniquely identified
by remote/local ids) that connects to a particular service endpoint.
Args:
destination: The service:command string, see ADB documentation.
timeout_ms: Timeout in milliseconds for the Open to succeed (or as a
PolledTimeout object).
Raises:
AdbProtocolError: Wrong local_id sent to us, or we didn't get a ready
response.
Returns:
An AdbStream object that can be used to read/write data to the specified
service endpoint, or None if the requested service couldn't be opened.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
stream_transport = self._make_stream_transport()
self.transport.write_message(
adb_message.AdbMessage(
command='OPEN',
arg0=stream_transport.local_id, arg1=0,
data=destination + '\0'),
timeout)
if not stream_transport.ensure_opened(timeout):
return None
return AdbStream(destination, stream_transport) | [
"def",
"open_stream",
"(",
"self",
",",
"destination",
",",
"timeout_ms",
"=",
"None",
")",
":",
"timeout",
"=",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
"stream_transport",
"=",
"self",
".",
"_make_stream_transport",
"(",
... | Opens a new stream to a destination service on the device.
Not the same as the posix 'open' or any other Open methods, this
corresponds to the OPEN message described in the ADB protocol
documentation mentioned above. It creates a stream (uniquely identified
by remote/local ids) that connects to a particular service endpoint.
Args:
destination: The service:command string, see ADB documentation.
timeout_ms: Timeout in milliseconds for the Open to succeed (or as a
PolledTimeout object).
Raises:
AdbProtocolError: Wrong local_id sent to us, or we didn't get a ready
response.
Returns:
An AdbStream object that can be used to read/write data to the specified
service endpoint, or None if the requested service couldn't be opened. | [
"Opens",
"a",
"new",
"stream",
"to",
"a",
"destination",
"service",
"on",
"the",
"device",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L648-L680 |
226,952 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection.close_stream_transport | def close_stream_transport(self, stream_transport, timeout):
"""Remove the given stream transport's id from our map of id's.
If the stream id is actually removed, we send a CLSE message to let the
remote end know (this happens when we are ack'ing a CLSE message we
received). The ADB protocol doesn't say this is a requirement, but ADB
does it, so we do too.
Args:
stream_transport: The stream transport to close.
timeout: Timeout on the operation.
Returns:
True if the id was removed and message sent, False if it was already
missing from the stream map (already closed).
"""
with self._stream_transport_map_lock:
if stream_transport.local_id in self._stream_transport_map:
del self._stream_transport_map[stream_transport.local_id]
# If we never got a remote_id, there's no CLSE message to send.
if stream_transport.remote_id:
self.transport.write_message(adb_message.AdbMessage(
'CLSE', stream_transport.local_id, stream_transport.remote_id),
timeout)
return True
return False | python | def close_stream_transport(self, stream_transport, timeout):
"""Remove the given stream transport's id from our map of id's.
If the stream id is actually removed, we send a CLSE message to let the
remote end know (this happens when we are ack'ing a CLSE message we
received). The ADB protocol doesn't say this is a requirement, but ADB
does it, so we do too.
Args:
stream_transport: The stream transport to close.
timeout: Timeout on the operation.
Returns:
True if the id was removed and message sent, False if it was already
missing from the stream map (already closed).
"""
with self._stream_transport_map_lock:
if stream_transport.local_id in self._stream_transport_map:
del self._stream_transport_map[stream_transport.local_id]
# If we never got a remote_id, there's no CLSE message to send.
if stream_transport.remote_id:
self.transport.write_message(adb_message.AdbMessage(
'CLSE', stream_transport.local_id, stream_transport.remote_id),
timeout)
return True
return False | [
"def",
"close_stream_transport",
"(",
"self",
",",
"stream_transport",
",",
"timeout",
")",
":",
"with",
"self",
".",
"_stream_transport_map_lock",
":",
"if",
"stream_transport",
".",
"local_id",
"in",
"self",
".",
"_stream_transport_map",
":",
"del",
"self",
".",... | Remove the given stream transport's id from our map of id's.
If the stream id is actually removed, we send a CLSE message to let the
remote end know (this happens when we are ack'ing a CLSE message we
received). The ADB protocol doesn't say this is a requirement, but ADB
does it, so we do too.
Args:
stream_transport: The stream transport to close.
timeout: Timeout on the operation.
Returns:
True if the id was removed and message sent, False if it was already
missing from the stream map (already closed). | [
"Remove",
"the",
"given",
"stream",
"transport",
"s",
"id",
"from",
"our",
"map",
"of",
"id",
"s",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L682-L707 |
226,953 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection.streaming_command | def streaming_command(self, service, command='', timeout_ms=None):
"""One complete set of packets for a single command.
Helper function to call open_stream and yield the output. Sends
service:command in a new connection, reading the data for the response. All
the data is held in memory, large responses will be slow and can fill up
memory.
Args:
service: The service on the device to talk to.
command: The command to send to the service.
timeout_ms: Timeout for the entire command, in milliseconds (or as a
PolledTimeout object).
Yields:
The data contained in the responses from the service.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
stream = self.open_stream('%s:%s' % (service, command), timeout)
if not stream:
raise usb_exceptions.AdbStreamUnavailableError(
'%s does not support service: %s', self, service)
for data in stream.read_until_close(timeout):
yield data | python | def streaming_command(self, service, command='', timeout_ms=None):
"""One complete set of packets for a single command.
Helper function to call open_stream and yield the output. Sends
service:command in a new connection, reading the data for the response. All
the data is held in memory, large responses will be slow and can fill up
memory.
Args:
service: The service on the device to talk to.
command: The command to send to the service.
timeout_ms: Timeout for the entire command, in milliseconds (or as a
PolledTimeout object).
Yields:
The data contained in the responses from the service.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
stream = self.open_stream('%s:%s' % (service, command), timeout)
if not stream:
raise usb_exceptions.AdbStreamUnavailableError(
'%s does not support service: %s', self, service)
for data in stream.read_until_close(timeout):
yield data | [
"def",
"streaming_command",
"(",
"self",
",",
"service",
",",
"command",
"=",
"''",
",",
"timeout_ms",
"=",
"None",
")",
":",
"timeout",
"=",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
"stream",
"=",
"self",
".",
"open_st... | One complete set of packets for a single command.
Helper function to call open_stream and yield the output. Sends
service:command in a new connection, reading the data for the response. All
the data is held in memory, large responses will be slow and can fill up
memory.
Args:
service: The service on the device to talk to.
command: The command to send to the service.
timeout_ms: Timeout for the entire command, in milliseconds (or as a
PolledTimeout object).
Yields:
The data contained in the responses from the service. | [
"One",
"complete",
"set",
"of",
"packets",
"for",
"a",
"single",
"command",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L709-L732 |
226,954 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection.read_for_stream | def read_for_stream(self, stream_transport, timeout_ms=None):
"""Attempt to read a packet for the given stream transport.
Will read packets from self.transport until one intended for the given
AdbStream is found. If another thread is already reading packets, this will
block until that thread reads a packet for this stream, or timeout expires.
Note that this method always returns None, but if a packet was read for the
given stream, it will have been added to that stream's message queue.
This is somewhat tricky to do - first we check if there's a message already
in our queue. If not, then we try to use our AdbConnection to read a
message for this stream.
If some other thread is already doing reads, then read_for_stream() will sit
in a tight loop, with a short delay, checking our message queue for a
message from the other thread.
Note that we must pass the queue in from the AdbStream, rather than looking
it up in the AdbConnection's map, because the AdbConnection may have
removed the queue from its map (while it still had messages in it). The
AdbStream itself maintains a reference to the queue to avoid dropping those
messages.
The AdbMessage read is guaranteed to be one of 'OKAY', 'WRTE', or 'CLSE'.
If it was a WRTE message, then it will have been automatically ACK'd with an
OKAY message, if it was a CLSE message it will have been ACK'd with a
corresponding CLSE message, and this AdbStream will be marked as closed.
Args:
stream_transport: The AdbStreamTransport for the stream that is reading
an AdbMessage from this AdbConnection.
timeout_ms: If provided, timeout, in milliseconds, to use. Note this
timeout applies to this entire call, not for each individual Read, since
there may be multiple reads if messages for other streams are read.
This argument may be a timeouts.PolledTimeout.
Returns:
AdbMessage that was read, guaranteed to be one of 'OKAY', 'CLSE', or
'WRTE' command.
Raises:
AdbTimeoutError: If we don't get a packet for this stream before
timeout expires.
AdbStreamClosedError: If the given stream has been closed.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
# Bail when the timeout expires, or when we no longer have the given stream
# in our map (it may have been closed, we don't want to leave a thread
# hanging in this loop when that happens).
while (not timeout.has_expired() and
stream_transport.local_id in self._stream_transport_map):
try:
# Block for up to 10ms to rate-limit how fast we spin.
return stream_transport.message_queue.get(True, .01)
except queue.Empty:
pass
# If someone else has the Lock, just keep checking our queue.
if not self._reader_lock.acquire(False):
continue
try:
# Now that we've acquired the Lock, we have to check the queue again,
# just in case someone had the Lock but hadn't yet added our message
# to the queue when we checked the first time. Now that we have the
# Lock ourselves, we're sure there are no potentially in-flight reads.
try:
return stream_transport.message_queue.get_nowait()
except queue.Empty:
pass
while not timeout.has_expired():
msg = self._handle_message_for_stream(
stream_transport, self.transport.read_message(timeout), timeout)
if msg:
return msg
finally:
self._reader_lock.release()
if timeout.has_expired():
raise usb_exceptions.AdbTimeoutError(
'Read timed out for %s', stream_transport)
# The stream is no longer in the map, so it's closed, but check for any
# queued messages.
try:
return stream_transport.message_queue.get_nowait()
except queue.Empty:
raise usb_exceptions.AdbStreamClosedError(
'Attempt to read from closed or unknown %s', stream_transport) | python | def read_for_stream(self, stream_transport, timeout_ms=None):
"""Attempt to read a packet for the given stream transport.
Will read packets from self.transport until one intended for the given
AdbStream is found. If another thread is already reading packets, this will
block until that thread reads a packet for this stream, or timeout expires.
Note that this method always returns None, but if a packet was read for the
given stream, it will have been added to that stream's message queue.
This is somewhat tricky to do - first we check if there's a message already
in our queue. If not, then we try to use our AdbConnection to read a
message for this stream.
If some other thread is already doing reads, then read_for_stream() will sit
in a tight loop, with a short delay, checking our message queue for a
message from the other thread.
Note that we must pass the queue in from the AdbStream, rather than looking
it up in the AdbConnection's map, because the AdbConnection may have
removed the queue from its map (while it still had messages in it). The
AdbStream itself maintains a reference to the queue to avoid dropping those
messages.
The AdbMessage read is guaranteed to be one of 'OKAY', 'WRTE', or 'CLSE'.
If it was a WRTE message, then it will have been automatically ACK'd with an
OKAY message, if it was a CLSE message it will have been ACK'd with a
corresponding CLSE message, and this AdbStream will be marked as closed.
Args:
stream_transport: The AdbStreamTransport for the stream that is reading
an AdbMessage from this AdbConnection.
timeout_ms: If provided, timeout, in milliseconds, to use. Note this
timeout applies to this entire call, not for each individual Read, since
there may be multiple reads if messages for other streams are read.
This argument may be a timeouts.PolledTimeout.
Returns:
AdbMessage that was read, guaranteed to be one of 'OKAY', 'CLSE', or
'WRTE' command.
Raises:
AdbTimeoutError: If we don't get a packet for this stream before
timeout expires.
AdbStreamClosedError: If the given stream has been closed.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
# Bail when the timeout expires, or when we no longer have the given stream
# in our map (it may have been closed, we don't want to leave a thread
# hanging in this loop when that happens).
while (not timeout.has_expired() and
stream_transport.local_id in self._stream_transport_map):
try:
# Block for up to 10ms to rate-limit how fast we spin.
return stream_transport.message_queue.get(True, .01)
except queue.Empty:
pass
# If someone else has the Lock, just keep checking our queue.
if not self._reader_lock.acquire(False):
continue
try:
# Now that we've acquired the Lock, we have to check the queue again,
# just in case someone had the Lock but hadn't yet added our message
# to the queue when we checked the first time. Now that we have the
# Lock ourselves, we're sure there are no potentially in-flight reads.
try:
return stream_transport.message_queue.get_nowait()
except queue.Empty:
pass
while not timeout.has_expired():
msg = self._handle_message_for_stream(
stream_transport, self.transport.read_message(timeout), timeout)
if msg:
return msg
finally:
self._reader_lock.release()
if timeout.has_expired():
raise usb_exceptions.AdbTimeoutError(
'Read timed out for %s', stream_transport)
# The stream is no longer in the map, so it's closed, but check for any
# queued messages.
try:
return stream_transport.message_queue.get_nowait()
except queue.Empty:
raise usb_exceptions.AdbStreamClosedError(
'Attempt to read from closed or unknown %s', stream_transport) | [
"def",
"read_for_stream",
"(",
"self",
",",
"stream_transport",
",",
"timeout_ms",
"=",
"None",
")",
":",
"timeout",
"=",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
"# Bail when the timeout expires, or when we no longer have the given st... | Attempt to read a packet for the given stream transport.
Will read packets from self.transport until one intended for the given
AdbStream is found. If another thread is already reading packets, this will
block until that thread reads a packet for this stream, or timeout expires.
Note that this method always returns None, but if a packet was read for the
given stream, it will have been added to that stream's message queue.
This is somewhat tricky to do - first we check if there's a message already
in our queue. If not, then we try to use our AdbConnection to read a
message for this stream.
If some other thread is already doing reads, then read_for_stream() will sit
in a tight loop, with a short delay, checking our message queue for a
message from the other thread.
Note that we must pass the queue in from the AdbStream, rather than looking
it up in the AdbConnection's map, because the AdbConnection may have
removed the queue from its map (while it still had messages in it). The
AdbStream itself maintains a reference to the queue to avoid dropping those
messages.
The AdbMessage read is guaranteed to be one of 'OKAY', 'WRTE', or 'CLSE'.
If it was a WRTE message, then it will have been automatically ACK'd with an
OKAY message, if it was a CLSE message it will have been ACK'd with a
corresponding CLSE message, and this AdbStream will be marked as closed.
Args:
stream_transport: The AdbStreamTransport for the stream that is reading
an AdbMessage from this AdbConnection.
timeout_ms: If provided, timeout, in milliseconds, to use. Note this
timeout applies to this entire call, not for each individual Read, since
there may be multiple reads if messages for other streams are read.
This argument may be a timeouts.PolledTimeout.
Returns:
AdbMessage that was read, guaranteed to be one of 'OKAY', 'CLSE', or
'WRTE' command.
Raises:
AdbTimeoutError: If we don't get a packet for this stream before
timeout expires.
AdbStreamClosedError: If the given stream has been closed. | [
"Attempt",
"to",
"read",
"a",
"packet",
"for",
"the",
"given",
"stream",
"transport",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L734-L823 |
226,955 | google/openhtf | openhtf/plugs/usb/adb_protocol.py | AdbConnection.connect | def connect(cls, transport, rsa_keys=None, timeout_ms=1000,
auth_timeout_ms=100):
"""Establish a new connection to a device, connected via transport.
Args:
transport: A transport to use for reads/writes from/to the device,
usually an instance of UsbHandle, but really it can be anything with
read() and write() methods.
rsa_keys: List of AuthSigner subclass instances to be used for
authentication. The device can either accept one of these via the sign
method, or we will send the result of get_public_key from the first one
if the device doesn't accept any of them.
timeout_ms: Timeout to wait for the device to respond to our CNXN
request. Actual timeout may take longer if the transport object passed
has a longer default timeout than timeout_ms, or if auth_timeout_ms is
longer than timeout_ms and public key auth is used. This argument may
be a PolledTimeout object.
auth_timeout_ms: Timeout to wait for when sending a new public key. This
is only relevant when we send a new public key. The device shows a
dialog and this timeout is how long to wait for that dialog. If used
in automation, this should be low to catch such a case as a failure
quickly; while in interactive settings it should be high to allow
users to accept the dialog. We default to automation here, so it's low
by default. This argument may be a PolledTimeout object.
Returns:
An instance of AdbConnection that is connected to the device.
Raises:
usb_exceptions.DeviceAuthError: When the device expects authentication,
but we weren't given any valid keys.
usb_exceptions.AdbProtocolError: When the device does authentication in an
unexpected way, or fails to respond appropriately to our CNXN request.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
if ADB_MESSAGE_LOG:
adb_transport = adb_message.DebugAdbTransportAdapter(transport)
else:
adb_transport = adb_message.AdbTransportAdapter(transport)
adb_transport.write_message(
adb_message.AdbMessage(
command='CNXN', arg0=ADB_VERSION, arg1=MAX_ADB_DATA,
data='host::%s\0' % ADB_BANNER),
timeout)
msg = adb_transport.read_until(('AUTH', 'CNXN'), timeout)
if msg.command == 'CNXN':
return cls(adb_transport, msg.arg1, msg.data)
# We got an AUTH response, so we have to try to authenticate.
if not rsa_keys:
raise usb_exceptions.DeviceAuthError(
'Device authentication required, no keys available.')
# Loop through our keys, signing the last 'banner' or token.
for rsa_key in rsa_keys:
if msg.arg0 != cls.AUTH_TOKEN:
raise usb_exceptions.AdbProtocolError('Bad AUTH response: %s', msg)
signed_token = rsa_key.sign(msg.data)
adb_transport.write_message(
adb_message.AdbMessage(
command='AUTH', arg0=cls.AUTH_SIGNATURE, arg1=0,
data=signed_token),
timeout)
msg = adb_transport.read_until(('AUTH', 'CNXN'), timeout)
if msg.command == 'CNXN':
return cls(adb_transport, msg.arg1, msg.data)
# None of the keys worked, so send a public key.
adb_transport.write_message(
adb_message.AdbMessage(
command='AUTH', arg0=cls.AUTH_RSAPUBLICKEY, arg1=0,
data=rsa_keys[0].get_public_key() + '\0'),
timeout)
try:
msg = adb_transport.read_until(
('CNXN',), timeouts.PolledTimeout.from_millis(auth_timeout_ms))
except usb_exceptions.UsbReadFailedError as exception:
if exception.is_timeout():
exceptions.reraise(usb_exceptions.DeviceAuthError,
'Accept auth key on device, then retry.')
raise
# The read didn't time-out, so we got a CNXN response.
return cls(adb_transport, msg.arg1, msg.data) | python | def connect(cls, transport, rsa_keys=None, timeout_ms=1000,
auth_timeout_ms=100):
"""Establish a new connection to a device, connected via transport.
Args:
transport: A transport to use for reads/writes from/to the device,
usually an instance of UsbHandle, but really it can be anything with
read() and write() methods.
rsa_keys: List of AuthSigner subclass instances to be used for
authentication. The device can either accept one of these via the sign
method, or we will send the result of get_public_key from the first one
if the device doesn't accept any of them.
timeout_ms: Timeout to wait for the device to respond to our CNXN
request. Actual timeout may take longer if the transport object passed
has a longer default timeout than timeout_ms, or if auth_timeout_ms is
longer than timeout_ms and public key auth is used. This argument may
be a PolledTimeout object.
auth_timeout_ms: Timeout to wait for when sending a new public key. This
is only relevant when we send a new public key. The device shows a
dialog and this timeout is how long to wait for that dialog. If used
in automation, this should be low to catch such a case as a failure
quickly; while in interactive settings it should be high to allow
users to accept the dialog. We default to automation here, so it's low
by default. This argument may be a PolledTimeout object.
Returns:
An instance of AdbConnection that is connected to the device.
Raises:
usb_exceptions.DeviceAuthError: When the device expects authentication,
but we weren't given any valid keys.
usb_exceptions.AdbProtocolError: When the device does authentication in an
unexpected way, or fails to respond appropriately to our CNXN request.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
if ADB_MESSAGE_LOG:
adb_transport = adb_message.DebugAdbTransportAdapter(transport)
else:
adb_transport = adb_message.AdbTransportAdapter(transport)
adb_transport.write_message(
adb_message.AdbMessage(
command='CNXN', arg0=ADB_VERSION, arg1=MAX_ADB_DATA,
data='host::%s\0' % ADB_BANNER),
timeout)
msg = adb_transport.read_until(('AUTH', 'CNXN'), timeout)
if msg.command == 'CNXN':
return cls(adb_transport, msg.arg1, msg.data)
# We got an AUTH response, so we have to try to authenticate.
if not rsa_keys:
raise usb_exceptions.DeviceAuthError(
'Device authentication required, no keys available.')
# Loop through our keys, signing the last 'banner' or token.
for rsa_key in rsa_keys:
if msg.arg0 != cls.AUTH_TOKEN:
raise usb_exceptions.AdbProtocolError('Bad AUTH response: %s', msg)
signed_token = rsa_key.sign(msg.data)
adb_transport.write_message(
adb_message.AdbMessage(
command='AUTH', arg0=cls.AUTH_SIGNATURE, arg1=0,
data=signed_token),
timeout)
msg = adb_transport.read_until(('AUTH', 'CNXN'), timeout)
if msg.command == 'CNXN':
return cls(adb_transport, msg.arg1, msg.data)
# None of the keys worked, so send a public key.
adb_transport.write_message(
adb_message.AdbMessage(
command='AUTH', arg0=cls.AUTH_RSAPUBLICKEY, arg1=0,
data=rsa_keys[0].get_public_key() + '\0'),
timeout)
try:
msg = adb_transport.read_until(
('CNXN',), timeouts.PolledTimeout.from_millis(auth_timeout_ms))
except usb_exceptions.UsbReadFailedError as exception:
if exception.is_timeout():
exceptions.reraise(usb_exceptions.DeviceAuthError,
'Accept auth key on device, then retry.')
raise
# The read didn't time-out, so we got a CNXN response.
return cls(adb_transport, msg.arg1, msg.data) | [
"def",
"connect",
"(",
"cls",
",",
"transport",
",",
"rsa_keys",
"=",
"None",
",",
"timeout_ms",
"=",
"1000",
",",
"auth_timeout_ms",
"=",
"100",
")",
":",
"timeout",
"=",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
"if",
... | Establish a new connection to a device, connected via transport.
Args:
transport: A transport to use for reads/writes from/to the device,
usually an instance of UsbHandle, but really it can be anything with
read() and write() methods.
rsa_keys: List of AuthSigner subclass instances to be used for
authentication. The device can either accept one of these via the sign
method, or we will send the result of get_public_key from the first one
if the device doesn't accept any of them.
timeout_ms: Timeout to wait for the device to respond to our CNXN
request. Actual timeout may take longer if the transport object passed
has a longer default timeout than timeout_ms, or if auth_timeout_ms is
longer than timeout_ms and public key auth is used. This argument may
be a PolledTimeout object.
auth_timeout_ms: Timeout to wait for when sending a new public key. This
is only relevant when we send a new public key. The device shows a
dialog and this timeout is how long to wait for that dialog. If used
in automation, this should be low to catch such a case as a failure
quickly; while in interactive settings it should be high to allow
users to accept the dialog. We default to automation here, so it's low
by default. This argument may be a PolledTimeout object.
Returns:
An instance of AdbConnection that is connected to the device.
Raises:
usb_exceptions.DeviceAuthError: When the device expects authentication,
but we weren't given any valid keys.
usb_exceptions.AdbProtocolError: When the device does authentication in an
unexpected way, or fails to respond appropriately to our CNXN request. | [
"Establish",
"a",
"new",
"connection",
"to",
"a",
"device",
"connected",
"via",
"transport",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/adb_protocol.py#L826-L912 |
226,956 | google/openhtf | examples/example_plugs.py | ExamplePlug.increment | def increment(self):
"""Increment our value, return the previous value."""
self.value += self.increment_size
return self.value - self.increment_size | python | def increment(self):
"""Increment our value, return the previous value."""
self.value += self.increment_size
return self.value - self.increment_size | [
"def",
"increment",
"(",
"self",
")",
":",
"self",
".",
"value",
"+=",
"self",
".",
"increment_size",
"return",
"self",
".",
"value",
"-",
"self",
".",
"increment_size"
] | Increment our value, return the previous value. | [
"Increment",
"our",
"value",
"return",
"the",
"previous",
"value",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/examples/example_plugs.py#L76-L79 |
226,957 | google/openhtf | openhtf/plugs/__init__.py | plug | def plug(update_kwargs=True, **plugs_map):
"""Creates a decorator that passes in plugs when invoked.
This function returns a decorator for a function that will replace positional
arguments to that function with the plugs specified. See the module
docstring for details and examples.
Note this decorator does not work with class or bound methods, but does work
with @staticmethod.
Args:
update_kwargs: If true, makes the decorated phase take this plug as a kwarg.
**plugs_map: Dict mapping name to Plug type.
Returns:
A PhaseDescriptor that will pass plug instances in as kwargs when invoked.
Raises:
InvalidPlugError: If a type is provided that is not a subclass of BasePlug.
"""
for a_plug in plugs_map.values():
if not (isinstance(a_plug, PlugPlaceholder)
or issubclass(a_plug, BasePlug)):
raise InvalidPlugError(
'Plug %s is not a subclass of plugs.BasePlug nor a placeholder '
'for one' % a_plug)
def result(func):
"""Wrap the given function and return the wrapper.
Args:
func: The function to wrap.
Returns:
A PhaseDescriptor that, when called will invoke the wrapped function,
passing plugs as keyword args.
Raises:
DuplicatePlugError: If a plug name is declared twice for the
same function.
"""
phase = openhtf.core.phase_descriptor.PhaseDescriptor.wrap_or_copy(func)
duplicates = (frozenset(p.name for p in phase.plugs) &
frozenset(plugs_map))
if duplicates:
raise DuplicatePlugError(
'Plugs %s required multiple times on phase %s' % (duplicates, func))
phase.plugs.extend([
PhasePlug(name, a_plug, update_kwargs=update_kwargs)
for name, a_plug in six.iteritems(plugs_map)])
return phase
return result | python | def plug(update_kwargs=True, **plugs_map):
"""Creates a decorator that passes in plugs when invoked.
This function returns a decorator for a function that will replace positional
arguments to that function with the plugs specified. See the module
docstring for details and examples.
Note this decorator does not work with class or bound methods, but does work
with @staticmethod.
Args:
update_kwargs: If true, makes the decorated phase take this plug as a kwarg.
**plugs_map: Dict mapping name to Plug type.
Returns:
A PhaseDescriptor that will pass plug instances in as kwargs when invoked.
Raises:
InvalidPlugError: If a type is provided that is not a subclass of BasePlug.
"""
for a_plug in plugs_map.values():
if not (isinstance(a_plug, PlugPlaceholder)
or issubclass(a_plug, BasePlug)):
raise InvalidPlugError(
'Plug %s is not a subclass of plugs.BasePlug nor a placeholder '
'for one' % a_plug)
def result(func):
"""Wrap the given function and return the wrapper.
Args:
func: The function to wrap.
Returns:
A PhaseDescriptor that, when called will invoke the wrapped function,
passing plugs as keyword args.
Raises:
DuplicatePlugError: If a plug name is declared twice for the
same function.
"""
phase = openhtf.core.phase_descriptor.PhaseDescriptor.wrap_or_copy(func)
duplicates = (frozenset(p.name for p in phase.plugs) &
frozenset(plugs_map))
if duplicates:
raise DuplicatePlugError(
'Plugs %s required multiple times on phase %s' % (duplicates, func))
phase.plugs.extend([
PhasePlug(name, a_plug, update_kwargs=update_kwargs)
for name, a_plug in six.iteritems(plugs_map)])
return phase
return result | [
"def",
"plug",
"(",
"update_kwargs",
"=",
"True",
",",
"*",
"*",
"plugs_map",
")",
":",
"for",
"a_plug",
"in",
"plugs_map",
".",
"values",
"(",
")",
":",
"if",
"not",
"(",
"isinstance",
"(",
"a_plug",
",",
"PlugPlaceholder",
")",
"or",
"issubclass",
"(... | Creates a decorator that passes in plugs when invoked.
This function returns a decorator for a function that will replace positional
arguments to that function with the plugs specified. See the module
docstring for details and examples.
Note this decorator does not work with class or bound methods, but does work
with @staticmethod.
Args:
update_kwargs: If true, makes the decorated phase take this plug as a kwarg.
**plugs_map: Dict mapping name to Plug type.
Returns:
A PhaseDescriptor that will pass plug instances in as kwargs when invoked.
Raises:
InvalidPlugError: If a type is provided that is not a subclass of BasePlug. | [
"Creates",
"a",
"decorator",
"that",
"passes",
"in",
"plugs",
"when",
"invoked",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L217-L269 |
226,958 | google/openhtf | openhtf/plugs/__init__.py | BasePlug.uses_base_tear_down | def uses_base_tear_down(cls):
"""Checks whether the tearDown method is the BasePlug implementation."""
this_tear_down = getattr(cls, 'tearDown')
base_tear_down = getattr(BasePlug, 'tearDown')
return this_tear_down.__code__ is base_tear_down.__code__ | python | def uses_base_tear_down(cls):
"""Checks whether the tearDown method is the BasePlug implementation."""
this_tear_down = getattr(cls, 'tearDown')
base_tear_down = getattr(BasePlug, 'tearDown')
return this_tear_down.__code__ is base_tear_down.__code__ | [
"def",
"uses_base_tear_down",
"(",
"cls",
")",
":",
"this_tear_down",
"=",
"getattr",
"(",
"cls",
",",
"'tearDown'",
")",
"base_tear_down",
"=",
"getattr",
"(",
"BasePlug",
",",
"'tearDown'",
")",
"return",
"this_tear_down",
".",
"__code__",
"is",
"base_tear_dow... | Checks whether the tearDown method is the BasePlug implementation. | [
"Checks",
"whether",
"the",
"tearDown",
"method",
"is",
"the",
"BasePlug",
"implementation",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L196-L200 |
226,959 | google/openhtf | openhtf/plugs/__init__.py | PlugManager.get_plug_mro | def get_plug_mro(self, plug_type):
"""Returns a list of names identifying the plug classes in the plug's MRO.
For example:
['openhtf.plugs.user_input.UserInput']
Or:
['openhtf.plugs.user_input.UserInput',
'my_module.advanced_user_input.AdvancedUserInput']
"""
ignored_classes = (BasePlug, FrontendAwareBasePlug)
return [
self.get_plug_name(base_class) for base_class in plug_type.mro()
if (issubclass(base_class, BasePlug) and
base_class not in ignored_classes)
] | python | def get_plug_mro(self, plug_type):
"""Returns a list of names identifying the plug classes in the plug's MRO.
For example:
['openhtf.plugs.user_input.UserInput']
Or:
['openhtf.plugs.user_input.UserInput',
'my_module.advanced_user_input.AdvancedUserInput']
"""
ignored_classes = (BasePlug, FrontendAwareBasePlug)
return [
self.get_plug_name(base_class) for base_class in plug_type.mro()
if (issubclass(base_class, BasePlug) and
base_class not in ignored_classes)
] | [
"def",
"get_plug_mro",
"(",
"self",
",",
"plug_type",
")",
":",
"ignored_classes",
"=",
"(",
"BasePlug",
",",
"FrontendAwareBasePlug",
")",
"return",
"[",
"self",
".",
"get_plug_name",
"(",
"base_class",
")",
"for",
"base_class",
"in",
"plug_type",
".",
"mro",... | Returns a list of names identifying the plug classes in the plug's MRO.
For example:
['openhtf.plugs.user_input.UserInput']
Or:
['openhtf.plugs.user_input.UserInput',
'my_module.advanced_user_input.AdvancedUserInput'] | [
"Returns",
"a",
"list",
"of",
"names",
"identifying",
"the",
"plug",
"classes",
"in",
"the",
"plug",
"s",
"MRO",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L338-L352 |
226,960 | google/openhtf | openhtf/plugs/__init__.py | PlugManager.initialize_plugs | def initialize_plugs(self, plug_types=None):
"""Instantiate required plugs.
Instantiates plug types and saves the instances in self._plugs_by_type for
use in provide_plugs().
Args:
plug_types: Plug types may be specified here rather than passed
into the constructor (this is used primarily for unit testing
phases).
"""
types = plug_types if plug_types is not None else self._plug_types
for plug_type in types:
# Create a logger for this plug. All plug loggers go under the 'plug'
# sub-logger in the logger hierarchy.
plug_logger = self.logger.getChild(plug_type.__name__)
if plug_type in self._plugs_by_type:
continue
try:
if not issubclass(plug_type, BasePlug):
raise InvalidPlugError(
'Plug type "%s" is not an instance of BasePlug' % plug_type)
if plug_type.logger != _LOG:
# They put a logger attribute on the class itself, overriding ours.
raise InvalidPlugError(
'Do not override "logger" in your plugs.', plug_type)
# Override the logger so that __init__'s logging goes into the record.
plug_type.logger = plug_logger
try:
plug_instance = plug_type()
finally:
# Now set it back since we'll give the instance a logger in a moment.
plug_type.logger = _LOG
# Set the logger attribute directly (rather than in BasePlug) so we
# don't depend on subclasses' implementation of __init__ to have it
# set.
if plug_instance.logger != _LOG:
raise InvalidPlugError(
'Do not set "self.logger" in __init__ in your plugs', plug_type)
else:
# Now the instance has its own copy of the test logger.
plug_instance.logger = plug_logger
except Exception: # pylint: disable=broad-except
plug_logger.exception('Exception instantiating plug type %s', plug_type)
self.tear_down_plugs()
raise
self.update_plug(plug_type, plug_instance) | python | def initialize_plugs(self, plug_types=None):
"""Instantiate required plugs.
Instantiates plug types and saves the instances in self._plugs_by_type for
use in provide_plugs().
Args:
plug_types: Plug types may be specified here rather than passed
into the constructor (this is used primarily for unit testing
phases).
"""
types = plug_types if plug_types is not None else self._plug_types
for plug_type in types:
# Create a logger for this plug. All plug loggers go under the 'plug'
# sub-logger in the logger hierarchy.
plug_logger = self.logger.getChild(plug_type.__name__)
if plug_type in self._plugs_by_type:
continue
try:
if not issubclass(plug_type, BasePlug):
raise InvalidPlugError(
'Plug type "%s" is not an instance of BasePlug' % plug_type)
if plug_type.logger != _LOG:
# They put a logger attribute on the class itself, overriding ours.
raise InvalidPlugError(
'Do not override "logger" in your plugs.', plug_type)
# Override the logger so that __init__'s logging goes into the record.
plug_type.logger = plug_logger
try:
plug_instance = plug_type()
finally:
# Now set it back since we'll give the instance a logger in a moment.
plug_type.logger = _LOG
# Set the logger attribute directly (rather than in BasePlug) so we
# don't depend on subclasses' implementation of __init__ to have it
# set.
if plug_instance.logger != _LOG:
raise InvalidPlugError(
'Do not set "self.logger" in __init__ in your plugs', plug_type)
else:
# Now the instance has its own copy of the test logger.
plug_instance.logger = plug_logger
except Exception: # pylint: disable=broad-except
plug_logger.exception('Exception instantiating plug type %s', plug_type)
self.tear_down_plugs()
raise
self.update_plug(plug_type, plug_instance) | [
"def",
"initialize_plugs",
"(",
"self",
",",
"plug_types",
"=",
"None",
")",
":",
"types",
"=",
"plug_types",
"if",
"plug_types",
"is",
"not",
"None",
"else",
"self",
".",
"_plug_types",
"for",
"plug_type",
"in",
"types",
":",
"# Create a logger for this plug. A... | Instantiate required plugs.
Instantiates plug types and saves the instances in self._plugs_by_type for
use in provide_plugs().
Args:
plug_types: Plug types may be specified here rather than passed
into the constructor (this is used primarily for unit testing
phases). | [
"Instantiate",
"required",
"plugs",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L362-L409 |
226,961 | google/openhtf | openhtf/plugs/__init__.py | PlugManager.update_plug | def update_plug(self, plug_type, plug_value):
"""Update internal data stores with the given plug value for plug type.
Safely tears down the old instance if one was already created, but that's
generally not the case outside unittests. Also, we explicitly pass the
plug_type rather than detecting it from plug_value to allow unittests to
override plugs with Mock instances.
Note this should only be used inside unittests, as this mechanism is not
compatible with RemotePlug support.
"""
self._plug_types.add(plug_type)
if plug_type in self._plugs_by_type:
self._plugs_by_type[plug_type].tearDown()
plug_name = self.get_plug_name(plug_type)
self._plugs_by_type[plug_type] = plug_value
self._plugs_by_name[plug_name] = plug_value
self._plug_descriptors[plug_name] = self._make_plug_descriptor(plug_type) | python | def update_plug(self, plug_type, plug_value):
"""Update internal data stores with the given plug value for plug type.
Safely tears down the old instance if one was already created, but that's
generally not the case outside unittests. Also, we explicitly pass the
plug_type rather than detecting it from plug_value to allow unittests to
override plugs with Mock instances.
Note this should only be used inside unittests, as this mechanism is not
compatible with RemotePlug support.
"""
self._plug_types.add(plug_type)
if plug_type in self._plugs_by_type:
self._plugs_by_type[plug_type].tearDown()
plug_name = self.get_plug_name(plug_type)
self._plugs_by_type[plug_type] = plug_value
self._plugs_by_name[plug_name] = plug_value
self._plug_descriptors[plug_name] = self._make_plug_descriptor(plug_type) | [
"def",
"update_plug",
"(",
"self",
",",
"plug_type",
",",
"plug_value",
")",
":",
"self",
".",
"_plug_types",
".",
"add",
"(",
"plug_type",
")",
"if",
"plug_type",
"in",
"self",
".",
"_plugs_by_type",
":",
"self",
".",
"_plugs_by_type",
"[",
"plug_type",
"... | Update internal data stores with the given plug value for plug type.
Safely tears down the old instance if one was already created, but that's
generally not the case outside unittests. Also, we explicitly pass the
plug_type rather than detecting it from plug_value to allow unittests to
override plugs with Mock instances.
Note this should only be used inside unittests, as this mechanism is not
compatible with RemotePlug support. | [
"Update",
"internal",
"data",
"stores",
"with",
"the",
"given",
"plug",
"value",
"for",
"plug",
"type",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L425-L442 |
226,962 | google/openhtf | openhtf/plugs/__init__.py | PlugManager.wait_for_plug_update | def wait_for_plug_update(self, plug_name, remote_state, timeout_s):
"""Wait for a change in the state of a frontend-aware plug.
Args:
plug_name: Plug name, e.g. 'openhtf.plugs.user_input.UserInput'.
remote_state: The last observed state.
timeout_s: Number of seconds to wait for an update.
Returns:
An updated state, or None if the timeout runs out.
Raises:
InvalidPlugError: The plug can't be waited on either because it's not in
use or it's not a frontend-aware plug.
"""
plug = self._plugs_by_name.get(plug_name)
if plug is None:
raise InvalidPlugError('Cannot wait on unknown plug "%s".' % plug_name)
if not isinstance(plug, FrontendAwareBasePlug):
raise InvalidPlugError('Cannot wait on a plug %s that is not an subclass '
'of FrontendAwareBasePlug.' % plug_name)
state, update_event = plug.asdict_with_event()
if state != remote_state:
return state
if update_event.wait(timeout_s):
return plug._asdict() | python | def wait_for_plug_update(self, plug_name, remote_state, timeout_s):
"""Wait for a change in the state of a frontend-aware plug.
Args:
plug_name: Plug name, e.g. 'openhtf.plugs.user_input.UserInput'.
remote_state: The last observed state.
timeout_s: Number of seconds to wait for an update.
Returns:
An updated state, or None if the timeout runs out.
Raises:
InvalidPlugError: The plug can't be waited on either because it's not in
use or it's not a frontend-aware plug.
"""
plug = self._plugs_by_name.get(plug_name)
if plug is None:
raise InvalidPlugError('Cannot wait on unknown plug "%s".' % plug_name)
if not isinstance(plug, FrontendAwareBasePlug):
raise InvalidPlugError('Cannot wait on a plug %s that is not an subclass '
'of FrontendAwareBasePlug.' % plug_name)
state, update_event = plug.asdict_with_event()
if state != remote_state:
return state
if update_event.wait(timeout_s):
return plug._asdict() | [
"def",
"wait_for_plug_update",
"(",
"self",
",",
"plug_name",
",",
"remote_state",
",",
"timeout_s",
")",
":",
"plug",
"=",
"self",
".",
"_plugs_by_name",
".",
"get",
"(",
"plug_name",
")",
"if",
"plug",
"is",
"None",
":",
"raise",
"InvalidPlugError",
"(",
... | Wait for a change in the state of a frontend-aware plug.
Args:
plug_name: Plug name, e.g. 'openhtf.plugs.user_input.UserInput'.
remote_state: The last observed state.
timeout_s: Number of seconds to wait for an update.
Returns:
An updated state, or None if the timeout runs out.
Raises:
InvalidPlugError: The plug can't be waited on either because it's not in
use or it's not a frontend-aware plug. | [
"Wait",
"for",
"a",
"change",
"in",
"the",
"state",
"of",
"a",
"frontend",
"-",
"aware",
"plug",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L477-L506 |
226,963 | google/openhtf | openhtf/plugs/__init__.py | PlugManager.get_frontend_aware_plug_names | def get_frontend_aware_plug_names(self):
"""Returns the names of frontend-aware plugs."""
return [name for name, plug in six.iteritems(self._plugs_by_name)
if isinstance(plug, FrontendAwareBasePlug)] | python | def get_frontend_aware_plug_names(self):
"""Returns the names of frontend-aware plugs."""
return [name for name, plug in six.iteritems(self._plugs_by_name)
if isinstance(plug, FrontendAwareBasePlug)] | [
"def",
"get_frontend_aware_plug_names",
"(",
"self",
")",
":",
"return",
"[",
"name",
"for",
"name",
",",
"plug",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_plugs_by_name",
")",
"if",
"isinstance",
"(",
"plug",
",",
"FrontendAwareBasePlug",
")",
"]"
... | Returns the names of frontend-aware plugs. | [
"Returns",
"the",
"names",
"of",
"frontend",
"-",
"aware",
"plugs",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/__init__.py#L508-L511 |
226,964 | google/openhtf | openhtf/output/servers/station_server.py | _wait_for_any_event | def _wait_for_any_event(events, timeout_s):
"""Wait for any in a list of threading.Event's to be set.
Args:
events: List of threading.Event's.
timeout_s: Max duration in seconds to wait before returning.
Returns:
True if at least one event was set before the timeout expired, else False.
"""
def any_event_set():
return any(event.is_set() for event in events)
result = timeouts.loop_until_timeout_or_true(
timeout_s, any_event_set, sleep_s=_WAIT_FOR_ANY_EVENT_POLL_S)
return result or any_event_set() | python | def _wait_for_any_event(events, timeout_s):
"""Wait for any in a list of threading.Event's to be set.
Args:
events: List of threading.Event's.
timeout_s: Max duration in seconds to wait before returning.
Returns:
True if at least one event was set before the timeout expired, else False.
"""
def any_event_set():
return any(event.is_set() for event in events)
result = timeouts.loop_until_timeout_or_true(
timeout_s, any_event_set, sleep_s=_WAIT_FOR_ANY_EVENT_POLL_S)
return result or any_event_set() | [
"def",
"_wait_for_any_event",
"(",
"events",
",",
"timeout_s",
")",
":",
"def",
"any_event_set",
"(",
")",
":",
"return",
"any",
"(",
"event",
".",
"is_set",
"(",
")",
"for",
"event",
"in",
"events",
")",
"result",
"=",
"timeouts",
".",
"loop_until_timeout... | Wait for any in a list of threading.Event's to be set.
Args:
events: List of threading.Event's.
timeout_s: Max duration in seconds to wait before returning.
Returns:
True if at least one event was set before the timeout expired, else False. | [
"Wait",
"for",
"any",
"in",
"a",
"list",
"of",
"threading",
".",
"Event",
"s",
"to",
"be",
"set",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/servers/station_server.py#L119-L135 |
226,965 | google/openhtf | openhtf/output/servers/station_server.py | StationWatcher._poll_for_update | def _poll_for_update(self):
"""Call the callback with the current test state, then wait for a change."""
test, test_state = _get_executing_test()
if test is None:
time.sleep(_WAIT_FOR_EXECUTING_TEST_POLL_S)
return
state_dict, event = self._to_dict_with_event(test_state)
self._update_callback(state_dict)
plug_manager = test_state.plug_manager
plug_events = [
plug_manager.get_plug_by_class_path(plug_name).asdict_with_event()[1]
for plug_name in plug_manager.get_frontend_aware_plug_names()
]
events = [event] + plug_events
# Wait for the test state or a plug state to change, or for the previously
# executing test to finish.
while not _wait_for_any_event(events, _CHECK_FOR_FINISHED_TEST_POLL_S):
new_test, _ = _get_executing_test()
if test != new_test:
break | python | def _poll_for_update(self):
"""Call the callback with the current test state, then wait for a change."""
test, test_state = _get_executing_test()
if test is None:
time.sleep(_WAIT_FOR_EXECUTING_TEST_POLL_S)
return
state_dict, event = self._to_dict_with_event(test_state)
self._update_callback(state_dict)
plug_manager = test_state.plug_manager
plug_events = [
plug_manager.get_plug_by_class_path(plug_name).asdict_with_event()[1]
for plug_name in plug_manager.get_frontend_aware_plug_names()
]
events = [event] + plug_events
# Wait for the test state or a plug state to change, or for the previously
# executing test to finish.
while not _wait_for_any_event(events, _CHECK_FOR_FINISHED_TEST_POLL_S):
new_test, _ = _get_executing_test()
if test != new_test:
break | [
"def",
"_poll_for_update",
"(",
"self",
")",
":",
"test",
",",
"test_state",
"=",
"_get_executing_test",
"(",
")",
"if",
"test",
"is",
"None",
":",
"time",
".",
"sleep",
"(",
"_WAIT_FOR_EXECUTING_TEST_POLL_S",
")",
"return",
"state_dict",
",",
"event",
"=",
... | Call the callback with the current test state, then wait for a change. | [
"Call",
"the",
"callback",
"with",
"the",
"current",
"test",
"state",
"then",
"wait",
"for",
"a",
"change",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/servers/station_server.py#L176-L199 |
226,966 | google/openhtf | openhtf/output/servers/station_server.py | StationWatcher._to_dict_with_event | def _to_dict_with_event(cls, test_state):
"""Process a test state into the format we want to send to the frontend."""
original_dict, event = test_state.asdict_with_event()
# This line may produce a 'dictionary changed size during iteration' error.
test_state_dict = data.convert_to_base_types(original_dict)
test_state_dict['execution_uid'] = test_state.execution_uid
return test_state_dict, event | python | def _to_dict_with_event(cls, test_state):
"""Process a test state into the format we want to send to the frontend."""
original_dict, event = test_state.asdict_with_event()
# This line may produce a 'dictionary changed size during iteration' error.
test_state_dict = data.convert_to_base_types(original_dict)
test_state_dict['execution_uid'] = test_state.execution_uid
return test_state_dict, event | [
"def",
"_to_dict_with_event",
"(",
"cls",
",",
"test_state",
")",
":",
"original_dict",
",",
"event",
"=",
"test_state",
".",
"asdict_with_event",
"(",
")",
"# This line may produce a 'dictionary changed size during iteration' error.",
"test_state_dict",
"=",
"data",
".",
... | Process a test state into the format we want to send to the frontend. | [
"Process",
"a",
"test",
"state",
"into",
"the",
"format",
"we",
"want",
"to",
"send",
"to",
"the",
"frontend",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/servers/station_server.py#L202-L210 |
226,967 | google/openhtf | openhtf/util/threads.py | _safe_lock_release_py2 | def _safe_lock_release_py2(rlock):
"""Ensure that a threading.RLock is fully released for Python 2.
The RLock release code is:
https://github.com/python/cpython/blob/2.7/Lib/threading.py#L187
The RLock object's release method does not release all of its state if an
exception is raised in the middle of its operation. There are three pieces of
internal state that must be cleaned up:
- owning thread ident, an integer.
- entry count, an integer that counts how many times the current owner has
locked the RLock.
- internal lock, a threading.Lock instance that handles blocking.
Args:
rlock: threading.RLock, lock to fully release.
Yields:
None.
"""
assert isinstance(rlock, threading._RLock)
ident = _thread.get_ident()
expected_count = 0
if rlock._RLock__owner == ident:
expected_count = rlock._RLock__count
try:
yield
except ThreadTerminationError:
# Check if the current thread still owns the lock by checking if we can
# acquire the underlying lock.
if rlock._RLock__block.acquire(0):
# Lock is clean, so unlock and we are done.
rlock._RLock__block.release()
elif rlock._RLock__owner == ident and expected_count > 0:
# The lock is still held up the stack, so make sure the count is accurate.
if rlock._RLock__count != expected_count:
rlock._RLock__count = expected_count
elif rlock._RLock__owner == ident or rlock._RLock__owner is None:
# The internal lock is still acquired, but either this thread or no thread
# owns it, which means it needs to be hard reset.
rlock._RLock__owner = None
rlock._RLock__count = 0
rlock._RLock__block.release()
raise | python | def _safe_lock_release_py2(rlock):
"""Ensure that a threading.RLock is fully released for Python 2.
The RLock release code is:
https://github.com/python/cpython/blob/2.7/Lib/threading.py#L187
The RLock object's release method does not release all of its state if an
exception is raised in the middle of its operation. There are three pieces of
internal state that must be cleaned up:
- owning thread ident, an integer.
- entry count, an integer that counts how many times the current owner has
locked the RLock.
- internal lock, a threading.Lock instance that handles blocking.
Args:
rlock: threading.RLock, lock to fully release.
Yields:
None.
"""
assert isinstance(rlock, threading._RLock)
ident = _thread.get_ident()
expected_count = 0
if rlock._RLock__owner == ident:
expected_count = rlock._RLock__count
try:
yield
except ThreadTerminationError:
# Check if the current thread still owns the lock by checking if we can
# acquire the underlying lock.
if rlock._RLock__block.acquire(0):
# Lock is clean, so unlock and we are done.
rlock._RLock__block.release()
elif rlock._RLock__owner == ident and expected_count > 0:
# The lock is still held up the stack, so make sure the count is accurate.
if rlock._RLock__count != expected_count:
rlock._RLock__count = expected_count
elif rlock._RLock__owner == ident or rlock._RLock__owner is None:
# The internal lock is still acquired, but either this thread or no thread
# owns it, which means it needs to be hard reset.
rlock._RLock__owner = None
rlock._RLock__count = 0
rlock._RLock__block.release()
raise | [
"def",
"_safe_lock_release_py2",
"(",
"rlock",
")",
":",
"assert",
"isinstance",
"(",
"rlock",
",",
"threading",
".",
"_RLock",
")",
"ident",
"=",
"_thread",
".",
"get_ident",
"(",
")",
"expected_count",
"=",
"0",
"if",
"rlock",
".",
"_RLock__owner",
"==",
... | Ensure that a threading.RLock is fully released for Python 2.
The RLock release code is:
https://github.com/python/cpython/blob/2.7/Lib/threading.py#L187
The RLock object's release method does not release all of its state if an
exception is raised in the middle of its operation. There are three pieces of
internal state that must be cleaned up:
- owning thread ident, an integer.
- entry count, an integer that counts how many times the current owner has
locked the RLock.
- internal lock, a threading.Lock instance that handles blocking.
Args:
rlock: threading.RLock, lock to fully release.
Yields:
None. | [
"Ensure",
"that",
"a",
"threading",
".",
"RLock",
"is",
"fully",
"released",
"for",
"Python",
"2",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/threads.py#L54-L97 |
226,968 | google/openhtf | openhtf/util/threads.py | loop | def loop(_=None, force=False): # pylint: disable=invalid-name
"""Causes a function to loop indefinitely."""
if not force:
raise AttributeError(
'threads.loop() is DEPRECATED. If you really like this and want to '
'keep it, file an issue at https://github.com/google/openhtf/issues '
'and use it as @loop(force=True) for now.')
def real_loop(fn):
@functools.wraps(fn)
def _proc(*args, **kwargs):
"""Wrapper to return."""
while True:
fn(*args, **kwargs)
_proc.once = fn # way for tests to invoke the function once
# you may need to pass in "self" since this may be unbound.
return _proc
return real_loop | python | def loop(_=None, force=False): # pylint: disable=invalid-name
"""Causes a function to loop indefinitely."""
if not force:
raise AttributeError(
'threads.loop() is DEPRECATED. If you really like this and want to '
'keep it, file an issue at https://github.com/google/openhtf/issues '
'and use it as @loop(force=True) for now.')
def real_loop(fn):
@functools.wraps(fn)
def _proc(*args, **kwargs):
"""Wrapper to return."""
while True:
fn(*args, **kwargs)
_proc.once = fn # way for tests to invoke the function once
# you may need to pass in "self" since this may be unbound.
return _proc
return real_loop | [
"def",
"loop",
"(",
"_",
"=",
"None",
",",
"force",
"=",
"False",
")",
":",
"# pylint: disable=invalid-name",
"if",
"not",
"force",
":",
"raise",
"AttributeError",
"(",
"'threads.loop() is DEPRECATED. If you really like this and want to '",
"'keep it, file an issue at http... | Causes a function to loop indefinitely. | [
"Causes",
"a",
"function",
"to",
"loop",
"indefinitely",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/threads.py#L101-L118 |
226,969 | google/openhtf | openhtf/util/threads.py | synchronized | def synchronized(func): # pylint: disable=invalid-name
"""Hold self._lock while executing func."""
@functools.wraps(func)
def synchronized_method(self, *args, **kwargs):
"""Wrapper to return."""
if not hasattr(self, '_lock'):
if func.__name__ in type(self).__dict__:
hint = ''
else:
hint = (' Might be missing call to super in %s.__init__?' %
type(self).__name__)
raise RuntimeError('Can\'t synchronize method `%s` of %s without '
'attribute `_lock`.%s' %
(func.__name__, type(self).__name__, hint))
with self._lock: # pylint: disable=protected-access
return func(self, *args, **kwargs)
return synchronized_method | python | def synchronized(func): # pylint: disable=invalid-name
"""Hold self._lock while executing func."""
@functools.wraps(func)
def synchronized_method(self, *args, **kwargs):
"""Wrapper to return."""
if not hasattr(self, '_lock'):
if func.__name__ in type(self).__dict__:
hint = ''
else:
hint = (' Might be missing call to super in %s.__init__?' %
type(self).__name__)
raise RuntimeError('Can\'t synchronize method `%s` of %s without '
'attribute `_lock`.%s' %
(func.__name__, type(self).__name__, hint))
with self._lock: # pylint: disable=protected-access
return func(self, *args, **kwargs)
return synchronized_method | [
"def",
"synchronized",
"(",
"func",
")",
":",
"# pylint: disable=invalid-name",
"@",
"functools",
".",
"wraps",
"(",
"func",
")",
"def",
"synchronized_method",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"\"\"\"Wrapper to return.\"\"\"",
... | Hold self._lock while executing func. | [
"Hold",
"self",
".",
"_lock",
"while",
"executing",
"func",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/threads.py#L241-L257 |
226,970 | google/openhtf | openhtf/util/threads.py | KillableThread.kill | def kill(self):
"""Terminates the current thread by raising an error."""
self._killed.set()
if not self.is_alive():
logging.debug('Cannot kill thread that is no longer running.')
return
if not self._is_thread_proc_running():
logging.debug("Thread's _thread_proc function is no longer running, "
'will not kill; letting thread exit gracefully.')
return
self.async_raise(ThreadTerminationError) | python | def kill(self):
"""Terminates the current thread by raising an error."""
self._killed.set()
if not self.is_alive():
logging.debug('Cannot kill thread that is no longer running.')
return
if not self._is_thread_proc_running():
logging.debug("Thread's _thread_proc function is no longer running, "
'will not kill; letting thread exit gracefully.')
return
self.async_raise(ThreadTerminationError) | [
"def",
"kill",
"(",
"self",
")",
":",
"self",
".",
"_killed",
".",
"set",
"(",
")",
"if",
"not",
"self",
".",
"is_alive",
"(",
")",
":",
"logging",
".",
"debug",
"(",
"'Cannot kill thread that is no longer running.'",
")",
"return",
"if",
"not",
"self",
... | Terminates the current thread by raising an error. | [
"Terminates",
"the",
"current",
"thread",
"by",
"raising",
"an",
"error",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/threads.py#L193-L203 |
226,971 | google/openhtf | openhtf/util/threads.py | KillableThread.async_raise | def async_raise(self, exc_type):
"""Raise the exception."""
# Should only be called on a started thread, so raise otherwise.
assert self.ident is not None, 'Only started threads have thread identifier'
# If the thread has died we don't want to raise an exception so log.
if not self.is_alive():
_LOG.debug('Not raising %s because thread %s (%s) is not alive',
exc_type, self.name, self.ident)
return
result = ctypes.pythonapi.PyThreadState_SetAsyncExc(
ctypes.c_long(self.ident), ctypes.py_object(exc_type))
if result == 0 and self.is_alive():
# Don't raise an exception an error unnecessarily if the thread is dead.
raise ValueError('Thread ID was invalid.', self.ident)
elif result > 1:
# Something bad happened, call with a NULL exception to undo.
ctypes.pythonapi.PyThreadState_SetAsyncExc(self.ident, None)
raise RuntimeError('Error: PyThreadState_SetAsyncExc %s %s (%s) %s' % (
exc_type, self.name, self.ident, result)) | python | def async_raise(self, exc_type):
"""Raise the exception."""
# Should only be called on a started thread, so raise otherwise.
assert self.ident is not None, 'Only started threads have thread identifier'
# If the thread has died we don't want to raise an exception so log.
if not self.is_alive():
_LOG.debug('Not raising %s because thread %s (%s) is not alive',
exc_type, self.name, self.ident)
return
result = ctypes.pythonapi.PyThreadState_SetAsyncExc(
ctypes.c_long(self.ident), ctypes.py_object(exc_type))
if result == 0 and self.is_alive():
# Don't raise an exception an error unnecessarily if the thread is dead.
raise ValueError('Thread ID was invalid.', self.ident)
elif result > 1:
# Something bad happened, call with a NULL exception to undo.
ctypes.pythonapi.PyThreadState_SetAsyncExc(self.ident, None)
raise RuntimeError('Error: PyThreadState_SetAsyncExc %s %s (%s) %s' % (
exc_type, self.name, self.ident, result)) | [
"def",
"async_raise",
"(",
"self",
",",
"exc_type",
")",
":",
"# Should only be called on a started thread, so raise otherwise.",
"assert",
"self",
".",
"ident",
"is",
"not",
"None",
",",
"'Only started threads have thread identifier'",
"# If the thread has died we don't want to ... | Raise the exception. | [
"Raise",
"the",
"exception",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/threads.py#L205-L225 |
226,972 | google/openhtf | openhtf/util/logs.py | get_record_logger_for | def get_record_logger_for(test_uid):
"""Return the child logger associated with the specified test UID."""
htf_logger = logging.getLogger(RECORD_LOGGER_PREFIX)
record_logger = HtfTestLogger('.'.join(((RECORD_LOGGER_PREFIX, test_uid))))
record_logger.parent = htf_logger
return record_logger | python | def get_record_logger_for(test_uid):
"""Return the child logger associated with the specified test UID."""
htf_logger = logging.getLogger(RECORD_LOGGER_PREFIX)
record_logger = HtfTestLogger('.'.join(((RECORD_LOGGER_PREFIX, test_uid))))
record_logger.parent = htf_logger
return record_logger | [
"def",
"get_record_logger_for",
"(",
"test_uid",
")",
":",
"htf_logger",
"=",
"logging",
".",
"getLogger",
"(",
"RECORD_LOGGER_PREFIX",
")",
"record_logger",
"=",
"HtfTestLogger",
"(",
"'.'",
".",
"join",
"(",
"(",
"(",
"RECORD_LOGGER_PREFIX",
",",
"test_uid",
"... | Return the child logger associated with the specified test UID. | [
"Return",
"the",
"child",
"logger",
"associated",
"with",
"the",
"specified",
"test",
"UID",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/logs.py#L161-L166 |
226,973 | google/openhtf | openhtf/util/logs.py | initialize_record_handler | def initialize_record_handler(test_uid, test_record, notify_update):
"""Initialize the record handler for a test.
For each running test, we attach a record handler to the top-level OpenHTF
logger. The handler will append OpenHTF logs to the test record, while
filtering out logs that are specific to any other test run.
"""
htf_logger = logging.getLogger(LOGGER_PREFIX)
htf_logger.addHandler(RecordHandler(test_uid, test_record, notify_update)) | python | def initialize_record_handler(test_uid, test_record, notify_update):
"""Initialize the record handler for a test.
For each running test, we attach a record handler to the top-level OpenHTF
logger. The handler will append OpenHTF logs to the test record, while
filtering out logs that are specific to any other test run.
"""
htf_logger = logging.getLogger(LOGGER_PREFIX)
htf_logger.addHandler(RecordHandler(test_uid, test_record, notify_update)) | [
"def",
"initialize_record_handler",
"(",
"test_uid",
",",
"test_record",
",",
"notify_update",
")",
":",
"htf_logger",
"=",
"logging",
".",
"getLogger",
"(",
"LOGGER_PREFIX",
")",
"htf_logger",
".",
"addHandler",
"(",
"RecordHandler",
"(",
"test_uid",
",",
"test_r... | Initialize the record handler for a test.
For each running test, we attach a record handler to the top-level OpenHTF
logger. The handler will append OpenHTF logs to the test record, while
filtering out logs that are specific to any other test run. | [
"Initialize",
"the",
"record",
"handler",
"for",
"a",
"test",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/logs.py#L169-L177 |
226,974 | google/openhtf | openhtf/util/logs.py | log_once | def log_once(log_func, msg, *args, **kwargs):
""""Logs a message only once."""
if msg not in _LOG_ONCE_SEEN:
log_func(msg, *args, **kwargs)
# Key on the message, ignoring args. This should fit most use cases.
_LOG_ONCE_SEEN.add(msg) | python | def log_once(log_func, msg, *args, **kwargs):
""""Logs a message only once."""
if msg not in _LOG_ONCE_SEEN:
log_func(msg, *args, **kwargs)
# Key on the message, ignoring args. This should fit most use cases.
_LOG_ONCE_SEEN.add(msg) | [
"def",
"log_once",
"(",
"log_func",
",",
"msg",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"msg",
"not",
"in",
"_LOG_ONCE_SEEN",
":",
"log_func",
"(",
"msg",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"# Key on the message, ignorin... | Logs a message only once. | [
"Logs",
"a",
"message",
"only",
"once",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/logs.py#L188-L193 |
226,975 | google/openhtf | openhtf/util/logs.py | configure_logging | def configure_logging():
"""One-time initialization of loggers. See module docstring for more info."""
# Define the top-level logger.
htf_logger = logging.getLogger(LOGGER_PREFIX)
htf_logger.propagate = False
htf_logger.setLevel(logging.DEBUG)
# By default, don't print any logs to the CLI.
if CLI_LOGGING_VERBOSITY == 0:
htf_logger.addHandler(logging.NullHandler())
return
if CLI_LOGGING_VERBOSITY == 1:
logging_level = logging.INFO
else:
logging_level = logging.DEBUG
# Configure a handler to print to the CLI.
cli_handler = KillableThreadSafeStreamHandler(stream=sys.stdout)
cli_handler.setFormatter(CliFormatter())
cli_handler.setLevel(logging_level)
cli_handler.addFilter(MAC_FILTER)
htf_logger.addHandler(cli_handler)
# Suppress CLI logging if the --quiet flag is used, or while CLI_QUIET is set
# in the console_output module.
cli_handler.addFilter(console_output.CliQuietFilter()) | python | def configure_logging():
"""One-time initialization of loggers. See module docstring for more info."""
# Define the top-level logger.
htf_logger = logging.getLogger(LOGGER_PREFIX)
htf_logger.propagate = False
htf_logger.setLevel(logging.DEBUG)
# By default, don't print any logs to the CLI.
if CLI_LOGGING_VERBOSITY == 0:
htf_logger.addHandler(logging.NullHandler())
return
if CLI_LOGGING_VERBOSITY == 1:
logging_level = logging.INFO
else:
logging_level = logging.DEBUG
# Configure a handler to print to the CLI.
cli_handler = KillableThreadSafeStreamHandler(stream=sys.stdout)
cli_handler.setFormatter(CliFormatter())
cli_handler.setLevel(logging_level)
cli_handler.addFilter(MAC_FILTER)
htf_logger.addHandler(cli_handler)
# Suppress CLI logging if the --quiet flag is used, or while CLI_QUIET is set
# in the console_output module.
cli_handler.addFilter(console_output.CliQuietFilter()) | [
"def",
"configure_logging",
"(",
")",
":",
"# Define the top-level logger.",
"htf_logger",
"=",
"logging",
".",
"getLogger",
"(",
"LOGGER_PREFIX",
")",
"htf_logger",
".",
"propagate",
"=",
"False",
"htf_logger",
".",
"setLevel",
"(",
"logging",
".",
"DEBUG",
")",
... | One-time initialization of loggers. See module docstring for more info. | [
"One",
"-",
"time",
"initialization",
"of",
"loggers",
".",
"See",
"module",
"docstring",
"for",
"more",
"info",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/logs.py#L319-L346 |
226,976 | google/openhtf | openhtf/util/logs.py | RecordHandler.emit | def emit(self, record):
"""Save a logging.LogRecord to our test record.
Logs carry useful metadata such as the logger name and level information.
We capture this in a structured format in the test record to enable
filtering by client applications.
Args:
record: A logging.LogRecord to record.
"""
try:
message = self.format(record)
log_record = LogRecord(
record.levelno, record.name, os.path.basename(record.pathname),
record.lineno, int(record.created * 1000), message,
)
self._test_record.add_log_record(log_record)
self._notify_update()
except Exception: # pylint: disable=broad-except
self.handleError(record) | python | def emit(self, record):
"""Save a logging.LogRecord to our test record.
Logs carry useful metadata such as the logger name and level information.
We capture this in a structured format in the test record to enable
filtering by client applications.
Args:
record: A logging.LogRecord to record.
"""
try:
message = self.format(record)
log_record = LogRecord(
record.levelno, record.name, os.path.basename(record.pathname),
record.lineno, int(record.created * 1000), message,
)
self._test_record.add_log_record(log_record)
self._notify_update()
except Exception: # pylint: disable=broad-except
self.handleError(record) | [
"def",
"emit",
"(",
"self",
",",
"record",
")",
":",
"try",
":",
"message",
"=",
"self",
".",
"format",
"(",
"record",
")",
"log_record",
"=",
"LogRecord",
"(",
"record",
".",
"levelno",
",",
"record",
".",
"name",
",",
"os",
".",
"path",
".",
"bas... | Save a logging.LogRecord to our test record.
Logs carry useful metadata such as the logger name and level information.
We capture this in a structured format in the test record to enable
filtering by client applications.
Args:
record: A logging.LogRecord to record. | [
"Save",
"a",
"logging",
".",
"LogRecord",
"to",
"our",
"test",
"record",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/logs.py#L269-L288 |
226,977 | google/openhtf | openhtf/util/logs.py | CliFormatter.format | def format(self, record):
"""Format the record as tersely as possible but preserve info."""
super(CliFormatter, self).format(record)
localized_time = datetime.datetime.fromtimestamp(record.created)
terse_time = localized_time.strftime(u'%H:%M:%S')
terse_level = record.levelname[0]
terse_name = record.name.split('.')[-1]
match = RECORD_LOGGER_RE.match(record.name)
if match:
# Figure out which OpenHTF subsystem the record came from.
subsys_match = SUBSYSTEM_LOGGER_RE.match(record.name)
if subsys_match:
terse_name = '<{subsys}: {id}>'.format(
subsys=subsys_match.group('subsys'),
id=subsys_match.group('id'))
else:
# Fall back to using the last five characters of the test UUID.
terse_name = '<test %s>' % match.group('test_uid')[-5:]
return '{lvl} {time} {logger} - {msg}'.format(lvl=terse_level,
time=terse_time,
logger=terse_name,
msg=record.message) | python | def format(self, record):
"""Format the record as tersely as possible but preserve info."""
super(CliFormatter, self).format(record)
localized_time = datetime.datetime.fromtimestamp(record.created)
terse_time = localized_time.strftime(u'%H:%M:%S')
terse_level = record.levelname[0]
terse_name = record.name.split('.')[-1]
match = RECORD_LOGGER_RE.match(record.name)
if match:
# Figure out which OpenHTF subsystem the record came from.
subsys_match = SUBSYSTEM_LOGGER_RE.match(record.name)
if subsys_match:
terse_name = '<{subsys}: {id}>'.format(
subsys=subsys_match.group('subsys'),
id=subsys_match.group('id'))
else:
# Fall back to using the last five characters of the test UUID.
terse_name = '<test %s>' % match.group('test_uid')[-5:]
return '{lvl} {time} {logger} - {msg}'.format(lvl=terse_level,
time=terse_time,
logger=terse_name,
msg=record.message) | [
"def",
"format",
"(",
"self",
",",
"record",
")",
":",
"super",
"(",
"CliFormatter",
",",
"self",
")",
".",
"format",
"(",
"record",
")",
"localized_time",
"=",
"datetime",
".",
"datetime",
".",
"fromtimestamp",
"(",
"record",
".",
"created",
")",
"terse... | Format the record as tersely as possible but preserve info. | [
"Format",
"the",
"record",
"as",
"tersely",
"as",
"possible",
"but",
"preserve",
"info",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/logs.py#L294-L315 |
226,978 | google/openhtf | openhtf/output/callbacks/mfg_inspector.py | _send_mfg_inspector_request | def _send_mfg_inspector_request(envelope_data, credentials, destination_url):
"""Send upload http request. Intended to be run in retry loop."""
logging.info('Uploading result...')
http = httplib2.Http()
if credentials.access_token_expired:
credentials.refresh(http)
credentials.authorize(http)
resp, content = http.request(destination_url, 'POST', envelope_data)
try:
result = json.loads(content)
except Exception:
logging.debug('Upload failed with response %s: %s', resp, content)
raise UploadFailedError(resp, content)
if resp.status != 200:
logging.debug('Upload failed: %s', result)
raise UploadFailedError(result['error'], result)
return result | python | def _send_mfg_inspector_request(envelope_data, credentials, destination_url):
"""Send upload http request. Intended to be run in retry loop."""
logging.info('Uploading result...')
http = httplib2.Http()
if credentials.access_token_expired:
credentials.refresh(http)
credentials.authorize(http)
resp, content = http.request(destination_url, 'POST', envelope_data)
try:
result = json.loads(content)
except Exception:
logging.debug('Upload failed with response %s: %s', resp, content)
raise UploadFailedError(resp, content)
if resp.status != 200:
logging.debug('Upload failed: %s', result)
raise UploadFailedError(result['error'], result)
return result | [
"def",
"_send_mfg_inspector_request",
"(",
"envelope_data",
",",
"credentials",
",",
"destination_url",
")",
":",
"logging",
".",
"info",
"(",
"'Uploading result...'",
")",
"http",
"=",
"httplib2",
".",
"Http",
"(",
")",
"if",
"credentials",
".",
"access_token_exp... | Send upload http request. Intended to be run in retry loop. | [
"Send",
"upload",
"http",
"request",
".",
"Intended",
"to",
"be",
"run",
"in",
"retry",
"loop",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/callbacks/mfg_inspector.py#L27-L48 |
226,979 | google/openhtf | openhtf/output/callbacks/mfg_inspector.py | send_mfg_inspector_data | def send_mfg_inspector_data(inspector_proto, credentials, destination_url):
"""Upload MfgEvent to steam_engine."""
envelope = guzzle_pb2.TestRunEnvelope()
envelope.payload = zlib.compress(inspector_proto.SerializeToString())
envelope.payload_type = guzzle_pb2.COMPRESSED_MFG_EVENT
envelope_data = envelope.SerializeToString()
for _ in xrange(5):
try:
result = _send_mfg_inspector_request(
envelope_data, credentials, destination_url)
return result
except UploadFailedError:
time.sleep(1)
logging.critical(
'Could not upload to mfg-inspector after 5 attempts. Giving up.')
return {} | python | def send_mfg_inspector_data(inspector_proto, credentials, destination_url):
"""Upload MfgEvent to steam_engine."""
envelope = guzzle_pb2.TestRunEnvelope()
envelope.payload = zlib.compress(inspector_proto.SerializeToString())
envelope.payload_type = guzzle_pb2.COMPRESSED_MFG_EVENT
envelope_data = envelope.SerializeToString()
for _ in xrange(5):
try:
result = _send_mfg_inspector_request(
envelope_data, credentials, destination_url)
return result
except UploadFailedError:
time.sleep(1)
logging.critical(
'Could not upload to mfg-inspector after 5 attempts. Giving up.')
return {} | [
"def",
"send_mfg_inspector_data",
"(",
"inspector_proto",
",",
"credentials",
",",
"destination_url",
")",
":",
"envelope",
"=",
"guzzle_pb2",
".",
"TestRunEnvelope",
"(",
")",
"envelope",
".",
"payload",
"=",
"zlib",
".",
"compress",
"(",
"inspector_proto",
".",
... | Upload MfgEvent to steam_engine. | [
"Upload",
"MfgEvent",
"to",
"steam_engine",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/callbacks/mfg_inspector.py#L51-L69 |
226,980 | google/openhtf | openhtf/output/callbacks/mfg_inspector.py | MfgInspector._convert | def _convert(self, test_record_obj):
"""Convert and cache a test record to a mfg-inspector proto."""
if self._cached_proto is None:
self._cached_proto = self._converter(test_record_obj)
return self._cached_proto | python | def _convert(self, test_record_obj):
"""Convert and cache a test record to a mfg-inspector proto."""
if self._cached_proto is None:
self._cached_proto = self._converter(test_record_obj)
return self._cached_proto | [
"def",
"_convert",
"(",
"self",
",",
"test_record_obj",
")",
":",
"if",
"self",
".",
"_cached_proto",
"is",
"None",
":",
"self",
".",
"_cached_proto",
"=",
"self",
".",
"_converter",
"(",
"test_record_obj",
")",
"return",
"self",
".",
"_cached_proto"
] | Convert and cache a test record to a mfg-inspector proto. | [
"Convert",
"and",
"cache",
"a",
"test",
"record",
"to",
"a",
"mfg",
"-",
"inspector",
"proto",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/callbacks/mfg_inspector.py#L174-L180 |
226,981 | google/openhtf | openhtf/output/callbacks/mfg_inspector.py | MfgInspector.save_to_disk | def save_to_disk(self, filename_pattern=None):
"""Returns a callback to convert test record to proto and save to disk."""
if not self._converter:
raise RuntimeError(
'Must set _converter on subclass or via set_converter before calling '
'save_to_disk.')
pattern = filename_pattern or self._default_filename_pattern
if not pattern:
raise RuntimeError(
'Must specify provide a filename_pattern or set a '
'_default_filename_pattern on subclass.')
def save_to_disk_callback(test_record_obj):
proto = self._convert(test_record_obj)
output_to_file = callbacks.OutputToFile(pattern)
with output_to_file.open_output_file(test_record_obj) as outfile:
outfile.write(proto.SerializeToString())
return save_to_disk_callback | python | def save_to_disk(self, filename_pattern=None):
"""Returns a callback to convert test record to proto and save to disk."""
if not self._converter:
raise RuntimeError(
'Must set _converter on subclass or via set_converter before calling '
'save_to_disk.')
pattern = filename_pattern or self._default_filename_pattern
if not pattern:
raise RuntimeError(
'Must specify provide a filename_pattern or set a '
'_default_filename_pattern on subclass.')
def save_to_disk_callback(test_record_obj):
proto = self._convert(test_record_obj)
output_to_file = callbacks.OutputToFile(pattern)
with output_to_file.open_output_file(test_record_obj) as outfile:
outfile.write(proto.SerializeToString())
return save_to_disk_callback | [
"def",
"save_to_disk",
"(",
"self",
",",
"filename_pattern",
"=",
"None",
")",
":",
"if",
"not",
"self",
".",
"_converter",
":",
"raise",
"RuntimeError",
"(",
"'Must set _converter on subclass or via set_converter before calling '",
"'save_to_disk.'",
")",
"pattern",
"=... | Returns a callback to convert test record to proto and save to disk. | [
"Returns",
"a",
"callback",
"to",
"convert",
"test",
"record",
"to",
"proto",
"and",
"save",
"to",
"disk",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/callbacks/mfg_inspector.py#L182-L202 |
226,982 | google/openhtf | openhtf/output/callbacks/mfg_inspector.py | MfgInspector.upload | def upload(self):
"""Returns a callback to convert a test record to a proto and upload."""
if not self._converter:
raise RuntimeError(
'Must set _converter on subclass or via set_converter before calling '
'upload.')
if not self.credentials:
raise RuntimeError('Must provide credentials to use upload callback.')
def upload_callback(test_record_obj):
proto = self._convert(test_record_obj)
self.upload_result = send_mfg_inspector_data(
proto, self.credentials, self.destination_url)
return upload_callback | python | def upload(self):
"""Returns a callback to convert a test record to a proto and upload."""
if not self._converter:
raise RuntimeError(
'Must set _converter on subclass or via set_converter before calling '
'upload.')
if not self.credentials:
raise RuntimeError('Must provide credentials to use upload callback.')
def upload_callback(test_record_obj):
proto = self._convert(test_record_obj)
self.upload_result = send_mfg_inspector_data(
proto, self.credentials, self.destination_url)
return upload_callback | [
"def",
"upload",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"_converter",
":",
"raise",
"RuntimeError",
"(",
"'Must set _converter on subclass or via set_converter before calling '",
"'upload.'",
")",
"if",
"not",
"self",
".",
"credentials",
":",
"raise",
"Run... | Returns a callback to convert a test record to a proto and upload. | [
"Returns",
"a",
"callback",
"to",
"convert",
"a",
"test",
"record",
"to",
"a",
"proto",
"and",
"upload",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/output/callbacks/mfg_inspector.py#L204-L219 |
226,983 | google/openhtf | openhtf/plugs/usb/shell_service.py | AsyncCommandHandle._writer_thread_proc | def _writer_thread_proc(self, is_raw):
"""Write as long as the stream is not closed."""
# If we're not in raw mode, do line-buffered reads to play nicer with
# potential interactive uses, max of MAX_ADB_DATA, since anything we write
# to the stream will get packetized to that size anyway.
#
# Loop until our stream gets closed, which will cause one of these
# operations to raise. Since we're in a separate thread, it'll just get
# ignored, which is what we want.
reader = self.stdin.read if is_raw else self.stdin.readline
while not self.stream.is_closed():
self.stream.write(reader(adb_protocol.MAX_ADB_DATA)) | python | def _writer_thread_proc(self, is_raw):
"""Write as long as the stream is not closed."""
# If we're not in raw mode, do line-buffered reads to play nicer with
# potential interactive uses, max of MAX_ADB_DATA, since anything we write
# to the stream will get packetized to that size anyway.
#
# Loop until our stream gets closed, which will cause one of these
# operations to raise. Since we're in a separate thread, it'll just get
# ignored, which is what we want.
reader = self.stdin.read if is_raw else self.stdin.readline
while not self.stream.is_closed():
self.stream.write(reader(adb_protocol.MAX_ADB_DATA)) | [
"def",
"_writer_thread_proc",
"(",
"self",
",",
"is_raw",
")",
":",
"# If we're not in raw mode, do line-buffered reads to play nicer with",
"# potential interactive uses, max of MAX_ADB_DATA, since anything we write",
"# to the stream will get packetized to that size anyway.",
"#",
"# Loop ... | Write as long as the stream is not closed. | [
"Write",
"as",
"long",
"as",
"the",
"stream",
"is",
"not",
"closed",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/shell_service.py#L132-L143 |
226,984 | google/openhtf | openhtf/plugs/usb/shell_service.py | AsyncCommandHandle._reader_thread_proc | def _reader_thread_proc(self, timeout):
"""Read until the stream is closed."""
for data in self.stream.read_until_close(timeout_ms=timeout):
if self.stdout is not None:
self.stdout.write(data) | python | def _reader_thread_proc(self, timeout):
"""Read until the stream is closed."""
for data in self.stream.read_until_close(timeout_ms=timeout):
if self.stdout is not None:
self.stdout.write(data) | [
"def",
"_reader_thread_proc",
"(",
"self",
",",
"timeout",
")",
":",
"for",
"data",
"in",
"self",
".",
"stream",
".",
"read_until_close",
"(",
"timeout_ms",
"=",
"timeout",
")",
":",
"if",
"self",
".",
"stdout",
"is",
"not",
"None",
":",
"self",
".",
"... | Read until the stream is closed. | [
"Read",
"until",
"the",
"stream",
"is",
"closed",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/shell_service.py#L145-L149 |
226,985 | google/openhtf | openhtf/plugs/usb/shell_service.py | AsyncCommandHandle.wait | def wait(self, timeout_ms=None):
"""Block until this command has completed.
Args:
timeout_ms: Timeout, in milliseconds, to wait.
Returns:
Output of the command if it complete and self.stdout is a StringIO
object or was passed in as None. Returns True if the command completed but
stdout was provided (and was not a StringIO object). Returns None if the
timeout expired before the command completed. Be careful to check the
return value explicitly for None, as the output may be ''.
"""
closed = timeouts.loop_until_timeout_or_true(
timeouts.PolledTimeout.from_millis(timeout_ms),
self.stream.is_closed, .1)
if closed:
if hasattr(self.stdout, 'getvalue'):
return self.stdout.getvalue()
return True
return None | python | def wait(self, timeout_ms=None):
"""Block until this command has completed.
Args:
timeout_ms: Timeout, in milliseconds, to wait.
Returns:
Output of the command if it complete and self.stdout is a StringIO
object or was passed in as None. Returns True if the command completed but
stdout was provided (and was not a StringIO object). Returns None if the
timeout expired before the command completed. Be careful to check the
return value explicitly for None, as the output may be ''.
"""
closed = timeouts.loop_until_timeout_or_true(
timeouts.PolledTimeout.from_millis(timeout_ms),
self.stream.is_closed, .1)
if closed:
if hasattr(self.stdout, 'getvalue'):
return self.stdout.getvalue()
return True
return None | [
"def",
"wait",
"(",
"self",
",",
"timeout_ms",
"=",
"None",
")",
":",
"closed",
"=",
"timeouts",
".",
"loop_until_timeout_or_true",
"(",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"timeout_ms",
")",
",",
"self",
".",
"stream",
".",
"is_closed... | Block until this command has completed.
Args:
timeout_ms: Timeout, in milliseconds, to wait.
Returns:
Output of the command if it complete and self.stdout is a StringIO
object or was passed in as None. Returns True if the command completed but
stdout was provided (and was not a StringIO object). Returns None if the
timeout expired before the command completed. Be careful to check the
return value explicitly for None, as the output may be ''. | [
"Block",
"until",
"this",
"command",
"has",
"completed",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/shell_service.py#L169-L189 |
226,986 | google/openhtf | openhtf/plugs/usb/shell_service.py | ShellService.command | def command(self, command, raw=False, timeout_ms=None):
"""Run the given command and return the output."""
return ''.join(self.streaming_command(command, raw, timeout_ms)) | python | def command(self, command, raw=False, timeout_ms=None):
"""Run the given command and return the output."""
return ''.join(self.streaming_command(command, raw, timeout_ms)) | [
"def",
"command",
"(",
"self",
",",
"command",
",",
"raw",
"=",
"False",
",",
"timeout_ms",
"=",
"None",
")",
":",
"return",
"''",
".",
"join",
"(",
"self",
".",
"streaming_command",
"(",
"command",
",",
"raw",
",",
"timeout_ms",
")",
")"
] | Run the given command and return the output. | [
"Run",
"the",
"given",
"command",
"and",
"return",
"the",
"output",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/shell_service.py#L223-L225 |
226,987 | google/openhtf | openhtf/plugs/usb/shell_service.py | ShellService.streaming_command | def streaming_command(self, command, raw=False, timeout_ms=None):
"""Run the given command and yield the output as we receive it."""
if raw:
command = self._to_raw_command(command)
return self.adb_connection.streaming_command('shell', command, timeout_ms) | python | def streaming_command(self, command, raw=False, timeout_ms=None):
"""Run the given command and yield the output as we receive it."""
if raw:
command = self._to_raw_command(command)
return self.adb_connection.streaming_command('shell', command, timeout_ms) | [
"def",
"streaming_command",
"(",
"self",
",",
"command",
",",
"raw",
"=",
"False",
",",
"timeout_ms",
"=",
"None",
")",
":",
"if",
"raw",
":",
"command",
"=",
"self",
".",
"_to_raw_command",
"(",
"command",
")",
"return",
"self",
".",
"adb_connection",
"... | Run the given command and yield the output as we receive it. | [
"Run",
"the",
"given",
"command",
"and",
"yield",
"the",
"output",
"as",
"we",
"receive",
"it",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/shell_service.py#L227-L231 |
226,988 | google/openhtf | openhtf/plugs/usb/shell_service.py | ShellService.async_command | def async_command(self, command, stdin=None, stdout=None, raw=False,
timeout_ms=None):
"""Run the given command on the device asynchronously.
Input will be read from stdin, output written to stdout. ADB doesn't
distinguish between stdout and stdin on the device, so they get interleaved
into stdout here. stdin and stdout should be file-like objects, so you
could use sys.stdin and sys.stdout to emulate the 'adb shell' commandline.
Args:
command: The command to run, will be run with /bin/sh -c 'command' on
the device.
stdin: File-like object to read from to pipe to the command's stdin. Can
be None, in which case nothing will be written to the command's stdin.
stdout: File-like object to write the command's output to. Can be None,
in which case the command's output will be buffered internally, and can
be access via the return value of wait().
raw: If True, run the command as per RawCommand (see above).
timeout_ms: Timeout for the command, in milliseconds.
Returns:
An AsyncCommandHandle instance that can be used to send/receive data to
and from the command or wait on the command to finish.
Raises:
AdbStreamUnavailableError: If the remote devices doesn't support the
shell: service.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
if raw:
command = self._to_raw_command(command)
stream = self.adb_connection.open_stream('shell:%s' % command, timeout)
if not stream:
raise usb_exceptions.AdbStreamUnavailableError(
'%s does not support service: shell', self)
if raw and stdin is not None:
# Short delay to make sure the ioctl to set raw mode happens before we do
# any writes to the stream, if we don't do this bad things happen...
time.sleep(.1)
return AsyncCommandHandle(stream, stdin, stdout, timeout, raw) | python | def async_command(self, command, stdin=None, stdout=None, raw=False,
timeout_ms=None):
"""Run the given command on the device asynchronously.
Input will be read from stdin, output written to stdout. ADB doesn't
distinguish between stdout and stdin on the device, so they get interleaved
into stdout here. stdin and stdout should be file-like objects, so you
could use sys.stdin and sys.stdout to emulate the 'adb shell' commandline.
Args:
command: The command to run, will be run with /bin/sh -c 'command' on
the device.
stdin: File-like object to read from to pipe to the command's stdin. Can
be None, in which case nothing will be written to the command's stdin.
stdout: File-like object to write the command's output to. Can be None,
in which case the command's output will be buffered internally, and can
be access via the return value of wait().
raw: If True, run the command as per RawCommand (see above).
timeout_ms: Timeout for the command, in milliseconds.
Returns:
An AsyncCommandHandle instance that can be used to send/receive data to
and from the command or wait on the command to finish.
Raises:
AdbStreamUnavailableError: If the remote devices doesn't support the
shell: service.
"""
timeout = timeouts.PolledTimeout.from_millis(timeout_ms)
if raw:
command = self._to_raw_command(command)
stream = self.adb_connection.open_stream('shell:%s' % command, timeout)
if not stream:
raise usb_exceptions.AdbStreamUnavailableError(
'%s does not support service: shell', self)
if raw and stdin is not None:
# Short delay to make sure the ioctl to set raw mode happens before we do
# any writes to the stream, if we don't do this bad things happen...
time.sleep(.1)
return AsyncCommandHandle(stream, stdin, stdout, timeout, raw) | [
"def",
"async_command",
"(",
"self",
",",
"command",
",",
"stdin",
"=",
"None",
",",
"stdout",
"=",
"None",
",",
"raw",
"=",
"False",
",",
"timeout_ms",
"=",
"None",
")",
":",
"timeout",
"=",
"timeouts",
".",
"PolledTimeout",
".",
"from_millis",
"(",
"... | Run the given command on the device asynchronously.
Input will be read from stdin, output written to stdout. ADB doesn't
distinguish between stdout and stdin on the device, so they get interleaved
into stdout here. stdin and stdout should be file-like objects, so you
could use sys.stdin and sys.stdout to emulate the 'adb shell' commandline.
Args:
command: The command to run, will be run with /bin/sh -c 'command' on
the device.
stdin: File-like object to read from to pipe to the command's stdin. Can
be None, in which case nothing will be written to the command's stdin.
stdout: File-like object to write the command's output to. Can be None,
in which case the command's output will be buffered internally, and can
be access via the return value of wait().
raw: If True, run the command as per RawCommand (see above).
timeout_ms: Timeout for the command, in milliseconds.
Returns:
An AsyncCommandHandle instance that can be used to send/receive data to
and from the command or wait on the command to finish.
Raises:
AdbStreamUnavailableError: If the remote devices doesn't support the
shell: service. | [
"Run",
"the",
"given",
"command",
"on",
"the",
"device",
"asynchronously",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/plugs/usb/shell_service.py#L234-L273 |
226,989 | google/openhtf | openhtf/util/conf.py | Configuration.load_flag_values | def load_flag_values(self, flags=None):
"""Load flag values given from command line flags.
Args:
flags: An argparse Namespace containing the command line flags.
"""
if flags is None:
flags = self._flags
for keyval in flags.config_value:
k, v = keyval.split('=', 1)
v = self._modules['yaml'].load(v) if isinstance(v, str) else v
# Force any command line keys and values that are bytes to unicode.
k = k.decode() if isinstance(k, bytes) else k
v = v.decode() if isinstance(v, bytes) else v
self._flag_values.setdefault(k, v) | python | def load_flag_values(self, flags=None):
"""Load flag values given from command line flags.
Args:
flags: An argparse Namespace containing the command line flags.
"""
if flags is None:
flags = self._flags
for keyval in flags.config_value:
k, v = keyval.split('=', 1)
v = self._modules['yaml'].load(v) if isinstance(v, str) else v
# Force any command line keys and values that are bytes to unicode.
k = k.decode() if isinstance(k, bytes) else k
v = v.decode() if isinstance(v, bytes) else v
self._flag_values.setdefault(k, v) | [
"def",
"load_flag_values",
"(",
"self",
",",
"flags",
"=",
"None",
")",
":",
"if",
"flags",
"is",
"None",
":",
"flags",
"=",
"self",
".",
"_flags",
"for",
"keyval",
"in",
"flags",
".",
"config_value",
":",
"k",
",",
"v",
"=",
"keyval",
".",
"split",
... | Load flag values given from command line flags.
Args:
flags: An argparse Namespace containing the command line flags. | [
"Load",
"flag",
"values",
"given",
"from",
"command",
"line",
"flags",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L255-L271 |
226,990 | google/openhtf | openhtf/util/conf.py | Configuration.declare | def declare(self, name, description=None, **kwargs):
"""Declare a configuration key with the given name.
Args:
name: Configuration key to declare, must not have been already declared.
description: If provided, use this as the description for this key.
**kwargs: Other kwargs to pass to the Declaration, only default_value
is currently supported.
"""
if not self._is_valid_key(name):
raise self.InvalidKeyError(
'Invalid key name, must begin with a lowercase letter', name)
if name in self._declarations:
raise self.KeyAlreadyDeclaredError(
'Configuration key already declared', name)
self._declarations[name] = self.Declaration(
name, description=description, **kwargs) | python | def declare(self, name, description=None, **kwargs):
"""Declare a configuration key with the given name.
Args:
name: Configuration key to declare, must not have been already declared.
description: If provided, use this as the description for this key.
**kwargs: Other kwargs to pass to the Declaration, only default_value
is currently supported.
"""
if not self._is_valid_key(name):
raise self.InvalidKeyError(
'Invalid key name, must begin with a lowercase letter', name)
if name in self._declarations:
raise self.KeyAlreadyDeclaredError(
'Configuration key already declared', name)
self._declarations[name] = self.Declaration(
name, description=description, **kwargs) | [
"def",
"declare",
"(",
"self",
",",
"name",
",",
"description",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"self",
".",
"_is_valid_key",
"(",
"name",
")",
":",
"raise",
"self",
".",
"InvalidKeyError",
"(",
"'Invalid key name, must begin w... | Declare a configuration key with the given name.
Args:
name: Configuration key to declare, must not have been already declared.
description: If provided, use this as the description for this key.
**kwargs: Other kwargs to pass to the Declaration, only default_value
is currently supported. | [
"Declare",
"a",
"configuration",
"key",
"with",
"the",
"given",
"name",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L337-L353 |
226,991 | google/openhtf | openhtf/util/conf.py | Configuration.reset | def reset(self):
"""Reset the loaded state of the configuration to what it was at import.
Note that this does *not* reset values set by commandline flags or loaded
from --config-file (in fact, any values loaded from --config-file that have
been overridden are reset to their value from --config-file).
"""
# Populate loaded_values with values from --config-file, if it was given.
self._loaded_values = {}
if self._flags.config_file is not None:
self.load_from_file(self._flags.config_file, _allow_undeclared=True) | python | def reset(self):
"""Reset the loaded state of the configuration to what it was at import.
Note that this does *not* reset values set by commandline flags or loaded
from --config-file (in fact, any values loaded from --config-file that have
been overridden are reset to their value from --config-file).
"""
# Populate loaded_values with values from --config-file, if it was given.
self._loaded_values = {}
if self._flags.config_file is not None:
self.load_from_file(self._flags.config_file, _allow_undeclared=True) | [
"def",
"reset",
"(",
"self",
")",
":",
"# Populate loaded_values with values from --config-file, if it was given.",
"self",
".",
"_loaded_values",
"=",
"{",
"}",
"if",
"self",
".",
"_flags",
".",
"config_file",
"is",
"not",
"None",
":",
"self",
".",
"load_from_file"... | Reset the loaded state of the configuration to what it was at import.
Note that this does *not* reset values set by commandline flags or loaded
from --config-file (in fact, any values loaded from --config-file that have
been overridden are reset to their value from --config-file). | [
"Reset",
"the",
"loaded",
"state",
"of",
"the",
"configuration",
"to",
"what",
"it",
"was",
"at",
"import",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L356-L366 |
226,992 | google/openhtf | openhtf/util/conf.py | Configuration.load_from_file | def load_from_file(self, yamlfile, _override=True, _allow_undeclared=False):
"""Loads the configuration from a file.
Parsed contents must be a single dict mapping config key to value.
Args:
yamlfile: The opened file object to load configuration from.
See load_from_dict() for other args' descriptions.
Raises:
ConfigurationInvalidError: If configuration file can't be read, or can't
be parsed as either YAML (or JSON, which is a subset of YAML).
"""
self._logger.info('Loading configuration from file: %s', yamlfile)
try:
parsed_yaml = self._modules['yaml'].safe_load(yamlfile.read())
except self._modules['yaml'].YAMLError:
self._logger.exception('Problem parsing YAML')
raise self.ConfigurationInvalidError(
'Failed to load from %s as YAML' % yamlfile)
if not isinstance(parsed_yaml, dict):
# Parsed YAML, but it's not a dict.
raise self.ConfigurationInvalidError(
'YAML parsed, but wrong type, should be dict', parsed_yaml)
self._logger.debug('Configuration loaded from file: %s', parsed_yaml)
self.load_from_dict(
parsed_yaml, _override=_override, _allow_undeclared=_allow_undeclared) | python | def load_from_file(self, yamlfile, _override=True, _allow_undeclared=False):
"""Loads the configuration from a file.
Parsed contents must be a single dict mapping config key to value.
Args:
yamlfile: The opened file object to load configuration from.
See load_from_dict() for other args' descriptions.
Raises:
ConfigurationInvalidError: If configuration file can't be read, or can't
be parsed as either YAML (or JSON, which is a subset of YAML).
"""
self._logger.info('Loading configuration from file: %s', yamlfile)
try:
parsed_yaml = self._modules['yaml'].safe_load(yamlfile.read())
except self._modules['yaml'].YAMLError:
self._logger.exception('Problem parsing YAML')
raise self.ConfigurationInvalidError(
'Failed to load from %s as YAML' % yamlfile)
if not isinstance(parsed_yaml, dict):
# Parsed YAML, but it's not a dict.
raise self.ConfigurationInvalidError(
'YAML parsed, but wrong type, should be dict', parsed_yaml)
self._logger.debug('Configuration loaded from file: %s', parsed_yaml)
self.load_from_dict(
parsed_yaml, _override=_override, _allow_undeclared=_allow_undeclared) | [
"def",
"load_from_file",
"(",
"self",
",",
"yamlfile",
",",
"_override",
"=",
"True",
",",
"_allow_undeclared",
"=",
"False",
")",
":",
"self",
".",
"_logger",
".",
"info",
"(",
"'Loading configuration from file: %s'",
",",
"yamlfile",
")",
"try",
":",
"parsed... | Loads the configuration from a file.
Parsed contents must be a single dict mapping config key to value.
Args:
yamlfile: The opened file object to load configuration from.
See load_from_dict() for other args' descriptions.
Raises:
ConfigurationInvalidError: If configuration file can't be read, or can't
be parsed as either YAML (or JSON, which is a subset of YAML). | [
"Loads",
"the",
"configuration",
"from",
"a",
"file",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L368-L397 |
226,993 | google/openhtf | openhtf/util/conf.py | Configuration.load_from_dict | def load_from_dict(self, dictionary, _override=True, _allow_undeclared=False):
"""Loads the config with values from a dictionary instead of a file.
This is meant for testing and bin purposes and shouldn't be used in most
applications.
Args:
dictionary: The dictionary containing config keys/values to update.
_override: If True, new values will override previous values.
_allow_undeclared: If True, silently load undeclared keys, otherwise
warn and ignore the value. Typically used for loading config
files before declarations have been evaluated.
"""
undeclared_keys = []
for key, value in self._modules['six'].iteritems(dictionary):
# Warn in this case. We raise if you try to access a config key that
# hasn't been declared, but we don't raise here so that you can use
# configuration files that are supersets of required configuration for
# any particular test station.
if key not in self._declarations and not _allow_undeclared:
undeclared_keys.append(key)
continue
if key in self._loaded_values:
if _override:
self._logger.info(
'Overriding previously loaded value for %s (%s) with value: %s',
key, self._loaded_values[key], value)
else:
self._logger.info(
'Ignoring new value (%s), keeping previous value for %s: %s',
value, key, self._loaded_values[key])
continue
# Force any keys and values that are bytes to unicode.
key = key.decode() if isinstance(key, bytes) else key
value = value.decode() if isinstance(value, bytes) else value
self._loaded_values[key] = value
if undeclared_keys:
self._logger.warning('Ignoring undeclared configuration keys: %s',
undeclared_keys) | python | def load_from_dict(self, dictionary, _override=True, _allow_undeclared=False):
"""Loads the config with values from a dictionary instead of a file.
This is meant for testing and bin purposes and shouldn't be used in most
applications.
Args:
dictionary: The dictionary containing config keys/values to update.
_override: If True, new values will override previous values.
_allow_undeclared: If True, silently load undeclared keys, otherwise
warn and ignore the value. Typically used for loading config
files before declarations have been evaluated.
"""
undeclared_keys = []
for key, value in self._modules['six'].iteritems(dictionary):
# Warn in this case. We raise if you try to access a config key that
# hasn't been declared, but we don't raise here so that you can use
# configuration files that are supersets of required configuration for
# any particular test station.
if key not in self._declarations and not _allow_undeclared:
undeclared_keys.append(key)
continue
if key in self._loaded_values:
if _override:
self._logger.info(
'Overriding previously loaded value for %s (%s) with value: %s',
key, self._loaded_values[key], value)
else:
self._logger.info(
'Ignoring new value (%s), keeping previous value for %s: %s',
value, key, self._loaded_values[key])
continue
# Force any keys and values that are bytes to unicode.
key = key.decode() if isinstance(key, bytes) else key
value = value.decode() if isinstance(value, bytes) else value
self._loaded_values[key] = value
if undeclared_keys:
self._logger.warning('Ignoring undeclared configuration keys: %s',
undeclared_keys) | [
"def",
"load_from_dict",
"(",
"self",
",",
"dictionary",
",",
"_override",
"=",
"True",
",",
"_allow_undeclared",
"=",
"False",
")",
":",
"undeclared_keys",
"=",
"[",
"]",
"for",
"key",
",",
"value",
"in",
"self",
".",
"_modules",
"[",
"'six'",
"]",
".",... | Loads the config with values from a dictionary instead of a file.
This is meant for testing and bin purposes and shouldn't be used in most
applications.
Args:
dictionary: The dictionary containing config keys/values to update.
_override: If True, new values will override previous values.
_allow_undeclared: If True, silently load undeclared keys, otherwise
warn and ignore the value. Typically used for loading config
files before declarations have been evaluated. | [
"Loads",
"the",
"config",
"with",
"values",
"from",
"a",
"dictionary",
"instead",
"of",
"a",
"file",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L405-L445 |
226,994 | google/openhtf | openhtf/util/conf.py | Configuration._asdict | def _asdict(self):
"""Create a dictionary snapshot of the current config values."""
# Start with any default values we have, and override with loaded values,
# and then override with flag values.
retval = {key: self._declarations[key].default_value for
key in self._declarations if self._declarations[key].has_default}
retval.update(self._loaded_values)
# Only update keys that are declared so we don't allow injecting
# un-declared keys via commandline flags.
for key, value in self._modules['six'].iteritems(self._flag_values):
if key in self._declarations:
retval[key] = value
return retval | python | def _asdict(self):
"""Create a dictionary snapshot of the current config values."""
# Start with any default values we have, and override with loaded values,
# and then override with flag values.
retval = {key: self._declarations[key].default_value for
key in self._declarations if self._declarations[key].has_default}
retval.update(self._loaded_values)
# Only update keys that are declared so we don't allow injecting
# un-declared keys via commandline flags.
for key, value in self._modules['six'].iteritems(self._flag_values):
if key in self._declarations:
retval[key] = value
return retval | [
"def",
"_asdict",
"(",
"self",
")",
":",
"# Start with any default values we have, and override with loaded values,",
"# and then override with flag values.",
"retval",
"=",
"{",
"key",
":",
"self",
".",
"_declarations",
"[",
"key",
"]",
".",
"default_value",
"for",
"key"... | Create a dictionary snapshot of the current config values. | [
"Create",
"a",
"dictionary",
"snapshot",
"of",
"the",
"current",
"config",
"values",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L448-L460 |
226,995 | google/openhtf | openhtf/util/conf.py | Configuration.help_text | def help_text(self):
"""Return a string with all config keys and their descriptions."""
result = []
for name in sorted(self._declarations.keys()):
result.append(name)
result.append('-' * len(name))
decl = self._declarations[name]
if decl.description:
result.append(decl.description.strip())
else:
result.append('(no description found)')
if decl.has_default:
result.append('')
quotes = '"' if type(decl.default_value) is str else ''
result.append(' default_value={quotes}{val}{quotes}'.format(
quotes=quotes, val=decl.default_value))
result.append('')
result.append('')
return '\n'.join(result) | python | def help_text(self):
"""Return a string with all config keys and their descriptions."""
result = []
for name in sorted(self._declarations.keys()):
result.append(name)
result.append('-' * len(name))
decl = self._declarations[name]
if decl.description:
result.append(decl.description.strip())
else:
result.append('(no description found)')
if decl.has_default:
result.append('')
quotes = '"' if type(decl.default_value) is str else ''
result.append(' default_value={quotes}{val}{quotes}'.format(
quotes=quotes, val=decl.default_value))
result.append('')
result.append('')
return '\n'.join(result) | [
"def",
"help_text",
"(",
"self",
")",
":",
"result",
"=",
"[",
"]",
"for",
"name",
"in",
"sorted",
"(",
"self",
".",
"_declarations",
".",
"keys",
"(",
")",
")",
":",
"result",
".",
"append",
"(",
"name",
")",
"result",
".",
"append",
"(",
"'-'",
... | Return a string with all config keys and their descriptions. | [
"Return",
"a",
"string",
"with",
"all",
"config",
"keys",
"and",
"their",
"descriptions",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L463-L481 |
226,996 | google/openhtf | openhtf/util/conf.py | Configuration.save_and_restore | def save_and_restore(self, _func=None, **config_values):
"""Decorator for saving conf state and restoring it after a function.
This decorator is primarily for use in tests, where conf keys may be updated
for individual test cases, but those values need to be reverted after the
test case is done.
Examples:
conf.declare('my_conf_key')
@conf.save_and_restore
def MyTestFunc():
conf.load(my_conf_key='baz')
SomeFuncUnderTestThatUsesMyConfKey()
conf.load(my_conf_key='foo')
MyTestFunc()
print conf.my_conf_key # Prints 'foo', *NOT* 'baz'
# Without the save_and_restore decorator, MyTestFunc() would have had the
# side effect of altering the conf value of 'my_conf_key' to 'baz'.
# Config keys can also be initialized for the context inline at decoration
# time. This is the same as setting them at the beginning of the
# function, but is a little clearer syntax if you know ahead of time what
# config keys and values you need to set.
@conf.save_and_restore(my_conf_key='baz')
def MyOtherTestFunc():
print conf.my_conf_key # Prints 'baz'
MyOtherTestFunc()
print conf.my_conf_key # Prints 'foo' again, for the same reason.
Args:
_func: The function to wrap. The returned wrapper will invoke the
function and restore the config to the state it was in at invocation.
**config_values: Config keys can be set inline at decoration time, see
examples. Note that config keys can't begin with underscore, so
there can be no name collision with _func.
Returns:
Wrapper to replace _func, as per Python decorator semantics.
"""
functools = self._modules['functools'] # pylint: disable=redefined-outer-name
if not _func:
return functools.partial(self.save_and_restore, **config_values)
@functools.wraps(_func)
def _saving_wrapper(*args, **kwargs):
saved_config = dict(self._loaded_values)
try:
self.load_from_dict(config_values)
return _func(*args, **kwargs)
finally:
self._loaded_values = saved_config # pylint: disable=attribute-defined-outside-init
return _saving_wrapper | python | def save_and_restore(self, _func=None, **config_values):
"""Decorator for saving conf state and restoring it after a function.
This decorator is primarily for use in tests, where conf keys may be updated
for individual test cases, but those values need to be reverted after the
test case is done.
Examples:
conf.declare('my_conf_key')
@conf.save_and_restore
def MyTestFunc():
conf.load(my_conf_key='baz')
SomeFuncUnderTestThatUsesMyConfKey()
conf.load(my_conf_key='foo')
MyTestFunc()
print conf.my_conf_key # Prints 'foo', *NOT* 'baz'
# Without the save_and_restore decorator, MyTestFunc() would have had the
# side effect of altering the conf value of 'my_conf_key' to 'baz'.
# Config keys can also be initialized for the context inline at decoration
# time. This is the same as setting them at the beginning of the
# function, but is a little clearer syntax if you know ahead of time what
# config keys and values you need to set.
@conf.save_and_restore(my_conf_key='baz')
def MyOtherTestFunc():
print conf.my_conf_key # Prints 'baz'
MyOtherTestFunc()
print conf.my_conf_key # Prints 'foo' again, for the same reason.
Args:
_func: The function to wrap. The returned wrapper will invoke the
function and restore the config to the state it was in at invocation.
**config_values: Config keys can be set inline at decoration time, see
examples. Note that config keys can't begin with underscore, so
there can be no name collision with _func.
Returns:
Wrapper to replace _func, as per Python decorator semantics.
"""
functools = self._modules['functools'] # pylint: disable=redefined-outer-name
if not _func:
return functools.partial(self.save_and_restore, **config_values)
@functools.wraps(_func)
def _saving_wrapper(*args, **kwargs):
saved_config = dict(self._loaded_values)
try:
self.load_from_dict(config_values)
return _func(*args, **kwargs)
finally:
self._loaded_values = saved_config # pylint: disable=attribute-defined-outside-init
return _saving_wrapper | [
"def",
"save_and_restore",
"(",
"self",
",",
"_func",
"=",
"None",
",",
"*",
"*",
"config_values",
")",
":",
"functools",
"=",
"self",
".",
"_modules",
"[",
"'functools'",
"]",
"# pylint: disable=redefined-outer-name",
"if",
"not",
"_func",
":",
"return",
"fun... | Decorator for saving conf state and restoring it after a function.
This decorator is primarily for use in tests, where conf keys may be updated
for individual test cases, but those values need to be reverted after the
test case is done.
Examples:
conf.declare('my_conf_key')
@conf.save_and_restore
def MyTestFunc():
conf.load(my_conf_key='baz')
SomeFuncUnderTestThatUsesMyConfKey()
conf.load(my_conf_key='foo')
MyTestFunc()
print conf.my_conf_key # Prints 'foo', *NOT* 'baz'
# Without the save_and_restore decorator, MyTestFunc() would have had the
# side effect of altering the conf value of 'my_conf_key' to 'baz'.
# Config keys can also be initialized for the context inline at decoration
# time. This is the same as setting them at the beginning of the
# function, but is a little clearer syntax if you know ahead of time what
# config keys and values you need to set.
@conf.save_and_restore(my_conf_key='baz')
def MyOtherTestFunc():
print conf.my_conf_key # Prints 'baz'
MyOtherTestFunc()
print conf.my_conf_key # Prints 'foo' again, for the same reason.
Args:
_func: The function to wrap. The returned wrapper will invoke the
function and restore the config to the state it was in at invocation.
**config_values: Config keys can be set inline at decoration time, see
examples. Note that config keys can't begin with underscore, so
there can be no name collision with _func.
Returns:
Wrapper to replace _func, as per Python decorator semantics. | [
"Decorator",
"for",
"saving",
"conf",
"state",
"and",
"restoring",
"it",
"after",
"a",
"function",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L483-L542 |
226,997 | google/openhtf | openhtf/util/conf.py | Configuration.inject_positional_args | def inject_positional_args(self, method):
"""Decorator for injecting positional arguments from the configuration.
This decorator wraps the given method, so that any positional arguments are
passed with corresponding values from the configuration. The name of the
positional argument must match the configuration key.
Keyword arguments are *NEVER* modified, even if their names match
configuration keys. Avoid naming keyword args names that are also
configuration keys to avoid confusion.
Additional positional arguments may be used that do not appear in the
configuration, but those arguments *MUST* be specified as keyword arguments
upon invocation of the method. This is to avoid ambiguity in which
positional arguments are getting which values.
Args:
method: The method to wrap.
Returns:
A wrapper that, when invoked, will call the wrapped method, passing in
configuration values for positional arguments.
"""
inspect = self._modules['inspect']
argspec = inspect.getargspec(method)
# Index in argspec.args of the first keyword argument. This index is a
# negative number if there are any kwargs, or 0 if there are no kwargs.
keyword_arg_index = -1 * len(argspec.defaults or [])
arg_names = argspec.args[:keyword_arg_index or None]
kwarg_names = argspec.args[len(arg_names):]
functools = self._modules['functools'] # pylint: disable=redefined-outer-name
# Create the actual method wrapper, all we do is update kwargs. Note we
# don't pass any *args through because there can't be any - we've filled
# them all in with values from the configuration. Any positional args that
# are missing from the configuration *must* be explicitly specified as
# kwargs.
@functools.wraps(method)
def method_wrapper(**kwargs):
"""Wrapper that pulls values from openhtf.util.conf."""
# Check for keyword args with names that are in the config so we can warn.
for kwarg in kwarg_names:
if kwarg in self:
self._logger.warning('Keyword arg %s not set from configuration, but '
'is a configuration key', kwarg)
# Set positional args from configuration values.
final_kwargs = {name: self[name] for name in arg_names if name in self}
for overridden in set(kwargs) & set(final_kwargs):
self._logger.warning('Overriding configuration value for kwarg %s (%s) '
'with provided kwarg value: %s', overridden,
self[overridden], kwargs[overridden])
final_kwargs.update(kwargs)
if inspect.ismethod(method):
name = '%s.%s' % (method.__self__.__class__.__name__, method.__name__)
else:
name = method.__name__
self._logger.debug('Invoking %s with %s', name, final_kwargs)
return method(**final_kwargs)
# We have to check for a 'self' parameter explicitly because Python doesn't
# pass it as a keyword arg, it passes it as the first positional arg.
if argspec.args[0] == 'self':
@functools.wraps(method)
def self_wrapper(self, **kwargs): # pylint: disable=invalid-name
"""Wrapper that pulls values from openhtf.util.conf."""
kwargs['self'] = self
return method_wrapper(**kwargs)
return self_wrapper
return method_wrapper | python | def inject_positional_args(self, method):
"""Decorator for injecting positional arguments from the configuration.
This decorator wraps the given method, so that any positional arguments are
passed with corresponding values from the configuration. The name of the
positional argument must match the configuration key.
Keyword arguments are *NEVER* modified, even if their names match
configuration keys. Avoid naming keyword args names that are also
configuration keys to avoid confusion.
Additional positional arguments may be used that do not appear in the
configuration, but those arguments *MUST* be specified as keyword arguments
upon invocation of the method. This is to avoid ambiguity in which
positional arguments are getting which values.
Args:
method: The method to wrap.
Returns:
A wrapper that, when invoked, will call the wrapped method, passing in
configuration values for positional arguments.
"""
inspect = self._modules['inspect']
argspec = inspect.getargspec(method)
# Index in argspec.args of the first keyword argument. This index is a
# negative number if there are any kwargs, or 0 if there are no kwargs.
keyword_arg_index = -1 * len(argspec.defaults or [])
arg_names = argspec.args[:keyword_arg_index or None]
kwarg_names = argspec.args[len(arg_names):]
functools = self._modules['functools'] # pylint: disable=redefined-outer-name
# Create the actual method wrapper, all we do is update kwargs. Note we
# don't pass any *args through because there can't be any - we've filled
# them all in with values from the configuration. Any positional args that
# are missing from the configuration *must* be explicitly specified as
# kwargs.
@functools.wraps(method)
def method_wrapper(**kwargs):
"""Wrapper that pulls values from openhtf.util.conf."""
# Check for keyword args with names that are in the config so we can warn.
for kwarg in kwarg_names:
if kwarg in self:
self._logger.warning('Keyword arg %s not set from configuration, but '
'is a configuration key', kwarg)
# Set positional args from configuration values.
final_kwargs = {name: self[name] for name in arg_names if name in self}
for overridden in set(kwargs) & set(final_kwargs):
self._logger.warning('Overriding configuration value for kwarg %s (%s) '
'with provided kwarg value: %s', overridden,
self[overridden], kwargs[overridden])
final_kwargs.update(kwargs)
if inspect.ismethod(method):
name = '%s.%s' % (method.__self__.__class__.__name__, method.__name__)
else:
name = method.__name__
self._logger.debug('Invoking %s with %s', name, final_kwargs)
return method(**final_kwargs)
# We have to check for a 'self' parameter explicitly because Python doesn't
# pass it as a keyword arg, it passes it as the first positional arg.
if argspec.args[0] == 'self':
@functools.wraps(method)
def self_wrapper(self, **kwargs): # pylint: disable=invalid-name
"""Wrapper that pulls values from openhtf.util.conf."""
kwargs['self'] = self
return method_wrapper(**kwargs)
return self_wrapper
return method_wrapper | [
"def",
"inject_positional_args",
"(",
"self",
",",
"method",
")",
":",
"inspect",
"=",
"self",
".",
"_modules",
"[",
"'inspect'",
"]",
"argspec",
"=",
"inspect",
".",
"getargspec",
"(",
"method",
")",
"# Index in argspec.args of the first keyword argument. This index... | Decorator for injecting positional arguments from the configuration.
This decorator wraps the given method, so that any positional arguments are
passed with corresponding values from the configuration. The name of the
positional argument must match the configuration key.
Keyword arguments are *NEVER* modified, even if their names match
configuration keys. Avoid naming keyword args names that are also
configuration keys to avoid confusion.
Additional positional arguments may be used that do not appear in the
configuration, but those arguments *MUST* be specified as keyword arguments
upon invocation of the method. This is to avoid ambiguity in which
positional arguments are getting which values.
Args:
method: The method to wrap.
Returns:
A wrapper that, when invoked, will call the wrapped method, passing in
configuration values for positional arguments. | [
"Decorator",
"for",
"injecting",
"positional",
"arguments",
"from",
"the",
"configuration",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/conf.py#L544-L616 |
226,998 | google/openhtf | openhtf/util/functions.py | call_once | def call_once(func):
"""Decorate a function to only allow it to be called once.
Note that it doesn't make sense to only call a function once if it takes
arguments (use @functools.lru_cache for that sort of thing), so this only
works on callables that take no args.
"""
argspec = inspect.getargspec(func)
if argspec.args or argspec.varargs or argspec.keywords:
raise ValueError('Can only decorate functions with no args', func, argspec)
@functools.wraps(func)
def _wrapper():
# If we haven't been called yet, actually invoke func and save the result.
if not _wrapper.HasRun():
_wrapper.MarkAsRun()
_wrapper.return_value = func()
return _wrapper.return_value
_wrapper.has_run = False
_wrapper.HasRun = lambda: _wrapper.has_run
_wrapper.MarkAsRun = lambda: setattr(_wrapper, 'has_run', True)
return _wrapper | python | def call_once(func):
"""Decorate a function to only allow it to be called once.
Note that it doesn't make sense to only call a function once if it takes
arguments (use @functools.lru_cache for that sort of thing), so this only
works on callables that take no args.
"""
argspec = inspect.getargspec(func)
if argspec.args or argspec.varargs or argspec.keywords:
raise ValueError('Can only decorate functions with no args', func, argspec)
@functools.wraps(func)
def _wrapper():
# If we haven't been called yet, actually invoke func and save the result.
if not _wrapper.HasRun():
_wrapper.MarkAsRun()
_wrapper.return_value = func()
return _wrapper.return_value
_wrapper.has_run = False
_wrapper.HasRun = lambda: _wrapper.has_run
_wrapper.MarkAsRun = lambda: setattr(_wrapper, 'has_run', True)
return _wrapper | [
"def",
"call_once",
"(",
"func",
")",
":",
"argspec",
"=",
"inspect",
".",
"getargspec",
"(",
"func",
")",
"if",
"argspec",
".",
"args",
"or",
"argspec",
".",
"varargs",
"or",
"argspec",
".",
"keywords",
":",
"raise",
"ValueError",
"(",
"'Can only decorate... | Decorate a function to only allow it to be called once.
Note that it doesn't make sense to only call a function once if it takes
arguments (use @functools.lru_cache for that sort of thing), so this only
works on callables that take no args. | [
"Decorate",
"a",
"function",
"to",
"only",
"allow",
"it",
"to",
"be",
"called",
"once",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/functions.py#L23-L45 |
226,999 | google/openhtf | openhtf/util/functions.py | call_at_most_every | def call_at_most_every(seconds, count=1):
"""Call the decorated function at most count times every seconds seconds.
The decorated function will sleep to ensure that at most count invocations
occur within any 'seconds' second window.
"""
def decorator(func):
try:
call_history = getattr(func, '_call_history')
except AttributeError:
call_history = collections.deque(maxlen=count)
setattr(func, '_call_history', call_history)
@functools.wraps(func)
def _wrapper(*args, **kwargs):
current_time = time.time()
window_count = sum(ts > current_time - seconds for ts in call_history)
if window_count >= count:
# We need to sleep until the relevant call is outside the window. This
# should only ever be the the first entry in call_history, but if we
# somehow ended up with extra calls in the window, this recovers.
time.sleep(call_history[window_count - count] - current_time + seconds)
# Append this call, deque will automatically trim old calls using maxlen.
call_history.append(time.time())
return func(*args, **kwargs)
return _wrapper
return decorator | python | def call_at_most_every(seconds, count=1):
"""Call the decorated function at most count times every seconds seconds.
The decorated function will sleep to ensure that at most count invocations
occur within any 'seconds' second window.
"""
def decorator(func):
try:
call_history = getattr(func, '_call_history')
except AttributeError:
call_history = collections.deque(maxlen=count)
setattr(func, '_call_history', call_history)
@functools.wraps(func)
def _wrapper(*args, **kwargs):
current_time = time.time()
window_count = sum(ts > current_time - seconds for ts in call_history)
if window_count >= count:
# We need to sleep until the relevant call is outside the window. This
# should only ever be the the first entry in call_history, but if we
# somehow ended up with extra calls in the window, this recovers.
time.sleep(call_history[window_count - count] - current_time + seconds)
# Append this call, deque will automatically trim old calls using maxlen.
call_history.append(time.time())
return func(*args, **kwargs)
return _wrapper
return decorator | [
"def",
"call_at_most_every",
"(",
"seconds",
",",
"count",
"=",
"1",
")",
":",
"def",
"decorator",
"(",
"func",
")",
":",
"try",
":",
"call_history",
"=",
"getattr",
"(",
"func",
",",
"'_call_history'",
")",
"except",
"AttributeError",
":",
"call_history",
... | Call the decorated function at most count times every seconds seconds.
The decorated function will sleep to ensure that at most count invocations
occur within any 'seconds' second window. | [
"Call",
"the",
"decorated",
"function",
"at",
"most",
"count",
"times",
"every",
"seconds",
"seconds",
"."
] | 655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09 | https://github.com/google/openhtf/blob/655e85df7134db7bdf8f8fdd6ff9a6bf932e7b09/openhtf/util/functions.py#L47-L73 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.