| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from __future__ import annotations |
| |
|
| | import csv |
| | import datetime |
| | import multiprocessing |
| | import os |
| | import sys |
| | import threading |
| | from collections import defaultdict, namedtuple |
| | from contextlib import contextmanager |
| | from functools import wraps |
| | from inspect import getframeinfo, stack |
| | from queue import Empty |
| | from time import perf_counter, perf_counter_ns |
| | from typing import TYPE_CHECKING, Any, cast |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from monai.utils import optional_import |
| |
|
| | if TYPE_CHECKING: |
| | from ignite.engine import Events |
| | else: |
| | Events = optional_import("ignite.engine", name="Events") |
| |
|
| | pd, has_pandas = optional_import("pandas") |
| |
|
| | __all__ = [ |
| | "torch_profiler_full", |
| | "torch_profiler_time_cpu_gpu", |
| | "torch_profiler_time_end_to_end", |
| | "PerfContext", |
| | "WorkflowProfiler", |
| | "ProfileHandler", |
| | "select_transform_call", |
| | ] |
| |
|
| |
|
| | def torch_profiler_full(func): |
| | """ |
| | A decorator which will run the torch profiler for the decorated function, |
| | printing the results in full. |
| | Note: Enforces a gpu sync point which could slow down pipelines. |
| | """ |
| |
|
| | @wraps(func) |
| | def wrapper(*args, **kwargs): |
| | with torch.autograd.profiler.profile(use_cuda=True) as prof: |
| | result = func(*args, **kwargs) |
| |
|
| | print(prof, flush=True) |
| |
|
| | return result |
| |
|
| | return wrapper |
| |
|
| |
|
| | def torch_profiler_time_cpu_gpu(func): |
| | """ |
| | A decorator which measures the execution time of both the CPU and GPU components |
| | of the decorated function, printing both results. |
| | Note: Enforces a gpu sync point which could slow down pipelines. |
| | """ |
| |
|
| | @wraps(func) |
| | def wrapper(*args, **kwargs): |
| | with torch.autograd.profiler.profile(use_cuda=True) as prof: |
| | result = func(*args, **kwargs) |
| |
|
| | cpu_time = prof.self_cpu_time_total |
| | gpu_time = sum(evt.self_cuda_time_total for evt in prof.function_events) |
| |
|
| | cpu_time = torch.autograd.profiler.format_time(cpu_time) |
| | gpu_time = torch.autograd.profiler.format_time(gpu_time) |
| |
|
| | print(f"cpu time: {cpu_time}, gpu time: {gpu_time}", flush=True) |
| |
|
| | return result |
| |
|
| | return wrapper |
| |
|
| |
|
| | def torch_profiler_time_end_to_end(func): |
| | """ |
| | A decorator which measures the total execution time from when the decorated |
| | function is called to when the last cuda operation finishes, printing the result. |
| | Note: Enforces a gpu sync point which could slow down pipelines. |
| | """ |
| |
|
| | @wraps(func) |
| | def wrapper(*args, **kwargs): |
| | torch.cuda.synchronize() |
| | start = perf_counter() |
| |
|
| | result = func(*args, **kwargs) |
| |
|
| | torch.cuda.synchronize() |
| | end = perf_counter() |
| |
|
| | total_time = (end - start) * 1e6 |
| | total_time_str = torch.autograd.profiler.format_time(total_time) |
| | print(f"End-to-end time: {total_time_str}", flush=True) |
| |
|
| | return result |
| |
|
| | return wrapper |
| |
|
| |
|
| | class PerfContext: |
| | """ |
| | Context manager for tracking how much time is spent within context blocks. This uses `time.perf_counter` to |
| | accumulate the total amount of time in seconds in the attribute `total_time` over however many context blocks |
| | the object is used in. |
| | """ |
| |
|
| | def __init__(self): |
| | self.total_time: float = 0 |
| | self.start_time: float | None = None |
| |
|
| | def __enter__(self): |
| | self.start_time = perf_counter() |
| | return self |
| |
|
| | def __exit__(self, exc_type, exc_value, exc_traceback): |
| | if self.start_time is not None: |
| | self.total_time += perf_counter() - self.start_time |
| | self.start_time = None |
| |
|
| |
|
| | |
| | ProfileResult = namedtuple("ProfileResult", ["name", "time", "filename", "lineno", "pid", "timestamp"]) |
| |
|
| |
|
| | def select_transform_call(frame): |
| | """Returns True if `frame` is a call to a `Transform` object's `_call__` method.""" |
| | from monai.transforms import Transform |
| |
|
| | self_obj = frame.f_locals.get("self", None) |
| | return frame.f_code.co_name == "__call__" and isinstance(self_obj, Transform) |
| |
|
| |
|
| | class WorkflowProfiler: |
| | """ |
| | Profiler for timing all aspects of a workflow. This includes using stack tracing to capture call times for |
| | all selected calls (by default calls to `Transform.__call__` methods), times within context blocks, times |
| | to generate items from iterables, and times to execute decorated functions. |
| | |
| | This profiler must be used only within its context because it uses an internal thread to read results from a |
| | multiprocessing queue. This allows the profiler to function across multiple threads and processes, though the |
| | multiprocess tracing is at times unreliable and not available in Windows at all. |
| | |
| | The profiler uses `sys.settrace` and `threading.settrace` to find all calls to profile, this will be set when |
| | the context enters and cleared when it exits so proper use of the context is essential to prevent excessive |
| | tracing. Note that tracing has a high overhead so times will not accurately reflect real world performance |
| | but give an idea of relative share of time spent. |
| | |
| | The tracing functionality uses a selector to choose which calls to trace, since tracing all calls induces |
| | infinite loops and would be terribly slow even if not. This selector is a callable accepting a `call` trace |
| | frame and returns True if the call should be traced. The default is `select_transform_call` which will return |
| | True for `Transform.__call__` calls only. |
| | |
| | Example showing use of all profiling functions: |
| | |
| | .. code-block:: python |
| | |
| | import monai.transform as mt |
| | from monai.utils import WorkflowProfiler |
| | import torch |
| | |
| | comp=mt.Compose([mt.ScaleIntensity(),mt.RandAxisFlip(0.5)]) |
| | |
| | with WorkflowProfiler() as wp: |
| | for _ in wp.profile_iter("range",range(5)): |
| | with wp.profile_ctx("Loop"): |
| | for i in range(10): |
| | comp(torch.rand(1,16,16)) |
| | |
| | @wp.profile_callable() |
| | def foo(): pass |
| | |
| | foo() |
| | foo() |
| | |
| | print(wp.get_times_summary_pd()) # print results |
| | |
| | Args: |
| | call_selector: selector to determine which calls to trace, use None to disable tracing |
| | """ |
| |
|
| | def __init__(self, call_selector=select_transform_call): |
| | self.results = defaultdict(list) |
| | self.parent_pid = os.getpid() |
| | self.read_thread: threading.Thread | None = None |
| | self.lock = threading.RLock() |
| | self.queue: multiprocessing.SimpleQueue = multiprocessing.SimpleQueue() |
| | self.queue_timeout = 0.1 |
| | self.call_selector = call_selector |
| |
|
| | def _is_parent(self): |
| | """Return True if this is the parent process.""" |
| | return os.getpid() == self.parent_pid |
| |
|
| | def _is_thread_active(self): |
| | """Return True if the read thread should be still active.""" |
| | return self.read_thread is not None or not self.queue.empty() |
| |
|
| | def _read_thread_func(self): |
| | """Read results from the queue and add to self.results in a thread stared by `__enter__`.""" |
| | while self._is_parent() and self._is_thread_active(): |
| | try: |
| | result = self.queue.get() |
| |
|
| | if result is None: |
| | break |
| |
|
| | self.add_result(result) |
| | except Empty: |
| | pass |
| |
|
| | if not (not self._is_parent() or self.queue.empty()): |
| | raise AssertionError |
| |
|
| | def _put_result(self, name, timedelta, filename, lineno): |
| | """Add a ProfileResult object to the queue.""" |
| | ts = str(datetime.datetime.now()) |
| | self.queue.put(ProfileResult(name, timedelta, filename, lineno, os.getpid(), ts)) |
| |
|
| | def _trace_call(self, frame, why, arg): |
| | """ |
| | Trace calls, when a call is encountered that is accepted by self.call_selector, create a new function to |
| | trace that call and measure the time from the call to a "return" frame. |
| | """ |
| | if why == "call": |
| | if self.call_selector(frame): |
| | calling_frame = frame |
| | start = perf_counter_ns() |
| |
|
| | def _call_profiler(frame, why, arg): |
| | """Defines a new inner trace function just for this call.""" |
| | if why == "return": |
| | diff = perf_counter_ns() - start |
| | f_code = calling_frame.f_code |
| | self_obj = calling_frame.f_locals.get("self", None) |
| | name = f_code.co_name |
| | if self_obj is not None: |
| | name = f"{type(self_obj).__name__}.{name}" |
| |
|
| | self._put_result(name, diff, f_code.co_filename, f_code.co_firstlineno) |
| |
|
| | |
| | |
| | |
| | |
| | return _call_profiler |
| | else: |
| | return self._trace_call |
| |
|
| | def __enter__(self): |
| | """Enter the context, creating the read thread and setting up tracing if needed.""" |
| | self.read_thread = threading.Thread(target=self._read_thread_func) |
| | self.read_thread.start() |
| |
|
| | if self.call_selector is not None: |
| | threading.settrace(self._trace_call) |
| | sys.settrace(self._trace_call) |
| |
|
| | return self |
| |
|
| | def __exit__(self, exc_type, exc_value, traceback): |
| | """Terminate the read thread cleanly and reset tracing if needed.""" |
| | if not self._is_parent(): |
| | raise AssertionError |
| |
|
| | self.queue.put(None) |
| |
|
| | read_thread = cast(threading.Thread, self.read_thread) |
| | self.read_thread = None |
| |
|
| | read_thread.join() |
| |
|
| | if self.call_selector is not None: |
| | threading.settrace(None) |
| | sys.settrace(None) |
| |
|
| | def add_result(self, result: ProfileResult) -> None: |
| | """Add a result in a thread-safe manner to the internal results dictionary.""" |
| | with self.lock: |
| | self.results[result.name].append(result) |
| |
|
| | def get_results(self): |
| | """Get a fresh results dictionary containing fresh tuples of ProfileResult objects.""" |
| | if not self._is_parent(): |
| | raise RuntimeError("Only parent process can collect results") |
| |
|
| | with self.lock: |
| | return {k: tuple(v) for k, v in self.results.items()} |
| |
|
| | @contextmanager |
| | def profile_ctx(self, name, caller=None): |
| | """Creates a context to profile, placing a timing result onto the queue when it exits.""" |
| | if caller is None: |
| | caller = getframeinfo(stack()[2][0]) |
| |
|
| | start = perf_counter_ns() |
| | try: |
| | yield |
| | finally: |
| | diff = perf_counter_ns() - start |
| | self._put_result(name, diff, caller.filename, caller.lineno) |
| |
|
| | def profile_callable(self, name=None): |
| | """ |
| | Decorator which can be applied to a function which profiles any calls to it. All calls to decorated |
| | callables must be done within the context of the profiler. |
| | """ |
| |
|
| | def _outer(func): |
| | _name = func.__name__ if name is None else name |
| | return self.profile_ctx(_name)(func) |
| |
|
| | return _outer |
| |
|
| | def profile_iter(self, name, iterable): |
| | """Wrapper around anything iterable to profile how long it takes to generate items.""" |
| |
|
| | class _Iterable: |
| |
|
| | def __iter__(_self): |
| | do_iter = True |
| | orig_iter = iter(iterable) |
| | caller = getframeinfo(stack()[1][0]) |
| |
|
| | while do_iter: |
| | try: |
| | start = perf_counter_ns() |
| | item = next(orig_iter) |
| | diff = perf_counter_ns() - start |
| | |
| | self._put_result(name, diff, caller.filename, caller.lineno) |
| | yield item |
| | except StopIteration: |
| | do_iter = False |
| |
|
| | return _Iterable() |
| |
|
| | def get_times_summary(self, times_in_s=True): |
| | """ |
| | Returns a dictionary mapping results entries to tuples containing the number of items, time sum, time average, |
| | time std dev, time min, and time max. |
| | """ |
| | result = {} |
| | for k, v in self.get_results().items(): |
| | timemult = 1e-9 if times_in_s else 1.0 |
| | all_times = [res.time * timemult for res in v] |
| |
|
| | timesum = sum(all_times) |
| | timeavg = timesum / len(all_times) |
| | timestd = np.std(all_times) |
| | timemin = min(all_times) |
| | timemax = max(all_times) |
| |
|
| | result[k] = (len(v), timesum, timeavg, timestd, timemin, timemax) |
| |
|
| | return result |
| |
|
| | def get_times_summary_pd(self, times_in_s=True): |
| | """Returns the same information as `get_times_summary` but in a Pandas DataFrame.""" |
| | import pandas as pd |
| |
|
| | summ = self.get_times_summary(times_in_s) |
| | suffix = "s" if times_in_s else "ns" |
| | columns = ["Count", f"Total Time ({suffix})", "Avg", "Std", "Min", "Max"] |
| |
|
| | df = pd.DataFrame.from_dict(summ, orient="index", columns=columns) |
| | df = df.sort_values(columns[1], ascending=False) |
| | return df |
| |
|
| | def dump_csv(self, stream=sys.stdout): |
| | """Save all results to a csv file.""" |
| | all_results = list(self.get_results().values()) |
| | writer = csv.DictWriter(stream, fieldnames=all_results[0][0]._asdict().keys()) |
| | writer.writeheader() |
| |
|
| | for rlist in all_results: |
| | for r in rlist: |
| | writer.writerow(r._asdict()) |
| |
|
| |
|
| | class ProfileHandler: |
| | """ |
| | Handler for Ignite Engine classes which measures the time from a start event ton an end event. This can be used to |
| | profile epoch, iteration, and other events as defined in `ignite.engine.Events`. This class should be used only |
| | within the context of a profiler object. |
| | |
| | Args: |
| | name: name of event to profile |
| | profiler: instance of WorkflowProfiler used by the handler, should be within the context of this object |
| | start_event: item in `ignite.engine.Events` stating event at which to start timing |
| | end_event: item in `ignite.engine.Events` stating event at which to stop timing |
| | """ |
| |
|
| | def __init__(self, name: str, profiler: WorkflowProfiler, start_event: Events, end_event: Events): |
| | self.name = name |
| | self.profiler = profiler |
| | self.start_event = start_event |
| | self.end_event = end_event |
| | self.ctx: Any = None |
| |
|
| | def attach(self, engine): |
| | engine.add_event_handler(self.start_event, self.start) |
| | engine.add_event_handler(self.end_event, self.end) |
| | return self |
| |
|
| | def start(self, engine): |
| | self.ctx = self.profiler.profile_ctx(self.name) |
| | self.ctx.__enter__() |
| |
|
| | def end(self, engine): |
| | self.ctx.__exit__(None, None, None) |
| | self.ctx = None |
| |
|