Skip to content

Commit

Permalink
rename default to torch
Browse files Browse the repository at this point in the history
  • Loading branch information
FindHao committed Oct 11, 2024
1 parent 4398557 commit 8b48eea
Show file tree
Hide file tree
Showing 3 changed files with 9 additions and 9 deletions.
8 changes: 4 additions & 4 deletions run.py
Original file line number Diff line number Diff line change
Expand Up @@ -477,15 +477,15 @@ def main() -> None:
)
parser.add_argument(
"--metrics-gpu-backend",
choices=["dcgm", "default", "nvml"],
default="default",
choices=["torch", "nvml", "dcgm"],
default="torch",
help="""
Specify the backend [dcgm, default, nvml] to collect metrics. In all modes,
Specify the backend [torch, nvml, dcgm] to collect metrics. In all modes,
the latency (execution time) is always collected using `time.time_ns()`. The CPU
and GPU peak memory usage metrics are optional. The CPU peak memory usage is
collected by `psutil.Process()` in all modes. In nvml mode, the GPU peak memory
usage is collected by the `nvml` library. In dcgm mode, the GPU peak memory usage is
collected by the `dcgm` library. In default mode, the GPU peak memory usage is collected
collected by the `dcgm` library. In torch mode, the GPU peak memory usage is collected
by `torch.cuda.max_memory_allocated()`.
""",
)
Expand Down
2 changes: 1 addition & 1 deletion torchbenchmark/util/experiment/metrics.py
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ def get_peak_memory(
num_iter=MEMPROF_ITER,
export_metrics_file="",
metrics_needed=[],
metrics_gpu_backend="default",
metrics_gpu_backend="torch",
cpu_monitored_pid=None,
) -> Tuple[Optional[float], Optional[str], Optional[float]]:
"Run one step of the model, and return the peak memory in MB."
Expand Down
8 changes: 4 additions & 4 deletions userbenchmark/triton/run.py
Original file line number Diff line number Diff line change
Expand Up @@ -94,13 +94,13 @@ def get_parser(args=None):
)
parser.add_argument(
"--metrics-gpu-backend",
choices=["default", "nvml"],
default="default",
choices=["torch", "nvml"],
default="torch",
help=(
"Specify the backend [default, nvml] to collect metrics. In all modes, the latency "
"Specify the backend [torch, nvml] to collect metrics. In all modes, the latency "
"(execution time) is always collected using `time.time_ns()`. The CPU peak memory "
"usage is collected by `psutil.Process()`. In nvml mode, the GPU peak memory usage "
"is collected by the `nvml` library. In default mode, the GPU peak memory usage is "
"is collected by the `nvml` library. In torch mode, the GPU peak memory usage is "
"collected by `torch.cuda.max_memory_allocated()`."
),
)
Expand Down

0 comments on commit 8b48eea

Please sign in to comment.