nvidia-tririon-inference-server-model-analyzer

環境

Ubuntu 22.04
trition inference server version : 23.08

在git clone下來的model_analyzer資料夾下啟動docker

1
2
3
4
5
docker run -it --gpus all \
-v /var/run/docker.sock:/var/run/docker.sock \
-v $(pwd)/examples/quick-start:$(pwd)/examples/quick-start \
-v $(pwd)/examples/quick-start/output-model-repo:$(pwd)/examples/quick-start/output-model-repo \
--net=host nvcr.io/nvidia/tritonserver:23.08-py3-sdk

假設現在所在路徑$(pwd)是/home/user/model_inference

啟動docker後會發現quick-start所在的的路徑會跟host主機一模一樣,因此下面的指令可以直接用跟host一模一樣的路徑去執行

1
2
3
4
5
model-analyzer profile \
--model-repository /home/user/model_inference/model_analyzer/examples/quick-start \
--profile-models add_sub --triton-launch-mode=docker \
--output-model-repository-path /home/user/model_inference/model_analyzer/examples/quick-start/output-model-repo/add \
--export-path profile_results

參考:

https://github.com/triton-inference-server/model_analyzer/blob/4b45d2daeb9f574d13ae0e774677c87c04ef2124/docs/quick_start.md

nvidia-tririon-inference-server-model-analyzer

https://jenhaoyang.github.io/uncategorized/f7bb94e239b3/

Author

Steven

Posted on

2025-07-11

Updated on

2025-07-11

Licensed under

Comments