From 6351e36d37f4bc6aba5b9a4e7bac79b79cc14838 Mon Sep 17 00:00:00 2001 From: Ilyas Moutawwakil <57442720+IlyasMoutawwakil@users.noreply.github.com> Date: Mon, 15 Jul 2024 17:15:44 +0200 Subject: [PATCH] Update readme (#228) --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 083b5221..780b4fbb 100644 --- a/README.md +++ b/README.md @@ -14,7 +14,8 @@ Optimum-Benchmark is a unified [multi-backend & multi-device](#backends--devices *News* 📰 - 🥳 PyPI package is now available for installation: `pip install optimum-benchmark` 🎉 [check it out](https://pypi.org/project/optimum-benchmark/) ! -- numactl support for Process and Torchrun launchers to control the NUMA nodes on which the benchmark runs 🧠 +- Model loading latency/memory/energy tracking for all backends in the inference scenario 🚀 +- numactl support for Process and Torchrun launchers to control the NUMA nodes on which the benchmark runs. - 4 minimal docker images (`cpu`, `cuda`, `rocm`, `cuda-ort`) in [packages](https://github.com/huggingface/optimum-benchmark/pkgs/container/optimum-benchmark) for testing, benchmarking and reproducibility 🐳 - vLLM backend for benchmarking [vLLM](https://github.com/vllm-project/vllm)'s inference engine 🚀 - Hosting the codebase of the [LLM-Perf Leaderboard](https://huggingface.co/spaces/optimum/llm-perf-leaderboard) 🥇