Skip to content

Commit

Permalink
Set is_distributed false by default in vllm (#266)
Browse files Browse the repository at this point in the history
  • Loading branch information
asesorov committed Sep 20, 2024
1 parent 1992de3 commit 99a90c9
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion optimum_benchmark/trackers/latency.py
Original file line number Diff line number Diff line change
Expand Up @@ -121,7 +121,8 @@ def __init__(self, device: str, backend: str):
self.device = device
self.backend = backend
self.is_asynchronous = self.backend == "pytorch" and self.device == "cuda"
self.is_distributed = is_torch_distributed_available() and torch.distributed.is_initialized()
self.is_distributed = (self.backend != "vllm" and
is_torch_distributed_available() and torch.distributed.is_initialized())

if self.is_asynchronous:
LOGGER.info("\t+ Tracking latency using Pytorch CUDA events")
Expand Down

0 comments on commit 99a90c9

Please sign in to comment.