Skip to content

Commit

Permalink
Update vllm/v1/worker/gpu_model_runner.py
Browse files Browse the repository at this point in the history
Co-authored-by: Cody Yu <[email protected]>
Signed-off-by: Chen Zhang <[email protected]>
  • Loading branch information
heheda12345 and comaniac committed Jan 14, 2025
1 parent f79dff2 commit 97176da
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion vllm/v1/worker/gpu_model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -858,7 +858,7 @@ def capture_model(self) -> None:

def initialize_kv_cache(self, kv_cache_config: KVCacheConfig) -> None:
if len(kv_cache_config.groups) > 1:
raise NotImplementedError("Multiple groups are not supported yet.")
raise NotImplementedError("Hybrid kv-cache groups are not supported yet.")

Check failure on line 861 in vllm/v1/worker/gpu_model_runner.py

View workflow job for this annotation

GitHub Actions / ruff (3.12)

Ruff (E501)

vllm/v1/worker/gpu_model_runner.py:861:81: E501 Line too long (86 > 80)

kv_caches: Dict[str, torch.Tensor] = {}

Expand Down

0 comments on commit 97176da

Please sign in to comment.