Skip to content

Commit 44f56fd

Browse files
comaniacrasmith
authored andcommitted
[Bugfix] Remove block size constraint (vllm-project#11723)
1 parent 4e70d7c commit 44f56fd

File tree

1 file changed

+0
-5
lines changed

1 file changed

+0
-5
lines changed

vllm/config.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1015,11 +1015,6 @@ def _verify_args(self) -> None:
10151015
raise ValueError(
10161016
"GPU memory utilization must be less than 1.0. Got "
10171017
f"{self.gpu_memory_utilization}.")
1018-
from vllm.platforms import current_platform
1019-
if (current_platform.is_cuda() and self.block_size is not None
1020-
and self.block_size > 32):
1021-
raise ValueError("CUDA Paged Attention kernel only supports "
1022-
f"block sizes up to 32. Got {self.block_size}.")
10231018

10241019
def _verify_cache_dtype(self) -> None:
10251020
if self.cache_dtype == "auto":

0 commit comments

Comments
 (0)