Skip to content

Commit f961d7f

Browse files
authored
[BugFix] Pass in correct VLLM config in FlashInfer backend (vllm-project#13207) (vllm-project#16973)
Signed-off-by: 苏政渊 <suzhengyuan@moonshot.cn> Co-authored-by: 苏政渊 <suzhengyuan@moonshot.cn>
1 parent d059110 commit f961d7f

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

vllm/attention/backends/flashinfer.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@
3737
is_block_tables_empty)
3838
from vllm.attention.layer import Attention
3939
from vllm.attention.ops.paged_attn import PagedAttention
40-
from vllm.config import VllmConfig, get_current_vllm_config
40+
from vllm.config import VllmConfig
4141
from vllm.logger import init_logger
4242
from vllm.utils import (async_tensor_h2d, get_kv_cache_torch_dtype,
4343
make_tensor_with_pad)
@@ -187,7 +187,7 @@ def __init__(self, runner):
187187
# Global hyperparameters shared by all attention layers
188188
self.global_hyperparameters: Optional[PerLayerParameters] = None
189189

190-
self.vllm_config = get_current_vllm_config()
190+
self.vllm_config = self.runner.vllm_config
191191

192192
def _get_workspace_buffer(self):
193193
if self._workspace_buffer is None:
@@ -613,7 +613,7 @@ def __init__(self, input_builder: "ModelInputForGPUBuilder"):
613613
# Global hyperparameters shared by all attention layers
614614
self.global_hyperparameters: Optional[PerLayerParameters] = None
615615

616-
self.vllm_config = get_current_vllm_config()
616+
self.vllm_config = self.runner.vllm_config
617617

618618
def prepare(self):
619619
self.slot_mapping: List[int] = []

0 commit comments

Comments
 (0)