File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change 3737 is_block_tables_empty )
3838from vllm .attention .layer import Attention
3939from vllm .attention .ops .paged_attn import PagedAttention
40- from vllm .config import VllmConfig , get_current_vllm_config
40+ from vllm .config import VllmConfig
4141from vllm .logger import init_logger
4242from vllm .utils import (async_tensor_h2d , get_kv_cache_torch_dtype ,
4343 make_tensor_with_pad )
@@ -187,7 +187,7 @@ def __init__(self, runner):
187187 # Global hyperparameters shared by all attention layers
188188 self .global_hyperparameters : Optional [PerLayerParameters ] = None
189189
190- self .vllm_config = get_current_vllm_config ()
190+ self .vllm_config = self . runner . vllm_config
191191
192192 def _get_workspace_buffer (self ):
193193 if self ._workspace_buffer is None :
@@ -613,7 +613,7 @@ def __init__(self, input_builder: "ModelInputForGPUBuilder"):
613613 # Global hyperparameters shared by all attention layers
614614 self .global_hyperparameters : Optional [PerLayerParameters ] = None
615615
616- self .vllm_config = get_current_vllm_config ()
616+ self .vllm_config = self . runner . vllm_config
617617
618618 def prepare (self ):
619619 self .slot_mapping : List [int ] = []
You can’t perform that action at this time.
0 commit comments