File tree Expand file tree Collapse file tree
tests/v1/kv_connector/unit
vllm/distributed/kv_transfer/kv_connector/v1 Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -128,7 +128,7 @@ def test_read_blocks_for_req_expands_remote_ids(
128128 from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .metadata import (
129129 NixlConnectorMetadata ,
130130 )
131- from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .transfer_plan import (
131+ from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .tp_mapping import (
132132 TPMapping ,
133133 )
134134 from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .worker import (
Original file line number Diff line number Diff line change 1111
1212import pytest
1313
14- from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .transfer_plan import (
14+ from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .tp_mapping import (
1515 TPMapping ,
1616 compute_tp_mapping ,
1717)
Original file line number Diff line number Diff line change @@ -821,7 +821,6 @@ def __init__(
821821 self .cache_config = vllm_config .cache_config
822822 self .kv_cache_config = kv_cache_config
823823 self .use_mla = self .model_config .use_mla
824- self ._physical_blocks_per_logical_kv_block = 1
825824 self ._sync_block_size_with_kernel ()
826825
827826 # Get the attention backend from the first layer
@@ -864,9 +863,6 @@ def _sync_block_size_with_kernel(self) -> None:
864863 kernel_block_size ,
865864 )
866865 assert self .block_size > kernel_block_size
867- self ._physical_blocks_per_logical_kv_block = (
868- self .block_size // kernel_block_size
869- )
870866 self .block_size = kernel_block_size
871867
872868 def __del__ (self ):
File renamed without changes.
Original file line number Diff line number Diff line change 4444from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .stats import (
4545 NixlKVConnectorStats ,
4646)
47- from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .transfer_plan import (
47+ from vllm .distributed .kv_transfer .kv_connector .v1 .nixl .tp_mapping import (
4848 ReadSpec ,
4949 TPMapping ,
5050 _is_attention_spec ,
You can’t perform that action at this time.
0 commit comments