Skip to content

Commit eccfb71

Browse files
[CI] Fix UT (#2452)
Make UT CI happy - vLLM version: v0.10.0 - vLLM main: vllm-project/vllm@d983769 --------- Signed-off-by: wangxiyuan <wangxiyuan1007@gmail.com> Signed-off-by: MengqingCao <cmq0113@163.com> Co-authored-by: MengqingCao <cmq0113@163.com>
1 parent 3f867ee commit eccfb71

File tree

5 files changed

+30
-27
lines changed

5 files changed

+30
-27
lines changed

tests/ut/core/test_scheduler.py

Lines changed: 15 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@
88
SchedulerConfig, SpeculativeConfig, VllmConfig)
99
from vllm.multimodal.inputs import PlaceholderRange
1010
from vllm.sampling_params import SamplingParams
11+
from vllm.v1.core.kv_cache_utils import (get_request_block_hasher,
12+
init_none_hash)
1113
from vllm.v1.core.sched.output import SchedulerOutput
1214
from vllm.v1.kv_cache_interface import (FullAttentionSpec, KVCacheConfig,
1315
KVCacheGroupSpec)
@@ -36,7 +38,10 @@ def create_requests(
3638
mm_positions: Optional[list[PlaceholderRange]] = None,
3739
max_tokens: int = 16,
3840
stop_token_ids: Optional[list[int]] = None,
41+
block_size: int = 3,
42+
hash_fn=hash,
3943
):
44+
init_none_hash(hash_fn)
4045
prompt_logprobs = PROMPT_LOGPROBS
4146
sampling_params = SamplingParams(ignore_eos=False,
4247
max_tokens=max_tokens,
@@ -46,16 +51,16 @@ def create_requests(
4651
for i in range(num_requests):
4752
mm_position = None
4853
mm_inputs = None
49-
request = Request(
50-
request_id=f"{i}",
51-
prompt_token_ids=[i] * num_tokens,
52-
sampling_params=sampling_params,
53-
multi_modal_kwargs=mm_inputs,
54-
multi_modal_placeholders=mm_position,
55-
multi_modal_hashes=None,
56-
eos_token_id=EOS_TOKEN_ID,
57-
pooling_params=None,
58-
)
54+
request = Request(request_id=f"{i}",
55+
prompt_token_ids=[i] * num_tokens,
56+
sampling_params=sampling_params,
57+
multi_modal_kwargs=mm_inputs,
58+
multi_modal_placeholders=mm_position,
59+
multi_modal_hashes=None,
60+
eos_token_id=EOS_TOKEN_ID,
61+
pooling_params=None,
62+
block_hasher=get_request_block_hasher(
63+
block_size, hash_fn))
5964
requests.append(request)
6065
return requests
6166

tests/ut/kv_connector/test_mooncake_connector.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1152,8 +1152,6 @@ def setUp(self):
11521152
MagicMock()),
11531153
patch.dict('sys.modules',
11541154
{'vllm_ascend.envs': self.envs_ascend_mock}),
1155-
patch('vllm_ascend.distributed.mooncake_connector.envs_ascend',
1156-
self.envs_ascend_mock),
11571155
]
11581156

11591157
for p in self.patches:

tests/ut/kv_connector/utils.py

Lines changed: 4 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,6 @@ def assert_scheduler_empty(scheduler: Scheduler):
5555

5656

5757
def create_vllm_config(
58-
model: str = "facebook/opt-125m",
5958
max_num_seqs: int = 16,
6059
max_num_batched_tokens: int = 1024,
6160
block_size: int = 128,
@@ -66,14 +65,11 @@ def create_vllm_config(
6665
max_num_batched_tokens=max_num_batched_tokens,
6766
max_model_len=max_num_batched_tokens,
6867
)
68+
fake_weight_path = os.path.join(os.path.dirname(__file__), "..",
69+
"fake_weight")
6970
model_config = ModelConfig(
70-
model=model,
71-
task="auto",
72-
tokenizer=model,
73-
tokenizer_mode="auto",
74-
trust_remote_code=True,
75-
dtype="float16",
76-
seed=42,
71+
model=fake_weight_path,
72+
skip_tokenizer_init=True,
7773
)
7874
# Cache config, optionally force APC
7975
cache_config = CacheConfig(

tests/ut/torchair/test_utils.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ def test_register_torchair_model(self, mock_model_registry):
5151
mock_model_registry.return_value = mock_registry
5252
utils.register_torchair_model()
5353

54-
self.assertEqual(mock_model_registry.register_model.call_count, 3)
54+
self.assertEqual(mock_model_registry.register_model.call_count, 5)
5555
call_args_list = mock_model_registry.register_model.call_args_list
5656

5757
expected_registrations = [
@@ -63,7 +63,11 @@ def test_register_torchair_model(self, mock_model_registry):
6363
),
6464
("DeepseekV3ForCausalLM",
6565
"vllm_ascend.torchair.models.torchair_deepseek_v3:TorchairDeepseekV3ForCausalLM"
66-
)
66+
),
67+
("Qwen2ForCausalLM",
68+
"vllm_ascend.torchair.models.qwen2:CustomQwen2ForCausalLM"),
69+
("Qwen3ForCausalLM",
70+
"vllm_ascend.torchair.models.qwen3_moe:CustomQwen3MoeForCausalLM")
6771
]
6872

6973
for i, (expected_name,

tests/ut/worker/test_input_batch.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
import pytest
2121
import torch
2222
from vllm.sampling_params import SamplingParams
23-
from vllm.utils import is_pin_memory_available, make_tensor_with_pad
23+
from vllm.utils import make_tensor_with_pad
2424
from vllm.v1.pool.metadata import PoolingMetadata
2525
from vllm.v1.sample.logits_processor import LogitsProcessors
2626
from vllm.v1.sample.metadata import SamplingMetadata
@@ -237,7 +237,7 @@ def test_sampling_metadata_in_input_batch(device: str, batch_size: int):
237237
max_model_len=1024,
238238
max_num_batched_tokens=1024,
239239
device=torch.device(device),
240-
pin_memory=is_pin_memory_available(),
240+
pin_memory=False,
241241
vocab_size=1024,
242242
block_sizes=[1],
243243
)
@@ -298,7 +298,7 @@ def same(t1: Optional[torch.Tensor], t2: Optional[torch.Tensor]) -> bool:
298298
assert (expected_sampling_metadata.output_token_ids ==
299299
sampling_metadata.output_token_ids)
300300
assert expected_sampling_metadata.no_penalties == \
301-
sampling_metadata.no_penalties
301+
sampling_metadata.no_penalties
302302
if sampling_metadata.allowed_token_ids_mask:
303303
assert torch.allclose(
304304
expected_sampling_metadata.allowed_token_ids_mask,
@@ -328,7 +328,7 @@ def test_swap_states_in_input_batch(device: str, batch_size: int,
328328
max_model_len=1024,
329329
max_num_batched_tokens=1024,
330330
device=torch.device(device),
331-
pin_memory=is_pin_memory_available(),
331+
pin_memory=False,
332332
vocab_size=1024,
333333
block_sizes=[1],
334334
)
@@ -337,7 +337,7 @@ def test_swap_states_in_input_batch(device: str, batch_size: int,
337337
max_model_len=1024,
338338
max_num_batched_tokens=1024,
339339
device=torch.device(device),
340-
pin_memory=is_pin_memory_available(),
340+
pin_memory=False,
341341
vocab_size=1024,
342342
block_sizes=[1],
343343
)

0 commit comments

Comments
 (0)