Skip to content

Commit 1a341b7

Browse files
committed
delete deepseek modeling
Signed-off-by: whx-sjtu <2952154980@qq.com>
1 parent 31dda3f commit 1a341b7

File tree

7 files changed

+43
-618
lines changed

7 files changed

+43
-618
lines changed

tests/ut/models/conftest.py

Lines changed: 1 addition & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -90,25 +90,11 @@ def mock_distributed():
9090
mock_vllm_config.scheduler_config = Mock(max_num_seqs=256)
9191
mock_vllm_config.model_config = Mock(max_model_len=2048, quant_config=None)
9292

93-
with patch("vllm_ascend.models.deepseek_v2.get_tensor_model_parallel_rank", return_value=0), \
94-
patch("vllm_ascend.models.deepseek_v2.get_tensor_model_parallel_world_size", return_value=1), \
95-
patch("vllm_ascend.models.deepseek_v2.get_tp_group", return_value=tp_group), \
96-
patch("vllm_ascend.models.deepseek_v2.get_pp_group", return_value=pp_group), \
97-
patch("vllm_ascend.models.deepseek_v2.get_pp_group",
98-
return_value=Mock(is_first_rank=False, is_last_rank=False)), \
99-
patch("vllm_ascend.ops.fused_moe.get_current_vllm_config", return_value=mock_vllm_config), \
93+
with patch("vllm_ascend.ops.fused_moe.get_current_vllm_config", return_value=mock_vllm_config), \
10094
patch("vllm_ascend.ops.moe.token_dispatcher.torch.distributed.get_rank", return_value=0), \
10195
patch("vllm_ascend.ops.moe.token_dispatcher.get_ascend_soc_version", return_value=None), \
10296
patch.dict("vllm.distributed.parallel_state.__dict__", _TP=tp_group, _EP=ep_group, _DP=dp_group,
10397
_PP=pp_group), \
10498
patch.dict("vllm_ascend.distributed.parallel_state.__dict__", _MC2=ep_group), \
10599
patch("torch.npu.current_device", return_value=0):
106100
yield
107-
108-
109-
@pytest.fixture
110-
def mock_forward_context():
111-
forward_context = Mock(in_profile_run=False, with_prefill=False)
112-
with patch("vllm_ascend.models.deepseek_v2.get_forward_context",
113-
return_value=forward_context):
114-
yield

tests/ut/models/test_deepseek_mtp.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,8 +37,6 @@ def setup_mtp_layer(self, mocker: MockerFixture, vllm_config: VllmConfig,
3737
mocker.patch(
3838
"vllm_ascend.ops.vocab_parallel_embedding.AscendVocabParallelEmbedding.__init__",
3939
return_value=None)
40-
mocker.patch("vllm_ascend.models.deepseek_v2.get_ascend_config",
41-
return_value=mocker.Mock())
4240

4341
mtp_layer = CustomDeepSeekMultiTokenPredictorLayer(config, "0", None)
4442
mocker_deepseek_v2_decode_layer.assert_called_once()

tests/ut/models/test_deepseek_v2.py

Lines changed: 0 additions & 107 deletions
This file was deleted.

tests/ut/ops/test_llm_head.py

Lines changed: 42 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,42 @@
1+
#
2+
# Licensed under the Apache License, Version 2.0 (the "License");
3+
# you may not use this file except in compliance with the License.
4+
# You may obtain a copy of the License at
5+
#
6+
# http://www.apache.org/licenses/LICENSE-2.0
7+
#
8+
# Unless required by applicable law or agreed to in writing, software
9+
# distributed under the License is distributed on an "AS IS" BASIS,
10+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
11+
# See the License for the specific language governing permissions and
12+
# limitations under the License.
13+
# This file is a part of the vllm-ascend project.
14+
#
15+
from unittest.mock import patch
16+
17+
import torch
18+
from vllm.model_executor.layers.logits_processor import LogitsProcessor
19+
from vllm.model_executor.layers.vocab_parallel_embedding import ParallelLMHead
20+
21+
22+
def test_deepseek_v2_lmhead():
23+
class SimpleConfig:
24+
25+
def __init__(self):
26+
self.vocab_size = 10000
27+
self.hidden_size = 128
28+
29+
config = SimpleConfig()
30+
31+
lmhead = ParallelLMHead(config.vocab_size, config.hidden_size)
32+
logits_processor = LogitsProcessor(config.vocab_size)
33+
34+
mock_output = torch.randn(2, 4, config.hidden_size)
35+
mock_logits = torch.randn(2, 4, config.vocab_size)
36+
37+
with patch.object(lmhead.quant_method, "apply", return_value=mock_logits):
38+
with patch.object(logits_processor,
39+
"_gather_logits",
40+
return_value=mock_logits):
41+
logits = logits_processor(lmhead, mock_output)
42+
assert logits.shape == (2, 4, config.vocab_size)

vllm_ascend/models/__init__.py

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -29,14 +29,6 @@ def register_model():
2929
"vllm_ascend.models.qwen2_5_vl_without_padding:AscendQwen2_5_VLForConditionalGeneration_Without_Padding"
3030
)
3131

32-
ModelRegistry.register_model(
33-
"DeepseekV2ForCausalLM",
34-
"vllm_ascend.models.deepseek_v2:CustomDeepseekV2ForCausalLM")
35-
36-
ModelRegistry.register_model(
37-
"DeepseekV3ForCausalLM",
38-
"vllm_ascend.models.deepseek_v2:CustomDeepseekV3ForCausalLM")
39-
4032
ModelRegistry.register_model(
4133
"DeepSeekMTPModel",
4234
"vllm_ascend.models.deepseek_mtp:CustomDeepSeekMTP")

0 commit comments

Comments
 (0)