@@ -88,36 +88,16 @@ def setUp(self):
88
88
self .mock_self .cos_sin_cache = self .cos_sin_cache
89
89
self .mock_self .is_neox_style = self .is_neox_style
90
90
91
- @patch ('vllm_ascend.ops.rotary_embedding.get_ascend_config' )
92
- def test_rope_forward_oot_torchair_enabled_base (self ,
93
- mock_get_ascend_config ):
94
- # Setup mock for torchair enabled
95
- mock_config = MagicMock ()
96
- mock_config .torchair_graph_config .enabled = True
97
- mock_get_ascend_config .return_value = mock_config
98
- with patch .object (self .layer ,
99
- "forward_native" ,
100
- return_value = (self .query ,
101
- self .key )) as mock_forward_native :
102
- result_q , result_k = self .layer .forward (self .positions , self .query ,
103
- self .key )
104
-
105
- mock_forward_native .assert_called_once ()
106
- self .assertTrue (torch .equal (result_q , self .query ))
107
- self .assertTrue (torch .equal (result_k , self .key ))
108
-
109
91
@patch ('torch.ops._C' )
110
- @patch ('vllm_ascend.ops.rotary_embedding.get_ascend_config' )
111
92
@patch ('vllm_ascend.ops.rotary_embedding.is_310p' , return_value = False )
112
93
@patch ('vllm_ascend.ops.rotary_embedding.custom_rotary_embedding_enabled' ,
113
94
return_value = True )
114
95
@patch ('torch.ops._npu_rotary_embedding' )
115
96
def test_rope_forward_oot_custom_kernel (self , mock_rotary_embedding ,
116
97
mock_custom_enabled , mock_is_310p ,
117
- mock_get_ascend_config , mock__c ):
98
+ mock__c ):
118
99
mock_config = MagicMock ()
119
100
mock_config .torchair_graph_config .enabled = False
120
- mock_get_ascend_config .return_value = mock_config
121
101
122
102
# Setup mock for custom kernel path
123
103
@@ -130,16 +110,13 @@ def test_rope_forward_oot_custom_kernel(self, mock_rotary_embedding,
130
110
self .assertEqual (result_q .shape , self .query .shape )
131
111
self .assertEqual (result_k .shape , self .key .shape )
132
112
133
- @patch ('vllm_ascend.ops.rotary_embedding.get_ascend_config' )
134
113
@patch ('vllm_ascend.ops.rotary_embedding.custom_rotary_embedding_enabled' ,
135
114
return_value = False )
136
115
@patch ('torch_npu._npu_rotary_embedding' )
137
116
def test_rope_forward_oot_contiguous (self , mock_npu_rotary ,
138
- mock_custom_enabled ,
139
- mock_get_ascend_config ):
117
+ mock_custom_enabled ):
140
118
mock_config = MagicMock ()
141
119
mock_config .torchair_graph_config .enabled = False
142
- mock_get_ascend_config .return_value = mock_config
143
120
144
121
# Test contiguous path when custom is disabled
145
122
non_contig_query = self .query .transpose (0 , 1 )
@@ -153,27 +130,22 @@ def test_rope_forward_oot_contiguous(self, mock_npu_rotary,
153
130
self .assertEqual (result_q .shape , non_contig_query .shape )
154
131
self .assertEqual (result_k .shape , non_contig_key .shape )
155
132
156
- @patch ('vllm_ascend.ops.rotary_embedding.get_ascend_config' )
157
- def test_rope_forward_oot_with_offsets (self , mock_get_ascend_config ):
133
+ def test_rope_forward_oot_with_offsets (self ):
158
134
mock_config = MagicMock ()
159
135
mock_config .torchair_graph_config .enabled = False
160
- mock_get_ascend_config .return_value = mock_config
161
136
162
137
# Test that NotImplementedError is raised when offsets is provided
163
138
offsets = torch .tensor ([1 , 2 , 3 ])
164
139
with self .assertRaises (NotImplementedError ):
165
140
self .layer .forward (self .positions , self .query , self .key , offsets )
166
141
167
- @patch ('vllm_ascend.ops.rotary_embedding.get_ascend_config' )
168
142
@patch ('vllm_ascend.ops.rotary_embedding.custom_rotary_embedding_enabled' ,
169
143
return_value = False )
170
144
@patch ('torch_npu._npu_rotary_embedding' )
171
145
def test_rope_forward_oot_neox_style_override (self , mock_npu_rotary ,
172
- mock_custom_enabled ,
173
- mock_get_ascend_config ):
146
+ mock_custom_enabled ):
174
147
mock_config = MagicMock ()
175
148
mock_config .torchair_graph_config .enabled = False
176
- mock_get_ascend_config .return_value = mock_config
177
149
178
150
# Test neox_style override
179
151
result_q , result_k = self .layer .forward (self .positions ,
0 commit comments