Skip to content

Commit 29099df

Browse files
committed
Rename output_* to return_*
1 parent 1773dea commit 29099df

File tree

3 files changed

+44
-44
lines changed

3 files changed

+44
-44
lines changed

ci/L0_additional_outputs_vllm/additional_outputs_test.py

Lines changed: 30 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -42,9 +42,9 @@ def _get_inputs(
4242
prompt,
4343
stream=True,
4444
sampling_parameters=None,
45-
output_finish_reason=None,
46-
output_cumulative_logprob=None,
47-
output_num_token_ids=None,
45+
return_finish_reason=None,
46+
return_cumulative_logprob=None,
47+
return_num_token_ids=None,
4848
):
4949
inputs = []
5050

@@ -64,21 +64,21 @@ def _get_inputs(
6464
)
6565
)
6666

67-
if output_finish_reason is not None:
68-
inputs.append(grpcclient.InferInput("output_finish_reason", [1], "BOOL"))
69-
inputs[-1].set_data_from_numpy(np.array([output_finish_reason], dtype=bool))
67+
if return_finish_reason is not None:
68+
inputs.append(grpcclient.InferInput("return_finish_reason", [1], "BOOL"))
69+
inputs[-1].set_data_from_numpy(np.array([return_finish_reason], dtype=bool))
7070

71-
if output_cumulative_logprob is not None:
71+
if return_cumulative_logprob is not None:
7272
inputs.append(
73-
grpcclient.InferInput("output_cumulative_logprob", [1], "BOOL")
73+
grpcclient.InferInput("return_cumulative_logprob", [1], "BOOL")
7474
)
7575
inputs[-1].set_data_from_numpy(
76-
np.array([output_cumulative_logprob], dtype=bool)
76+
np.array([return_cumulative_logprob], dtype=bool)
7777
)
7878

79-
if output_num_token_ids is not None:
80-
inputs.append(grpcclient.InferInput("output_num_token_ids", [1], "BOOL"))
81-
inputs[-1].set_data_from_numpy(np.array([output_num_token_ids], dtype=bool))
79+
if return_num_token_ids is not None:
80+
inputs.append(grpcclient.InferInput("return_num_token_ids", [1], "BOOL"))
81+
inputs[-1].set_data_from_numpy(np.array([return_num_token_ids], dtype=bool))
8282

8383
return inputs
8484

@@ -104,12 +104,12 @@ def _assert_text_output_valid(self):
104104
assert len(text_output) > 0, "output is empty"
105105
assert text_output.count(" ") > 4, "output is not a sentence"
106106

107-
def _assert_finish_reason(self, output_finish_reason):
107+
def _assert_finish_reason(self, return_finish_reason):
108108
for i in range(len(self._responses)):
109109
result, error = self._responses[i]["result"], self._responses[i]["error"]
110110
assert error is None
111111
finish_reason_np = result.as_numpy(name="finish_reason")
112-
if output_finish_reason is None or output_finish_reason == False:
112+
if return_finish_reason is None or return_finish_reason == False:
113113
assert finish_reason_np is None
114114
continue
115115
finish_reason = finish_reason_np[0].decode("utf-8")
@@ -118,25 +118,25 @@ def _assert_finish_reason(self, output_finish_reason):
118118
else:
119119
assert finish_reason == "length"
120120

121-
def _assert_cumulative_logprob(self, output_cumulative_logprob):
121+
def _assert_cumulative_logprob(self, return_cumulative_logprob):
122122
prev_cumulative_logprob = 0.0
123123
for response in self._responses:
124124
result, error = response["result"], response["error"]
125125
assert error is None
126126
cumulative_logprob_np = result.as_numpy(name="cumulative_logprob")
127-
if output_cumulative_logprob is None or output_cumulative_logprob == False:
127+
if return_cumulative_logprob is None or return_cumulative_logprob == False:
128128
assert cumulative_logprob_np is None
129129
continue
130130
cumulative_logprob = cumulative_logprob_np[0].astype(float)
131131
assert cumulative_logprob != prev_cumulative_logprob
132132
prev_cumulative_logprob = cumulative_logprob
133133

134-
def _assert_num_token_ids(self, output_num_token_ids):
134+
def _assert_num_token_ids(self, return_num_token_ids):
135135
for response in self._responses:
136136
result, error = response["result"], response["error"]
137137
assert error is None
138138
num_token_ids_np = result.as_numpy(name="num_token_ids")
139-
if output_num_token_ids is None or output_num_token_ids == False:
139+
if return_num_token_ids is None or return_num_token_ids == False:
140140
assert num_token_ids_np is None
141141
continue
142142
num_token_ids = num_token_ids_np[0].astype(int)
@@ -160,26 +160,26 @@ def _assert_num_token_ids(self, output_num_token_ids):
160160
assert num_token_ids >= 0
161161

162162
@pytest.mark.parametrize("stream", [True, False])
163-
@pytest.mark.parametrize("output_finish_reason", [None, True, False])
164-
@pytest.mark.parametrize("output_cumulative_logprob", [None, True, False])
165-
@pytest.mark.parametrize("output_num_token_ids", [None, True, False])
163+
@pytest.mark.parametrize("return_finish_reason", [None, True, False])
164+
@pytest.mark.parametrize("return_cumulative_logprob", [None, True, False])
165+
@pytest.mark.parametrize("return_num_token_ids", [None, True, False])
166166
def test_additional_outputs(
167167
self,
168168
stream,
169-
output_finish_reason,
170-
output_cumulative_logprob,
171-
output_num_token_ids,
169+
return_finish_reason,
170+
return_cumulative_logprob,
171+
return_num_token_ids,
172172
):
173173
inputs = self._get_inputs(
174174
self._prompt,
175175
stream=stream,
176176
sampling_parameters=self._sampling_parameters,
177-
output_finish_reason=output_finish_reason,
178-
output_cumulative_logprob=output_cumulative_logprob,
179-
output_num_token_ids=output_num_token_ids,
177+
return_finish_reason=return_finish_reason,
178+
return_cumulative_logprob=return_cumulative_logprob,
179+
return_num_token_ids=return_num_token_ids,
180180
)
181181
self._llm_infer(inputs)
182182
self._assert_text_output_valid()
183-
self._assert_finish_reason(output_finish_reason)
184-
self._assert_cumulative_logprob(output_cumulative_logprob)
185-
self._assert_num_token_ids(output_num_token_ids)
183+
self._assert_finish_reason(return_finish_reason)
184+
self._assert_cumulative_logprob(return_cumulative_logprob)
185+
self._assert_num_token_ids(return_num_token_ids)

docs/additional_outputs.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -43,7 +43,7 @@ The reason why the sequence is finished. See
4343
[here](https://github.com/vllm-project/vllm/blob/v0.6.3.post1/vllm/outputs.py#L26)
4444
for more details.
4545

46-
To enable, set `output_finish_reason` input tensor to `True`. The reason will be
46+
To enable, set `return_finish_reason` input tensor to `True`. The reason will be
4747
sent as a string on the `finish_reason` output tensor.
4848

4949
Supported since r24.11.
@@ -54,7 +54,7 @@ The cumulative log probability of the generated output text. See
5454
[here](https://github.com/vllm-project/vllm/blob/v0.6.3.post1/vllm/outputs.py#L22)
5555
for more details.
5656

57-
To enable, set `output_cumulative_logprob` input tensor to `True`. The floating
57+
To enable, set `return_cumulative_logprob` input tensor to `True`. The floating
5858
point value will be sent on the `cumulative_logprob` output tensor.
5959

6060
Supported since r24.11.
@@ -68,7 +68,7 @@ presumed to be zero. See
6868
[here](https://github.com/vllm-project/vllm/blob/v0.6.3.post1/vllm/outputs.py#L21)
6969
for more details on the token IDs of the generated output text.
7070

71-
To enable, set `output_num_token_ids` input tensor to `True`. The unsigned
71+
To enable, set `return_num_token_ids` input tensor to `True`. The unsigned
7272
integer value will be sent on the `num_token_ids` output tensor.
7373

7474
Supported since r24.11.
@@ -88,7 +88,7 @@ inputs[-1].set_data_from_numpy(
8888
np.array(["example prompt".encode("utf-8")], dtype=np.object_)
8989
)
9090

91-
inputs.append(grpcclient.InferInput("output_finish_reason", [1], "BOOL"))
91+
inputs.append(grpcclient.InferInput("return_finish_reason", [1], "BOOL"))
9292
inputs[-1].set_data_from_numpy(np.array([True], dtype=bool))
9393

9494
def callback(result, error):

src/model.py

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -89,19 +89,19 @@ def _auto_complete_inputs_and_outputs(auto_complete_model_config):
8989
"optional": True,
9090
},
9191
{
92-
"name": "output_finish_reason",
92+
"name": "return_finish_reason",
9393
"data_type": "TYPE_BOOL",
9494
"dims": [1],
9595
"optional": True,
9696
},
9797
{
98-
"name": "output_cumulative_logprob",
98+
"name": "return_cumulative_logprob",
9999
"data_type": "TYPE_BOOL",
100100
"dims": [1],
101101
"optional": True,
102102
},
103103
{
104-
"name": "output_num_token_ids",
104+
"name": "return_num_token_ids",
105105
"data_type": "TYPE_BOOL",
106106
"dims": [1],
107107
"optional": True,
@@ -348,11 +348,11 @@ def _get_input_tensors(self, request):
348348
else:
349349
parameters = request.parameters()
350350

351-
# output_finish_reason, output_cumulative_logprob, output_num_token_ids
351+
# return_finish_reason, return_cumulative_logprob, return_num_token_ids
352352
additional_outputs = {
353-
"output_finish_reason": None,
354-
"output_cumulative_logprob": None,
355-
"output_num_token_ids": None,
353+
"return_finish_reason": None,
354+
"return_cumulative_logprob": None,
355+
"return_num_token_ids": None,
356356
}
357357
for tensor_name in additional_outputs.keys():
358358
tensor = pb_utils.get_input_tensor_by_name(request, tensor_name)
@@ -445,7 +445,7 @@ def _create_response(
445445
)
446446

447447
# finish_reason
448-
if additional_outputs["output_finish_reason"]:
448+
if additional_outputs["return_finish_reason"]:
449449
finish_reason = [
450450
str(output.finish_reason) for output in request_output.outputs
451451
]
@@ -456,7 +456,7 @@ def _create_response(
456456
)
457457

458458
# cumulative_logprob
459-
if additional_outputs["output_cumulative_logprob"]:
459+
if additional_outputs["return_cumulative_logprob"]:
460460
cumulative_logprob = [
461461
output.cumulative_logprob for output in request_output.outputs
462462
]
@@ -468,7 +468,7 @@ def _create_response(
468468
)
469469

470470
# num_token_ids
471-
if additional_outputs["output_num_token_ids"]:
471+
if additional_outputs["return_num_token_ids"]:
472472
if prev_request_output is None:
473473
# this is the first response
474474
prev_lens = [0] * len(request_output.outputs)

0 commit comments

Comments
 (0)