File tree Expand file tree Collapse file tree 1 file changed +5
-5
lines changed Expand file tree Collapse file tree 1 file changed +5
-5
lines changed Original file line number Diff line number Diff line change 30
30
31
31
** LATEST RELEASE: You are currently on the main branch which tracks
32
32
under-development progress towards the next release. The current release branch
33
- is [ r23.11 ] ( https://github.com/triton-inference-server/vllm_backend/tree/r23.11 )
34
- and which corresponds to the 23.11 container release on
33
+ is [ r23.12 ] ( https://github.com/triton-inference-server/vllm_backend/tree/r23.12 )
34
+ and which corresponds to the 23.12 container release on
35
35
[ NVIDIA GPU Cloud (NGC)] ( https://catalog.ngc.nvidia.com/orgs/nvidia/containers/tritonserver ) .**
36
36
37
37
# vLLM Backend
@@ -96,9 +96,9 @@ A sample command to build a Triton Server container with all options enabled is
96
96
--endpoint=grpc
97
97
--endpoint=sagemaker
98
98
--endpoint=vertex-ai
99
- --upstream-container-version=23.10
100
- --backend=python:r23.10
101
- --backend=vllm:r23.10
99
+ --upstream-container-version=23.12
100
+ --backend=python:r23.12
101
+ --backend=vllm:r23.12
102
102
```
103
103
104
104
### Option 3. Add the vLLM Backend to the Default Triton Container
You can’t perform that action at this time.
0 commit comments