@@ -113,6 +113,14 @@ func (m *ModelService) SetSecondaryServerURL(url string) *ModelService {
113
113
return m
114
114
}
115
115
116
+ // SetAPIRequestTimeout set api request timeout
117
+ func (m * ModelService ) SetAPIRequestTimeout (timeout time.Duration ) * ModelService {
118
+ if m .TritonService != nil {
119
+ m .TritonService .SetAPIRequestTimeout (timeout )
120
+ }
121
+ return m
122
+ }
123
+
116
124
// SetJsonEncoder set json encoder
117
125
func (m * ModelService ) SetJsonEncoder (encoder utils.JSONMarshal ) * ModelService {
118
126
m .TritonService .SetJSONEncoder (encoder )
@@ -130,55 +138,43 @@ func (m *ModelService) SetJsonDecoder(decoder utils.JSONUnmarshal) *ModelService
130
138
//////////////////////////////////////////// Triton Service API Function ////////////////////////////////////////////
131
139
132
140
// CheckServerReady check server is ready.
133
- func (m * ModelService ) CheckServerReady (requestTimeout time. Duration ) (bool , error ) {
134
- return m .TritonService .CheckServerReady (requestTimeout )
141
+ func (m * ModelService ) CheckServerReady () (bool , error ) {
142
+ return m .TritonService .CheckServerReady ()
135
143
}
136
144
137
145
// CheckServerAlive check server is alive.
138
- func (m * ModelService ) CheckServerAlive (requestTimeout time. Duration ) (bool , error ) {
139
- return m .TritonService .CheckServerAlive (requestTimeout )
146
+ func (m * ModelService ) CheckServerAlive () (bool , error ) {
147
+ return m .TritonService .CheckServerAlive ()
140
148
}
141
149
142
150
// CheckModelReady check model is ready.
143
- func (m * ModelService ) CheckModelReady (
144
- modelName , modelVersion string , requestTimeout time.Duration ,
145
- ) (bool , error ) {
146
- return m .TritonService .CheckModelReady (modelName , modelVersion , requestTimeout )
151
+ func (m * ModelService ) CheckModelReady (modelName , modelVersion string ) (bool , error ) {
152
+ return m .TritonService .CheckModelReady (modelName , modelVersion )
147
153
}
148
154
149
155
// GetServerMeta get server meta.
150
- func (m * ModelService ) GetServerMeta (
151
- requestTimeout time.Duration ,
152
- ) (* nvidia_inferenceserver.ServerMetadataResponse , error ) {
153
- return m .TritonService .ServerMetadata (requestTimeout )
156
+ func (m * ModelService ) GetServerMeta () (* nvidia_inferenceserver.ServerMetadataResponse , error ) {
157
+ return m .TritonService .ServerMetadata ()
154
158
}
155
159
156
160
// GetModelMeta get model meta.
157
- func (m * ModelService ) GetModelMeta (
158
- modelName , modelVersion string , requestTimeout time.Duration ,
159
- ) (* nvidia_inferenceserver.ModelMetadataResponse , error ) {
160
- return m .TritonService .ModelMetadataRequest (modelName , modelVersion , requestTimeout )
161
+ func (m * ModelService ) GetModelMeta (modelName , modelVersion string ) (* nvidia_inferenceserver.ModelMetadataResponse , error ) {
162
+ return m .TritonService .ModelMetadataRequest (modelName , modelVersion )
161
163
}
162
164
163
165
// GetAllModelInfo get all model info.
164
- func (m * ModelService ) GetAllModelInfo (
165
- repoName string , isReady bool , requestTimeout time.Duration ,
166
- ) (* nvidia_inferenceserver.RepositoryIndexResponse , error ) {
167
- return m .TritonService .ModelIndex (repoName , isReady , requestTimeout )
166
+ func (m * ModelService ) GetAllModelInfo (repoName string , isReady bool ) (* nvidia_inferenceserver.RepositoryIndexResponse , error ) {
167
+ return m .TritonService .ModelIndex (repoName , isReady )
168
168
}
169
169
170
170
// GetModelConfig get model config.
171
- func (m * ModelService ) GetModelConfig (
172
- modelName , modelVersion string , requestTimeout time.Duration ,
173
- ) (interface {}, error ) {
174
- return m .TritonService .ModelConfiguration (modelName , modelVersion , requestTimeout )
171
+ func (m * ModelService ) GetModelConfig (modelName , modelVersion string ) (interface {}, error ) {
172
+ return m .TritonService .ModelConfiguration (modelName , modelVersion )
175
173
}
176
174
177
175
// GetModelInferStats get model infer stats.
178
- func (m * ModelService ) GetModelInferStats (
179
- modelName , modelVersion string , requestTimeout time.Duration ,
180
- ) (* nvidia_inferenceserver.ModelStatisticsResponse , error ) {
181
- return m .TritonService .ModelInferStats (modelName , modelVersion , requestTimeout )
176
+ func (m * ModelService ) GetModelInferStats (modelName , modelVersion string ) (* nvidia_inferenceserver.ModelStatisticsResponse , error ) {
177
+ return m .TritonService .ModelInferStats (modelName , modelVersion )
182
178
}
183
179
184
180
//////////////////////////////////////////// Triton Service API Function ////////////////////////////////////////////
0 commit comments