Description
I've loaded a model via v2/repository/models/simple/load endpoint.
But when querying v2/repository/index endpoint I get a [] as a responce.
Triton Information
What version of Triton are you using?
2.42.0
Are you using the Triton container or did you build it yourself?
Triton container, version nvcr.io/nvidia/tritonserver:24.01-py3
To Reproduce
- I've took this model: https://github.com/triton-inference-server/server/tree/main/docs/examples/model_repository/simple
- Loaded it with with python script using
tritonclient
model_name = "simple"
config_path = models_repository[model_name]["config"]
model_path = models_repository[model_name]["model"]
with open(model_path, "rb") as f:
model_bytes = f.read()
json_obj = _pbtxt_to_json(config_path)
triton_client.load_model(
model_name=model_name,
config=json_obj,
files={
"file:1/model.graphdef": model_bytes,
},
)
- Then
triton_client.get_model_repository_index()
# returns: []
Describe the models (framework, inputs, outputs), ideally include the model configuration file (if using an ensemble include the model configuration file for that as well).
Model mentioned above
Expected behavior
I expect than this code:
triton_client.get_model_repository_index()
will return responce according to this specification
$repository_index_response =
[
{
"name" : $string,
"version" : $string #optional,
"state" : $string,
"reason" : $string
},
…
]
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/protocol/extension_model_repository.html
Description
I've loaded a model via
v2/repository/models/simple/loadendpoint.But when querying
v2/repository/indexendpoint I get a[]as a responce.Triton Information
What version of Triton are you using?
2.42.0Are you using the Triton container or did you build it yourself?
Triton container, version
nvcr.io/nvidia/tritonserver:24.01-py3To Reproduce
tritonclientDescribe the models (framework, inputs, outputs), ideally include the model configuration file (if using an ensemble include the model configuration file for that as well).
Model mentioned above
Expected behavior
I expect than this code:
will return responce according to this specification
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/protocol/extension_model_repository.html