You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As per my conversation with @adriangonz on slack there seem to be a bug in accessing the endpoints of the inference graph nodes where some nodes of the inference graph are prepackaged Triton servers. For example in single node setting metadata of a triton node is accessible but the same endpoint is not accessible when the Triton node is part of an inference graph.
To reproduce
To compare the two situation, you can deploy following single node Triton server:
Describe the bug
As per my conversation with @adriangonz on slack there seem to be a bug in accessing the endpoints of the inference graph nodes where some nodes of the inference graph are prepackaged Triton servers. For example in single node setting metadata of a triton node is accessible but the same endpoint is not accessible when the Triton node is part of an inference graph.
To reproduce
To compare the two situation, you can deploy following single node Triton server:
and the metadata endpoint is accessible at:
output as expected:
However inside the graph:
The endpoint
curl localhost:32000/seldon/default/gpt2/v2/models/gpt2
will result in{"error":"Model gpt2 not found"}
Other combinations like
curl localhost:32000/seldon/default/gpt2/gpt2/v2/models/gpt2
won't work either.Expected behaviour
Intermediate triton endpoint be accessible.
Environment
The text was updated successfully, but these errors were encountered: