Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Best practice for multiple servings from one model #915

Closed
dtaniwaki opened this issue Oct 6, 2019 · 2 comments
Closed

Best practice for multiple servings from one model #915

dtaniwaki opened this issue Oct 6, 2019 · 2 comments
Milestone

Comments

@dtaniwaki
Copy link
Contributor

I'd like to serve multiple predictions from one model. I currently create multiple model files using one model and create multiple containers that serve a model from a different model file. In this method, the memory footprint of the loaded model becomes multiple of one model serving.

I came up with an idea of using metadata so I can call different methods in the predict method of a model. However, I think an endpoint serving different kinds of outputs is weird and hard to maintain the graphs.

Could someone tell me your idea or known best practice for this kind of case?

@ukclivecox
Copy link
Contributor

Its much harder to manage SeldonDeployments if the underlying resources are shared. It certainly could be that we implement some form of sharing for model servers that allow it but I would think this is more when 2 separate SeldonDeployments happen to use the same model. This seems to be your use case.

You could create create 3 SeldonDeployments for your use case. 1 there has the core model and 2 that act as proxy servers to the core model for each external use case. Would this fir your need?

@ukclivecox ukclivecox added this to the 0.5.x milestone Oct 8, 2019
@ukclivecox
Copy link
Contributor

Please reopen with specific issue if this is still causing a blocker for you @dtaniwaki

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants