-
Notifications
You must be signed in to change notification settings - Fork 835
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU support with SERVICE_TYPE Model #590
Comments
Can you provide more details on the error at "parsing" stage? |
No, the version for seldon-core and seldon-core-crd both are 0.2.5. Installed with helm locally. The seldon-core-apiserver reports an error below:
|
Are you able to try this with the latest from master? |
Not yet, the latest version is really hard to be installed for the Ambassador stuff, so I followed the guide in the example which uses version 0.2.5. |
When I say "blocked in parsing stage", I mean I can saw the deployment name with |
If using 0.2.5 can you check the logs of the cluster-manager? What problems are you having with Ambassador. In master you would install the official Ambassador helm chart. The issue you are having I think is due to parsing of Quantity in the protobuffer specs. This should be fixed in the version in master which was why I was hoping you could test with latest? |
Yes, you are correct, I just found a similar scenario in issue #45 . I checked the cluster-manager, it is for the Quantity parsing.
However, the cluster-manager reports another error:
Looks like unable to find the image, but apparently it was hosted. What else could be make this happen? |
The name in the |
Exactly! Changing the container name resolves my problem. Now I can see a deployment is being created. @cliveseldon Thanks very much for your patient! |
There should no issue. As long as your model image and Pod is correctly setup. We'd love to have an example in this area so happy to help you get everything working. |
* add prestop to raw yaml files * check in helm changes for scaling down * make default cpu for mlserver / triton to 1 * add scale endpoint to server CR * changes from HPA PR#83 (from clive) * update to CRDs * add default triton cpu request * remove init containers as we dont use them * autogenerate server resource * add agent and rclone cpu / memory requests * update helm * remove init container from autogen file * k6 runner fix * update scaling logic + tests * fix draining empty server + adding tests * when failedscheduling we can have available repls * reduce cool down timer to 1 minute * lint * Add autoscaling docs * add terminationGracePeriodSeconds to helm * add autogen files * improve agent logging around scaling events * move scaling logs to debug level * Update docs/source/contents/kubernetes/autoscaling/index.md Co-authored-by: Alex Rakowski <[email protected]> * Update docs/source/contents/kubernetes/autoscaling/index.md Co-authored-by: Alex Rakowski <[email protected]> * Update docs/source/contents/kubernetes/autoscaling/index.md Co-authored-by: Alex Rakowski <[email protected]> * docs changes * trim space from helm parameter value * tidy up comment in reconciler * update raw yaml files Co-authored-by: Alex Rakowski <[email protected]>
Hi, I was trying to deploy a
SeldonDeployment
to the cluster, which asks for gpu resource and CUDA. I writes the .yaml as the official doc suggests, however the deployment was blocked at "parsing" the CRD stage, which results in no deployment or service was created. It was totally OK to deploy models without gpu required.I didn't find any example on using gpus, so, my question is: does Seldon-core support GPU? Or does anyone has succeeded in deploying a model with gpu required?
This is a part from my .yaml:
The text was updated successfully, but these errors were encountered: