Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

typo in "Pretrained GPT2 Model Deployment Example" notebook #4080

Closed
saeid93 opened this issue May 6, 2022 · 3 comments · Fixed by #4197
Closed

typo in "Pretrained GPT2 Model Deployment Example" notebook #4080

saeid93 opened this issue May 6, 2022 · 3 comments · Fixed by #4197
Assignees
Labels

Comments

@saeid93
Copy link
Contributor

saeid93 commented May 6, 2022

Describe the bug

I think there is a typo in the following notebook Pretrained GPT2 Model Deployment Example

!mc mb minio-seldon/onnx-gpt2 -p
!mc cp ./model.onnx minio-seldon/onnx-gpt2/gpt2/1/

Should be

!mc mb minio-seldon/onnx-gpt2 -p
!mc cp ./model.onnx minio-seldon/onnx-gpt2

as in the:

%%writefile gpt2-deploy.yaml
apiVersion: machinelearning.seldon.io/v1alpha2
kind: SeldonDeployment
metadata:
  name: gpt2
spec:
  predictors:
  - graph:
      implementation: TRITON_SERVER
      logger:
        mode: all
      modelUri: s3://onnx-gpt2
      envSecretRefName: seldon-init-container-secret
      name: gpt2
      type: MODEL
    name: default
    replicas: 1
  protocol: kfserving

In the current format in the website init container was not able to find the model from Minio but putting the model in the root directory of the minio bucket solved the problem.

@saeid93 saeid93 added the bug label May 6, 2022
@saeid93 saeid93 changed the title type in "Pretrained GPT2 Model Deployment Example" notebook typo in "Pretrained GPT2 Model Deployment Example" notebook May 6, 2022
@saeid93 saeid93 mentioned this issue May 9, 2022
@axsaucedo
Copy link
Contributor

Fixed via #4086

@RafalSkolasinski
Copy link
Contributor

I wonder if the typo may be rather in mc mb ... such that it is missing further subpath? Triton is very strict on folder structure and the actual model binaries are expected to be subfolders (sub-buckets?) of s3://onnx-gpt2.

@saeid93
Copy link
Contributor Author

saeid93 commented May 11, 2022

I reran the notebook with the corrected path and exec into the classsifier container, I can confirm that the model exist in the container in the following path:

triton-server@gpt2-default-0-gpt2-5798b7cbb9-xwc67:/mnt/models$ ls
model.onnx
triton-server@gpt2-default-0-gpt2-5798b7cbb9-xwc67:/mnt/models$ pwd
/mnt/models

Is it how it should be?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants