-
I am trying to use openbmb/MiniCPM3-4B-GGUF, but the model doesn't stop generating output. After checking the documentation, I learned that I need to customize the chat wrapper. But how should I define the format? Can you provide some help? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
I just tested this model and it worked as expected. npx -y node-llama-cpp chat --prompt 'Hi there!' https://huggingface.co/openbmb/MiniCPM3-4B-GGUF/blob/main/minicpm3-4b-q4_k_m.gguf If you could share the code you've used and the result of running the I recommend you to try to scaffold a new project and run it with this model. |
Beta Was this translation helpful? Give feedback.
I just tested this model and it worked as expected.
You can run this command to test it yourself:
npx -y node-llama-cpp chat --prompt 'Hi there!' https://huggingface.co/openbmb/MiniCPM3-4B-GGUF/blob/main/minicpm3-4b-q4_k_m.gguf
If you could share the code you've used and the result of running the
npx -y node-llama-cpp inspect gpu
command then I can try to help you find what the issue is.I recommend you to try to scaffold a new project and run it with this model.