Skip to content

bentoml/BentoMoshi

Repository files navigation

Self-host Moshi with BentoML

This is a BentoML example project, showing you how to serve and deploy Moshi with BentoML.

See here for a full list of BentoML example projects.

Prerequisites

If you want to test the Service locally, we recommend you use a Nvidia GPU with at least 48G VRAM.

Instructions

git clone https://github.com/bentoml/BentoMoshi.git && cd BentoMoshi

# option 1: bentoml serve [RECOMMENDED]
bentoml serve . --debug

# option 2: uv
uvx --from . server

To use the client, specify the URL envvar on bentocloud:

# option 1: uv [RECOMMENDED]
URL=<bentocloud-endpoint> uvx --from . client

# option 2: using python
URL=<bentocloud-endpoint> python bentomoshi/client.py

Note

If you are hosting this on your own server, make sure to include the port the model is served into the URL, for example:

URL=http://localhost:3000 uvx --from . client

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages