You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If a user wants to run dolphin on LM studio with the airoboros wrapper:
export OPENAI_API_BASE=http://127.0.0.1:1234
export BACKEND_TYPE=lmstudio
memgpt run --model airoboros_xxx
Config (when using a local model)
model is "local", or can be "airoboros_xxx" in which case model == wrapper
model_endpoint stores the IP from OPENAI_API_BASE
[defaults]
model = local
model_endpoint = http://localhost:1234
Proposed setup (with memgpt run)
User does not specify and ENV variables, it's all in config
Add a --wrapper arg and config variable
If a user wants to run dolphin on LM studio with the airoboros wrapper:
memgpt run --wrapper airoboros_xxx --endpoint http://localhost:1234 --endpoint_type lmstudio
For almost all backends, it's OK for the model to be unspecified, because what model is running is determined by the backend. The only exception to this is Ollama, which requires you to pass the model name in the POST request. This is already a special case in our documentation: https://memgpt.readthedocs.io/en/latest/ollama/ (currently, we ask the user to set an additional environment variable).
Special Ollama case:
memgpt run --model dolphin_xxx --wrapper airoboros_xxx --endpoint http://localhost:11434 --endpoint_type ollama
Proposed setup (with memgpt configure, then memgpt run)
If the user says no to OpenAI, no to Azure, then:
Ask for their endpoint type (lmstudio, ollama, etc)
Ask for their endpoint IP
We should do input checking / sanitation on the IP they provide (http prefix? hanging /v1/?)
Ask what prompt formatter / wrapper they want to use
IMO I think we should hide this, make it default to the default, but can override with memgpt run --wrapper
Config (when using a local model)
[defaults]
model = optional for non-Ollama (default None), for Ollama this is the real model name (eg dolphin-2.2.1-mistral7b)
model_endpoint = http://localhost:1234
model_endpoint_type = lmstudio
Special case where the user wants to use OpenAI, but swap the endpoint to a proxy
export OPENAI_API_BASE="<proxy_address>"
memgpt run
Config
We do NOT set model_endpoint to this proxy address, instead let openai-python handle this for us (on our end we act like nothing changed, it's just openai):
Current setup
If a user wants to run dolphin on LM studio with the airoboros wrapper:
Config (when using a local model)
OPENAI_API_BASE
Proposed setup (with
memgpt run
)--wrapper
arg and config variableIf a user wants to run dolphin on LM studio with the airoboros wrapper:
For almost all backends, it's OK for the
model
to be unspecified, because what model is running is determined by the backend. The only exception to this is Ollama, which requires you to pass themodel
name in the POST request. This is already a special case in our documentation: https://memgpt.readthedocs.io/en/latest/ollama/ (currently, we ask the user to set an additional environment variable).Special Ollama case:
Proposed setup (with
memgpt configure
, thenmemgpt run
)http
prefix? hanging/v1/
?)memgpt run --wrapper
Config (when using a local model)
lmstudio:
ollama:
Special case where the user wants to use OpenAI, but swap the endpoint to a proxy
Config
We do NOT set
model_endpoint
to this proxy address, instead letopenai-python
handle this for us (on our end we act like nothing changed, it's justopenai
):The text was updated successfully, but these errors were encountered: