Skip to content

This is a sample example repo on how to use Vapi with your custom-llm

Notifications You must be signed in to change notification settings

VapiAI/advanced-concepts-custom-llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Vapi Integration with Custom LLM

Welcome to the Vapi CustomLLM sample project! This guide will walk you through integrating Vapi with your custom language model.

Getting Started

To set up this project:

  1. Create a Python Virtual Environment

    python -m venv .venv
    source .venv/bin/activate
    
  2. Set Up Environment Variables

    • Create a .env file in your repository:

      cp .env.example .env
    • Add your OPENAI_API_KEY to the .env file.

  3. Install Dependencies

    pip install --upgrade pip &&\ 
        pip install -r requirements.txt    
  4. Run the FastAPI Server

    fastapi dev ./app/main.py 

Exposing Localhost to the Internet

To make your local server accessible over the internet, you can use ngrok. Follow these steps:

  1. Install ngrok
  2. Expose Localhost
    ngrok http 8000   

Conclusion

This sample project demonstrates how to integrate a custom language model with Vapi. By following these steps, developers can create a more versatile and responsive voice assistant tailored to their users' unique needs.

For more help and detailed documentation, please refer to the official Vapi documentation.

About

This is a sample example repo on how to use Vapi with your custom-llm

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published