Skip to content

Latest commit

 

History

History
18 lines (14 loc) · 686 Bytes

readme.md

File metadata and controls

18 lines (14 loc) · 686 Bytes

LLaMa.cpp Gemma Web-UI

This project uses llama.cpp to load model from a local file, delivering fast and memory-efficient inference.
The project is currently designed for Google Gemma, and will support more models in the future.

Deployment

Prerequisites

Installation

  1. Download Gemma model from Google repository.
  2. Edit the model-path config.yaml, this should point to the actual model path.
  3. Start the web-ui by command:
    screen -S "webui" bash ./start-ui.sh