This Flow Launcher Plugin allows you to interact with a local Ollama instance and your favorite and private LLMs (e.g. Llama 3, Gemma 2, Phi-3, TinyLlama, ...). It provides a convenient way to access the power of these language models directly from Flow Launcher. The plugin offers two interaction options, copying the answer directly to the clipboard and writing the conversation (question + response) to a text file, which can be opened directly via the plugin.
To be able to communicate with a local LLM and use this plugin, you need a running Ollama instance (server). There are two different installation variants here, which you can also select depending on the underlying OS:
- Installation via Docker (OS independent)
- Go to Install Docker Engine and follow the installation instructions
- Pull the Ollama Image and start the Container
- Optional: If the container and the Large Language Model are to be operated using a graphics card, follow the next steps
- Download and installation of the NVIDIA CUDA Toolkit
- Download and Installation of the NVIDIA Container Toolkit
- Here is an example of a Docker-Compose that you can use to deploy Ollama with an NVIDIA graphics card:
--- services: ollama: image: ollama/ollama:latest container_name: ollama ports: - "11434:11434" restart: always volumes: - $HOME/.ollama:/root/.ollama deploy: resources: reservations: devices: - driver: nvidia count: 1 capabilities: [gpu]
- Direct installation OS
- Go to Ollama Download and select your operating system
- Follow the installation instructions and then ensure that the instance is running
- Download and install Flow Launcher.
- Open the Flow Launcher settings
- Go to the Plugin Store tab
- Search for 'Ollama'
- Click and Install Ollama
- Flow Launcher should restart automatically. If not, restart Flow Launcher manually
- Re-open the Flow Launcher settings and head to the Plugins tab
- Customize the Settings according to your configuration
- Run the 'Save Settings' command in Flow Launcher
Setting | Default | Description |
---|---|---|
Action keyword | ol | Keyword to activate the plugin. |
Ollama Host | http://localhost:11434 | URL of the local Ollama instance to communicate via API. |
Ollama Model | llama3.2:1b | The LLM to be used (Ollama model library). |
Automatic Model Download | [ ] - false | Download LLM automatically if not already installed. Be careful - the download may take some time and storage on your disk. |
Save Chat to File | [x] - true | Should the chat be saved as a text file? This allows it to be opened directly in a text editor. |
Chat preview preserve newline | [ ] - false | Should the chat preview retain the line breaks or output them as continuous text. If true, the heading 'Copy Response to Clipboard' can be moved outside the visible area. However the text is still always copied to the clipboard with the correct formatting. |
Chat preview length | 100 | Length of the chat preview, freely selectable. |
Prompt Stop | || | Characters to indicate end of prompt. This saves computing time, as otherwise the LLM is executed every time a key is pressed. |
Log Level | ERROR | The Log Level can be adjusted for error analysis. Normally not of interest for users. |