Build your automation workflow with the power of AI and your personal knowledge base.
Create powerful workflows with just drag and drop, without any programming.
VectorVein is a no-code AI workflow software inspired by LangChain and langflow, designed to combine the powerful capabilities of large language models and enable users to easily achieve intelligent and automated workflows for various daily tasks.
You can experience VectorVein's online version here, with no need to download or install.
Official website Online Documentation
After downloading VectorVein from Release, the program will create a "data" folder in the installation directory to store the database and static file resources.
VectorVein is built using pywebview, based on the webview2 kernel, so you need to install the webview2 runtime. If the software cannot be opened, you may need to download the webview2 runtime manually from https://developer.microsoft.com/en-us/microsoft-edge/webview2/
Most workflows and agents in the software involve the use of AI large language models, so you should at least provide a usable configuration for a large language model. For workflows, you can see which large language models are being used in the interface, as shown in the image below.
After the software opens normally, click the settings button, and enter the API Key for OpenAI/Moonshot/Zhipu AI/Anthropic in the "Large Language Models" tab to use non-local AI features.
If using a local large language model, fill in the local model configuration information in the Local Large Language Model
tab. Currently supports OpenAI-compatible interfaces, such as LM-Studio, Ollama, vLLM, etc.
The API Base for LM-Studio is usually http://localhost:1234/v1/
The API Base for Ollama is usually http://localhost:11434/v1/
After completing the configuration, click
Save Model Family
first, and then clickSave Settings
.
Currently, the speech recognition services of OpenAI/Deepgram are supported. For OpenAI services, you can use the same configuration as the large language model or set up a speech recognition service compatible with the OpenAI API (such as Groq).
When you need to perform vector searches using vector data, you have the option to use embedding services provided by OpenAI or configure local embedding services in the Embedding Model
settings. Currently, supported local embedding services require you to set up text-embeddings-inference yourself.
For ease of daily use, you can configure shortcuts to quickly initiate voice conversations with the Agent. By launching through the shortcut, you can directly interact with the Agent via speech recognition. It is important to ensure that the speech recognition service is correctly configured beforehand.
Include Screenshot means that while starting the conversation, a screenshot of the screen will be taken and uploaded as an attachment to the conversation.
To use your own local Stable Diffusion API, you need to add the parameter --api to the startup item of webui-user.bat, that is
set COMMANDLINE_ARGS=--api
A workflow represents a work task process, including input, output, and how input is processed to reach the output result.
Examples:
- Translation Workflow: The input is an English Word document, and the output is also a Word document. You can design a workflow to translate the input Chinese document and generate a Chinese document output.
- Mind Map Workflow: If the output of the translation workflow is changed to a mind map, you can get a workflow that reads an English Word document and summarizes it into a Chinese mind map.
- Web Article Summary Workflow: If the input of the mind map workflow is changed to a URL of a web article, you can get a workflow that reads a web article and summarizes it into a Chinese mind map.
- Automatic Classification of Customer Complaints Workflow: The input is a table containing complaint content, and you can customize the keywords that need to be classified, so that the complaints can be automatically classified. The output is an automatically generated Excel table containing the classification results.
Each workflow has a User Interface and an Editor Interface. The user interface is used for daily workflow operations, and the editor interface is used for workflow editing. Usually, after designing a workflow, you only need to run it in the user interface and do not need to modify it in the editor interface.
The user interface is shown above and is divided into three parts: input, output, and trigger (usually a run button). You can directly enter content for daily use, click the run button to see the output result.
To view the executed workflow, click Workflow Run Records, as shown in the following figure.
You can add our official templates to your workflow or create a new one. It is recommended to familiarize yourself with the use of workflows using official templates at the beginning.
The workflow editor interface is shown above. You can edit the name, tags, and detailed description at the top. The left side is the node list of the workflow, and the right is the canvas of the workflow. You can drag the desired node from the left side to the canvas, and then connect the node through the wire to form a workflow.
You can view a tutorial on creating a simple crawler + AI summary mind map workflow here.
You can also try this online interactive tutorial.
-
Backend
- Python 3.8 ~ Python 3.11
- PDM installed
-
Frontend
- Vue3
- Vite
Run the following command in the backend directory to install dependencies:
pdm install
pdm install -G mac
Normally, PDM will automatically find the system's Python and create a virtual environment and install dependencies.
After installation, run the following command to start the backend development server and see the running effect:
pdm run dev
If you need to modify the frontend code, you need to run the following command in the frontend directory to install dependencies:
pnpm install
When pulling the project code for the first time, you also need to run
pnpm install
to install the front-end dependencies.If you don't need to develop any front-end code at all, you can directly copy the
web
folder from the release version into thebackend
folder.
After the frontend dependencies are installed, you need to compile the frontend code into the static file directory of the backend. A shortcut instruction has been provided in the project. Run the following command in the backend directory to pack and copy the frontend resources:
pdm run build-front
The project uses pyinstaller for packaging. Run the following command in the backend directory to package it into an executable file:
pdm run build
After packaging, the executable file will be generated in thebackend/dist directory.
VectorVein is an open-source software that supports personal non-commercial use. Please refer to LICENSE for specific agreements.