Code of AAAI2025 Paper 《VIoTGPT: Learning to Schedule Vision Tools in LLMs towards Intelligent Video Internet of Things》.
Video Internet of Things (VIoT) has shown full potential in collecting an unprecedented volume of video data. How to schedule the domain-specific perceiving models and analyze the collected videos uniformly, efficiently, and especially intelligently to accomplish complicated tasks is challenging. To address the challenge, we build VIoTGPT, the framework based on LLMs to correctly interact with humans, query knowledge videos, and invoke vision models to analyze multimedia data collaboratively. To support VIoTGPT and related future works, we meticulously crafted the VIoT-Tool dataset, including the training dataset and the benchmark involving 11 representative vision models across three categories based on semi-automatic annotations. To guide LLM to act as the intelligent agent towards intelligent VIoT, we resort to the ReAct instruction tuning method based on VIoT-Tool to learn the tool capability. Quantitative and qualitative experiments and analyses demonstrate the effectiveness of VIoTGPT. We believe VIoTGPT contributes to improving human-centered experiences in VIoT applications.
Download the pre-trained LLama-2-7b and Vicuna-7b, then download the LoRA weights from the following links.
torch==2.0.1
transformers==4.31.0
deepspeed==0.10.0
langchain==0.0.101
gradio==3.23.0
- Train
cd train
train_vicuna.sh
- Test
test_vicuna.sh
@inproceedings{zhong2025VIoTGPT,
title={VIoTGPT: Learning to Schedule Vision Tools in LLMs towards Intelligent Video Internet of Things},
author={Zhong, Yaoyao and Qi, Mengshi and Wang, Rui and Qiu, Yuhan and Zhang, Yang and Ma, Huadong},
booktitle={AAAI},
year={2025}
}