There are many 1000s of Gradio apps on Hugging Face Spaces. This library puts them at the tips of your LLM's fingers 🦾
Specifically, gradio-tools
is a Python library for converting Gradio apps into tools that can be leveraged by a large language model (LLM)-based agent to complete its task. For example, an LLM could use a Gradio tool to transcribe a voice recording it finds online and then summarize it for you. Or it could use a different Gradio tool to apply OCR to a document on your Google Drive and then answer questions about it.
Currently supported libraries for agents are:
gradio-tools
comes with a set of pre-built tools you can leverage immediately! These include:
- StableDiffusionTool - Generate an image from a given prompt using the open source stable diffusion demo hosted on HuggingFace spaces
- ImageCaptionTool - Caption an image by providing a filepath based on Niels Rogge's HuggingFace Space
- ImageToMusicTool - Create an audio clip that matches the style of a given image file based on Sylvain Filoni's HuggingFace Space
- StableDiffusionPromptGeneratorTool - Use this tool to improve a prompt for stable diffusion and other image generators based on this HuggingFace Space
- TextToVideoTool - A tool for creating short videos from text. Based on this HuggingFace Space
- WhisperAudioTranscriptionTool - A tool for transcribing audio with Whisper. Based on this HuggingFace Space
- ClipInterrogatorTool - A tool for reverse engineering a prompt from a source image. Based on this HuggingFace Space
- DocQueryDocumentAnsweringTool - A tool for answering questions about a document from the from the image of the document. Based on this HuggingFace Space
- BarkTextToSpeechTool - A tool for text-to-speech. Based on this HuggingFace Space
We welcome more contributions!
Simply import the desired tools from gradio_tools
(or create your own!) and pass to initialize_agent
from LangChain.
In this example, we use some pre-built tools to generate images, caption them, and create a video!
Read the How It Works section to learn how to create your own tools! We welcome any new tools to the library!
from gradio_tools import (StableDiffusionTool, ImageCaptioningTool, StableDiffusionPromptGeneratorTool,
TextToVideoTool)
from langchain.agents import initialize_agent
from langchain.llms import OpenAI
from langchain.memory import ConversationBufferMemory
llm = OpenAI(temperature=0)
memory = ConversationBufferMemory(memory_key="chat_history")
tools = [StableDiffusionTool().langchain, ImageCaptioningTool().langchain,
StableDiffusionPromptGeneratorTool().langchain, TextToVideoTool().langchain]
agent = initialize_agent(tools, llm, memory=memory, agent="conversational-react-description", verbose=True)
output = agent.run(input=("Please create a photo of a dog riding a skateboard "
"but improve my prompt prior to using an image generator."
"Please caption the generated image and create a video for it using the improved prompt."))
Gradio.Tools.+.LLM.Agents.mp4
See the /examples
directory for more complete code examples.
The core abstraction is the GradioTool
, which lets you define a new tool for your LLM as long as you implement a standard interface:
class GradioTool(BaseTool):
def __init__(self, name: str, description: str, src: str) -> None:
@abstractmethod
def create_job(self, query: str) -> Job:
pass
@abstractmethod
def postprocess(self, output: Tuple[Any] | Any) -> str:
pass
The requirements are:
- The name for your tool
- The description for your tool. This is crucial! Agents decide which tool to use based on their description. Be precise and be sure to inclue example of what the input and the output of the tool should look like.
- The url or space id, e.g.
freddyaboulton/calculator
, of the Gradio application. Based on this value,gradio_tools
will create a gradio client instance to query the upstream application via API. Be sure to click the link and learn more about the gradio client library if you are not familiar with it. - create_job - Given a string, this method should parse that string and return a job from the client. Most times, this is as simple as passing the string to the
submit
function of the client. More info on creating jobs here - postprocess - Given the result of the job, convert it to a string the LLM can display to the user.
- Optional - Some libraries, e.g. MiniChain, may need some info about the underlying gradio input and output types used by the tool. By default, this will return gr.Textbox() but
if you'd like to provide more accurate info, implement the
_block_input(self, gr)
and_block_output(self, gr)
methods of the tool. Thegr
variable is the gradio module (the result ofimport gradio as gr
). It will be automatically imported by theGradiTool
parent class and passed to the_block_input
and_block_output
methods.
And that's it!
A LangChain agent is a Large Language Model (LLM) that takes user input and reports an output based on using one of many tools at its disposal.
Gradio is the defacto standard tool for building Machine Learning Web Applications and sharing them with the world - all with just python! 🐍