Deploy almost any Text Embedding and Reranker models with high throughput OpenAI-compatible Endpoints on RunPod Serverless, powered by the fastest embedding inference engine, built for serving - Infinity
When using torch
backend, you can deploy any models supported by the sentence-transformers library.
This also means that you can deploy any model from the Massive Text Embedding Benchmark (MTEB) Leaderboard, which is currently the most popular and comprehensive leaderboard for embedding models.
Note
We are adding a UI for deployment similar to Worker vLLM, but for now, you can manually create the endpoint with the regular serverless configurator.
We offer a pre-built Docker Image for the Infinity Embedding Serverless Worker that you can configure entirely with Environment Variables when creating the Endpoint:
You can directly use the following docker images and configure them via Environment Variables.
CUDA Version | Stable (Latest Release) | Development (Latest Commit) | Note |
---|---|---|---|
11.8.0 | runpod/worker-infinity-embedding:stable-cuda11.8.0 |
runpod/worker-infinity-embedding:dev-cuda11.8.0 |
Available on all RunPod Workers without additional selection needed. |
12.1.0 | runpod/worker-infinity-embedding:stable-cuda12.1.0 |
runpod/worker-infinity-embedding:dev-cuda12.1.0 |
When creating an Endpoint, select CUDA Version 12.4, 12.3, 12.2 and 12.1 in the filter. About 10% less total available machines than 11.8.0, but higher performance. |
[NOTE] Latest image version (pre) runpod/worker-infinity-text-embedding:0.0.1-cuda12.1.0
-
MODEL_NAMES
HuggingFace repo of a single model or multiple models separated by semicolon.
- Examples:
- Single Model:
BAAI/bge-small-en-v1.5
- Multiple Models:
BAAI/bge-small-en-v1.5;intfloat/e5-large-v2;
- Single Model:
- Examples:
-
BATCH_SIZES
Batch Size for each model separated by semicolon.
- Default:
32
- Default:
-
BACKEND
Backend for all models.
- Options:
torch
optimum
ctranslate2
- Default:
torch
- Options:
-
DTYPES
Precision for each model separated by semicolon.
- Options:
auto
fp16
fp8
(New! Only compatible with H100 and L40S)
- Default:
auto
- Options:
-
INFINITY_QUEUE_SIZE
How many requests can be queued in the Infinity Engine.
- Default:
48000
- Default:
-
RUNPOD_MAX_CONCURRENT_REQUESTS
How many requests can be processed concurrently by the RunPod Worker.
- Default:
300
- Default:
Coming soon!
There are two ways to use the endpoint - OpenAI Compatibility matching how you would use OpenAI API, and Standard Usage with the RunPod API. Note that reranking is only available with Standard Usage.
- Install OpenAI Python SDK
pip install openai
- Initialize OpenAI client and set the API Key to your RunPod API Key, and base URL to
https://api.runpod.ai/v2/YOUR_ENDPOINT_ID/openai/v1
, whereYOUR_ENDPOINT_ID
is the ID of your endpoint, e.g.elftzf0lld1vw1
from openai import OpenAI
client = OpenAI(
api_key=RUNPOD_API_KEY,
base_url="https://api.runpod.ai/v2/YOUR_ENDPOINT_ID/openai/v1"
)
-
Define the input
You may embed a single text or a list of texts
- Single Text
embedding_input = "Hello, world!"
- List of Texts
embedding_input = ["Hello, world!", "This is a test."]
- Single Text
-
Get the embeddings
client.embeddings.create( model="YOUR_DEPLOYED_MODEL_NAME", input=embedding_input )
Where
YOUR_DEPLOYED_MODEL_NAME
is the name of one of the models you deployed to the worker.
You may use /run
(asynchronous, start job and return job ID) or /runsync
(synchronous, wait for job to finish and return result)
Inputs:
model
: name of one of the deployed models.input
: single text string or list of texts to embed
Inputs:
model
: name of one of the deployed modelsquery
: query text (single string)docs
: list of documents to rerank by queryreturn_docs
: whether to return the reranked documents or not
We'd like to thank Michael Feil for creating the Infinity Embedding Engine and actively being involved in the development of this worker!