-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[RFC]: Enhancing LoRA Management for Production Environments in vLLM #6275
Comments
I'm in favor of all these! Please also make sure it is well documented. |
Yes, this all makes sense. Let's make sure to ensure that performance doesn't degrade too much with loading from remote storage. |
Yes, the issue you have noted prevent us from running vLLM. I would also include the ability to apply (merge) more than one adapter simultaneously to a single request. I am looking forward to these features making itself into vLLM. |
Hi Jeff, Thank you for sharing the RFC on Lora. I noticed my feature request was included, which is appreciated. Want to check whether there are plans to implement the load/unload API for the base model? Thanks in advance for your attention to this matter. |
I would love to add the following feature in this RFC. LOAD ADAPTERS FROM S3 SUPPORTED STORAGESLoRAX has this feature https://loraexchange.ai/models/adapters/#s3 This brings a new challenge in vLLM. Including type of the source (huggingface or s3). This is handled in LoRAX by providing the default 'adapter-source'. It needs to support storages which supports S3 schema (like Cloudflare R2) https://github.com/predibase/lorax/blob/main/server/lorax_server/utils/sources/s3.py
|
This RFC proposes improvements to the management of Low-Rank Adaptation (LoRA) in vLLM to make it more suitable for production environments. This proposal aims to address several pain points observed in the current implementation. Feedback and discussions are welcome, and we hope to gather input and refine the proposal based on community insights.
Motivation.
This RFC proposes improvements to the management of Low-Rank Adaptation (LoRA) in vLLM to make it more suitable for production environments. This proposal aims to address several pain points observed in the current implementation. Feedback and discussions are welcome, and we hope to gather input and refine the proposal based on community insights.
Motivation
LoRA integration in production environments faces several challenges that need to be addressed to ensure smooth and efficient deployment and management. The main issues observed include:
Visibility of LoRA Information: Currently, the relationship between LoRA and base models is not exposed clearly by the engine. The
/v1/models
endpoint does not display this information. Related issues: [Feature]: Expose Lora lineage information from /v1/models #6274Dynamic Loading and Unloading: LoRA adapters cannot be dynamically loaded or unloaded after the server has started. Related issues: Multi-LoRA - Support for providing /load and /unload API #3308 [Feature]: Allow LoRA adapters to be specified as in-memory dict of tensors #4068 [Feature]: load/unload API to run multiple LLMs in a single GPU instance #5491
Remote Registry Support: LoRA adapters cannot be pulled from remote model repositories during runtime, making it cumbersome to manage artifacts locally. Related issues: [Feature]: Support loading lora adapters from HuggingFace in runtime #6233 [Bug]: relative path doesn't work for Lora adapter model #6231
Observability: There is a lack of metrics and observability enhancements related to LoRA, making it difficult to monitor and manage.
Cluster level Support: Information about LoRA is not easily accessible to resource managers, hindering support for service discovery, load balancing, and scheduling in cluster environments. Related issues: [RFC]: Add control panel support for vLLM #4873
Proposed Change.
1. Support Dynamically Loading or Unloading LoRA Adapters
To enhance flexibility and manageability, we propose introducing the ability to dynamically load and unload LoRA adapters at runtime.
/v1/add_adapter
and/v1/remove_adapter
inapi_server.py
.LoraRequest
, otherwise, we should let theengine
to load the lora vialora_manager
explicitly.2. Load LoRA Adapters from Remote Storage
Enabling LoRA adapters to be loaded from remote storage during runtime will simplify artifact management and deployment processes. The technical detail could be adding
get_adapter_absolute_path
,loral_local_path
tolocal_path
3. Build Better LoRA Model Lineage
To improve the visibility and management of LoRA models, we propose building a more robust model lineage metadata. This system will:
LoRAParserAction
to support json , we need to ask user to explicitly specify the base modelhttps://github.com/Jeffwan/vllm/blob/dd793d1de59b5efad25f4794b68cb935824c7a11/vllm/entrypoints/openai/cli_args.py#L16-L23BaseModelPath
to replaceserved_model_names
https://github.com/Jeffwan/vllm/blob/dd793d1de59b5efad25f4794b68cb935824c7a11/vllm/entrypoints/openai/serving_engine.py#L33. It would be great to pass the model path and model names separatelyshow_available_models
to updateroot
andparent
https://github.com/Jeffwan/vllm/blob/dd793d1de59b5efad25f4794b68cb935824c7a11/vllm/entrypoints/openai/serving_engine.py#L61-L624. Lora Observability enhancement
Improving observability by adding metrics specific to LoRA will help in better monitoring and management. Proposed metrics include:
5. Control Plane support(service discovery, load balancing, scheduling) for Loras
Since vLLM community focus more on the inference engine, the cluster level features would be a separate design I am working on in Kubernetes WG-Serving. I will link back to this issue shortly.
PR List
Feedback Period.
No response
CC List.
@simon-mo @Yard1
Any Other Things.
No response
The text was updated successfully, but these errors were encountered: