-
-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add Internlm2 #2666
Add Internlm2 #2666
Conversation
Hi, @Leymore. Thanks for following my last work. And i also came up with the same implementation with you today. |
I am just a busboy and not familiar with the origin of this |
Hi there. In our effort to optimize training efficiency, we've consolidated the |
@gaoyang07 Thanks for your quick reply! But i meant the packed |
For more details on the conversion process, please refer to https://github.com/InternLM/InternLM/blob/3599ddd0e48968faced0831a4f32a44389d61d40/tools/convert2llama.py#L59-L70. Moreover, within our designed |
I cannot load model with merge this commit,the rror message is below: I have a nvidia RTX with 24G RAM,with same configuration I can load internlm-chat-7b and can load internlm2-chat-7b with transformers but cannot load it with vllm[this commit],any can help? |
I am not a expert on this issue, but I managed to run the following code with llm = LLM(model="internlm/internlm2-chat-7b", trust_remote_code=True, gpu_memory_utilization=0.25, max_model_len=2048) |
Hi, @esmeetu , recap this issue, is there anything more I need to do to have this PR merged? |
It works! thank you! |
I copied most of the codes from #2527, with the following changes:
einops
internlm2-chat-20b
Here are my test scripts and results:
For
internlm/internlm2-chat-7b
:For
internlm/internlm2-chat-20b
:For
internlm/internlm2-chat-20b
with TP=2: