-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Demo is broken #3
Comments
Problem located. THU-GLM has changed pos and attention mask in https://huggingface.co/THUDM/chatglm-6b/commit/373fd6b9d484841b490856a5570d6c450f20c22c Thus, swiching from latest impl would solve the issue # from modeling_chatglm import ChatGLMForConditionalGeneration
# from transformers import AutoTokenizer, GenerationConfig
# model = ChatGLMForConditionalGeneration.from_pretrained("THUDM/chatglm-6b").float()
# tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True)
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True)
model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True).float() |
The upstream is keeping up with breaking changes, while I wrongly setup my huggingface space without pinning the upstream. Will keep this issue updated. |
Have updated the repo but not fully assured everything is working now. As the model weights are not changed at all I am like 90% confident. Will keep this issue open for a while as I am testing with the new environment. |
It seems the huggingface's demo is broken now
After cloning the local and running some simple tests, the issues should be related to the attention / tokenizer (I guess?)
The text was updated successfully, but these errors were encountered: