-
-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: The output of Aria model is not correct #12241
Comments
I found that I can't load the model in transformers v4.48 if |
I've updated the hf repo. Can you give it a try? @DarkLight1337 |
Still doesn't work. Maybe we should just remove |
I've removed |
It works now, thanks! |
One bug I've found is that the vision tower's output is a tensor with all zero values. That is because the aria model uses vllm/vllm/model_executor/models/aria.py Lines 57 to 70 in b1d9205
The current checkpoint in hf repo after https://huggingface.co/rhymes-ai/Aria/discussions/11 includes the weights for The transformers impl of Aria avoids this issue by using the output before the Sadly, even after fixing the issue, the output is still nonsensical 😂 |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
The issue was observed in the comment posted by @DarkLight1337 #12207 (comment). I've tested the output of different versions of vllm and transformers and have some observations:
The inference code
The inference code
The inference code
cc @DarkLight1337 @Isotr0py
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: