Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

different inference result #453

Open
xd2333 opened this issue May 12, 2024 · 2 comments
Open

different inference result #453

xd2333 opened this issue May 12, 2024 · 2 comments
Labels
currently fixing Am fixing now!

Comments

@xd2333
Copy link

xd2333 commented May 12, 2024

hi unslothai, i got different inference result when using unsloth, i'v tested qwen1.5-chat and tinyllama-chat and got same issue, generate by unsloth always get a bad result compare with transformers and dont know why

and here is my case:
https://colab.research.google.com/drive/1dxGKB-c3U8BYX-m2rQie8R12--0-JQMs?usp=sharing

@danielhanchen danielhanchen added the currently fixing Am fixing now! label May 13, 2024
@danielhanchen
Copy link
Contributor

You're correct! It seems like max_seq_length's default of 4096 is auto scaling TinyLlama, causing bad outputs - I'll fix this asap - thanks for the report!

@xd2333
Copy link
Author

xd2333 commented May 13, 2024

You're correct! It seems like max_seq_length's default of 4096 is auto scaling TinyLlama, causing bad outputs - I'll fix this asap - thanks for the report!

Hi unslothai, thx for fixing that! tinyllama-chat seems better not but i found Qwen1.5-7B-Chat still not well

and here is the case too:
https://colab.research.google.com/drive/1dxGKB-c3U8BYX-m2rQie8R12--0-JQMs?usp=sharing#scrollTo=47OE5BgPB6Wm

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
currently fixing Am fixing now!
Projects
None yet
Development

No branches or pull requests

2 participants