Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[CogVLM-chat-v1.1] LM weights are different with vicuna-7b-v1.5 #466

Open
minostauros opened this issue Apr 24, 2024 · 3 comments
Open

[CogVLM-chat-v1.1] LM weights are different with vicuna-7b-v1.5 #466

minostauros opened this issue Apr 24, 2024 · 3 comments
Assignees

Comments

@minostauros
Copy link

minostauros commented Apr 24, 2024

While CogVLM is trained, LM weights are fronzen.

From my observation however, the LM weights of cogvlm are different with Vicuna

Vicuna: https://huggingface.co/lmsys/vicuna-7b-v1.5/tree/main
CogVLM: cogvlm-chat-v1.1 (both from HF or SAT)

Can I ask why or the proper source of the language model?

  • CogVLM-Chat-v1.1 (SAT)
image
  • Vicuna-7B-v1.5
image
@zRzRzRzRzRzRzR zRzRzRzRzRzRzR self-assigned this Apr 25, 2024
@antigone660
Copy link

I have migrated the CogVLM-chat-hf to mindspore and found the modle works well when the input including image and text, but if there is only text query without image, the performance is not so good, it may relate to this issue I guess.
e88da6f7dd39b4446b47f4b5411003b

@minostauros
Copy link
Author

minostauros commented Apr 29, 2024

I have migrated the CogVLM-chat-hf to mindspore and found the modle works well when the input including image and text, but if there is only text query without image, the performance is not so good, it may relate to this issue I guess.

@antigone660 In text-only mode, the prompt template is different. Did you use the following prompt for text-only query?

text_only_template = "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: {} ASSISTANT:"

In my case, text-only mode works well regardless of this issue

@antigone660
Copy link

@minostauros Thanks for your reply I did not use the template before and it works now : )

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants