-
Notifications
You must be signed in to change notification settings - Fork 383
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CogVLM源代码是否支持多轮对话训练 #475
Comments
我用的grounding模型。 目的是实现以下样式的多轮对话: |
能否解答一下,非常感谢 |
支持,我们开源的数据集就有多轮对话,但是grounding 模型应该不行 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
多轮对话是采用这个吗 chat_old_history_to_prompt
如果数据标签为图文对:问1,答1,问2,答2
chat_old_history_to_prompt生成prompt=问1,答1,问2,预测结果与答2计算loss ?
这样的一条数据:问1,答1,问2,答2。要在网络里面训练几次?
第一次:训练 prompt=问1,第二次训练prompt=问1,答1,问2 ?
对于dataset.py有应该如何读取多轮对话数据标签:
使用sat格式微调,如何进行多轮对话训练。
主要修改哪部分代码可以实现多轮对话训练。
是要调用chat_old_history_to_prompt吗
应该基于哪个模型训练自有多轮对话数据集
多论对话和单轮对话数据集读取方面有差异吗
The text was updated successfully, but these errors were encountered: