Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss calculation always 0 #207

Open
sanipanwala opened this issue Feb 27, 2024 · 4 comments
Open

Loss calculation always 0 #207

sanipanwala opened this issue Feb 27, 2024 · 4 comments
Assignees

Comments

@sanipanwala
Copy link

Hello,

I'm trying to fine-tune the 34B model but during fine-tuning, I always get a loss 0. While I was able to fine-tune 7B and 13B models but not 34B.

Let me know if I'm overlooking this or please give me suggestions.

Thanks.

@jgehring
Copy link
Contributor

Hi @sanipanwala, we don't provide support for fine-tuning in this repository. Which tools are you using for this? Are you sure they support the 34B model well? The exact same setting works for 7B and 13B? In any case, a loss of 0 at the start of training is a good indication that something's going wrong.

@jgehring jgehring self-assigned this Feb 28, 2024
@sanipanwala
Copy link
Author

@jgehring I mean I'm using "codellama/CodeLlama-34b-hf" model and running a normal Python script and yes same configuration works with 7B and 13B.

Thanks.

@sssszh
Copy link

sssszh commented Apr 3, 2024

@sanipanwala
Hi, have you solved this problem yet?

I found the same problem when trying to peft fine-tune CodeLLama-7B (using LlamaForSequenceClassification), the Loss is always 0 during the fine-tuning.

Thanks!

@sanipanwala
Copy link
Author

Hi @sssszh ,

No, I haven't found any solution yet.

Thanks,
Sani

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants