Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Use e5-mistral-7b-instruct as body model for text classification #485

Open
matallanas opened this issue Feb 5, 2024 · 2 comments
Open

Comments

@matallanas
Copy link

I wanted to ask, if anyone has used the intfloat/e5-mistral-7b-instruct as base model for a multiclass classification task. I am trying to use but I have a problem with the GPU memory and I don't know if anyone has been able to use it properly.
Thank you so much in advance.

@tomaarsen
Copy link
Member

Hello!

I'm afraid that intfloat/e5-mistral-7b-instruct is not currently compatible with Sentence Transformers due to its left-padding tokenizer.
I also would not commonly recommend a model of this size for classification, especially because it'll take a lot of memory to finetune it.

A good alternative & strong model is BAAI/bge-large-en-v1.5.

  • Tom Aarsen

@matallanas
Copy link
Author

My problem is that I wanted to create a intent classification with prompts from a chat. The length of each text is variable and sometimes with more than 512 tokens and in different languages. That was reason to use intfloat/e5-mistral-7b-instruct, any other model that you can recommend for this problem?
Thanks again.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants