Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question]: I am using Question Generator feature from llama-index , i want to know which LLM is best for this ? #13121

Open
1 task done
nikhitaKanoj opened this issue Apr 26, 2024 · 3 comments
Labels
question Further information is requested

Comments

@nikhitaKanoj
Copy link

Question Validation

  • I have searched both the documentation and discord for an answer.

Question

I am using Question generator from llama index , here they use Open AI LLM , I want to use open source LLM , which is preferred? Interms of getting high accurate questions from the document uploaded :
https://docs.llamaindex.ai/en/stable/examples/evaluation/QuestionGeneration/

@nikhitaKanoj nikhitaKanoj added the question Further information is requested label Apr 26, 2024
Copy link

dosubot bot commented Apr 26, 2024

For generating high-accuracy questions from documents using an open-source Large Language Model (LLM), there isn't a one-size-fits-all recommendation. However, based on community discussions and issues within the LlamaIndex repository, models like GPT-J, OPT, and FLAN-T5 have been explored, with FLAN-T5 (in its xl and xxl versions) showing promising results despite limitations on input length. It's important to note that the effectiveness of these models can vary based on the specific requirements of your application, such as the need for parsing output, generating questions with guidance, or handling sub-questions.

The choice of an open-source LLM should consider the functionalities you're leveraging in LlamaIndex's Question Generator, including the ability to generate, parse, and refine questions. As the landscape of open-source LLMs is rapidly evolving, keeping an eye on the latest developments and community discussions within repositories like LlamaIndex can provide insights into new and potentially more capable models as they emerge.

If you have specific features or requirements in mind for the LLM in your question generation application, sharing those might help narrow down the options further.

Details

Sources


About Dosu
This response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research.

Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant.

To continue the conversation, mention @dosu.

@nikhitaKanoj
Copy link
Author

@dosu, My application is : When the Document is uploaded, all the possible questions has to be listed from that document.
I used Nous-Hemers-Llama2 13B , the performance is very poor , whereas ofcource OpenAI LLM was veyr much accurate ,
I want to achieve atleast nearly accurate questions using Open source LLM's .

Copy link

dosubot bot commented Apr 26, 2024

For enhancing the accuracy of question generation from documents using open-source LLMs, considering your experience with Nous-Hemers-Llama2 13B, you might want to explore other models like GPT-J, OPT, or FLAN-T5. These models have shown promise in various applications, though direct comparisons with OpenAI's models in terms of question generation accuracy might be challenging due to the proprietary nature of OpenAI's technology.

To potentially achieve near OpenAI-level accuracy with open-source models, you could experiment with adjusting the configurations specific to your application's needs. This includes optimizing the input data processing, tweaking the model parameters, and refining the post-generation question filtering process. Leveraging the latest versions of these models and staying updated with the community's findings can also provide incremental improvements.

The LlamaIndex's Question Generator's design, with features like customizable prompt templates and output parsing, supports the integration of various LLMs. By experimenting with different models and configurations, you can identify the best fit for your application's requirements.

Remember, the effectiveness of an LLM in question generation can significantly depend on how well it's tuned for the specific type of documents and questions you're targeting. Continuous testing and iteration will be key to improving question generation accuracy with open-source LLMs.

Details

Sources


About Dosu
This response is meant to be useful and save you time. It is not meant to be a precise solution, but rather a starting point for your own research.

Help me be more useful! Please leave a 👍 if this is helpful and 👎 if it is irrelevant.

To continue the conversation, mention @dosu.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

1 participant