Skip to content

Falcon LLM with Chat UI using LangChain and Chainlit

Notifications You must be signed in to change notification settings

progone/falcon-langchain

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Falcon LLM with LangChain Chat UI

This repository contains the necessary files and instructions to run Falcon LLM 7b with LangChain and interact with a chat user interface using Chainlit. Follow the steps below to set up and run the chat UI.

Prerequisites

  • Python 3.10 or higher
  • Operating System: macOS or Linux

Steps to Run the Chat UI

  1. Fork this repository or create a code space in GitHub.

  2. Install the required Python packages by running the following command in your terminal:

    pip install -r requirements.txt
    
  3. Create a .env file in the project directory. You can use the example.env file as a reference. Add your Hugging Face API token to the .env file in the following format:

    HUGGINGFACEHUB_API_TOKEN=your_huggingface_token
    
  4. Run the following command in your terminal to start the chat UI:

    chainlit run app.py -w
    

    This will launch the chat UI, allowing you to interact with the Falcon LLM model using LangChain.

Note: Ensure that you have provided a valid Hugging Face API token in the .env file, as mentioned in step 3. Without a valid token, the chat UI will not function properly.

If you encounter any issues or have questions, please reach out to me on Twitter

Enjoy using Falcon LLM with LangChain!

About

Falcon LLM with Chat UI using LangChain and Chainlit

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 80.8%
  • Python 19.2%