The application illustrates use of Langchain retrievers, generators and prompt templates to create a RAG pipeline LLAMA3 8B is used as the LLM to process the input query
The code is structured into 4 components :
- App/The driver code : Has the streamlit calls, along with instantiation of other components. The chat history is also preserved here in memory
- The Retriever : This creates the loader for various input formats such as documents,confluence, webpages etc converts the doccuments into word embeddings and then stores the same in a vector database for quick retrieval
- Generator : The generator is reponsible for creating the prompt using the chat history, query and data from retriever.
- LLM interface * : Currently kept in the most basic format , this will house the LLM APIs, access tokens etc.
- pip install -r requirements.txt
- Put relevant documents in data folder
- change app code to point to the appropiate path in the data folder
- streamlit run app.py
- Confluence links are removed for security reasons, you can use the confluence links of your organization to get it working with confluence data
- Disclaimer : This is still work in progress , so use with caution ;)
Ankur Neog /[email protected]