- Research on Open Source LLMs - Falcon/MPT/Llama etc
- Study LangChain
- Study how to fine tune LLMs - see LORA framework
- Pick variant based on:
- a. Size (First smaller variant e.g. 7b, then later larger 30b-e.g. Falcon-7b followed by Falcon-30B-b)
- b. Training mechanism: Instruction preferred, Completion may not meet our need (e.g. MPT-30B-Instruct)
- POC: Two main tracks:
- a. Finetuning: Fine tune using "corporate" data (we will use dummy data and build framework)
- b. ReAct framework: Thought -> Reasoning -> Action - iterative
- Implement a Proof of Concept (POC) with two main tracks:
- a. Finetuning: Fine-tune the selected LLM using "corporate" data (dummy data will be used initially).
- b. ReAct Framework: Explore the Thought -> Reasoning -> Action iterative framework using the LLM.
- bloom-1b
- langchain
##Frameworks
- LoRA
- Document loading (from pdf, txt,word etc.)
- Document embedding, using LLM embeddings
- Train the retrival model
- query the document