Skip to content

ncdisrup-ai/Virtual-Assistant-LLAMA2-OpenAI-with-specific-knowledge-RAG-

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Virtual-Assistant LLAMA2/OpenAI with specific knowledge (RAG)

Open-Source RAG with LLaMa 13B (4 bits for less GPU memory), Faiss, HuggingFace and Langchain or with OpenAI. With gradio UI

🤔 What is this?

Description:
In this Poc we'll create a RAG Open-Source solution with Llama-13b-chat with HuggingFace embedings, Faiss (Vector DB), all orchestrated by LangChain. Or we could parametrize with OpenAI.With gradio UI.

In terms of struture of the solution, we have the main UI in file RAG_QAw_Parametrization.ipynb that import all the parametrization (which model, temperature, chain...) from parametrization.ipynb and the core RAG functions from RAGQA.ipynb. RAGQA.ipynb import also Parametrization.ipynb.

Retrieval Augmented Generation (RAG) is an advanced Natural Language Processing (NLP) technique that combines both retrieval and generation elements to enhance AI language models' capabilities.

📚 Data

We can import your own knowledge database with the solution.

🚀 Quick Install

Due to the power of GPU needed i advise you to use colab with RAG_QAw_Parametrization.ipynb(for Llama2). Need to insert your huggingface key

📖 Documentation / UI

🧮 Virtual Assistant Parametrization UI:

🧮 Virtual Assistant Question Anwsering UI:

Please see the description in .ipynb about this project.

🚀 Results

LLama2 with 4096 token on window prompt show a natural improvement in VA Open Source

About

Virtual-Assistant LLAMA2/OpenAI with specific knowledge (RAG) in gradio UI.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published