Skip to content

badeok0716/llm_rag_chatbot

Repository files navigation

A Tutorial for LLM Chat Streaming via RAG

환경 설치

''' source install_env.sh '''

test vllm

''' conda activate vllm python test_vllm.py '''

RAG chatbot via langchain

''' python chat.py --modelname gpt-4o '''

Distillation

질문 생성

''' python make_questions.py '''

답변 생성

''' python make_answers.py '''

finetuning

''' python finetune_chat.py '''

TODO for efficiency

  • Implement .batch() in langchain ChatOpenAI to support batch inference in OpenAI API.
  • vLLM inference of unsloth finetuned model.

참고

About

llm_rag_chatbot

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published