Skip to content

Latest commit

 

History

History
64 lines (40 loc) · 2.39 KB

README.md

File metadata and controls

64 lines (40 loc) · 2.39 KB

cover picture

Exploring interactions with LLMs : Practical insights with Spring AI

LLM (Large Language Model) concept is widely used to build chatbots, question-answering systems, and other conversational applications.

In this workshop, we will discover how to interact with a LLM as we craft a conversational agent step-by-step. Throughout this journey, we will understand the capabilities and limitations of LLMs, and explore prompt engineering techniques to elevate their standard behavior.

For the practical part, we will learn about Spring AI's chat completion features to interact with LLMs. Spring AI serves as an abstraction layer able to interact with several LLM providers. For the workshop's purposes, we will use Ollama, a solution for running LLMs locally.

🎁 Quick start

If you don't want to save your work, you can start the workshop directly in a cloud environment.

Open in GitHub Codespaces

Open in Gitpod

🔧 Environment setup

🔮 Install Mistral 7B model

Execute this command only once to pull the model.

docker exec -it ollama sh -c "ollama pull mistral:7b"

🏗️ Build the project

Execute this command after code update.

mvn clean install

🏃‍♀️ Run the application

mvn spring-boot:run

🚀 Hands-on

  1. Zero-shot prompting (15min)
  2. Conversational memory (15min)
  3. Information extraction (20min)
  4. Retrieval Augmented Generation (RAG) (30min)
  5. Conclusion

⭐ Contributors

This workshop is proposed by AXA France's Development Guild.

Maintainers:

  • Clément GIGUEL :octocat:
  • Lamine DIAKHABY :octocat: