Compare outputs from 3 locally hosted LLMs (via Ollama) or 3 cloud-hosted LLMs (via Groq API). This tool helps QA professionals validate AI outputs for test automation, bug triage, and more.
- Compare 3 LLMs at once (local or cloud)
- Toggle between local (Ollama) and cloud (Groq API) models in the UI
- Fully offline with Ollama, or use Groq for cloud LLMs
- Fast, lightweight, and private
- Simple, responsive UI with QA-focused tools
For Local LLMs:
- Ollama (for running models locally)
For Cloud LLMs:
- Groq API Key (free tier available)
- Node.js (for the API proxy server)
For Both:
- VS Code + Live Server extension
- Generate test cases from requirements
- Convert scenarios to Gherkin
- Summarize test failures or logs
- Generate assertion code from plain language
- Compare coverage across models
- Explain automation errors (stack traces, type errors)
git clone https://github.com/your-username/llm-qa-assistant.git
cd llm-qa-assistant
ollama pull phi3:medium-128k
ollama pull deepseek-r1:8b
ollama pull qwen:1.8b
ollama serve
Open index.html
with Live Server in VS Code.
Type a QA prompt and click "Compare Models".
☁️ Cloud LLMs (Groq API)
- Create a
.env
file in the project root:GROQ_API_KEY=your-groq-api-key-here
- Install dependencies:
npm install
- Start the Groq API proxy server:
npm start
- Open
index.html
with Live Server in VS Code. - Check the "Use Groq-hosted LLMs" box in the UI before comparing.