Official implementation for "GLaPE: Gold Label-agnostic Prompt Evaluation and Optimization for Large Language Models" (stay tuned & more will be updated)
-
Updated
Feb 6, 2024 - Python
Official implementation for "GLaPE: Gold Label-agnostic Prompt Evaluation and Optimization for Large Language Models" (stay tuned & more will be updated)
The prompt engineering, prompt management, and prompt evaluation tool for Python
The prompt engineering, prompt management, and prompt evaluation tool for TypeScript, JavaScript, and NodeJS.
The prompt engineering, prompt management, and prompt evaluation tool for Ruby.
The prompt engineering, prompt management, and prompt evaluation tool for Java.
An AI-driven system to automatically generate, evaluate, and rank prompts using Monte-Carlo and Elo Ranking system for enterprise-grade Retrieval Augmented Generation (RAG) systems.
The prompt engineering, prompt management, and prompt evaluation tool for Kotlin.
The prompt engineering, prompt management, and prompt evaluation tool for C# and .NET
A few prompts that I am storing in a repo for the purpose of running controlled experiments comparing and benchmarking different LLMs for defined use-cases
Runs two simple test prompts against 5 Anthropic models. Visually compares speed, capability, costs.
Add a description, image, and links to the prompt-evaluation topic page so that developers can more easily learn about it.
To associate your repository with the prompt-evaluation topic, visit your repo's landing page and select "manage topics."