Skip to content
@langwatch

LangWatch

LangWatch 🏰

LangWatch Logo

Observe, Evaluate & Optimize your LLM performance LangWatch is an end-to-end evaluation and observability platform, helping teams ship their AI agents reliably and >8x faster!

Get started (free!) | Documentation | LangEvals Documentation

Welcome to LangWatch, the all-in-one open-source LLMops platform

LangWatch allows you to track, monitor, guardrail and evaluate your LLMs apps for measuring quality and alert on issues.

For domain experts, it allows you to easily sift through conversations, see topics being discussed and annotate and score messages for improvement in a collaborative manner with the development team.

For developers, it allows you to debug, build datasets, prompt engineer on the playground and run batch evaluations or DSPy experiments to continuously improve the product.

Finally, for the business, it allows you to track conversation metrics and give full user and quality analytics, cost tracking, build custom dashboards and even integrate it back on your own platform for reporting to your customers.

LangWatch Optimization Studio

You can sign up and already start the integration on our free tier by following the guides bellow:

🎬 Getting Started

Language of choice missing? 📧 Email us, or 💬 join our Discord and let us know!

Getting running

  1. Clone the relevant repository
  • git clone https://github.com/langwatch/langwatch.git
  • # or if you use the GitHub CLI
    gh repo clone langwatch/langwatch
  1. Follow the documentation for setup and usage
  2. Contribute by opening issues, submitting pull requests, or discussing ideas.

🔑 Key projects

  • LangWatch The core platform for LLM Ops, integrating monitoring, analytics, and optimization tools.
  • LangEvals A unified framework for evaluating language models, aggregating multiple scoring methods and LLM guardrails.
  • Docs Comprehensive documentation to help users set up and utilize LangWatch tools.

🎸 Demo

📺 Short video (3 min) for a sneak peak of LangWatch and a brief introduction to the concepts.

🤝 Contributing

Contributions are what make the open-source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.

Please read our Contribution Guidelines for details on our code of conduct, and the process for submitting pull requests.

🛟 Support

If you have questions or need help, join our community:

Popular repositories Loading

  1. langwatch langwatch Public

    The ultimate LLM Ops platform - Monitoring, Analytics, Evaluations, Datasets and Prompt Optimization ✨

    TypeScript 1.3k 76

  2. create-agent-app create-agent-app Public

    Helping you select an AI agent framework

    Python 95 6

  3. scenario scenario Public

    Agent testing library that uses an agent to test your agent

    Python 71 4

  4. langevals langevals Public

    LangEvals aggregates various language model evaluators into a single platform, providing a standard interface for a multitude of scores and LLM guardrails, for you to protect and benchmark your LLM…

    Jupyter Notebook 50 8

  5. data-simulator data-simulator Public

    Synthetic Data Generation

    Jupyter Notebook 4

  6. scenario-go scenario-go Public

    Agent testing library that uses an agent to test your agent, in Go.

    Go 2

Repositories

Showing 9 of 9 repositories
  • langwatch Public

    The ultimate LLM Ops platform - Monitoring, Analytics, Evaluations, Datasets and Prompt Optimization ✨

    langwatch/langwatch’s past year of commit activity
    TypeScript 1,291 76 14 (1 issue needs help) 16 Updated Apr 22, 2025
  • create-agent-app Public

    Helping you select an AI agent framework

    langwatch/create-agent-app’s past year of commit activity
    Python 95 MIT 6 0 0 Updated Apr 21, 2025
  • cookbooks Public

    example projects that use langwatch's features.

    langwatch/cookbooks’s past year of commit activity
    Jupyter Notebook 1 MIT 0 0 0 Updated Apr 21, 2025
  • scenario Public

    Agent testing library that uses an agent to test your agent

    langwatch/scenario’s past year of commit activity
    Python 71 4 0 1 Updated Apr 21, 2025
  • scenario-go Public

    Agent testing library that uses an agent to test your agent, in Go.

    langwatch/scenario-go’s past year of commit activity
    Go 2 MIT 0 0 0 Updated Apr 21, 2025
  • langevals Public

    LangEvals aggregates various language model evaluators into a single platform, providing a standard interface for a multitude of scores and LLM guardrails, for you to protect and benchmark your LLM models and pipelines.

    langwatch/langevals’s past year of commit activity
    Jupyter Notebook 50 MIT 8 2 (1 issue needs help) 2 Updated Apr 20, 2025
  • docs Public template Forked from mintlify/starter

    📖 The starter kit for your Mintlify docs

    langwatch/docs’s past year of commit activity
    MDX 0 272 0 1 Updated Apr 18, 2025
  • data-simulator Public

    Synthetic Data Generation

    langwatch/data-simulator’s past year of commit activity
    Jupyter Notebook 4 MIT 0 0 0 Updated Apr 15, 2025
  • .github Public
    langwatch/.github’s past year of commit activity
    0 0 0 0 Updated Mar 14, 2025

People

This organization has no public members. You must be a member to see who’s a part of this organization.

Most used topics

Loading…