Newsletter
Get notified when new AI tools are added
Join the community.
Lunary is an observability and evaluation tool for AI applications built on large language models. It helps teams understand how their AI behaves in production by collecting key metrics, logs, and user behavior.
Lunary tracks model requests and responses, errors, and latency so developers can see where quality changes over time and where users get stuck.
Store and version prompts, compare performance, and run A/B tests to iterate faster on prompt wording and model configurations.
Combine automated and manual evaluations, label conversations, and analyze quality by scenario. Product metrics help connect LLM behavior to business outcomes.