Latest Tech News

Stay updated with the latest in technology, AI, cybersecurity, and more

Filtered by: evaluation Clear Filter

Launch HN: Confident AI (YC W25) – Open-source evaluation framework for LLM apps

Hi HN - we're Jeffrey and Kritin, and we're building Confident AI ( https://confident-ai.com ). This is the cloud platform for DeepEval ( https://github.com/confident-ai/deepeval ), our open-source package that helps engineers evaluate and unit-test LLM applications. Think Pytest for LLMs. We spent the past year building DeepEval with the goal of providing the best LLM evaluation developer experience, growing it to run over 600K evaluations daily in CI/CD pipelines of enterprises like BCG, Astr