Deepchecks Monitoring

Deepchecks Monitoring

Open Source Monitoring for AI & ML

4.9
โ€ข26 reviewsโ€ข

586 followers

Deepchecks Monitoring takes the open source testing experience all the way to production: enabling you to send data over time, explore system status and receive alerts on problems that arise over time.

This is the 3rd launch from Deepchecks Monitoring. View more

Deepchecks LLM Evaluation

Validate, monitor, and safeguard LLM-based apps
Continuously validate LLM-based applications including LLM hallucinations, performance metrics, and potential pitfalls throughout the entire lifecycle from pre-deployment and internal experimentation to production.๐Ÿš€
Deepchecks LLM Evaluation gallery image
Deepchecks LLM Evaluation gallery image
Deepchecks LLM Evaluation gallery image
Deepchecks LLM Evaluation gallery image
Deepchecks LLM Evaluation gallery image
Deepchecks LLM Evaluation gallery image
Free Options
Launch Team

What do you think? โ€ฆ

Shir Chorev
Thanks @kevin for hunting our LLM Evaluation solution ๐Ÿ˜Š ๐Ÿ‘‹ Hey, ProductHunt community I am Shir, co-founder and CTO of Deepchecks. At Deepchecks, weโ€™ve built a pretty special solution for LLM Evaluation and are thrilled to launch it today on ProductHunt! When we launched our open-source testing package last year, we quickly received an overwhelming response with over 3K stars ๐ŸŒŸ and more than 900K downloads! After the launch of our NLP package in June, we noticed that an incredible amount of the feedback calls we were having about the NLP package were asking for help with evaluating LLM-based apps. ๐Ÿคฏ After creating an initial POC and getting feedback from various companies, we gained the confidence we needed to dive deeply into the LLM Evaluation space. And yes, turns out itโ€™s a pretty big deal. ๐Ÿš€ As we began working on the LLM Evaluation module, weโ€™ve arrived at some important learnings that teams are struggling to figure out answers to these questions while deploying their LLM apps: - Is it good? ๐Ÿ‘ (accuracy, relevance, usefulness, grounded in context, etc.) - Is it not problematic? ๐Ÿ‘Ž (bias, toxicity, PII leakage, straying from company policy, etc.) - Evaluating and comparing versions (that differ in their prompts, basemodels, or any other change in the pipeline) - Efficiently building a process for automatically estimating the quality of the LLM interactions and annotating them - Deployment lifecycle management from experimentations/development, staging/beta testing, to production. Deepchecks LLM Evaluation solution helps with- โœ… Simply and clearly assess "How good is your LLM application?" ๐Ÿ”€ Track and compare different combinations of prompts, models, and code. ๐Ÿ” Gain direct visibility into the functioning of your LLM-based application. โš ๏ธ Reduce the risk during the deployment of LLM-based applications. ๐Ÿ› Simplify compliance with AI-related policies and regulations. We're also hosting a launch event today at 8.30 AM PST today, feel free to sign up to interact with the Deepchecks team and see a live demo: https://www.linkedin.com/events/... Apply for Deepcheks LLM evaluation access: https://deepchecks.com/solutions... ๐Ÿ˜Š Would appreciate any questions, and hope to see you there!
Eyal Shulman
Congrats on your launch ๐Ÿš€ great stuff!
Garen D Orchyan
@shirch Congratulations on the launch team, best of luck today โ™ฅ๏ธ๐Ÿฆ„
Hay Day
so promising...
philip tannor
@hay_day3 thanks my friend!
Divyansh Chaurasia๐Ÿ‘จ๐Ÿปโ€๐Ÿ’ป
Excited for the launch! ๐ŸŽ‰
Shir Chorev
@asdivyansh Such a pleasure to have you with us on this journey
philip tannor
@asdivyansh yup itโ€™s a big deal โค๏ธ
On Freund
I've been loving every release from this team. Can't wait to try this one out.
Shir Chorev
@on thanks so much! Looking forward to hear your thoughts :)
philip tannor
@on canโ€™t wait for the feedback!

Do you use Deepchecks Monitoring?

ยฉ 2025 Product Hunt