Datadog LLM Observability secures generative AI applications


Datadog announced LLM Observability, which allows AI application developers and ML engineers to efficiently monitor, improve and secure large language model (LLM) applications.

With LLM Observability, companies can accelerate the deployment of generative AI applications to production environments and scale them reliably.

Organizations across all industries are racing to release generative AI features in a cost-effective way, but implementing and bringing them to production can present several challenges due to the complexity of LLM chains, their non-deterministic nature and the security risks they pose.

Datadog LLM Observability helps customers overcome these challenges so they can confidently deploy and monitor their generative AI applications. This new product provides visibility into each step of the LLM chain to easily identify the root cause of errors and unexpected responses such as hallucinations.

Users can also monitor operational metrics like latency and token usage to optimize performance and cost, and can evaluate the quality of their AI applications—such as topic relevance or toxicity—and gain insights to mitigate security and privacy risks with out-of-the-box quality and safety evaluations.

Unlike traditional tools and point solutions, Datadog’s LLM Observability offers prompt and response clustering, seamless integration with Datadog Application Performance Monitoring (APM), and out-of-the-box evaluation and sensitive data scanning capabilities to enhance the performance, accuracy and security of generative AI applications while helping to keep data private and secure.

“WHOOP Coach is powered by the latest and greatest in LLM AI. Datadog’s LLM Observability allows our engineering teams to evaluate performance of model changes, monitor production performance and increase quality of Coach interactions. LLM Observability allows WHOOP to provide and maintain coaching for all our members 24/7,” said Bobby Johansen, Senior Director Software at WHOOP.

“The Datadog LLM Observability solution helps our team understand, debug and evaluate the usage and performance of our GenAI applications. With it, we are able to address real-world issues, including monitoring response quality to prevent negative interactions and performance degradations, while ensuring we are providing our end users with positive experiences,” said Kyle Triplett, VP of Product at AppFolio.

“There’s a rush to adopt new LLM-based technologies, but organizations of all sizes and industries are finding it difficult to do so in a way that is both cost effective and doesn’t negatively impact the end user experience,” said Yrieix Garnier, VP of Product at Datadog. “Datadog LLM Observability provides the deep visibility needed to help teams manage and understand performance, detect drifts or biases, and resolve issues before they have a significant impact on the business or end-user experience.”

LLM Observability helps organizations:

  • Evaluate inference quality: Visualize the quality and effectiveness of LLM applications’ conversations—such as failure to answer—to monitor any hallucinations, drifts and the overall experience of the apps’ end users.
  • Identify root causes: Quickly pinpoint the root cause of errors and failures in the LLM chain with full visibility into end-to-end traces for each user request.
  • Improve costs and performance: Efficiently monitor key operational metrics for applications across all major platforms—including OpenAI, Anthropic, Azure OpenAI, Amazon Bedrock, Vertex AI and more—in a unified dashboard to uncover opportunities for performance and cost optimization.
  • Protect against security threats: Safeguard applications against prompt hacking and help prevent leaks of sensitive data, such as PII, emails and IP addresses, using built-in security and privacy scanners powered by Datadog Sensitive Data Scanner.

Datadog LLM Observability is generally available now.



Source link