SentryML
// SentryML

ML observability & MLOps — model monitoring, drift detection, debugging in production.

Engineering-focused coverage of ML observability and MLOps. Model monitoring, drift detection, training/serving skew, debugging production model failures, evaluation pipelines, and the tooling that actually works at scale.

MITRE ATLAS LLM detection runbook visualization
// Featured experiment

Detection Engineering for LLM Apps: A MITRE ATLAS-Mapped Runbook for Prompt Injection Alerting

Mapping LLM application telemetry to MITRE ATLAS techniques. Concrete log shapes, alerting heuristics, and a runbook structure that scales beyond ad-hoc grep rules.

May 7, 2026 [defense]
monitoring

A Lean 4 stability proof for tool-mediated LLM agents, and what it means for your runbook

A new arXiv paper certifies controllability and ISS robustness for an LLM-driven SOC agent using Lean 4. The MLOps takeaway is simpler than the math: monitor the action catalog, not the model.

deep-dive

The Authority Gap Is an Observability Problem: What MLOps Teams Should Actually Instrument

Orchid Security's framing of agent governance as a delegation problem lands in the lap of ML observability teams. The instrumentation we already own decides whether the authority graph is real or theatre.

monitoring

Embedding-Based Agent Monitoring Has a Blind Spot. Here's What to Watch Instead.

A new paper demonstrates three attack patterns — Slow Drift, Benign Wrapper, Chaos Seeding — that defeat embedding-based detection of malicious agents in LLM multi-agent systems. The fix requires monitoring logit-level confidence, not just output embeddings.

// Earlier entries

Subscribe

SentryML — in your inbox

ML observability & MLOps — model monitoring, drift detection, debugging in production. — delivered when there's something worth your inbox.

No spam. Unsubscribe anytime.