5 comments

  • sanghyunp55 minutes ago
    Runbook automation is one of those things every team says they'll build internally and never does. After 6 years on backend teams, our "runbooks" were always a Notion page nobody updated. The hard part is always the boundary between what can be automated and what still needs human judgment.
    • behat48 minutes ago
      Yes! That boundary between what can be automated and what still needs human judgement has shifted so much this last year. Things like &#x27;go check this dashboard&#x27; can now be automated.<p>ROI on runbooks (or good documentation in general) is much higher now if you have AI agents running them autonomously in the background. Makes it worth it to write&#x2F;maintain runbooks.
  • hrimfaxi2 hours ago
    How does this differ from cursor cloud agents where I can hook up MCPs, etc and even launch the agent in my own cloud to connect directly to internal hosts like dbs?
    • behat2 hours ago
      Thanks. Yeah, Cursor &#x2F; Claude code + MCP is powerful. We differentiate on two fronts, mainly:<p>1) Greater accuracy with our specialized tools: Most MCP tools allow agents to query data, or run *ql queries - this overwhelms context windows given the scale of telemetry data. Raw data is also not great for reasoning - we’ve designed our tools to ensure that models get data in the right format, enriched with statistical summaries, baselines, and correlation data, so LLMs can focus on reasoning.<p>2) Product UX: You’ll also find that text based outputs from general purpose agents are not sufficient for this task - our notebook UX offers a great way to visualize the underlying data so you can review and build trust with the AI.
      • hrimfaxi2 hours ago
        To be clear, are the main differentiators basically better built-in MCPs and better UX? Not knocking just trying to understand the differences.<p>I have had incredible success debugging issues by just hooking up Datadog MCP and giving agents access to it. Claude&#x2F;cursor don&#x27;t seem to have any issues pulling in the raw data they need in amounts that don&#x27;t overload their context.<p>Do you consider this a tool to be used in addition to something like cursor cloud agents or to replace it?
        • behat2 hours ago
          For the debugging workflow you described, we would be a standalone replacement for cursor or other agents. We don&#x27;t yet write code so can&#x27;t replace your cursor agents entirely.<p>Re: diffentiation - yes, faster, more accurate and more consistent. Partially because of better tools and UX, and partially because we anchor on runbooks. On-call engineers can quickly map out that the AI ran so-and-so steps, and here&#x27;s what it found for each, and here&#x27;s the time series graph that supports this.<p>Interesting that you have had great success with Datadog MCP. Do you mainly look at logs?
    • esafak1 hour ago
      They claim a 12% lead (from 36% to 48%) over Opus 4.6 in a RCA benchmark: <a href="https:&#x2F;&#x2F;www.relvy.ai&#x2F;blog&#x2F;relvy-improves-claude-accuracy-by-12pp-openrca-benchmark">https:&#x2F;&#x2F;www.relvy.ai&#x2F;blog&#x2F;relvy-improves-claude-accuracy-by-...</a>
      • behat1 hour ago
        heh, I was just about to post the following on your previous comment re: reproducible benchmark results. Thanks for posting the blog.<p>With the docker images that we offer, in theory, people can re-run the benchmark themselves with our agent. But we should document and make that easier.<p>At the end of it, you really would have to evaluate on your own production alerts. Hopefully the easy install + set up helps.
  • rishav1 hour ago
    Woohoo!!! Congrats on the big launch y&#x27;all
  • ramon1562 hours ago
    Congrats on the launch! I dig the concept, seems like a good tool :)
    • behat2 hours ago
      Thank you :)
  • adamsilvacons49 minutes ago
    [dead]