4 comments

  • marshavoidance100 days ago
    this adds a tool to “grade” robot demo episodes by analyzing blur, collisions, and movement smoothness, then filters the bad ones out of the dataset. Seems like a pragmatic way to tackle the data quality problem in robotics would be great to see how much it moves the needle in real training runs.
    • machinelearning89 days ago
      Training is expensive so I wouldn&#x27;t necessarily call it &quot;pragmatic&quot;.<p>- the tool&#x27;s goal is actually to provide a lightweight, practical way to avoid wasting training cycles on bad data.<p>Evals for robotics are also expensive.<p>- validation loss is a poor proxy of robot performance because success is underconstrained by imitation learning data<p>- most robot evals today are either done in sim (which at best serves as a proxy) or by a human scoring success in the real world (which is expensive).<p>It&#x27;s great if you have evals and want to backtrack (we&#x27;re building tools for that too) but you definitely don&#x27;t want to discover you have bad data after all that effort (learned that the hard way, multiple times).<p>The metrics the tool scores vary from tedious to impossible for a human to sanity check so there&#x27;s some non-obvious practical value in automating some of it.
  • sherinjosephroy99 days ago
    [flagged]
  • fighterhao102 days ago
    [flagged]