Hide table of contents

I’m trying to make actuator-boundary safety claims in embodied AI reproducible rather than anecdotal.

Once a model (or ROS stack, or planner) can issue actuator commands, failures become motion. At that point, prompts and policy layers aren’t really a safety boundary — what matters is what reaches the motor.

I don’t think we have a shared, reproducible way to both:

  • Enforce simple physical caps (velocity, acceleration, effort, position) independently of upstream logic
  • Generate artifacts that a third party can rerun to verify that enforcement actually worked

I’ve been prototyping a small hardware interposer (“Sentinel”) that sits between controller and actuator. The hardware itself isn’t especially novel. What I care about more is the standard + proof workflow around it.

The rough structure looks like this:

1) A machine-readable safety contract (SSC)

Defines units, modes (e.g., development vs. field), stop behavior, and required evidence fields.

2) A conformance harness

Includes enforcement tests, malformed traffic handling, and fuzzing / anti-wedge robustness.

3) Evidence packs

Artifacts containing config + build ID, enforcement outcomes, wedge counts, and latency distributions (P50/P95/P99 within a declared envelope), plus a simple verifier.

The goal isn’t “this makes robots safe.” It’s narrower than that.

The goal is: if someone claims actuator-level enforcement works, there should be a reproducible way to test and audit that claim.

Right now, robotics experimentation is easy. But safety evidence at the actuation boundary is rarely standardized or shared in a way that different labs could compare.

One way to frame this is epistemic rather than engineering: if enforcement claims cannot be independently rerun using shared tests and shared artifacts, then “safety” remains largely narrative.

My main uncertainties are:

  • Is this layer actually leverageful compared to perception or planning risks?
  • Would enforcement create nuisance stops that incentivize bypass?
  • Is “evidence pack” the right abstraction, or is there a better way to formalize reproducibility here?

I’ve put up a Manifund proposal to fund the first public milestone (SSC + harness + evidence tooling), but I’m more interested in whether this direction makes sense before pushing harder on it.

If you work on evaluation, governance, robotics, or safety standards, I’d appreciate thoughts — especially pointers to prior art I might be missing.

Happy to clarify technical details if helpful.

1

0
0

Reactions

0
0
Comments
No comments on this post yet.
Be the first to respond.
Curated and popular this week
Relevant opportunities