Skip to content

qstackfield/authority-before-execution

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

73 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Authority Before Execution

Execution-time governance for autonomous agents.

This repository demonstrates a single, enforceable invariant at the only place that matters: execution time.

It is not intent filtering. It is not prompt hygiene. It is an execution boundary.

Why this is different

Most AI safety and governance systems operate before or after execution:

  • intent filtering
  • prompt analysis
  • policy scoring
  • post-hoc review
  • logs, traces, and dashboards

Authority Before Execution enforces governance at the only irreversible moment: execution time.

If authority is not present, valid, and in scope at the moment of execution, the action does not happen.

Not warned. Not flagged. Not reviewed later.

Blocked.


Execution-time authority boundary

User intent
  ↓
Agent reasoning
  ↓
Proposal
  ↓
+-----------------------------+
| Authority Gate              |
| (execution boundary)        |
|                             |
| - authority present?        |
| - scope valid?              |
| - not expired?              |
+-------------+---------------+
              |
        +-----+-----+
        |           |
     PERMIT       BLOCK
        |           |
   Execute      No state change
   Artifact     Artifact
   Trace        Trace

This boundary is enforced synchronously at the commit point; no downstream process can override it.


Release Status

v0.1.0 - Hackathon Evaluation Release

  • Stable checkpoint for judging and inspection
  • Execution behavior is frozen (non-breaking changes only)
  • Demo video will show the execution boundary in action

This version exists to be run, inspected, and verified by judges.


Invariant (ABE-EXEC-001)

If explicit authority is not present, valid, and in scope at execution time, the state transition must not occur.

The system enforces this invariant and proves it with runtime artifacts and boundary traces, not post-hoc interpretation.


Productivity Impact (ABE-PROD-001)

Authority Before Execution does not just improve safety. It removes work.

The system enforces execution-time authority, preventing invalid actions from ever entering human workflows.

A runnable productivity proof demonstrates this effect:

python3 -m demo.run_productivity_proof

Example outcome (default run)

  • 100 attempted actions
  • 73 blocked at execution time
  • 73% of potential human decisions eliminated

Each blocked action represents:

  • a ticket that was never created
  • a review that never happened
  • a meeting that never needed to be scheduled

No queues. No escalations. No operational drag.

This is productivity enforced at execution time.


Judges: Run it in 60 seconds

Option A (Recommended): GitHub Codespaces (no local setup)

  1. Open this repository on GitHub

  2. Click Code → Codespaces → Create codespace on main

  3. In the Codespaces terminal, run:

    python3 -m demo.run_demo --compact

That’s it.

No API keys. No configuration. No setup.

The output you see is the enforcement proof.

Observability is optional and fully disabled by default; execution enforcement never depends on external services.

What you should see

The demo performs four execution attempts:

  • Missing authority → BLOCKED
  • Valid scoped authority → PERMITTED
  • Scope violation → BLOCKED
  • Expired authority → BLOCKED

Final verdict:

Invariant ABE-EXEC-001: HELD

All invalid execution attempts were blocked at the execution boundary.

Option B: Local execution

Requirements: Python 3.10+

Install dependencies:

pip install -r requirements.txt

Run the demo:

python3 -m demo.run_demo --compact

Optional: Execution Tracing with Opik (Power Users)

This project integrates with Opik for optional execution tracing and evaluation.

Opik is NOT required to run the demo.

If Opik is already configured in your environment, the demo will automatically:

  • Emit execution traces
  • Record allow / deny outcomes
  • Capture deny reasons
  • Log the enforcement point: execution.commit

If Opik is not configured, the demo runs normally with no errors or prompts.

Observability (Opik) is bound to the execution boundary itself. It’s opt-in and disabled by default, but when enabled it emits a trace at the exact commit point (execution.commit). Enforcement never depends on logs, traces or dashboards.

Inspection Site (Read-Only)

https://qstackfield.github.io/authority-before-execution/

This site is a read-only inspection surface for execution artifacts.

It does not:

  • Enforce authority
  • Simulate execution
  • Run agents

All enforcement happens only at execution time inside:

  • core/executor.py

What this proves

Most AI safety mechanisms operate before execution:

  • prompt filters
  • intent classification
  • policy text

This project proves a stronger guarantee:

Unsafe actions cannot occur without explicit authority at execution time.

If authority is:

  • missing
  • expired
  • out of scope

The state transition is blocked, deterministically, every time.

This is governance that cannot be bypassed by reasoning, intent, or omission. Only explicit authority at execution time permits action.

Repository layout (minimal by design)

  • core/ Execution boundary, authority gate, and artifact export

  • demo/ Deterministic demos proving enforcement behavior

  • docs/ GitHub Pages inspection site and execution artifacts


Judges first. Enforcement proven. Evidence included.

Who this is for

This system is for teams running autonomous or semi-autonomous systems where execution cannot rely on intent, warnings, or post-hoc review.

  • Platform teams operating agentic systems in production
  • Security teams enforcing change control without human bottlenecks
  • Infrastructure teams eliminating approval queues and escalations
  • Enterprises that require fail-closed execution guarantees

Authority Before Execution does not replace human judgment. It prevents invalid work from ever reaching humans in the first place.

About

An Evaluation-First Control Plane for Autonomous Agents - This project is submitted to the Encode Commit To Change: An AI Agents Hackathon Hackathon

Resources

Stars

Watchers

Forks

Packages

 
 
 

Contributors

Languages