Two sleek, futuristic unmanned aircraft flying above a body of water during sunset.

Calibrated Autonomy

Building autonomous systems that operators can trust in contested environments.

5

Active Projects in Calibrated Autonomy

The Focus

Autonomous systems deployed in degraded, denied, and chaotic environments must satisfy three properties simultaneously: they perform against mission objectives within their operational design domain, degrade gracefully when conditions shift outside it, and detect when they are operating beyond their competence boundary. Critically, these ODDs cannot be fully pre-specified — real operational environments are open-world, unstructured, and unpredictable. We believe systems must reason over open vocabularies and ingest unstructured data, not just perform within tightly enumerated conditions.

We are interested in frameworks that characterize and guarantee these properties from the subsystem to the system level. Performance and robustness make a system capable. Calibration makes it supervisable. Together, they are what make 1:N human command of autonomous systems possible.

Core Objectives

Contact Us

Seeing and sensing when satellites disappear

Fusing onboard perception to maintain precision in hostile signal environments

Autonomous aircraft cannot depend on GPS as a permanent crutch. In contested theaters, navigation systems must function under jamming, spoofing, and total denial. Visual-inertial navigation addresses this by tightly coupling camera-based perception with inertial measurement, building a continuously updated estimate of position and motion using onboard sensing alone. Instead of treating GPS as a primary reference, it becomes just one optional input among many — a convenience, not a requirement.

The technical challenge is maintaining long-term accuracy while operating under dynamic lighting, terrain ambiguity, vibration, and aggressive maneuvering. Robust navigation requires algorithms that reconcile noisy visual features with high-rate inertial data in real time, correcting drift without introducing instability. The system must remain computationally efficient, resilient to occlusion and adversarial conditions, and reliable across radically different environments — ocean, desert, urban terrain, and low-visibility weather — all without external infrastructure.

Turning commander intent into machine action

Bridging human language and autonomous mission logic

Operators communicate goals in terms of intent, priorities, and constraints — not in code. Translating that intent into executable behavior requires systems that can interpret ambiguous human language and map it to structured mission plans. The challenge is not simply understanding words, but capturing operational meaning: risk tolerance, sequencing, conditional objectives, and rules of engagement must survive the translation intact.

This demands architectures that combine language understanding with formal planning frameworks. The autonomy stack must transform high-level directives into verifiable action trees that adapt as conditions evolve. Misinterpretation is unacceptable; ambiguity must trigger clarification rather than silent failure. The system acts as an intelligent intermediary, ensuring that human strategy is faithfully represented in machine execution while preserving transparency and operator authority.

Swarms that think together without a single point of failure

Distributed intelligence for collaborative autonomy

Multi-vehicle operations cannot depend on a fragile central controller. In contested environments, communications degrade, nodes fail, and adversaries actively disrupt coordination. Decentralized planning distributes decision-making across agents so that the system remains functional even when partially disconnected. Each vehicle maintains local awareness while contributing to a shared mission objective.

The challenge lies in balancing independence with cooperation. Agents must negotiate task allocation, avoid conflicts, share intent, and replan dynamically without saturating limited bandwidth. Coordination algorithms must scale from small teams to large swarms while remaining predictable and verifiable. The result is a collective system that behaves coherently under pressure — resilient not because it avoids disruption, but because it is designed to absorb it.

Autonomy that degrades gracefully instead of collapsing

Maintaining mission capability under active opposition

Contested environments introduce uncertainty at every layer: sensing is corrupted, communications are intermittent, and adversaries adapt in real time. Resilient autonomy requires systems that anticipate degradation and continue operating within safe and effective bounds. Rather than assuming ideal conditions, the architecture must be built around failure tolerance and recovery.

This involves layered redundancy, adaptive planning, and self-monitoring behavior that detects when confidence drops below acceptable thresholds. The system must recognize when to continue autonomously, when to request human intervention, and when to fall back to conservative strategies. Resilience is not a single feature; it is a design philosophy that treats uncertainty as inevitable and prepares the autonomy stack to operate through it.

Keeping humans in command while machines carry the load

Designing collaboration instead of replacement

Autonomy succeeds only when it strengthens human decision-making rather than obscuring it. Effective teaming frameworks define clear boundaries between automated execution and human authority. The system must present intent, rationale, and predicted outcomes in a form operators can understand quickly, especially under stress. Transparency is not a luxury — it is a prerequisite for operational trust.

The technical challenge is building interaction models that scale with complexity. One operator managing multiple autonomous assets requires interfaces that compress information without hiding critical detail. The autonomy must anticipate operator needs, surface decisions at meaningful moments, and remain interruptible at all times. When designed correctly, the human and machine function as a single adaptive system, combining computational speed with strategic judgment.

Proving autonomy deserves confidence

Quantifying reliability before it reaches the field

Trust cannot be demanded; it must be measured. Autonomous systems require rigorous verification frameworks that evaluate behavior across simulated, synthetic, and real-world conditions. Metrics must capture not only performance but predictability, failure modes, and alignment with mission rules. The objective is to replace anecdotal confidence with quantifiable assurance.

Verification at this scale demands large simulation ecosystems, adversarial testing, and formal analysis methods that probe edge cases humans would never manually enumerate. Operators must be able to understand why the system behaves as it does and what its limits are. By formalizing trust as an engineering deliverable, autonomy transitions from experimental capability to deployable infrastructure.

Applications

Adaptive Airborne Enterprise (A2E)

Autonomous drone swarms operating 100+ nautical miles offshore to find, fix, track, and target surface vessels. Visual-based navigation enables operations in GPS-jammed environments. Decentralized planning allows swarms to replan dynamically under threat without a continuous datalink.

AI Copilots and Decision Assistants

AI-on-the-loop systems that synthesize information, reduce cognitive load, and provide actionable courses of action to pilots. Demonstrated in simulation and VR studies with quantitative improvements in decision quality. TPS Learjet flight test conducted in September 2025.

Multi-CCA Mission Command

Enabling a single mission commander to orchestrate 20+ collaborative combat aircraft by translating high-level intent into autonomous execution. The system handles routine tasks autonomously and queries the operator only when necessary.

Two F-15 fighter jets flying in formation above the wing of another aircraft.

What We're Looking For

The AI Studio doesn’t issue RFPs; instead, we rely on researchers and technologists working on problems in this space who see potential alignment to reach out to us. If you're working on something adjacent to these areas, reach out today!

GPS-denied navigation across diverse flight regimes and conditions

Natural language processing for intent translation

Multi-agent coordination without continuous communications

Resilient autonomy under adversarial actions

Human-machine teaming and trust frameworks

Methods to verify and validate autonomous behavior

Contact Us

Explore other lines of effort

Explore our Thesis

Simulation & Design for Mass

Use simulation to mitigate the design-to-fly loop and optimize design for first-shot deployment, reducing critical development timelines.

Two modern fighter jets overlaid with blue digital wireframe models against a clear blue sky.

Efficient Compute for AI at the Edge

Enable computationally expensive algorithms to run on embedded computers with resource/energy constraints.

Side view of a sleek drone with an open compartment revealing electronic components and dual propellers at the rear.

Frequently Asked Questions

What makes autonomy “trustworthy” in a defense context?

Trust is the bridge between commander's intent and machine execution. A system is trustworthy when it performs reliably in contested and denied environments – reacting to adversarial interference or environmental obstacles without needing to “phone home” for instructions.

How do you encode “Commander's Intent” into an autonomous system?

We focus on moving away from rigid, scripted behavior. Our research focuses on a system's ability to understand high-level mission objectives (the “what” and “why”) so it can autonomously determine the “how” in real time, even when communication with the operator is severed.

Why is autonomy necessary for “Air Power at Mass”?

Current systems often require multiple people to operate a single vehicle. To achieve mass, we must flip that ratio. Autonomy allows a single operator to command a fleet, preventing operator saturation and ensuring that the human remains a decision-maker at the strategic level, not a remote pilot at the tactical level.

How does the system handle adversarial actions without operator input?

Our focus is on resilient intelligence. We develop algorithms that enable the system to detect degradation in its own sensors or interference from an adversary and adapt its behavior to continue the mission or return safely, all without constant queries to a human controller.

Soldiers in camouflage uniforms and backpacks boarding a military transport aircraft on an airfield with helicopters in the background.

Operators

DAF-Stanford AI Studio exists to solve real operational problems alongside the people facing them. If you are encountering capability gaps at the edge and need autonomy that works under pressure, we are ready to partner with you. Our focus is to turn urgent field needs into deployable systems that directly strengthen the mission.

Contact Us
Stone colonnade at Stanford University with arched openings and decorative columns under a wooden-beamed ceiling with hanging lights.

Academia

DAF-Stanford AI Studio exists to translate frontier research into real operational impact. If you are advancing new methods in autonomy, AI, or simulation and want to see that work tested in high-stakes environments, we are ready to collaborate. Our goal is to move promising ideas from the lab into systems that matter.

Contact Us
Three engineers discussing a project inside a mechanical frame with electronic components and a steering wheel.

Industry

DAF-Stanford AI Studio exists to build deployable autonomy with people who solve hard technical problems. If you are developing AI systems that must perform under real-world constraints — power limits, contested environments, imperfect data — we are ready to work alongside you and turn that capability into operational systems.

Contact Us

Contact