Case Study · Defense UX · LUXCE

When the Software
Gets in the Way,
the Mission Fails

Modernizing mission planning software for the US Navy and Air Force — and what 13 active duty pilots taught us about the cost of software friction.

UX Designer II
US Navy
NAS Oceana, VA
Unified Planner
Strike Package Alpha
Route Plan — Bravo
Threat COA — Eastern
NAS Oceana · UAG Session
SUS Score
61
13 Participants

NAS Oceana:
13 pilots, one classroom

I arrived at Naval Air Station Oceana in Virginia Beach, badged in with my CAC card, and met with the 2Circle team — a veteran-owned defense technology firm — to plan the day. Our goal was specific: we weren't testing usability in the traditional sense. We were testing discoverability.

Thirteen participants filed in. F-15, F-22, and E-2 pilots alongside intelligence officers. They were handed a high-stakes strike planning scenario and tasked with building a mission plan from scratch using the Unified Planner.

I sat directly behind the pilots as they worked. My job was to watch, not to help. I tracked every moment of friction — where they paused, where they backtracked, where they leaned forward looking for something that wasn't where they expected it.

One thing I didn't anticipate: how quickly the pilots apologized. When the software confused them, their first instinct was to assume they had done something wrong. We reassured them constantly. That dynamic alone told us something important.

Constraint
Closed facility — no internet-connected devices, no recording equipment allowed inside
Constraint
Pre-deployment software — SUS scores were the primary quantitative instrument available
Method
Think-aloud protocol with guided debrief — 13 participants across two plan cells
Direct Quote · Participant
"I have no idea how to navigate this platform."
— Active duty pilot, mid-session
Direct Quote · Participant
"If all I have to do is validate the routing... that's huge."
— On the ATO route automation feature
Researcher Note
"Lots of head nods during the brief — seemingly lots of understanding."
— Field note, pre-task observation

What 13 pilots told us
without saying it

1

Deletion was consistently broken

Users struggled to remove shapes, threats, and entities across multiple contexts. One participant renamed an item "delete me" as a workaround — then still couldn't find the deletion function and moved on.

"I want to delete this shape but might be deleting multiple other things from other users."

Deletion interaction patterns were flagged as a priority redesign. Right-click context menus and clearer destructive action affordances were scoped for Q1 implementation.

2

Platform logic competed with mission logic

Users were forced to learn the system's internal structure before they could begin planning. Prescribed step sequences didn't match how pilots actually think about building a mission.

"Why do we have to do this before I can do that? Can I skip this?"

Workflow flexibility became a core design principle — users needed to start from what they knew and fill in details as the picture developed, not follow a software-imposed sequence.

3

ATO automation was the session's standout moment

When automated route generation from ATO data worked, the reaction was immediate and genuine. The capability existed — but buried under friction that prevented most users from discovering it.

"That's awesome... wow. If all I have to do is validate the routing, that's huge."

The ATO import flow was identified as a high-value feature requiring a discoverability redesign — not a rebuild, but a clearer path to a capability users genuinely valued when they found it.

4

Blue and red force data should never be grouped

The system grouped friendly and threat COA data in ways that violated how operators mentally separate these categories. This wasn't a preference — it was a fundamental IA mismatch.

"I would never tie them together intuitively."

Blue force vs. red force separation was added to the IA backlog as a structural issue affecting multiple areas of the system — not just the COA builder.

5

Bandwidth constraints were an underweighted risk

Multiple users raised low-bandwidth deployment scenarios — ships, forward operating locations. Tools relying on heavy data transfer become effectively worthless at sea. This dimension wasn't accounted for in the design environment.

Bandwidth-sensitive design considerations were escalated as a pre-deployment risk — particularly for features relying on real-time data sync and map rendering.

6

The timeline tool needed a full redesign

The timeline was the single most consistently criticized element across all participants. Multiple users couldn't complete timeline-related tasks without help. 2–3 participants had nearly identical failure experiences.

"Without help, I would never have done it. Intent was great — but the set-up was painful."

The timeline tool was scoped for a full redesign in Q1 — one of two direct, named design actions that resulted from the NAS Oceana session findings.

A question that
reframed everything

The post-session guided discussion sharpened what observation had surfaced. The most pointed moment came from the 2Circle lead running the event.

Key Question

How many flight suits need to tell us before we act?

The debrief surfaced a pointed question that lingered: at what point does repeated user feedback become undeniable enough to change product direction? Thirteen participants surfaced the same themes independently. The signal was consistent.

What Worked

Collaboration and real-time visibility were genuinely valued

The ability to reference and build on another planner's work was called out positively. Real-time visibility into what other users were changing was meaningful. The ATO route automation, once discovered, was the session's clearest moment of delight.

Researcher Observation

Elite users blamed themselves, not the software

F-22 pilots apologized when the software confused them. This pattern — expert users assuming personal error rather than system failure — is a strong signal that feedback and affordances are not meeting the moment.

From findings
to roadmap

In a pre-deployment government contracting environment, impact looks different than a consumer product dashboard. The pilots who struggled at NAS Oceana influenced design decisions made in reviews weeks later.

Direct Actions from the Session

Timeline tool — full redesign scoped for Q1. The most consistently criticized element across all 13 participants. Named directly in the post-session report.

Map inconsistencies — dedicated task force formed to address cross-cutting rendering and interaction issues surfaced during the session.

New capabilities and user stories written based on synthesized findings, presented to the full LUXCE team and scoped for Q1 implementation.

Research Delivered

Observational notes compiled across all 13 participants covering navigation struggles, task failures, moments of delight, and verbatim quotes.

Guided debrief facilitated with active duty pilots — structured discussion to surface themes not visible during observation alone.

Findings synthesized and presented to the broader LUXCE cross-organizational team, with prioritized design implications and open questions flagged.

What this
project taught me

This project shaped how I think about the relationship between user expertise and software design in ways a consumer product project never could have. Elite users are often the worst served by complex software — because their expertise gets blamed for their confusion. When an F-22 pilot can't find a target selection tool, nobody assumes the software failed them.

It also taught me something about research craft. Restraint is a skill. Sitting behind a pilot who is visibly frustrated and not intervening — because their struggle is the data — requires discipline you only develop by doing it. The temptation to help is the enemy of insight.

Working within a cross-organizational team across MTI, MITRE, and Northrop Grumman — each with different processes, priorities, and communication styles — was its own education in navigating complexity. Good design in that environment isn't just about the work. It's about alignment, communication, and knowing when to push and when to listen.

UX Research Usability Testing Information Architecture Prototyping · Adobe XD UI Component Design Defense Software Cross-org Collaboration
Part of a larger body of work
LUXCE: Embedded UX in
Defense Software
The full story of 3.5 years embedded across the Unified Planner, LCAC debriefing software, Boeing Tapestry, and SBIR programs.
← Back to LUXCE Case Study