The Association of the United States Army's 2025 Annual Meeting showcased a fundamental shift in ground warfare doctrine—autonomous systems aren't supplements anymore, they're core combat elements.
Walking the exhibit floor at AUSA 2025 this week, it's clear the conversation has moved past "should we integrate robots?" to "how fast can we field them?" The Army isn't experimenting anymore—they're writing doctrine, building formations, and signing contracts. And unlike the usual conference hype, this time there's actual hardware on display and realistic timelines attached.
The focus isn't on futuristic battlefield concepts. It's on deploying human-machine integrated formations by 2027-2028, with production contracts already in negotiation. This is the government contracting world I know: when acquisition timelines get this specific, money is about to move.
The Army's concept of Human-Machine Integrated Formations (HMIF) represents a fundamental doctrinal shift. We're not talking about adding a few unmanned systems to existing units. We're talking about restructuring formations from the ground up to integrate autonomous platforms as organic combat elements.
What This Actually Means:
At the company and battalion level, units will operate with:
The briefings emphasized ratio concepts: for every manned platform, units might operate 2-3 robotic systems. This isn't a technology demo—this is force structure planning.
The Robotic Combat Vehicle program has three variants, and for once, the Army seems to be learning from past acquisition mistakes by pursuing incremental capability rather than chasing perfection.
The RCV-Light is designed to operate ahead of formations, providing ISR and drawing fire to reveal enemy positions. It's expendable by design—better to lose a $2M robot than a $30M+ Bradley with four soldiers.
This is the workhorse of the HMIF concept. The RCV-M is meant to operate alongside manned infantry fighting vehicles, providing additional firepower while keeping soldiers out of direct fire zones.
The heavy variant has faced the most scrutiny. Do you need a tank-weight autonomous platform when you have actual tanks? The answer seems to be shifting toward specialized roles like breach operations where attrition is expected.
One of the more valuable sessions this week cut through vendor marketing to discuss actual autonomy levels being fielded:
Level 2-3 (Current Fielding):
Level 4 (2027-2028 Target):
Level 5 (Research Horizon):
The Army is being pragmatic here. They're not waiting for sci-fi autonomy. They're fielding Level 2-3 systems now and planning incremental capability upgrades. That's a sustainable acquisition approach.
Here's where the rubber really meets the road, and where I saw the most uncomfortable shifting in seats during Q&A sessions.
The Bandwidth Problem:
The Operator Burden:
The Interface Problem:
The most honest assessment came from a National Guard brigade commander: "We're going to field these systems, learn the hard way what works, and iterate. That's what we always do." That's military honesty, and probably accurate.
While RCVs dominated the exhibit hall, the most operationally significant autonomy might be in logistics:
Autonomous Convoy Vehicles:
Autonomous Resupply UAVs:
This is where autonomy makes immediate operational sense. Nobody needs a PhD in military ethics to approve an autonomous supply truck. Just program waypoints, execute mission, deliver ammo and water. The risk-reward calculation is obvious.
There were three separate panel discussions on "ethical AI" and "human-on-the-loop" requirements. What struck me wasn't what was said—it was what wasn't.
The Official Position:
The Unstated Reality:
One panel member, a former JAG officer, put it bluntly: "We're designing policy for the wars we wish we were fighting, not the wars we're actually preparing for."
That's the tension. The Pentagon's stated policy requires human authorization for lethal autonomous weapons. But the operational reality of contested environments with degraded communications and machine-speed threats is pushing toward greater autonomy.
My Take: The Army will field Level 3-4 autonomy with strict ROE requirements. Then, after the first high-intensity conflict where communications are disrupted and humans can't respond fast enough, policy will quietly evolve. We've seen this pattern before with armed UAVs—policy follows operational necessity, not the other way around.
The usual suspects are all over RCV development, but there are some interesting dynamics:
Traditional Primes:
Non-Traditional Entrants:
The Integration Layer:
What's interesting is the bifurcation: traditional primes build platforms, while non-traditional firms provide the autonomy brains. That creates integration risk but also potentially faster capability iteration.
The Army's HMIF concept is more aggressive than the Marine Corps' approach, which has focused on:
Allied Developments:
The U.S. Army's approach is the most ambitious in terms of formation-level integration. Whether that's visionary or overreach depends on whether they can solve the C2 and operator burden problems.
The Army is pursuing multiple parallel paths, which is smart but creates its own coordination challenges:
Middle Tier Acquisition (MTA):
Other Transaction Authority (OTA):
Traditional Programs of Record:
My read: The Army will use MTA/OTA to get initial capability fielded fast, then transition successful systems to programs of record for sustained production. It's a hybrid approach that makes sense if executed well—big "if."
Hardware development gets all the attention and funding. Training gets mentioned in passing. But here's the reality:
Current Force Has No Organic RCV Experience:
Simulation and Synthetic Training:
This is where the Army could fall short. You can throw money at platform development and accelerate timelines. You can't accelerate institutional knowledge development the same way. It took a decade+ to build true UAV operational expertise. RCV expertise will take similar time.
If you're in the defense technology space, here's what matters:
Software and Autonomy > Hardware Platforms: The differentiator isn't the vehicle chassis—it's the autonomy stack, sensor fusion, and AI decision-making algorithms. That's where the intellectual property and long-term value lives.
Integration is the Real Challenge: Platforms from multiple vendors need to work together in contested, degraded environments. If you can solve cross-platform interoperability and provide middleware that works, that's a sustainable business.
Sustainment Will Be Massive: Once these systems field, they'll need software updates, hardware refreshes, and continuous capability improvements. The sustainment contracts will dwarf the initial development contracts.
Don't Sleep on Simulation and Training: There's a massive opportunity in realistic RCV training systems. The Army hasn't prioritized it yet, but they'll need it desperately in 2-3 years.
Ethics and Policy Will Create Friction: Any company working on lethal autonomous systems needs lawyers who understand LOAC, DoD AI ethics principles, and how to navigate approval processes. Technical capability alone won't win contracts.
AUSA 2025 made it clear: human-machine integrated formations aren't a research project anymore. They're an acquisition program with timelines, funding, and operational urgency driven by near-peer competition.
Will the Army meet its 2027-2028 fielding timelines? Probably not completely—defense programs rarely do. But they'll field something, learn from it, and iterate. That's better than the analysis-paralysis that's killed previous modernization efforts.
The unresolved questions—bandwidth constraints, operator burden, autonomous engagement authorities—will be answered through operational necessity, not policy deliberation. That makes some people uncomfortable, and it should. But it's also the reality of how military technology evolves.
For those of us building technology in this space, the message is clear: the Army is buying. Build platforms that integrate easily, scale efficiently, and solve real operational problems. Skip the science fiction and focus on robust, mission-ready systems that work in degraded environments.
That's what wins contracts. That's what gets fielded. That's what matters.
Photo: Unsplash / [Military exercise simulation]
About the Author: Amyn Porbanderwala is Director of Innovation at Navaide, focusing on AI/ML integration for defense systems. He's a Marine Corps veteran who's worked on Navy ERP modernization, financial systems, and audit readiness. He builds mission-ready technology that lasts.