An artist’s rendition of a broad, synchronized cyber and psychological operations attack
An artist’s rendition of a broad, synchronized cyber and psychological operations attack. Image: AI by Gerardo Mena/Army University Press

Imagine a battlefield where AI predicts enemy movements before human planners even have time to react. This isn’t science fiction — it’s already happening in pockets of the US Department of Defense (DoD).

Yet, while AI has been in use for years, the Pentagon is only scratching the surface of what’s possible.

Historically, AI applications in the DoD have focused on math-based, deterministic models — tools that excel at specific tasks but don’t fully leverage the massive data flowing through military systems.

The next frontier lies with large language models (LLMs), which can digest vast information, provide predictive insights, and support both strategic planning and day-to-day operations like logistics and human resources.

But realizing this potential won’t happen automatically. Two critical steps must come first: establishing a trusted foundation for AI decision-making and overcoming cross-branch policy hurdles.

Step 1: Building Trust in AI Decisions

For the DoD, every decision must be grounded in data that leaders can understand and defend. While AI systems rely on data to generate insights, their internal decision-making processes are often opaque, producing conclusions without clear visibility into how they were reached.

In military environments, where the consequences of error are significant, this lack of transparency presents a real challenge. 

A machine cannot be allowed to determine mission-critical outcomes without meaningful human oversight. AI, and especially large language models, must operate within a broader decision-making framework where human judgment, reasoning, and accountability remain central.

In practice, this means integrating AI explainability directly into operational deployments. 

Tools such as SHAP (SHapley Additive exPlanations), LIME (Local Interpretable Model-agnostic Explanations), or custom audit layers embedded within LLM pipelines can help illuminate the factors influencing model outputs and provide leaders with greater confidence in AI-assisted recommendations.

Establishing a trusted AI foundation also requires clear operational guardrails. This includes implementing human-on-the-loop or human-in-the-loop protocols for mission-critical scenarios, requiring transparency documentation such as model cards and data sheets before deployment, and embedding interpretability dashboards into command interfaces so decision-makers can explore how inputs shape AI conclusions.

Equally important is investing in AI literacy across the force, enabling personnel to understand when to rely on AI outputs and when to challenge them.

Not every AI application will require the same level of scrutiny, but in any scenario where human lives are at stake, maintaining a human role in the decision-making process is essential.

US Army soldiers conduct drone test flights and software troubleshooting. Image: Micah Wilson/US Army

Step 2: Overcoming Policy Barriers Across Branches

Each military service operates with its own missions, priorities, and risk tolerances, and each must protect its data at the highest levels. 

While these safeguards are essential, they often result in limited data sharing across the department, constraining the effectiveness of AI systems that depend on broad, timely access to information.

Recent executive initiatives have called on federal agencies to redefine AI governance in ways that enable both innovation and safety. 

For the DoD, this presents an opportunity to align policies across services without compromising security. The department can establish guidelines that define what information should always be shared, what should be shared selectively, and what must remain isolated to support mission requirements.

Large language models cannot deliver value without access to data, and in operational environments, that data must often be real-time. 

Enabling LLMs to interact with trusted tools and live information sources is essential to unlocking their diagnostic and predictive capabilities.

At the same time, the tools that support LLM data collection and inference must meet the highest security standards and be capable of evolving as threats change.

This calls for a more standardized approach to AI security across the department. 

Core measures should include adopting Zero Trust Architecture as the baseline for all AI-enabled systems, embedding automated compliance checks aligned with frameworks such as NIST and the Risk Management Framework, and ensuring end-to-end encryption and data lineage tracking for both training and inference. 

Secure or air-gapped inference environments will remain necessary for handling classified information, while real-time monitoring and anomaly detection can help identify adversarial manipulation or model drift.

Regular red-teaming of AI systems should be treated as an operational necessity, not a one-time exercise.

Operational concept of AI-enabled human-agent teaming and machine learning. Image: US Army

The Path Forward

Our adversaries are already using AI, making the safe and effective adoption of these technologies essential to maintaining US security. 

Over the past five years, defense leadership has made meaningful progress in expanding AI use and approving new applications, but fully realizing its potential will require deliberate action.

By establishing trusted, transparent AI systems and consistent governance and security standards across the services, the Department of Defense can scale AI responsibly and accelerate innovation. 

The opportunity is clear: AI can enhance warfighting, optimize logistics, and improve decision-making at every level, but only if these foundational steps are taken first. 

The technology is ready. Now the department must be, too.


Headshot Greg Porter

Greg Porter is Principal Solutions Architect at Sev1Tech.


The views and opinions expressed here are those of the author and do not necessarily reflect the editorial position of Military AI.

You May Also Like

Moving Fast, Failing Faster: Why AI Governance Matters in Defense

Rushing AI and commercial software into defense systems without supply chain transparency turns modernization into a mission-threatening risk.

Why Trump Can’t Ignore China’s Digital Silk Road in the AI Race

While China uses AI to expand its global influence, Trump’s policies leave the US unprepared to compete in this new technological battlefield.

Congress Is About to Make a Big AI Mistake

The US should maintain open commercial channels for non-sensitive technology and keep the world dependent on its platforms.

America’s AI Advantage Relies on Leverage

In the AI race, leverage beats isolation and strategy outperforms broad restrictions.