Edge-Native MLLMs: Real-Time Threat Identification on Low-SWaP Devices
by Bo Layer, CTO | March 18, 2025

The modern battlefield is drowning in sensor data. The key isn't more data, but better, faster processing at the tactical edge. This article examines the latest breakthroughs in deploying multimodal large language models (MLLMs) directly onto low-power, edge-native hardware. I explore the model quantization techniques and architectural innovations that allow powerful AI to fuse imagery, signals, and text on a device the size of a credit card—providing instant threat identification without a fragile link to the cloud.
For years, the dream of AI has been tied to the cloud. We would collect massive amounts of data at the tactical edge, send it back to a massive data center for processing, and then wait for the answer. This model is slow, inefficient, and, in a contested environment, dangerously fragile. The future of military AI is not in the cloud; it's at the edge. It's about putting the power of AI directly into the hands of the warfighter, on small, low-power devices that can operate in a completely disconnected environment. And the key to this is a new generation of edge-native, multimodal large language models (MLLMs).
An MLLM is a type of AI that can understand and reason about multiple types of data at the same time. It can 'see' a video feed, 'listen' to a radio transmission, and 'read' a text message, and then fuse all of this information together to create a single, coherent picture of the battlespace. This is a level of situational awareness that has never before been possible at the tactical edge.
But running a powerful MLLM on a low-power, edge device is a monumental challenge. These models are massive, and they require a huge amount of computational power. That's why we are pioneering a suite of techniques for model compression and optimization. We are using techniques like quantization, which reduces the precision of the model's calculations without a significant loss in accuracy. We are using pruning, which removes the unnecessary connections in the neural network. And we are using knowledge distillation, which involves training a smaller, more efficient model to mimic the behavior of a larger, more powerful one.
The result is a new generation of MLLMs that are small enough and efficient enough to run on a device the size of a credit card. This will allow us to deploy powerful AI capabilities on everything from a soldier's helmet to a small drone. It will give our warfighters the ability to perform real-time threat identification, to fuse data from multiple sensors, and to make smarter, faster decisions in the most demanding environments.
This is the future of military AI. It's a future where the AI is not a distant, cloud-based service, but a trusted partner that is always there, always on, and always ready to help. It's a future that we are building today, at ROE Defense.