If you’ve played first-person shooters like Halo or Titanfall, you’ve seen a little bit of the potential future for troops and law enforcement on the ground. Those games give the player a virtual “heads-up” display of the game world around their avatar, with maps that show enemies and other essential data without having to look away from whatever’s down the barrel of their weapon.

The US military has been experimenting with heads-up data displays for years, but now the services are looking to cheaper alternatives to custom-built systems in the consumer electronics realm. In the not-too-distant future, when an Army squad leader wants to know what’s up ahead, he or she may reach for a smartphone or a Google Glass device instead of binoculars.

The US Army’s Maneuver Battle Lab at Fort Benning, Georgia, recently demonstrated one example of this adoption of off-the-shelf mobile and wearable tech. The Army is evaluating the use of off-the-shelf smartphone technology as a handheld “situational awareness tool” by using an app to connect to video feeds from robots on the ground and in the air.

Drones for eyes

The Small Unit Leader Situational Awareness Tool (SULSAT), an experiment of the Army’s Concepts Experimentation Program, got its first field demonstration at Fort Benning’s McKenna Military Operations Center on the Urban Terrain training area on May 13. Using a wireless connection to a tactical robot's or a small drone’s control system, SULSAT pulls video to a squad or platoon commander where they are—instead of them having to look over the shoulder of the operator or having to pass requests for instructions over a radio.

The $1 million SULSAT program, according to Army research and development budget documents, is part of an effort to better integrate robotics systems at the lowest level within the Army’s command structure—giving sergeants a pocketable tool that allows them to get a handle on what’s over the hill or around the corner. It can be used to clear routes for a squad or platoon, to identify potential improvised explosive devices, or just to get a handle on unfamiliar surroundings.

The main drawback of the current prototype is the limited range of its wireless connection. First Lieutenant Brandon Slusher, an officer with A Company, 1st Battalion, 29th Infantry Regiment, praised the tool, but told the Army News Service that the platoon commander still has to stay close to the tactical robotic control operator to get the video feeds.

Other goals of the program, beyond the live video feeds, include 3D virtual reality views built from air and ground sensors. In March, the Army posted a request for a system that can take in LiDAR and other sensor data from ground robots and drones to build a visualization of terrain, internal and external structures of buildings, and of potential threats—including automatic recognition of moving targets. This data would be pushed out to soldiers on their handheld devices.

Ok, Glass—bombs away

The Army isn’t the only service looking to give troops in the field a digitally augmented view of the world. The Air Force Research Laboratory’s 711th Human Performance Wing is testing a Google Glass-based system as part of a program with the improbable (and slightly forced) acronym BATMAN II—Battlefield Air Targeting Man Aided kNowledge (sic). AFRL obtained the Google Glass devices through Google’s Glass Explorer program.

There are two current target user populations for the BATMAN Glasses: the Air Force’s Pararescue jumpers (special forces operators with emergency medical training); and joint terminal air/attack controllers, the people on the ground who talk to attack aircraft and point to where the bombs should be dropped and hell otherwise unleashed.

AFRL is looking at how to use the sensors on Google Glass and its interface to help with tasks ranging from passing medical information back and forth to giving the look that kills—using where a person is looking to designate an air target. The original BATMAN platform was a ruggedized wearable computer system with a display monocle that jutted from a full headset. The AFRL has adopted Google Glass for the next phase of development both because of the reduction in weight of the device and its wealth of built-in sensors—as well as its dramatically lower price tag.

Google Glass is also being picked up on by DOD contractors for other potential uses. At the Special Operations Forces Industry Conference on May 20 in Tampa, Florida, the Battelle Memorial Institute—which operates the National Biodefense Analysis and Countermeasure Center and a number of other national labs—demonstrated Google Glass as part of a chemical-biological-radiological detection suite. The Glass apps, called Tactical Augmented Reality Applications (TARA), use video recognition and a wireless interface with chemical, biological, and radiological/nuclear sensors to give soldiers and other first responders a form of augmented reality—a heads-up display that presents information based on video analysis and sensor data.

In an interview with Ars, Darren Wolfe, of Battelle, said that the wireless interface to Glass could allow someone testing an area for potential hazards to quickly assess if there were hazardous agents present. “Instead of having to look at a readout, it gives them a simple color indicator to tell them if the instrument tests positive for an agent,” he said. “It will flash up red on the Glass if there’s a hazard.” This allows the person on scene to continue to focus on the task of containing or disposing of contaminated objects.

Another feature of TARA, useful on both the battlefield and in possible hazardous materials incidents, is its video recognition capability. The video from the Glass camera can be used with facial recognition software to alert soldiers or law enforcement officers when a person of interest is spotted, and to identify improvised explosive devices, hazardous chemical placards, and equipment potentially used for the construction of weapons of mass destruction. The Glass display can then provide a checklist of how to deal with the IED, or other items, based on visual analysis.

TARA can be used in combination with another application built with consumer-targeted hardware—the Oculus Rift virtual reality headset—as part of the Battelle Immersive Training Environment (BITE). Using Oculus’ built-in head sensors and a pair of handheld sensors, a user can interact with virtual objects in HAZMAT, IED, and WMD training scenarios and get simulated data from the Google Glass display.