The project is an attempt to bridge the gap between programming a robot to move and allowing it to move for itself. With a neural network in place, movement is given a loose degree of flexibility, what the researchers are calling "chaos." Alter's arm movement, head and posture will adjust and change on the system's own volition. The neural network ticking behind the scenes offers multiple movement modes, switching between a longer movement mode and a more random "chaos" mode. The decision to switch is influenced by the sensors dotted around the base and take in what's happening around Alter: proximity, humidity, noise and temperature. These sensors operate like the robot's version of skin, copying our own senses, even if the system is far, far simpler. If the proximity sensors detect a lot of people nearby, for example, the torso shudders as the robot's body reacts to its environment.

Alta the robot isn't terrifying at all. #alta #robot A video posted by Mat (@thtmtsmth) on Jul 29, 2016 at 12:40am PDT

Alter also sings -- in a horrific, nightmare-inducing way. The haunting melody that comes from the machine are sine waves vocalizing how the robot's fingers move. (The team apparently tested other noises and melodies but decided to keep things simple for this early model.)

The theory behind the CPG is based on one of the simplest artificial models for neurons, the Izhikevich neuron, which reacts in a way that's called "spiking and burst behavior": Something builds up, and the robot's system creates a signal spike, which chains together with other neurons. Professor Ikeue from Tokyo University describes the central pattern generator as "coupled pendulums" -- one bumps into another into another and a movement in formed. While not an equal, balanced rhythm, this becomes Alter's own rhythm. The researchers didn't make the movement; the robot made it itself.

Osaka University's Kouhei Ogawa, who worked on previous humanoids at the Ishiguro lab, added: "This time, Alter doesn't look like a human. It doesn't really move like human. However, it certainly has a presence." It's true. It feels like there's something alive in there that's neither human nor robot. Movement seems random -- even if it's nonsensical.

"Until now," Ogawa says, "making androids talk or interact for 10 minutes was an incredible amount of hard work -- simply to program something to react for that long. Alter, moving for itself, can do so easily." The robot will be on display to the public for a week while the Tokyo and Osaka teams hope interactions will inspire new ideas on what they should teach Alter next.