In the Siggraph 2018 paper Learning to Dress: Synthesizing Human Dressing Motion via Deep Reinforcement Learning, a Georgia Institute of Technology/Google Brain research team describe how they taught body-shame to an AI, leaving it with an unstoppable compulsion to clothe itself before the frowning mien of God.



The AI uses machine learning tools to "automatically discover robust dressing techniques," and manages to train a robust get-dressed model despite the high computational expense of simulating cloth.





The paper's fascinating: the secret to getting an AI to get dressed is haptics — a sense of touch that is used to dynamically retune the AI's coordination to adjust to the rippling, slithering, treacherous textiles — the model incorporates the ripping point of the cloth and penalizes AIs that rent their garments asunder while clothing themselves.



In this task, a t-shirt is initialized on the character's shoulders with

the character's neck contained within the collar. To randomize the

initial garment state, we apply a random impulse force with fixed

magnitude to all the garment vertices at the beginning of the simu-

lation. We allow the garment to settle for 1s before the character

begins to move. The first control policy completes the task of moving the right

end effector into gripping range of the specified grip feature. The

policy attempts to match a given position and orientation target

in the garment feature space. Once the error threshold is reached,

control transitions to an alignment policy designed to "tuck" the

left end effector and forearm under the waist feature of the garment

in preparation for dressing the arm. This policy attempts to contain

the arm within the triangle formed by the gripping hand and the

shoulders. This heuristic approximates the opening of the garment

waist feature. In addition, this policy is rewarded for contact with

the garment interior and penalized for geodesic distance from a

selected point on the interior of the garment. Once interior contact

is detected, and the arm is within the heuristic triangle, control is

transitioned to the left sleeve dressing controller which attempts to

minimize the end effector contact geodesic distance from the end

feature of the sleeve and maximize the containment depth of the arm

within the sleeve entrance feature. A task vector is provided which

indicates the direction the end effector should move to decrease its

contact geodesic distance (or points to the garment feature if not in contact). Once the limb has passed a threshold distance through

the sleeve, the re-grip controller directs the hands together into

position to exchange grip from right hand to left. Once the left

hand is within a threshold distance of its gripping target, the grip

exchange is triggered and control is transitioned to the second "tuck"

control policy with the same purpose and transition criteria as the

first. The second sleeve policy is then run to pass the right arm

through the right sleeve. At this point, the seventh and final policy

is used to guide the character back to its start pose while avoiding

garment tearing.



Learning to Dress: Synthesizing Human Dressing Motion via Deep

Reinforcement Learning [Alexander Clegg, Wenhao Yu, Jie Tan, C. Karen Liu And Greg Turk/ACM Transactions on Graphics, Vol. 37, No. 6]

(via JWZ)