It might be easier to build an AI that deserves our sympathy than to build an AI that is aligned with us. Is that a plausible plan B?