"We don’t want this kind of behavior on Facebook — and we’re investing heavily in both people and technology to keep bad content off our services. We have more than doubled the number of people working on safety and security this year and now have more than 20,000 people around the world. We use reports from our community and technology like machine learning and artificial intelligence to detect bad behavior and take action more quickly," it said.