Facebook on Monday said no users reported the video of the New Zealand mosque shootings while it was still live.

This was disputed by a reporter for Right Wing Watch, who said he was alerted to the live video and raised the alarm immediately.

"I definitely remember reporting this but there's no record of it in Facebook. It's very frustrating," Jared Holt told Business Insider.

The shooter streamed the attack on Facebook, and despite copies of the video spreading like wildfire, the company said the live broadcast was viewed fewer than 200 times.

Facebook has provided further details of its response to the video of the mosque shooting that left 50 people dead in Christchurch, New Zealand, on Friday.

The attack was streamed live on Facebook via what is thought to be a body-mounted camera. A link to the stream was posted on 8Chan, an imageboard site known for a being a hotbed of far-right extremist views.

In a blog post published Monday, Facebook said the live video was viewed fewer than 200 times. After the livestream ended, the video was viewed a further 3,800 times before being taken down, Facebook said.

Read more: YouTube's human moderators couldn't stem the deluge of Christchurch massacre videos, so YouTube benched them

Facebook said that no one reported the video while it was live and that the first user report came in 12 minutes after the broadcast ended, 29 minutes after it began.

This has been challenged. Jared Holt, a reporter for Right Wing Watch, said he was alerted to the livestream and reported it during the attack.

"I was sent a link to the 8chan post by someone who was scared shortly after it was posted. I followed the Facebook link shared in the post. It was mid-attack and it was horrifying. I reported it," Holt tweeted. "Either Facebook is lying or their system wasn't functioning properly."

Holt then checked and could find no record of his report on Facebook's internal tool for listing the reports users send off.

"I definitely remember reporting this but there's no record of it in Facebook. It's very frustrating," Holt told Business Insider. "I don't know that I believe Facebook would lie about this, especially given the fact law enforcement is likely asking them for info, but I'm so confused as to why the system appears not to have processed my flag."

Facebook declined to comment when contacted by Business Insider. It said on Sunday that it removed 1.5 million versions of the video that were uploaded to the platform.