Credit: Pixabay/CC0 Public Domain

Facebook on Wednesday beefed up automated tools to assist group moderators striving to keep exchanges civil in a time of clashing viewpoints.

The leading social network hammered by critics over vitriolic content in news feeds has played up groups as enclaves where people with conflicting views on hot-button issues can bond based on share interests from music to hobbies or parenthood.

"Some of these groups are millions of people," Facebook engineering vice president Tom Alison said while briefing AFP on for moderators.

More than 1.8 billion people use groups every month, and there are more than 70 million administrators working to "keep conversations healthy" in the forums, according to Alison.

While much content is not public in groups, Facebook has been looking at ways to control hateful and abusive content in these forums.

Automated systems at Facebook already check for posts in groups that violate the social network's rules about what content is acceptable.

A new "Admin Assist" feature lets moderators set criteria for what is considered acceptable at the group and then have posts or comments automatically checked for violations.

Moderators can also employ software to eliminate comments with links to unwanted content; slow down heated conversations to let heads cool, or require that people be members of the social network or a group for a while before being able to join in conversations.

"What these tools do is automate things admins did manually, but not expose anything they didn't have access to before," Alison said.

Facebook is testing that can watch for indications of conversations getting nasty, perhaps with rapid-fire replies or controversial content involved, then alert moderators.

"The AI looks at things associated with threads that have conflicts," Alison said.

"Some admins welcome people having debates; others don't want contentious conversations."

A group called Dads With Daughters by nonprofit Fathering Together is among those given early access to test the tools, according to moderator Brian Anderson.

The online community for fathers sharing advice, resources, and pride in raising daughters has more than 127,000 members.

"In the beginning there was really nothing, you just put rules up there and what a group is about," Anderson said of moderation in the early days.

Tools added by Facebook have reduce the number of moderators needed at the group, according to Anderson.

Hard stands by moderators against "toxic masculine tropes" such as toting shotguns to protect daughters from suitors was credited with helping establish the tenor in the community.

"You can tell the groups that are really putting in the effort to keep it a civil space versus posts that just get nasty right away."