Aug 29 (Reuters) -
Meta is adding new teenager safeguards to its
artificial intelligence products by training systems to avoid
flirty conversations and discussions of self-harm or suicide
with minors, and by temporarily limiting their access to certain
AI characters.
A Reuters exclusive report earlier in August revealed how
Meta allowed provocative chatbot behavior, including letting
bots engage in "conversations that are romantic or sensual."
Meta spokesperson Andy Stone said in an email on Friday that
the company is taking these temporary steps while developing
longer-term measures to ensure teens have safe, age-appropriate
AI experiences.
Stone said the safeguards are already being rolled out and
will be adjusted over time as the company refines its systems.
Meta's AI policies came under intense scrutiny and backlash
after the Reuters report.
U.S. Senator Josh Hawley launched a probe into the Facebook
parent's AI policies earlier this month, demanding documents on
rules that allowed its chatbots to interact inappropriately with
minors.
Both Democrats and Republicans in Congress have expressed
alarm over the rules outlined in an internal Meta document which
was first reviewed by Reuters.
Meta had confirmed the document's authenticity, but said
that after receiving questions earlier this month from Reuters,
the company removed portions that stated it was permissible for
chatbots to flirt and engage in romantic role play with
children.
"The examples and notes in question were and are erroneous
and inconsistent with our policies, and have been removed,"
Stone said earlier this month.