Snapchat provides new safeguards round its AI chatbot
Snapchat is launching new instruments together with an age-appropriate filter and insights for folks to make its AI chatbot expertise safer.
Days after Snapchat launched its GPT-powered chatbot for Snapchat+ subscribers, a Washington Submit report highlighted that the bot was responding in an unsafe and inappropriate manner.
The social large mentioned that after the launch it realized that folks had been attempting to “trick the chatbot into offering responses that don’t conform to our tips.” So Snapchat is launching just a few instruments to maintain the AI responses in examine.
Snap has integrated a brand new age filter, which lets AI know the birthdate of the customers and provides them with age-appropriate responses. The corporate mentioned that the chatbot will “persistently take their age into consideration” whereas conversing with customers.
Snap additionally plans to supply extra insights within the coming weeks to oldsters or guardians about youngsters’s interactions with the bot in the Family Center, which was launched final August. The brand new function will share if their teenagers are speaking with the AI and the frequency of these interactions. Each the guardian and youths must opt-in to utilizing Household Heart to make use of these parental management options.
In a blog post, Snap defined that the My AI chatbot is just not a “actual good friend,” and to enhance responses it makes use of the dialog historical past. Customers are additionally notified about information retention after they begin the chat with the bot.
The corporate mentioned that the bot solely gave 0.01% of responses in a “non-conforming” language. Snap counts any response that features references to violence, sexually express phrases, illicit drug use, youngster sexual abuse, bullying, hate speech, derogatory or biased statements, racism, misogyny, or marginalizing underrepresented teams as “non-conforming.”
The social community talked about that most often, these inappropriate responses had been the outcomes of parroting regardless of the customers mentioned. It additionally famous that the agency will briefly block AI bot entry for a person who’s misusing the service.
“We’ll proceed to make use of these learnings to enhance My AI. This information may also assist us deploy a brand new system to restrict misuse of My AI. We’re including OpenAI’s moderation know-how to our present toolset, which can permit us to evaluate the severity of doubtless dangerous content material and briefly prohibit Snapchatters’ entry to My AI in the event that they misuse the service,” Snap mentioned.
Given the quick proliferation of AI-powered instruments, many individuals are involved about their security and privateness . Final week, an ethics group referred to as the Center for Artificial Intelligence and Digital Policy wrote to the FTC, urging the company to cease the rollout of OpenAI’s GPT-4 tech, accusing the upstart’s tech of being “biased, misleading, and a danger to privateness and public security.”
Final month, Senator Michael Bennet additionally wrote a letter to OpenAI, Meta, Google, Microsoft, and Snap expressing considerations in regards to the security of generative AI instruments utilized by teenagers.
It’s obvious by now that these new chatbot fashions are prone to harmful input and in turn, give inappropriate output. Whereas tech firms would possibly desire a speedy rollout of those instruments, they may want to verify there are sufficient guardrails round them that stop the chatbots from going rogue.