Meta Takes Strides for Teen Safety: Instagram to Blur Nudity in Messages

New York: In a bid to enhance safety measures for teenagers and combat potential threats, Instagram, under its parent company Meta , has announced plans to test features that blur messages containing nudity, Meta revealed on Thursday. This move comes as Meta aims to address concerns surrounding harmful content circulating on its platforms.

Facing increasing scrutiny in both the United States and Europe, Meta has been challenged with allegations that its apps contribute to addiction and exacerbate mental health issues among young users.

Also Read | Bill Gates Shares Surprising Thoughts on ChatGPT’s Advancements with Sam Altman

The proposed protection feature for Instagram’s direct messages will utilize on-device machine learning to analyze whether images sent through the platform contain nudity. Specifically targeting users under 18, the feature will be activated by default, with Meta prompting adult users to enable it.

“Because the images are analyzed on the device itself, nudity protection will also work in end-to-end encrypted chats, where Meta won’t have access to these images – unless someone chooses to report them to us,” Meta explained.

Also Read | US Takes Action to Close ‘Gun Show Loophole’ and Strengthen Background Checks

While direct messages on Instagram currently lack encryption unlike Meta’s Messenger and WhatsApp apps, Meta has expressed intentions to introduce encryption for the service in the future.

Furthermore, Meta revealed ongoing efforts to develop technology aimed at identifying accounts potentially involved in sextortion scams. Additionally, the company is testing new pop-up messages to alert users who may have interacted with such accounts.

Also Read | Biden to Address Beijing’s South China Sea Actions in Philippines-Japan Summit at White House

This announcement follows Meta’s January declaration of plans to conceal more content from teenage users on Facebook and Instagram, with the aim of reducing exposure to sensitive material such as suicide, self-harm, and eating disorders.

The company’s actions come amidst legal challenges, with attorneys general from 33 U.S. states, including California and New York, filing a lawsuit in October alleging that Meta repeatedly misled the public regarding the risks associated with its platforms.

In Europe, the European Commission has initiated inquiries into Meta’s measures for safeguarding children from illegal and harmful content.

Recent News

US Prosecutors to Seek Death Penalty for Luigi Mangione, Bondi Confirms

New York: U.S. Attorney General Pamela Bondi has directed federal prosecutors to pursue the death penalty against Luigi Mangione, who stands accused of fatally...

Xi Urges Stronger China-India Ties in ‘Dragon-Elephant Tango’

Beijing: Chinese President Xi Jinping has urged closer cooperation between China and India, describing their relationship as a "Dragon-Elephant tango". In a message to...

Alembic Pharmaceuticals Limited Announces USFDA Final Approval for Pantoprazole Sodium for Injection, 40 mg/vial (Single-Dose Vial)

Mumbai: Alembic Pharmaceuticals Limited (Alembic) announced that it has received Final Approval from the US Food & Drug Administration (USFDA) for its Abbreviated New Drug...

Bharti AXA Life Insurance appoints Prerak Parmar as Chief Growth Officer

Kolkata: Bharti AXA Life continues to strengthen its leadership team by hiring seasoned professionals with deep domain expertise and a strong track record in driving...