top of page

Instagram Tightens Up: New Feature to Shield Teens from Explicit Content

Instagram is well-known for its vibrant and engaging content, but with the growing concerns over inappropriate material, particularly among teenagers, the platform has taken a significant step forward. Recognizing the need to protect its younger users, Instagram has developed a new feature using machine learning technology to filter out explicit content such as nudity and other harmful material from being visible to teenagers.

The new system, which is currently being tested, utilizes on-device machine learning to analyze content and block it from appearing on the feeds of users under the age of 18. When users attempt to send photos or videos containing nudity, a warning alert prompts them to reconsider before sending. This proactive approach aims to curb the exposure of teens to such content. Criticism has been mounting from users in the United States and Europe about the rise of explicit content on the platform and its accessibility to the youth. With this new feature, Instagram is tightening its controls to ensure a safer environment for all its users. The feature will automatically activate for users under 18, while those over 18 will have the option to turn it on manually. This testing phase promises to lead to a broader rollout soon, ensuring that Instagram remains a safe and positive space for all ages.