Overview
- Starting next week, notifications will roll out in the United States, United Kingdom, Australia and Canada, with expansion to more countries by year-end.
- Alerts will be sent to parents enrolled in Instagram’s supervision program via email, SMS, WhatsApp and in-app notifications, and they include expert guidance.
- Instagram already blocks these searches on teen accounts and redirects users to resources, and the new alerts flag persistent attempts to find such content.
- Meta says the alert threshold was set with its Suicide and Self-Harm Advisory Group to balance catching real risk against over-notifying families.
- The announcement comes during active cases in Los Angeles and New Mexico, as filings cite internal data (>19% of 13–15-year-olds saw unwanted sexual images; ~9% saw self-harm content) and Zuckerberg acknowledges age verification came too slowly.