Australia’s online watchdog has criticised the world’s biggest social platforms of failing to properly enforce the country’s ban on under-16s using their platforms, despite legislation that came into force in December. The eSafety Commissioner, Julie Inman Grant, has expressed “significant concerns” about adherence by Facebook, Instagram, Snapchat, TikTok and YouTube, citing poor practices including allowing banned users to repeatedly attempt age verification and insufficient measures to prevent new accounts. In its initial compliance assessment since the prohibition came into force, the regulator identified multiple shortcomings and has now moved from monitoring to active enforcement, cautioning that platforms must demonstrate they have implemented “appropriate systems and processes” to prevent children under 16 from accessing their services.
Regulatory Breaches Uncovered in Initial Significant Review
Australia’s eSafety Commissioner has outlined a troubling pattern of failure to comply amongst the world’s most prominent social media platforms in her first formal review following the ban took effect on 10 December. The report reveals that Meta, Snap, TikTok, YouTube and Snapchat have jointly failed to implement sufficient safeguards to stop minors from using their services. Julie Inman Grant expressed particular concern about systemic weaknesses in age verification systems, noting that some platforms have allowed children who originally stated themselves under 16 to subsequently claim they were older, effectively circumventing the law’s intent.
The findings represent a notable intensification in the regulatory response, with the eSafety Commissioner moving beyond monitoring to active enforcement. The regulator has emphasised that simply showing some children still hold accounts is insufficient; platforms must instead furnish substantive proof that they have put in place comprehensive systems and procedures intended to stop under-16s from creating accounts in the first place. This shift signals the government’s determination to hold tech giants accountable, with possible sanctions looming for companies that do not meet the statutory obligations.
- Allowing previously banned users to confirm again their age and restore account access
- Enabling multiple tries at the identical verification process with no repercussions
- Weak safeguards to prevent accounts for under-16s from being created
- Insufficient notification systems for parents and the general public
- Shortage of publicly available information about compliance actions and account deletions
The Extent of the Challenge
The considerable scale of social media activity amongst young Australians highlights the compliance challenge facing both the government and the platforms themselves. With millions of accounts already removed or restricted since the implementation of the ban, the figures paint a picture of widespread initial non-compliance. The eSafety Commissioner’s conclusions suggest that the operational and technical barriers to implementing age restrictions have proven far more complex than anticipated, with platforms having difficulty to distinguish genuine age declarations from fraudulent ones. This complexity has placed enforcement authorities grappling with the fundamental question of whether existing age verification systems are sufficient for the purpose.
Beyond the technical obstacles lies a wider issue about the willingness of platforms to prioritise compliance over user growth. Social media companies have consistently opposed strict identity verification requirements, citing data protection worries and the genuine difficulty of confirming age online. However, the regulatory report suggests that some platforms may not be making sufficient effort to implement the systems mandated legally. The shift towards active enforcement represents a pivotal moment: either platforms will substantially upgrade their regulatory systems, or they risk facing substantial fines that could reshape their business models in Australia and potentially influence compliance frameworks internationally.
What the Numbers Reveal
In the opening month after the ban’s launch, Australian officials indicated that 4.7 million accounts had been limited or deleted. Whilst this figure initially looked to show regulatory success, further investigation reveals a more complex picture. The sheer volume of account removals implies that many under-16s had managed to establish accounts in the first place, indicating that preventive controls were insufficient. Furthermore, the data casts doubt about whether deleted profiles represent authentic compliance or merely users deleting their pages voluntarily in in light of the updated rules.
The restricted transparency surrounding these figures has troubled independent observers seeking to assess the ban’s true effectiveness. Platforms have revealed scant details about their compliance procedures, effectiveness metrics, or the profile of deleted profiles. This opacity makes it difficult for regulators and the public to determine whether the ban is operating as planned or whether younger users are simply finding alternative ways to reach social media. The Commissioner’s push for thorough documentation of systematic compliance measures reflects increasing concern with platforms’ unwillingness to share full information.
Industry Response and Opposition
The major tech platforms have addressed the regulatory enforcement measures with a mixture of compliance assurances and doubts regarding the practical feasibility of the ban. Meta, which operates Facebook and Instagram, stressed its dedication to adhering to Australian law whilst at the same time contending that precise age verification continues to be a significant industry-wide challenge. The company has advocated for a different approach, suggesting that strong age verification systems and parental consent requirements implemented at the app store level would be more efficient than platform-level enforcement. This position reflects broader industry concerns that the current regulatory framework puts an impractical burden on individual platforms.
Snap, the developer of Snapchat, has adopted a more assertive public position, stating that it had locked 450,000 accounts since the ban took effect and claiming to continue locking more daily. However, sector analysts dispute whether such figures demonstrate genuine compliance or simply represent reactive account management. The fundamental tension between platforms’ business models—which traditionally depended on maximising user engagement and expansion—and the regulatory requirement to actively exclude an entire age demographic remains unresolved. Companies have long resisted stringent age verification, citing privacy issues and technical constraints, creating a standoff between authorities and platforms over who bears responsibility for implementation.
- Meta maintains age verification ought to take place at app store level instead of on individual platforms
- Snap states to have locked 450,000 user accounts following the ban’s implementation in December
- Industry groups point to privacy issues and technical challenges as barriers to effective age verification
- Platforms assert they are making their best effort whilst challenging the ban’s overall effectiveness
More Extensive Questions About the Prohibition’s Impact
As Australia’s under-16 online platform ban enters its enforcement phase, key concerns persist about whether the law will accomplish its stated objectives or merely drive young users towards unregulated platforms. The regulator’s initial compliance assessment reveals that following implementation, substantial gaps exist—children keep discovering ways to bypass age verification mechanisms, and platforms have had difficulty stop new underage accounts from being established. Critics contend that the ban’s effectiveness depends not merely on regulatory vigilance but on whether young people will genuinely abandon major social networks or simply migrate to alternative services, encrypted messaging applications, or VPNs designed to conceal their age and location.
The ban’s worldwide effects increase the complexity of assessments of its impact. Countries such as the United Kingdom, Canada, and multiple European countries are observing Australia’s experiment closely, considering similar legislation for their own citizens. If the ban proves ineffective at reducing children’s digital engagement or cannot protect them from dangerous online content, it could weaken the case for equivalent legislation elsewhere. Conversely, if regulation becomes sufficiently robust to effectively limit underage access, it may inspire other nations to pursue similar approaches. The conclusion will likely influence global regulatory trends for the foreseeable future, making Australia’s implementation efforts scrutinised far beyond its borders.
Who Gains and Those Who Suffer
Mental health campaigners and child safety organisations have championed the ban as a necessary intervention against algorithmic manipulation and exposure to harmful content. Parents and educators contend that taking young Australians off platforms designed to maximise engagement could lower anxiety levels, improve sleep patterns, and reduce exposure to cyberbullying. Tech companies’ own research has recognised the mental health risks linked to social media use amongst adolescents, adding weight to these concerns. However, the ban also removes legitimate uses of social media for young people—maintaining friendships, obtaining educational material, and engaging with online communities around common interests. The regulatory framework assumes harm outweighs benefit, a calculation that some young people and their families challenge.
The ban’s practical impact extends beyond individual users to affect content creators, small businesses, and community organisations dependent on social media platforms. Young people who might have followed creative careers through platforms like TikTok or Instagram now face legal barriers to participation. Small Australian businesses that depend on social media marketing lose access to younger demographic audiences. Community groups, charities, and educational organisations find it difficult to engage young people through channels they previously employed effectively. Meanwhile, the ban unintentionally favours large technology companies with resources to create age verification infrastructure, possibly reinforcing their market dominance rather than reducing it. These unforeseen effects suggest the ban’s effects go well past the simple goal of child protection.
What Follows for Compliance Monitoring
Australia’s eSafety Commissioner has announced a notable transition from hands-off observation to direct intervention, marking a pivotal moment in the implementation of the youth access prohibition. The watchdog will now gather evidence to determine whether services have neglected to implement “reasonable steps” to prevent underage access, a statutory benchmark that extends beyond simply recording that children remain on these systems. This approach requires tangible verification that companies have established proper safeguards and processes meant to keep out minors. The Commissioner’s office has indicated it will launch probes carefully, developing arguments that could trigger considerable sanctions for non-compliance. This shift from observation to action demonstrates increasing dissatisfaction with the platforms’ current efforts and signals that consensual engagement by itself is insufficient.
The rollout phase raises critical issues about the adequacy of penalties and the concrete procedures for ensuring platform accountability. Australia’s statutory provisions provides compliance mechanisms, but their efficacy hinges on the eSafety Commissioner’s commitment to initiate regulatory enforcement and the platforms’ capability to adjust substantively. International observers, especially regulators in the UK and EU, will keenly observe Australia’s implementation tactics and outcomes. A successful enforcement campaign could establish a blueprint for other nations considering similar bans, whilst inadequate results might compromise the comprehensive regulatory system. The next phase will be critical whether Australia’s pioneering regulatory approach translates into substantive defence for young people or becomes largely performative in its influence.
