Australia enforces under-16 social media ban as new rules took effect

Australia has finally introduced the world’s first nationwide prohibition on social media use for under-16s, forcing platforms to delete millions of accounts and prevent new registrations.

Instagram, TikTok, Facebook, YouTube, Snapchat, Reddit, Twitch, Kick and Threads are removing accounts held by younger users. At the same time, Bluesky has agreed to apply the same standard despite not being compelled to do so. The only central platform yet to confirm compliance is X.

The measure follows weeks of age-assurance checks, which have not been flawless, with cases of younger teenagers passing facial-verification tests designed to keep them offline.

Families are facing sharply different realities. Some teenagers feel cut off from friends who managed to bypass age checks, while others suddenly gain a structure that helps reduce unhealthy screen habits.

A small but vocal group of parents admit they are teaching their children how to use VPNs and alternative methods instead of accepting the ban, arguing that teenagers risk social isolation when friends remain active.

Supporters of the legislation counter that Australia imposes clear age limits in other areas of public life for reasons of well-being and community standards, and the same logic should shape online environments.

Regulators are preparing to monitor the transition closely.

The eSafety Commissioner will demand detailed reports from every platform covered by the law, including the volume of accounts removed, evidence of efforts to stop circumvention and assessments of whether reporting and appeals systems are functioning as intended.

Companies that fail to take reasonable steps may face significant fines. A government-backed academic advisory group will study impacts on behaviour, well-being, learning and unintended shifts towards more dangerous corners of the internet.

Global attention is growing as several countries weigh similar approaches. Denmark, Norway and Malaysia have already indicated they may replicate Australia’s framework, and the EU has endorsed the principle in a recent resolution.

Interest from abroad signals a broader debate about how societies should balance safety and autonomy for young people in digital spaces, instead of relying solely on platforms to set their own rules.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Teen chatbot use surges across the US

Nearly a third of US teenagers engage with AI chatbots each day, according to new Pew data. Researchers say nearly 70% have tried a chatbot, reflecting growing dependence on digital tools during schoolwork and leisure time. Concerns remain over exposure to mature content and possible mental health harms.

Pew surveyed almost 1,500 US teens aged 13 to 17, finding broadly similar usage patterns across gender and income. Older teens reported higher engagement, while Black and Hispanic teens showed slightly greater adoption than White peers.

Experts warn that frequent chatbot use may hinder development or encourage cheating in academic settings. Safety groups have urged parents to limit access to companion-like AI tools, citing risks posed by romantic or intimate interactions with minors.

Companies are now rolling out safeguards in response to public scrutiny and legal pressure. OpenAI and Character.AI have tightened controls, while Meta says it has adjusted policies following reports of inappropriate exchanges.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Teens worldwide divided over Australia’s under-16 social media ban

As Australia prepares to enforce the world’s first nationwide under-16 social-media ban on 10 December 2025, young people across the globe are voicing sharply different views about the move.

Some teens view it as an opportunity for a digital ‘detox’, a chance to step back from the constant social media pressure. Others argue the law is extreme, unfair, and likely to push youth toward less regulated corners of the internet.

In Mumbai, 19-year-old Pratigya Jena said the debate isn’t simple: ‘nothing is either black or white.’ She acknowledged that social media can help young entrepreneurs, but also warned that unrestricted access exposes children to inappropriate content.

Meanwhile, in Berlin, 13-year-old Luna Drewes expressed cautious optimism; she felt the ban might help reduce the pressure to conform to beauty standards that are often amplified online. Another teen, 15-year-old Enno Caro Brandes, said he understood the motivation but admitted he couldn’t imagine giving up social media altogether.

In Doha, older teens voiced more vigorous opposition. Sixteen-year-old Firdha Razak called the ban ‘really stupid,’ while sixteen-year-old Youssef Walid argued that it would be trivial to bypass using VPNs. Both said they feared losing vital social and communication outlets.

Some, like 15-year-old Mitchelle Okinedo from Lagos, suggested the ban ignored how deeply embedded social media is in modern life: ‘We were born with it,’ she said, hinting that simply cutting access may be unrealistic. Others noted the role of social media in self-expression, especially in areas where offline spaces are limited.

Even within Australia, opinions diverge. A 15-year-old named Layton Lewis said he doubted the ban would have significant effects. His mother, Emily, meanwhile, welcomed the change, hoping it might encourage more authentic offline friendships rather than ‘illusory’ online interactions.

The variety of reactions underscores how the law is approaching a stark test: while some see potential mental health or safety gains, many worry about the rights of teens, enforcement effectiveness, and whether simply banning access truly addresses the underlying risks.

As commentary and activism ramp up around digital-age regulation, few expect consensus, but many do expect the debate to shape future policy beyond Australia.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Google faces renewed EU scrutiny over AI competition

The European Commission has opened a formal antitrust investigation into whether AI features embedded in online search are being used to unfairly squeeze competitors in newly emerging digital markets shaped by generative AI.

The probe targets Alphabet-owned Google, focusing on allegations that the company imposes restrictive conditions on publishers and content creators while giving its own AI-driven services preferential placement over rival technologies and alternative search offerings.

Regulators are examining products such as AI Overviews and AI Mode, assessing how publisher content is reused within AI-generated summaries and whether media organisations are compensated in a clear, fair, and transparent manner.

EU competition chief Teresa Ribera said the European Commission’s action reflects a broader effort to protect online media and preserve competitive balance as artificial intelligence increasingly shapes how information is produced, discovered, and monetised.

The case adds to years of scrutiny by the European Commission over Google’s search and advertising businesses, even as the company proposes changes to its ad tech operations and continues to challenge earlier antitrust rulings.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Trump allows Nvidia to sell chips to approved Chinese customers

US President Donald Trump has allowed Nvidia to sell H200 AI chips to approved customers in China, marking a shift in export controls. The decision also covers firms such as AMD and follows continued lobbying by Nvidia chief executive Jensen Huang.

Nvidia had been barred from selling advanced chips to Beijing, but a partial reversal earlier required the firm to pay a share of its Chinese revenues to the US government. China later ordered firms to stop buying Nvidia products, pushing them towards domestic semiconductors.

Analysts suggest the new policy may buy time for negotiations over rare earth supplies, as China dominates processing of these minerals. Access to H200 chips may aid China’s tech sector, but experts warn they could also strengthen military AI capabilities.

Nvidia welcomed the announcement, saying the decision strikes a balance that benefits American industry. Shares rose slightly after the news, although the arrangement is expected to face scrutiny from national security advocates.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Canada-EU digital partnership expands cooperation on AI and security

The European Union and Canada have strengthened their digital partnership during the first Digital Partnership Council in Montreal. Both sides outlined a joint plan to enhance competitiveness and innovation, while supporting smaller firms through targeted regulation.

Senior representatives reconfirmed that cooperation with like-minded partners will be essential for economic resilience.

A new Memorandum of Understanding on AI placed a strong emphasis on trustworthy systems, shared standards and wider adoption across strategic sectors.

The two partners will exchange best practices to support sectors such as healthcare, manufacturing, energy, culture and public services.

They also agreed to collaborate on large-scale AI infrastructures and access to computing capacity, while encouraging scientific collaboration on advanced AI models and climate-related research.

A meeting that also led to an agreement on a structured dialogue on data spaces.

A second Memorandum of Understanding covered digital credentials and trust services. The plan includes joint testing of digital identity wallets, pilot projects and new use cases aimed at interoperability.

The EU and Canada also intend to work more closely on the protection of independent media, the promotion of reliable information online and the management of risks created by generative AI.

Both sides underlined their commitment to secure connectivity, with cooperation on 5G, subsea cables and potential new Arctic routes to strengthen global network resilience. Further plans aim to deepen collaboration on quantum technologies, semiconductors and high-performance computing.

A renewed partnership that reflects a shared commitment to resilient supply chains and secure cloud infrastructure as both regions prepare for future technological demands.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Survey reveals split views on AI in academic peer review

Growing use of generative AI within peer review is creating a sharp divide among physicists, according to a new survey by the Institute of Physics Publishing.

Researchers appear more informed and more willing to express firm views, with a notable rise in those who see a positive effect and a large group voicing strong reservations. Many believe AI tools accelerate early reading and help reviewers concentrate on novelty instead of routine work.

Others fear that reviewers might replace careful evaluation with automated text generation, undermining the value of expert judgement.

A sizeable proportion of researchers would be unhappy if AI-shaped assessments of their own papers, even though many quietly rely on such tools when reviewing for journals. Publishers are now revisiting their policies, yet they aim to respect authors who expect human-led scrutiny.

Editors also report that AI-generated reports often lack depth and fail to reflect domain expertise. Concerns extend to confidentiality, with organisations such as the American Physical Society warning that uploading manuscripts to chatbots can breach author trust.

Legal disputes about training data add further uncertainty, pushing publishers to approach policy changes with caution.

Despite disagreements, many researchers accept that AI will remain part of peer review as workloads increase and scientific output grows. The debate now centres on how to integrate new tools in a way that supports researchers instead of weakening the foundations of scholarly communication.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Creatives warn that AI is reshaping their jobs

AI is accelerating across creative fields, raising concerns among workers who say the technology is reshaping livelihoods faster than anyone expected.

A University of Cambridge study recently found that more than two-thirds of creative professionals fear AI has undermined their job security, and many now describe the shift as unavoidable.

One of them is Norwich-based artist Aisha Belarbi, who says the rise of image-generation tools has made commissions harder to secure as clients ‘can just generate whatever they want’. Although she works in both traditional and digital media, Belarbi says she increasingly struggles to distinguish original art from AI output. That uncertainty, she argues, threatens the value of lived experience and the labour behind creative work.

Others are embracing the change. Videographer JP Allard transformed his Milton Keynes production agency after discovering the speed and scale of AI-generated video. His company now produces multilingual ‘digital twins’ and fully AI-generated commercials, work he says is quicker and cheaper than traditional filming. Yet he acknowledges that the pace of change can leave staff behind and says retraining has not kept up with the technology.

For musician Ross Stewart, the concern centres on authenticity. After listening to what he later discovered was an AI-generated blues album, he questioned the impact of near-instant song creation on musicians’ livelihoods and exposure. He believes audiences will continue to seek human performance, but worries that the market for licensed music is already shifting towards AI alternatives.

Copywriter Niki Tibble has experienced similar pressures. Returning from maternity leave, she found that AI tools had taken over many entry-level writing tasks. While some clients still prefer human writers for strategy, nuance and brand voice, Tibble’s work has increasingly shifted toward reviewing and correcting AI-generated copy. She says the uncertainty leaves her unsure whether her role will exist in a decade.

Across these stories, creative workers describe a sector in rapid transition. While some see new opportunities, many fear the speed of adoption and a future where AI replaces the very work that has long defined their craft.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

New interview study tracks how workers adapt to AI

Anthropic has unveiled Anthropic Interviewer, an AI-driven tool for large-scale workplace interviews. The system used Claude to conduct 1,250 structured interviews with professionals across the general workforce, creative fields and scientific research.

In surveys, 86 percent said AI saves time and 65 percent felt satisfied with its role at work. Workers often hoped to automate routine tasks while preserving responsibilities that define their professional identity.

Creative workers reported major time savings and quality gains yet faced stigma and economic anxiety around AI use. Many hid AI tools from colleagues, feared market saturation and still insisted on retaining creative control.

Across groups, professionals imagined careers where humans oversee AI systems rather than perform every task themselves. Anthropic plans to keep using Anthropic Interviewer to track attitudes and inform future model design.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Japan weighs easing rules on personal data use

Japan is preparing to relax restrictions on personal data use to support rapid AI development. Government sources say a draft bill aims to expand third-party access to sensitive information.

Plans include allowing medical histories and criminal records to be obtained without consent for statistical purposes. Japanese officials argue such access could accelerate research while strengthening domestic competitiveness.

New administrative fines would target companies that profit from unlawfully acquired data affecting large groups. Penalties would match any gains made through misconduct, reflecting growing concern over privacy abuses.

A government panel has reviewed the law since 2023 and intends to present reforms soon. Debate is expected to intensify as critics warn of increased risks to individual rights if support for AI development in this regard continues.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot