Australia enforces under-16 social media ban as new rules took effect

Australia has finally introduced the world’s first nationwide prohibition on social media use for under-16s, forcing platforms to delete millions of accounts and prevent new registrations.

Instagram, TikTok, Facebook, YouTube, Snapchat, Reddit, Twitch, Kick and Threads are removing accounts held by younger users. At the same time, Bluesky has agreed to apply the same standard despite not being compelled to do so. The only central platform yet to confirm compliance is X.

The measure follows weeks of age-assurance checks, which have not been flawless, with cases of younger teenagers passing facial-verification tests designed to keep them offline.

Families are facing sharply different realities. Some teenagers feel cut off from friends who managed to bypass age checks, while others suddenly gain a structure that helps reduce unhealthy screen habits.

A small but vocal group of parents admit they are teaching their children how to use VPNs and alternative methods instead of accepting the ban, arguing that teenagers risk social isolation when friends remain active.

Supporters of the legislation counter that Australia imposes clear age limits in other areas of public life for reasons of well-being and community standards, and the same logic should shape online environments.

Regulators are preparing to monitor the transition closely.

The eSafety Commissioner will demand detailed reports from every platform covered by the law, including the volume of accounts removed, evidence of efforts to stop circumvention and assessments of whether reporting and appeals systems are functioning as intended.

Companies that fail to take reasonable steps may face significant fines. A government-backed academic advisory group will study impacts on behaviour, well-being, learning and unintended shifts towards more dangerous corners of the internet.

Global attention is growing as several countries weigh similar approaches. Denmark, Norway and Malaysia have already indicated they may replicate Australia’s framework, and the EU has endorsed the principle in a recent resolution.

Interest from abroad signals a broader debate about how societies should balance safety and autonomy for young people in digital spaces, instead of relying solely on platforms to set their own rules.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

G7 ministers meet in Montreal to boost industrial cooperation

Canada has opened the G7 Industry, Digital and Technology Ministers’ Meeting in Montreal, bringing together ministers, industry leaders, and international delegates to address shared industrial and technological challenges.

The meeting is being led by Industry Minister Melanie Joly and AI and Digital Innovation Minister Evan Solomon, with discussions centred on strengthening supply chains, accelerating innovation, and boosting industrial competitiveness across advanced economies.

Talks will focus on building resilient economies, expanding trusted digital infrastructure, and supporting growth while aligning industrial policy with economic security and national security priorities shared among G7 members.

The agenda builds on outcomes from the recent G7 leaders’ summit in Kananaskis, Canada, including commitments on quantum technologies, critical minerals cooperation, and a shared statement on AI and prosperity.

Canadian officials said closer coordination among trusted partners is essential amid global uncertainty and rapid technological change, positioning innovation-driven industry as a long-term foundation for economic growth, productivity, and shared prosperity.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

EU AI Act changes aim to ease high-risk compliance pressure

The European Commission has proposed a series of amendments to the EU AI Act to ensure a timely, smooth, and proportionate rollout of the bloc’s landmark AI rules.

Set out in the Digital Omnibus on AI published in November, the changes would delay some of the most demanding obligations of the AI Act, particularly for high-risk AI systems, linking compliance deadlines to the availability of supporting standards and guidance.

The proposal also introduces new grace periods for certain transparency requirements, especially for generative AI and deepfake systems, while leaving existing prohibitions on manipulative or exploitative uses of AI fully intact.

Other revisions include removing mandatory AI literacy requirements for providers and deployers and expanding the powers of the European AI Office, allowing it to directly supervise some general-purpose AI systems and AI embedded in large online platforms.

While the package includes simplification measures designed to ease burdens on smaller firms and encourage innovation, the amendments now face a complex legislative process, adding uncertainty for companies preparing to comply with the AI Act’s long-term obligations.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Teens worldwide divided over Australia’s under-16 social media ban

As Australia prepares to enforce the world’s first nationwide under-16 social-media ban on 10 December 2025, young people across the globe are voicing sharply different views about the move.

Some teens view it as an opportunity for a digital ‘detox’, a chance to step back from the constant social media pressure. Others argue the law is extreme, unfair, and likely to push youth toward less regulated corners of the internet.

In Mumbai, 19-year-old Pratigya Jena said the debate isn’t simple: ‘nothing is either black or white.’ She acknowledged that social media can help young entrepreneurs, but also warned that unrestricted access exposes children to inappropriate content.

Meanwhile, in Berlin, 13-year-old Luna Drewes expressed cautious optimism; she felt the ban might help reduce the pressure to conform to beauty standards that are often amplified online. Another teen, 15-year-old Enno Caro Brandes, said he understood the motivation but admitted he couldn’t imagine giving up social media altogether.

In Doha, older teens voiced more vigorous opposition. Sixteen-year-old Firdha Razak called the ban ‘really stupid,’ while sixteen-year-old Youssef Walid argued that it would be trivial to bypass using VPNs. Both said they feared losing vital social and communication outlets.

Some, like 15-year-old Mitchelle Okinedo from Lagos, suggested the ban ignored how deeply embedded social media is in modern life: ‘We were born with it,’ she said, hinting that simply cutting access may be unrealistic. Others noted the role of social media in self-expression, especially in areas where offline spaces are limited.

Even within Australia, opinions diverge. A 15-year-old named Layton Lewis said he doubted the ban would have significant effects. His mother, Emily, meanwhile, welcomed the change, hoping it might encourage more authentic offline friendships rather than ‘illusory’ online interactions.

The variety of reactions underscores how the law is approaching a stark test: while some see potential mental health or safety gains, many worry about the rights of teens, enforcement effectiveness, and whether simply banning access truly addresses the underlying risks.

As commentary and activism ramp up around digital-age regulation, few expect consensus, but many do expect the debate to shape future policy beyond Australia.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Mitigated ads personalisation coming to Meta platforms in the EU

Meta has agreed to introduce a less personalised ads option for Facebook and Instagram users in the EU, as part of efforts to comply with the bloc’s Digital Markets Act and address concerns over data use and user consent.

Under the revised model, users will be able to access Meta’s social media platforms without agreeing to extensive personal data processing for fully personalised ads. Instead, they can opt for an alternative experience based on significantly reduced data inputs, resulting in more limited ad targeting.

The option is set to roll out across the EU from January 2026. It marks the first time Meta has offered users a clear choice between highly personalised advertising and a reduced-data model across its core platforms.

The change follows months of engagement between Meta and Brussels after the European Commission ruled in April that the company had breached the DMA. Regulators stated that Meta’s previous approach had failed to provide users with a genuine and effective choice over how their data was used for advertising.

Once implemented, the Commission said it will gather evidence and feedback from Meta, advertisers, publishers, and other stakeholders. The goal is to assess the extent to which the new option is adopted and whether it significantly reshapes competition and data practices in the EU digital advertising market.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Google faces renewed EU scrutiny over AI competition

The European Commission has opened a formal antitrust investigation into whether AI features embedded in online search are being used to unfairly squeeze competitors in newly emerging digital markets shaped by generative AI.

The probe targets Alphabet-owned Google, focusing on allegations that the company imposes restrictive conditions on publishers and content creators while giving its own AI-driven services preferential placement over rival technologies and alternative search offerings.

Regulators are examining products such as AI Overviews and AI Mode, assessing how publisher content is reused within AI-generated summaries and whether media organisations are compensated in a clear, fair, and transparent manner.

EU competition chief Teresa Ribera said the European Commission’s action reflects a broader effort to protect online media and preserve competitive balance as artificial intelligence increasingly shapes how information is produced, discovered, and monetised.

The case adds to years of scrutiny by the European Commission over Google’s search and advertising businesses, even as the company proposes changes to its ad tech operations and continues to challenge earlier antitrust rulings.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Russia moves forward with a nationwide plan for generative AI

A broad plan to integrate generative AI across public administration and key sectors of the economy is being prepared by Russia.

Prime Minister Mikhail Mishustin explained that the new framework seeks to extend modern AI tools across regions and major industries in order to strengthen national technological capacity.

The president has already underlined the need for fully domestic AI products as an essential element of national sovereignty. Moscow intends to rely on locally developed systems instead of foreign platforms, an approach aimed at securing long-term independence and resilience.

A proposal created by the government and the Presidential Administration has been submitted for approval to establish a central headquarters that will guide the entire deployment effort.

The new body will set objectives, track progress and coordinate work across ministries and agencies while supporting broader access to advanced capabilities.

Officials in Russia view the plan as a strategic investment intended to reinforce national competitiveness in a rapidly changing technological environment.

Greater use of generative systems is expected to improve administrative efficiency, support regional development and encourage innovation across multiple sectors.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Canada-EU digital partnership expands cooperation on AI and security

The European Union and Canada have strengthened their digital partnership during the first Digital Partnership Council in Montreal. Both sides outlined a joint plan to enhance competitiveness and innovation, while supporting smaller firms through targeted regulation.

Senior representatives reconfirmed that cooperation with like-minded partners will be essential for economic resilience.

A new Memorandum of Understanding on AI placed a strong emphasis on trustworthy systems, shared standards and wider adoption across strategic sectors.

The two partners will exchange best practices to support sectors such as healthcare, manufacturing, energy, culture and public services.

They also agreed to collaborate on large-scale AI infrastructures and access to computing capacity, while encouraging scientific collaboration on advanced AI models and climate-related research.

A meeting that also led to an agreement on a structured dialogue on data spaces.

A second Memorandum of Understanding covered digital credentials and trust services. The plan includes joint testing of digital identity wallets, pilot projects and new use cases aimed at interoperability.

The EU and Canada also intend to work more closely on the protection of independent media, the promotion of reliable information online and the management of risks created by generative AI.

Both sides underlined their commitment to secure connectivity, with cooperation on 5G, subsea cables and potential new Arctic routes to strengthen global network resilience. Further plans aim to deepen collaboration on quantum technologies, semiconductors and high-performance computing.

A renewed partnership that reflects a shared commitment to resilient supply chains and secure cloud infrastructure as both regions prepare for future technological demands.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Survey reveals split views on AI in academic peer review

Growing use of generative AI within peer review is creating a sharp divide among physicists, according to a new survey by the Institute of Physics Publishing.

Researchers appear more informed and more willing to express firm views, with a notable rise in those who see a positive effect and a large group voicing strong reservations. Many believe AI tools accelerate early reading and help reviewers concentrate on novelty instead of routine work.

Others fear that reviewers might replace careful evaluation with automated text generation, undermining the value of expert judgement.

A sizeable proportion of researchers would be unhappy if AI-shaped assessments of their own papers, even though many quietly rely on such tools when reviewing for journals. Publishers are now revisiting their policies, yet they aim to respect authors who expect human-led scrutiny.

Editors also report that AI-generated reports often lack depth and fail to reflect domain expertise. Concerns extend to confidentiality, with organisations such as the American Physical Society warning that uploading manuscripts to chatbots can breach author trust.

Legal disputes about training data add further uncertainty, pushing publishers to approach policy changes with caution.

Despite disagreements, many researchers accept that AI will remain part of peer review as workloads increase and scientific output grows. The debate now centres on how to integrate new tools in a way that supports researchers instead of weakening the foundations of scholarly communication.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Instacart deepens partnership with OpenAI for real-time AI shopping

OpenAI and Instacart are expanding their longstanding collaboration by introducing a fully integrated grocery shopping experience inside ChatGPT.

Users can receive meal inspiration, browse products and place orders in one continuous conversation instead of switching across separate platforms.

A service that brings together Instacart’s real-time retail network with OpenAI’s most advanced models to produce an experience that feels like a direct link between a simple request and completed delivery.

The Instacart app becomes the first service to offer a full checkout flow inside ChatGPT by using the Agentic Commerce Protocol. When users mention food, ingredients or recipe ideas, ChatGPT can surface the app immediately.

Once the user connects an Instacart account, the system selects suitable items from nearby retailers and builds a complete cart that can be reviewed before payment. Users then pay securely inside the chat while Instacart manages collection and delivery through its established network.

The update also reflects broader cooperation between the two companies. Instacart continues to rely on OpenAI APIs to support personalised suggestions and real time guidance across its customer experience.

ChatGPT Enterprise assists internal teams, while Codex powers an internal coding agent that shortens development cycles instead of slowing them down with manual tasks. The partnership builds on Instacart’s early involvement in the Operator research preview, where it helped refine emerging agentic technologies.

A renewed partnership that strengthens OpenAI’s growing enterprise ecosystem. The company already works with major global brands across sectors such as retail, financial services and telecommunications.

The Instacart integration offers a view of how conversational agents may act as a bridge between everyday intent and immediate real-world action.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!