European Commission launches Apply AI and AI in Science strategies

Countries are racing to harness AI, and the European Commission has unveiled two strategies to maintain Europe’s competitiveness. Apply AI targets faster adoption across industries and the public sector, while AI in Science focuses on boosting Europe’s research leadership.

Commission President Ursula von der Leyen stated that Europe must shape AI’s future by balancing innovation and safety. The European Commission is mobilising €1 billion to boost adoption in healthcare, manufacturing, energy, defence, and culture, while supporting SMEs.

Measures include creating AI-powered screening centres for healthcare, backing frontier models, and upgrading testing infrastructure. An Apply AI Alliance will unite industry, academia, civil society, and public bodies to coordinate action, while an AI Observatory will monitor sector trends and impacts.

The AI in Science Strategy centres on RAISE, a new virtual institute to pool and coordinate resources for applying AI in research. Investments include €600 million in compute power through Horizon Europe and €58 million for talent networks, alongside plans to double annual AI research funding to over €3 billion.

The EU aims to position itself as a global hub for trustworthy and innovative AI by linking infrastructure, data, skills, and investment. Upcoming events, such as the AI in Science Summit in Copenhagen, will showcase new initiatives as Europe pushes to translate its AI ambitions into tangible outcomes.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

OpenAI expands ChatGPT Go to 16 new Asian markets

The US startup OpenAI has broadened access to its affordable ChatGPT Go plan, now available in 16 additional countries across Asia, including Malaysia, Vietnam, the Philippines, Pakistan, and Thailand.

Priced at under $5 per month, the plan offers local currency payments in select regions, while others will pay in USD with tax-adjusted variations.

ChatGPT Go gives users higher message and image-generation limits, increased upload capacity, and double the memory of the free plan.

A move that follows significant regional growth (Southeast Asia’s weekly active users increasing fourfold) and builds on earlier launches in India and Indonesia, where paid subscriptions have already doubled.

The expansion intensifies competition with Google, which recently introduced its Google AI Plus plan in more than 40 countries. Both companies are vying to attract users in fast-growing markets with low-cost AI access, each blending productivity and creative tools into subscription offerings.

At OpenAI’s DevDay 2025 in San Francisco, CEO Sam Altman announced that ChatGPT’s global weekly active users have reached 800 million.

OpenAI is also introducing in-chat applications from partners like Spotify, Zillow, and Coursera, signalling a shift toward transforming ChatGPT into a broader AI platform ecosystem.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Opal’s global rollout spans 15 countries with major feature upgrades

Google is expanding access to Opal, its no-code AI mini-app builder. Introduced two months ago within Google Labs, the tool enables users to create AI-powered mini-apps through natural language prompts, eliminating the need for coding.

According to Megan Li, Senior Product Manager at Google Labs, the expansion follows strong early engagement from creators. Users can access Opal at opal.withgoogle.com and join its builder community through Discord.

New debugging features aim to make workflows more transparent and efficient. Users can now run workflows step by step in a visual editor or adjust specific steps in the console, with real-time error reporting.

Performance upgrades have been introduced to speed up app creation, while parallel run capabilities enable simultaneous workflow steps. The rollout covers India, Canada, Japan, South Korea, Vietnam, Indonesia, Brazil, Singapore, Colombia, El Salvador, Costa Rica, Panamá, Honduras, and Argentina.

Meanwhile, Google DeepMind has launched Gemini 2.5 Computer Use, a specialised model capable of interacting with user interfaces. Available in preview through the Gemini API, it can be accessed via Google AI Studio and Vertex AI Studio.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Kazakhstan turns to AI to fight the shadow economy

Kazakhstan’s Prime Minister Olzhas Bektenov has directed the full implementation of AI across government agencies to meet President Kassym-Jomart Tokayev’s goal of reducing the shadow economy’s share in GDP to 15 percent in 2025.

At a government session, Bektenov said progress must go beyond reports and correspondence, calling for structural reforms in taxation, digitalisation, and business regulation. He urged ministries to pursue a ‘transparent economy’ through comprehensive AI and data integration initiatives.

The State Revenue Committee of Kazakhstan will lead the digital transformation, supported by a new Data Processing Centre established by the Ministry of Artificial Intelligence and Digital Development.

Bektenov stressed that digitalisation projects such as cashless payments and the digital tenge have already proven effective in curbing unrecorded transactions and improving financial oversight.

AI will also be deployed in customs risk profiling and cargo inspection analysis to detect fraud and reduce corruption.

The Ministries of Finance, Justice, Trade, and National Economy were instructed to integrate databases under the Smart Data Finance system and to finalise an automated risk management system for company registration by 25 November.

Deputy Prime Minister Serik Zhumangarin will oversee coordination.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Power grid spending surges as US braces for data centre and AI boom

US electric utilities are set to spend nearly $208 billion on the power grid in 2025 and more than $1.1 trillion over the next five years, according to the Edison Electric Institute. The surge in investment reflects rising demand from data centres, artificial intelligence, and wider electrification across the economy.

EEI data shows that investor-owned utilities spent $765 billion on capital projects in the five years to 2024. The new spending represents a significant increase and is aimed at upgrading and expanding infrastructure to keep pace with the accelerating demand for electricity.

The growing investment comes as demand from energy-intensive technologies continues to rise. Data centres and AI workloads are driving sustained growth in US power consumption, placing unprecedented pressure on existing infrastructure and prompting utilities to scale up their spending plans.

David Weeks, supply chain industry practice lead at Moody’s, warned that the escalating energy crisis could become a limiting factor across multiple industries. He said grid constraints and permitting delays must be factored into corporate supply chain strategies to avoid future disruptions.

As electrification spreads across the economy, grid reliability and capacity are becoming critical considerations for companies. The planned investment underscores the urgency of modernising the power grid to support economic growth while adapting to new technological demands.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

The global struggle to regulate children’s social media use

Finding equilibrium in children’s use of social media

Social media has become a defining part of modern childhood. Platforms like Instagram, TikTok, Snapchat and YouTube offer connection, entertainment and information at an unprecedented scale.

Yet concerns have grown about their impact on children’s mental health, education, privacy and safety. Governments, parents and civil society increasingly debate whether children should access these spaces freely, with restrictions, or not at all.

The discussion is no longer abstract. Across the world, policymakers are moving beyond voluntary codes to legal requirements, some proposing age thresholds or even outright bans for minors.

Supporters argue that children face psychological harm and exploitation online, while critics caution that heavy restrictions can undermine rights, fail to solve root problems and create new risks.

The global conversation is now at a turning point, where choices about social media regulation will shape the next generation’s digital environment.

Why social media is both a lifeline and a threat for youth

The influence of social media on children is double-edged. On the one side, these platforms enable creativity, allow marginalised voices to be heard, and provide educational content. During the pandemic, digital networks offered a lifeline of social interaction when schools were closed.

multiracial group of school kids using touchpads and listening to a teacher during computer class

Children and teens can build communities around shared interests, learn new skills, and sometimes even gain economic opportunities through digital platforms.

On the other side, research has linked heavy use of social media with increased rates of anxiety, depression, disrupted sleep and body image issues among young users. Recommendation algorithms often push sensational or harmful content, reinforcing vulnerabilities rather than mitigating them.

Cyberbullying, exposure to adult material, and risks of predatory contact are persistent challenges. Instead of strengthening resilience, platforms often prioritise engagement metrics that exploit children’s attention and emotional responses.

The scale of the issue is enormous. Billions of children around the world hold smartphones before the age of 12. With digital life inseparable from daily routines, even well-meaning parents struggle to set boundaries.

Governments face pressure to intervene, but approaches vary widely, reflecting different cultural norms, levels of trust in technology firms, and political attitudes toward child protection.

The Australian approach

Australia is at the forefront of regulation. In recent years, the country has passed strong online safety laws, led by its eSafety Commissioner. These rules include mandatory age verification for certain online services and obligations for platforms to design products with child safety in mind.

Most notably, Australia has signalled its willingness to explore outright bans on general social media access for children under 16. The government has pointed to mounting evidence of harm, from cyberbullying to mental health concerns, and has emphasised the need for early intervention.

australian social media laws for children safety

Instead of leaving responsibility entirely to parents, the state argues that platforms themselves must redesign the way they serve children.

Critics highlight several problems. Age verification requires identity checks, which can endanger privacy and create surveillance risks. Bans may also drive children to use less-regulated spaces or fake their ages, undermining the intended protections.

Others argue that focusing only on prohibition overlooks the need for broader digital literacy education. Yet Australia’s regulatory leadership has sparked a wider debate, prompting other countries to reconsider their own approaches.

Greece’s strong position

Last week, Greece reignited the global debate with its own strong position on restricting youth access to social media.

Speaking at the United Nations General Assembly during an event hosted by Australia on digital child safety, PM Kyriakos Mitsotakis said his government was prepared to consider banning social media for children under 16.

sweden social media ban for children

Mitsotakis warned that societies are conducting the ‘largest uncontrolled experiment on children’s minds’ by allowing unrestricted access to social media platforms. He cautioned that while the long-term effects of the experiment remain uncertain, they are unlikely to be positive.

Additionally, the prime minister pointed to domestic initiatives already underway, such as the ban on mobile phones in schools, which he claimed has already transformed the educational experience.

Mitsotakis acknowledged the difficulties of enforcing such regulations but insisted that complexity cannot be an excuse for inaction.

Across the whole world, similar conversations are gaining traction. Let’s review some of them.

National initiatives across the globe

UK

The UK introduced its Online Safety Act in 2023, one of the most comprehensive frameworks for regulating online platforms. Under the law, companies must assess risks to children and demonstrate how they mitigate harms.

Age assurance is required for certain services, including those hosting pornography or content promoting suicide or self-harm. While not an outright ban, the framework places a heavy responsibility on platforms to restrict harmful material and tailor their products to younger users.

EU

The EU has not introduced a specific social media ban, but its Digital Services Act requires major platforms to conduct systemic risk assessments, including risks to minors.

However, the European Commission has signalled that it may support stricter measures on youth access to social media, keeping the option of a bloc-wide ban under review.

Commission President Ursula von der Leyen has recently endorsed the idea of a ‘digital majority age’ and pledged to gather experts by year’s end to consider possible actions.

The Commission has pointed to the Digital Services Act as a strong baseline but argued that evolving risks demand continued vigilance.

EU

Companies must show regulators how algorithms affect young people and must offer transparency about their moderation practices.

In parallel, several EU states are piloting age verification measures for access to certain platforms. France, for example, has debated requiring parental consent for children under 15 to use social media.

USA

The USA lacks a single nationwide law, but several states are acting independently, although there are some issues with the Supreme Court and the First Amendment.

Florida, Texas, Utah, and Arkansas have passed laws requiring parental consent for minors to access social media, while others are considering restrictions.

The federal government has debated child online safety legislation, although political divides have slowed progress. Instead of a ban, American initiatives often blend parental rights, consumer protection, and platform accountability.

Canada

The Canadian government has introduced Bill C-63, the Online Harms Act, aiming to strengthen online child protection and limit the spread of harmful content.

Justice Minister Arif Virani said the legislation would ensure platforms take greater responsibility for reducing risks and preventing the amplification of content that incites hate, violence, or self-harm.

The framework would apply to platforms, including livestreaming and adult content services.

canada flag is depicted on the screen with the program code 1

They would be obliged to remove material that sexually exploits children or shares intimate content without consent, while also adopting safety measures to limit exposure to harmful content such as bullying, terrorism, and extremist propaganda.

However, the legislation also does not impose a complete social media ban for minors.

China

China’s cyberspace regulator has proposed restrictions on children’s smartphone use. The draft rules limit use to a maximum of two hours daily for those under 18, with stricter limits for younger age groups.

The Cyberspace Administration of China (CAC) said devices should include ‘minor mode’ programmes, blocking internet access for children between 10 p.m. and 6 a.m.

Teenagers aged 16 to 18 would be allowed two hours a day, those between eight and 16 just one hour, and those under eight years old only eight minutes.

It is important to add that parents could opt out of the restrictions if they wish.

India

In January, India proposed new rules to tighten controls on children’s access to social media, sparking a debate over parental empowerment and privacy risks.

The draft rules required parental consent before minors can create accounts on social media, e-commerce, or gaming platforms.

Verification would rely on identity documents or age data already held by providers.

Supporters argue the measures will give parents greater oversight and protect children from risks such as cyberbullying, harmful content, and online exploitation.

Singapore

PM Lawrence Wong has warned of the risks of excessive screen time while stressing that children must also be empowered to use technology responsibly. The ultimate goal is the right balance between safety and digital literacy.

In addition, researchers suggest schools should not ban devices out of fear but teach children how to manage them, likening digital literacy to learning how to swim safely. Such a strategy highlights that no single solution fits all societies.

Balancing rights and risks

Bans and restrictions raise fundamental rights issues. Children have the right to access information, to express themselves, and to participate in culture and society.

Overly strict bans can exclude them from opportunities that their peers elsewhere enjoy. Critics argue that bans may create inequalities between children whose families find workarounds and those who comply.

social media ban for under 16s

At the same time, the rights to health, safety and privacy must also be protected. The difficulty lies in striking a balance. Advocates of stronger regulation argue that platforms have failed to self-regulate effectively, and that states must step in.

Opponents argue that bans may create unintended harms and encourage authoritarian tendencies, with governments using child safety as a pretext for broader control of online spaces.

Instead of choosing one path, some propose hybrid approaches: stronger rules for design and data collection, combined with investment in education and digital resilience. Such approaches aim to prepare children to navigate online risks while making platforms less exploitative.

The future of social media and child protection

Looking forward, the global landscape is unlikely to converge on a single model. Some countries will favour bans and strict controls, others will emphasise parental empowerment, and still others will prioritise platform accountability.

What is clear is that the status quo is no longer acceptable to policymakers or to many parents.

Technological solutions will also evolve. Advances in privacy-preserving age verification may ease some concerns, although sceptics warn that surveillance risks will remain. At the same time, platforms may voluntarily redesign products for younger audiences, either to comply with regulations or to preserve trust.

Ultimately, the challenge is not whether to regulate, but how. Instead of focusing solely on prohibition, governments and societies may need to build layered protections: legal safeguards, technological checks, educational investments and cultural change.

If these can align, children may inherit a safer digital world that still allows them to learn, connect and create. If they cannot, the risks of exclusion or exploitation will remain unresolved.

black woman hands and phone for city map location gps or social media internet search in new york

In conclusion, the debate over banning or restricting social media for children reflects broader tensions between freedom, safety, privacy, and responsibility. Around the globe, governments are experimenting with different balances of control and empowerment.

Australia, as we have already shown, represents one of the boldest approaches, while others, from the UK and Greece to China and Singapore, are testing different variations.

What unites them is the recognition that children cannot simply be left alone in a digital ecosystem designed for profit rather than protection.

The next decade will determine whether societies can craft a sustainable balance, where technology serves the needs of the young instead of exploiting them.

In the end, it is our duty as human beings and responsible citizens.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

New report finds IT leaders unprepared for evolving cyber threats

A new global survey by 11:11 Systems highlights growing concerns among IT leaders over cyber incident recovery. More than 800 senior IT professionals across North America, Europe, and the Asia Pacific report a rising strain from evolving threats, staffing gaps, and limited clean-room infrastructure.

Over 80% of respondents experienced at least one major cyberattack in the past year, with more than half facing multiple incidents. Nearly half see recovery planning complexity as their top challenge, while over 80% say their organisations are overconfident in their recovery capabilities.

The survey also reveals that 74% believe integrating AI could increase cyberattack vulnerability. Despite this, 96% plan to invest in cyber incident recovery within the next 12 months, underlining its growing importance in budget strategies.

The financial stakes are high. Over 80% of respondents reported spending at least six figures during just one hour of downtime, with the top 5% incurring losses of over one million dollars per hour. Yet 30% of businesses do not test their recovery plans annually, despite these risks.

11:11 Systems’ CTO Justin Giardina said organisations must adopt a proactive, AI-driven approach to recovery. He emphasised the importance of advanced platforms, secure clean rooms, and tailored expertise to enhance cyber resilience and expedite recovery after incidents.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Policy hackathon shapes OpenAI proposals ahead of EU AI strategy

OpenAI has published 20 policy proposals to speed up AI adoption across the EU. Released shortly before the European Commission’s Apply AI Strategy, the report outlines practical steps for member states, businesses, and the public sector to bridge the gap between ambition and deployment.

The proposals originate from Hacktivate AI, a Brussels hackathon with 65 participants from EU institutions, governments, industry, and academia. They focus on workforce retraining, SME support, regulatory harmonisation, and public sector collaboration, highlighting OpenAI’s growing policy role in Europe.

Key ideas include Individual AI Learning Accounts to support workers, an AI Champions Network to mobilise SMEs, and a European GovAI Hub to share resources with public institutions. OpenAI’s Martin Signoux said the goal was to bridge the divide between strategy and action.

Europe already represents a major market for OpenAI tools, with widespread use among developers and enterprises, including Sanofi, Parloa, and Pigment. Yet adoption remains uneven, with IT and finance leading, manufacturing catching up, and other sectors lagging behind, exposing a widening digital divide.

The European Commission is expected to unveil its Apply AI Strategy within days. OpenAI’s proposals act as a direct contribution to the policy debate, complementing previous initiatives such as its EU Economic Blueprint and partnerships with governments in Germany and Greece.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

EU digital laws simplified by CEPS Task Force to boost innovation

The Centre for European Policy Studies (CEPS) Task Force, titled ‘Next Steps for EU Law and Regulation for the Digital World’, aims to refine and simplify the EU’s digital rulebook.

This rulebook now covers key legislation, including the Digital Markets Act (DMA), Digital Services Act (DSA), GDPR, Data Act, AI Act, Data Governance Act (DGA), and Cyber Resilience Act (CRA).

While these laws position Europe as a global leader in digital regulation, they also create complexity, overlaps, and legal uncertainty.

The Task Force focuses on enhancing coherence, efficiency, and consistency across digital acts while maintaining strong protections for consumers and businesses.

The CEPS Task Force emphasises targeted reforms to reduce compliance burdens, especially for SMEs, and strengthen safeguards.

It also promotes procedural improvements, including robust impact assessments, independent ex-post evaluations, and the adoption of RegTech solutions to streamline compliance and make regulation more adaptive.

Between November 2025 and January 2026, the Task Force will hold four workshops addressing: alignment of the DMA with competition law, fine-tuning the DSA, improving data governance, enhancing GDPR trust, and ensuring AI Act coherence.

The findings will be published in a Final Report in March 2026, outlining a simpler, more agile EU digital regulatory framework that fosters innovation, reduces regulatory burdens, and upholds Europe’s values.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

OpenAI expands into personal investing via acquisition of Roi

OpenAI has acquired the personal investing startup Roi, which promises AI-driven insights, education, and guidance for individual investors. The Verge reports that the acquisition marks OpenAI’s official entry into the personal finance space.

Following the deal, Roi will shut down its service on October 15 and delete all user data. Its offerings included traditional investing options alongside crypto and NFTs. The company cited this transition in its announcement.

OpenAI did not publicly disclose the purchase price. With this move, OpenAI takes a step beyond content, tools and agents, toward embedding financial services into its AI ecosystem. It questions how AI platforms may offer personalised wealth management or advisory services someday.

The acquisition also draws regulatory, ethical and trust considerations. Mixing AI with finance means issues like explainability, bias, fiduciary responsibility, data privacy and risk management become immediately relevant. Whether users will embrace AI financial advice depends as much on trust and governance as algorithmic accuracy.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!