Meta has announced that third-party AI chatbots will again be allowed to operate through WhatsApp in Europe, reversing restrictions introduced earlier this year.
The decision follows pressure from the European Commission, which had warned it could impose interim competition measures.
Earlier in 2026, Meta limited access to rival chatbot services on the messaging platform, prompting regulators to examine whether the move unfairly restricted competition in the rapidly expanding AI market.
WhatsApp remains one of the most widely used messaging applications across European countries, making platform access critical for emerging AI services.
Under the new arrangement, companies will be able to distribute general-purpose AI chatbots via the WhatsApp Business API for 12 months.
The change is intended to give European regulators time to complete their investigation while allowing competing AI services to operate within the platform ecosystem.
Meta has also indicated that businesses offering chatbots through WhatsApp will be required to pay fees to access the system.
The European Commission is now assessing whether these adjustments sufficiently address competition concerns surrounding the integration of AI services inside major digital platforms.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Pressure is growing in New Zealand to strengthen the Privacy Act following several high-profile data breaches. Debate in New Zealand intensified after a cyberattack exposed medical records from the Manage My Health patient portal.
The breach in New Zealand affected about 120,000 patients and involved threats to release documents on the dark web. Another incident forced the MediMap medication platform offline after unauthorised changes were detected in patient records.
Privacy specialists argue that current enforcement powers are too weak to deter serious failures. The Privacy Act allows only limited financial penalties, with fines generally capped at NZD10,000.
Officials are now considering reforms, including stronger penalties for privacy violations. Policymakers also warn that failure to strengthen the law could threaten the country’s EU data adequacy status.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
The European Commission has convened a new expert panel tasked with examining how children can be better protected across digital platforms, including social media, gaming environments and AI tools.
The initiative reflects growing concern across Europe regarding the psychological and safety risks associated with young users’ online behaviour.
Specialists from health, computer science, child rights and digital literacy will work alongside youth representatives to assess current research and policy responses.
Discussions during the first meeting centred on platform responsibility, including age-appropriate safety-by-design features, algorithmic amplification and addictive product design.
An initiative that also addresses digital literacy for children, parents and educators, while considering how regulatory measures can reduce risks without undermining the benefits of online participation.
The panel’s work complements the enforcement of the Digital Services Act and related European policies designed to strengthen protections for minors online.
Among the tools under development is an EU age-verification application currently tested in several member states, intended to support privacy-preserving checks compatible with the future EU digital identity framework.
The panel is expected to deliver policy recommendations to the Commission by summer 2026.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
AI is beginning to reshape corporate strategy as organisations shift from isolated technology experiments to broader operational transformation.
According to OpenAI, businesses that treat AI as a collection of disconnected pilots risk missing the bigger structural change that the technology enables.
A new framework describes five value models through which AI can gradually reshape companies. The first stage focuses on workforce empowerment, where tools such as ChatGPT spread AI capabilities across teams and improve everyday productivity.
Once employees develop fluency, organisations can introduce AI-native distribution models that transform how customers discover products and interact with digital services.
More advanced stages involve specialised systems. Expert capability integrates AI into research, creative production, and domain-specific analysis, allowing professionals to explore a wider range of ideas and experiments.
Meanwhile, systems and dependency management introduce AI tools capable of safely updating interconnected digital environments, including codebases, documentation, and operational processes.
The final stage involves full process re-engineering through autonomous agents. In such environments, AI systems coordinate complex workflows across departments while maintaining governance, accountability, and auditability.
Organisations that successfully progress through these stages may eventually redesign their business models rather than merely improving efficiency within existing structures.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
The Google Threat Intelligence Group (GTIG) has identified a powerful exploit toolkit, Coruna, that targets Apple iPhones running iOS versions 13.0 to 17.2.1.
The toolkit contains five complete exploit chains and 23 exploits designed to compromise devices using previously unseen techniques and mitigation bypasses.
Parts of the exploit chain were first detected in early 2025, when a client of a commercial surveillance vendor used them. Later investigations revealed the same framework in highly targeted attacks against Ukrainian users linked to a suspected Russian espionage group.
Toward the end of the year, the toolkit resurfaced in large-scale campaigns linked to financially motivated actors operating from China.
Coruna relies on a sophisticated JavaScript framework that identifies iPhone models and their iOS versions before delivering the appropriate WebKit remote code execution exploit and additional bypass techniques.
Several vulnerabilities exploited by the toolkit had previously been treated as zero-day flaws, highlighting the growing circulation of advanced cyber-attack tools among multiple threat actors.
Google warned that the payload can steal sensitive data, including financial and cryptocurrency wallet information, and allows attackers to deploy additional modules remotely.
The company has added related malicious domains to Safe Browsing and urged users to install the latest iOS updates, noting that the exploit kit does not affect the newest version of Apple’s operating system.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
The European Union’s data protection watchdog has urged stronger safeguards as negotiations continue with the US over access to biometric databases. European Data Protection Supervisor Wojciech Wiewiórowski said limits must ensure Europeans’ data is used only for agreed purposes.
Talks between the EU and the US involve potential arrangements that would allow US authorities to query national biometric systems. Databases across the EU contain sensitive information, including fingerprints and facial recognition data.
Past transatlantic data-sharing agreements between the two have faced legal challenges due to insufficient safeguards. European regulators are closely monitoring the Data Privacy Framework amid ongoing concerns about oversight.
Officials also warned that emerging AI technologies could create new surveillance risks linked to US data access. European authorities said they must negotiate as a unified bloc when dealing with the US.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
South Korea’s ruling Democratic Party and the Financial Services Commission have agreed to cap major shareholder stakes in domestic crypto exchanges at 20%. Exceptions of up to 34% would apply to new businesses to support early-stage operators.
Large exchanges like Upbit and Bithumb will have 3 years to comply, while smaller platforms will receive an additional 3-year grace period.
Current ownership exceeds the proposed cap, with Upbit at 25.5%, Bithumb at 73.6%, and Coinone at 53.4%. Korbit’s pending acquisition would give Mirae Asset Consulting 92% ownership, highlighting the extent of concentrated holdings in the market.
The cap seeks to curb governance risks from concentrated shareholding, following the FSC’s January 2026 proposal. The move gained urgency after Bithumb’s accidental $43 billion Bitcoin transfer, which raised concerns about internal controls.
The ownership limit will likely be included in South Korea’s upcoming Digital Asset Basic Act, alongside rules on stablecoins and crypto ETFs.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Chinese authorities have introduced new rules to classify online content that could affect the health and well-being of minors. Set to take effect on 1 March, the measures aim to adapt to a rapidly evolving internet landscape.
Top government bodies, including those in cyberspace, education, publishing, film, culture, tourism, public security, and radio and television, jointly released the initiative. Together, they outlined four categories of content that could negatively impact minors and specified their key characteristics.
Recent issues, such as the misuse of minors’ images, have been integrated into the regulatory framework. Authorities also established preventive guidelines to manage risks from emerging technologies, including algorithmic recommendations and generative AI.
Internet platforms and content producers are now required to take both proactive and corrective measures against harmful content. The rules emphasise that platforms must monitor, block, or remove information that could affect minors’ well-being.
The Cyberspace Administration of China pledged to continue purifying the online environment. Authorities will urge platforms to assume their primary responsibilities and strengthen governance of content affecting young users, aiming to create a safer and healthier digital space for children.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Governments across the world are increasingly treating AI as a strategic capability that shapes economic development, public services and national security. Momentum behind the idea of ‘sovereign AI’ is growing as countries reassess who controls the chips, cloud infrastructure, data and models powering modern technology.
Complete control over the entire AI stack remains unrealistic for most economies because of the enormous financial and technological costs involved. Global infrastructure continues to rely heavily on US technology firms, which still operate a large share of data centres and AI systems worldwide.
Policy makers are therefore exploring different approaches to sovereignty across the AI ecosystem rather than pursuing total independence. Strategies range from building domestic computing capacity to adapting global AI models for national languages, regulations and public services.
Several countries already illustrate different approaches. The EU is investing billions in AI infrastructure, Canada protects sensitive computing resources while using global models, and India prioritises applications that serve its multilingual population through public digital systems.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Scripts, manual rules, and layered software tools traditionally ran telecom networks. A new collaboration between Google Cloud and Nokia suggests a shift: software agents can respond to goals rather than just detailed instructions.
The companies are integrating agent-based AI into Nokia’s Network as Code platform, which exposes telecom capabilities through application programming interfaces (APIs). The system allows developers to build applications that interact directly with network features such as connectivity quality, device location checks, or network slicing.
The Google-Nokia partnership introduces an AI layer that enables software agents to determine which network functions to use to achieve a goal. Such changes make development more efficient, as the AI agent can interpret instructions, automatically select the appropriate network capabilities, and reduce the need for developers to call APIs one step at a time manually.
Such automation is increasingly being explored as telecom infrastructure grows more complex with 5G, edge computing, and billions of connected devices. New features such as network slicing provide flexibility for industrial applications, private enterprise networks, and specialised connectivity, but also add operational complexity for operators.
Industry groups, including the GSMA and the 3rd Generation Partnership Project, are developing frameworks to support network APIs and automation. While agent-based AI could help networks operate more like programmable platforms, telecom operators must still address questions around reliability, security, and interoperability before large-scale deployment becomes feasible.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!