EU moves closer to decision on ChatGPT oversight

The European Commission plans to decide by early 2026 whether OpenAI’s ChatGPT should be classified as a vast online platform under the Digital Services Act.

OpenAI’s tool reported 120.4 million average monthly users in the EU back in October, a figure far above the 45-million threshold that triggers more onerous obligations instead of lighter oversight.

Officials said the designation procedure depends on both quantitative and qualitative assessments of how a service operates, together with input from national authorities.

The Commission is examining whether a standalone AI chatbot can fall within the scope of rules usually applied to platforms such as social networks, online marketplaces and significant search engines.

ChatGPT’s user data largely stems from its integrated online search feature, which prompts users to allow the chatbot to search the web. The Commission noted that OpenAI could voluntarily meet the DSA’s risk-reduction obligations while the formal assessment continues.

The EU’s latest wave of designations included Meta’s WhatsApp, though the rules applied only to public channels, not private messaging.

A decision on ChatGPT that will clarify how far the bloc intends to extend its most stringent online governance framework to emerging AI systems.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

France targets X over algorithm abuse allegations

The cybercrime unit of the Paris prosecutor has raided the French office of X as part of an expanding investigation into alleged algorithm manipulation and illicit data extraction.

Authorities said the probe began in 2025 after a lawmaker warned that biassed algorithms on the platform might have interfered with automated data systems. Europol supported the operation together with national cybercrime officers.

Prosecutors confirmed that the investigation now includes allegations of complicity in circulating child sex abuse material, sexually explicit deepfakes and denial of crimes against humanity.

Elon Musk and former chief executive Linda Yaccarino have been summoned for questioning in April in their roles as senior figures of the company at the time.

The prosecutor’s office also announced its departure from X in favour of LinkedIn and Instagram, rather than continuing to use the platform under scrutiny.

X strongly rejected the accusations and described the raid as politically motivated. Musk claimed authorities should focus on pursuing sex offenders instead of targeting the company.

The platform’s government affairs team said the investigation amounted to law enforcement theatre rather than a legitimate examination of serious offences.

Regulatory pressure increased further as the UK data watchdog opened inquiries into both X and xAI over concerns about Grok producing sexualised deepfakes. Ofcom is already conducting a separate investigation that is expected to take months.

The widening scrutiny reflects growing unease around alleged harmful content, political interference and the broader risks linked to large-scale AI systems.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

AI legal tool rattles European data stocks

European data and legal software stocks fell sharply after US AI startup Anthropic launched a new tool for corporate legal teams. The company said the software can automate contract reviews, compliance workflows, and document triage, while clarifying that it does not offer legal advice.

Investors reacted swiftly, sending shares in Pearson, RELX, Sage, Wolters Kluwer, London Stock Exchange Group, and Experian sharply lower. Thomson Reuters also suffered a steep decline, reflecting concern that AI tools could erode demand for traditional data-driven services.

Market commentators warned that broader adoption of AI in professional services could compress margins or bypass established providers altogether. Morgan Stanley flagged intensifying competition, while AJ Bell pointed to rising investor anxiety across the sector.

The sell-off also revived debate over AI’s impact on employment, particularly in legal and other office-based roles. Recent studies suggest the UK may face greater disruption than other large economies as companies adopt AI tools, even as productivity gains continue to rise.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Innovation and security shape the UAE’s tech strategy

The United Arab Emirates is strengthening its global tech role by treating advanced innovation as a pillar of sovereignty rather than a standalone growth driver. National strategy increasingly links technology with long-term economic resilience, security, and geopolitical relevance.

A key milestone was the launch of the UAE Advanced Technology Centre with the Technology Innovation Institute and the World Economic Forum, announced alongside the Davos gathering.

The initiative highlights the UAE’s transition from technology consumer to active participant in shaping global governance frameworks for emerging technologies.

The centre focuses on policy and governance for areas including artificial intelligence, quantum computing, biotechnology, robotics, and space-based payment systems.

Backed by a flexible regulatory environment, the UAE is promoting regulatory experimentation and translating research into real-world applications through institutions such as the Mohamed bin Zayed University of Artificial Intelligence and innovation hubs like Masdar City.

Alongside innovation, authorities are addressing rising digital risks, particularly deepfake technologies that threaten financial systems, public trust, and national security.

By combining governance, ethical standards, and international cooperation, the UAE is advancing a model of digital sovereignty that prioritises security, shared benefits, and long-term strategic independence.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Porto summit highlights growing risks to undersea internet cables

The Second International Submarine Cable Resilience Summit opened this week in Porto, Portugal, bringing together senior officials from governments, international organisations, and industry to address the growing risks facing the underwater cables that carry most of the world’s internet traffic. The event highlighted how submarine cables have become critical infrastructure for the global digital economy, especially as societies grow more dependent on cloud services, AI, and cross-border data flows.

Opening the summit, Ambassador João Mira Gomes, Permanent Representative of Portugal to the United Nations Office at Geneva, explained that Portugal’s infrastructure minister was absent due to ongoing storm recovery efforts, underlining the real-world pressures facing critical infrastructure today. He recalled Portugal’s long history in global connectivity, noting that one of the earliest submarine cables linking Portugal and the United Kingdom was built to support the port wine trade, a reminder that communication networks and economic exchange have long evolved together.

Professor Sandra Maximiano, co-chair of the International Advisory Body for Submarine Cable Resilience, placed the discussions in a broader historical context. She pointed to the creation of the International Telecommunication Union in 1865 as the first global organisation dedicated to managing international communications, stressing that cooperation on submarine cables has always been a ‘positive-sum game’ in which all countries benefit from shared rules and coordination.

Maximiano also highlighted Portugal’s strategic role as a cable hub, citing its extensive coastline, large exclusive economic zone, and favourable landing conditions connecting Europe, the Americas, Africa, and Asia. She outlined key projects such as the Atlantic CAM system linking mainland Portugal with Madeira and the Azores using a resilient ring design and smart cable technology that combines telecommunications with seismic and oceanographic monitoring. Existing and planned systems, she said, are not just data pipelines but foundations for innovation, scientific cooperation, and strategic autonomy.

A major outcome of the summit was the adoption of the Porto Declaration on Submarine Cable Resilience, developed with input from more than 150 experts worldwide. The declaration sets out practical guidance to improve permitting and repair processes, strengthen legal frameworks, promote route diversity and risk mitigation, and enhance capacity-building, with special attention to the needs of small island states and developing countries.

ITU Secretary-General Doreen Bogdan-Martin framed these efforts within a rapidly changing digital landscape, announcing that 2026 will be designated the ‘year of resilience.’ She warned that the scale of global digital dependence has transformed the impact of cable disruptions, as even minor outages can ripple across AI systems, cloud platforms, and autonomous services. Resilience, she argued, now depends as much on international coordination and preparedness as on cable design itself.

From the European Union perspective, European Commission Vice-President Henna Virkkunen outlined upcoming EU measures, including a submarine cable security toolbox and targeted funding through the Connecting Europe Facility. She stressed the importance of regional coordination and praised Portugal’s active role in aligning EU initiatives with global efforts led by the ITU.

Closing the opening session, Ambassador Gomes linked cable resilience to broader goals of development and peace, warning that digital divides fuel inequality and instability, and reaffirming Portugal’s commitment to international cooperation and capacity-building as the summit moves the global conversation from policy to action.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

ChatGPT restored after global outage disrupts users worldwide

OpenAI faced a wave of global complaints after many users struggled to access ChatGPT.

Reports began circulating in the US during the afternoon, with outage cases climbing to more than 12.000 in less than half an hour. Social media quickly filled with questions from people trying to determine whether the disruption was widespread or a local glitch.

Also, users in the UK reported complete failure to generate responses, yet access returned when they switched to a US-based VPN.

Other regions saw mixed results, as VPNs in Ireland, Canada, India and Poland allowed ChatGPT to function, although replies were noticeably slower instead of consistent.

OpenAI later confirmed that several services were experiencing elevated errors. Engineers identified the source of the disruption, introduced mitigations and continued monitoring the recovery.

The company stressed that users in many regions might still experience intermittent problems while the system stabilises rather than operating at full capacity.

In the following update, OpenAI announced that its systems were fully operational again.

The status page indicated that the affected services had recovered, and engineers were no longer aware of active issues. The company added that the underlying fault was addressed, with further safeguards being developed to prevent similar incidents.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

WordPress introduces rules for responsible AI use

WordPress has released new guidelines to shape how AI is used across plugins, themes, documentation and media assets. The framework focuses on transparency, accountability and maintaining the project’s open source foundations.

Contributors remain fully responsible for AI-assisted work and are expected to disclose meaningful AI use during submissions. Reviewers are encouraged to assess such contributions with awareness of how automated tools influenced the output.

Strong emphasis is placed on licensing, with all AI-generated material required to remain compatible with GPLv2 or later. Tools that restrict redistribution or reproduce incompatible code are explicitly ruled out.

The guidance also targets so-called AI slop, including untested code, fabricated references and unnecessarily complex solutions. Maintainers are authorised to reject low-quality submissions that lack apparent human oversight.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

AI skills are gaining momentum among college students

AI tools are already widely used in higher education, with more than half of surveyed students required to use them in coursework and nearly two-thirds using them for assignments. However, the survey suggests that students are largely learning to use AI on their own, relying mainly on informal experimentation rather than structured university-led training.

At the same time, awareness and participation in formal AI education remain limited. Only 31% of students said they were aware of AI-related courses offered by their college or university, and fewer than 20% had taken one, highlighting a gap between widespread use and institutional teaching.

Despite this, many students recognise AI’s growing importance for their careers. Around half believe proficiency with AI tools will be important in the future, reflecting expectations that AI skills will be increasingly valued in the workplace.

Overall, the findings point to an opportunity for universities to strengthen AI education by integrating practical, advanced, and ethical AI training into curricula, helping students move beyond basic use toward workplace-ready skills.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

New AI safety report highlights control concerns

A major international AI safety report warns that AI systems are advancing rapidly, with sharp gains in reasoning, coding and scientific tasks. Researchers say progress remains uneven, leaving systems powerful yet unreliable.

The report highlights rising concerns over deepfakes, cyber misuse and emotional reliance on AI companions in the UK and the US. Experts note growing difficulty in distinguishing AI generated content from human work.

Safeguards against biological, chemical and cyber risks have improved, though oversight challenges persist in the UK and the US. Analysts warn advanced models are becoming better at evading evaluation and controls.

The impact of AI on jobs in the UK and the US remains uncertain, with mixed evidence across sectors. Researchers say labour disruption could accelerate if systems gain greater autonomy.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Chinese AI firms offer cash rewards to boost chatbot adoption

Technology firms in China are rolling out large cash incentive campaigns to attract users to their AI chatbots ahead of the expected launch of new AI models later this month.

Alibaba Group has earmarked CNY 3 billion for users of its Qwen AI app, with the promotion beginning on 6 February to coincide with Lunar New Year celebrations.

Tencent Holdings and Baidu have announced similar offers, together committing around CNY 1.5 billion in cash rewards and consumer electronics, including smartphones and televisions.

To qualify for prizes, users must register on the platforms and interact with the chatbots during the promotional period by asking questions or completing everyday planning tasks.

The incentives reflect intensifying competition with global developers such as Google and OpenAI, while also strengthening efforts to position China-based firms as potential local AI partners for Apple in the Chinese market.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!