Microsoft restores Exchange and Teams after Microsoft 365 disruption

The US tech giant, Microsoft, investigated a service disruption affecting Exchange Online, Teams and other Microsoft 365 services after users reported access and performance problems.

An incident that began late on Wednesday affected core communication tools used by enterprises for daily operations.

Engineers initially focused on diagnosing the fault, with Microsoft indicating that a potential third-party networking issue may have interfered with access to Outlook and Teams.

During the disruption, users experienced intermittent connectivity failures, latency and difficulties signing in across parts of the Microsoft 365 ecosystem.

Microsoft later confirmed that service access had been restored, although no detailed breakdown of the outage scope was provided.

The incident underlined the operational risks associated with cloud productivity platforms and the importance of transparency and resilience in enterprise digital infrastructure.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

From chips to jobs: Huang’s vision for AI at Davos 2026

AI is evolving into a foundational economic system rather than a standalone technology, according to NVIDIA chief executive Jensen Huang, who described AI as a five-layer infrastructure spanning energy, hardware, data centres, models and applications.

Speaking at the World Economic Forum in Davos, Huang argued that building and operating each layer is triggering what he called the most significant infrastructure expansion in human history, with job creation stretching from power generation and construction to cloud operations and software development.

Investment patterns suggest a structural shift instead of a speculative cycle. Venture capital funding in 2025 reached record levels, largely flowing into AI-native firms across healthcare, manufacturing, robotics and financial services.

Huang stressed that the application layer will deliver the most significant economic return as AI moves from experimentation to core operational use across industries.

Concerns around job displacement were framed as misplaced. AI automates tasks rather than replacing professional judgement, enabling workers to focus on higher-value activities.

In healthcare, productivity gains from AI-assisted diagnostics and documentation are already increasing demand for radiologists and nurses rather than reducing headcount, as improved efficiency enables institutions to treat more patients.

Huang positioned AI as critical national infrastructure, urging governments to develop domestic capabilities aligned with local language, culture and industrial strengths.

He described AI literacy as an essential skill, comparable to leadership or management, while arguing that accessible AI tools could narrow global technology divides rather than widen them.

Diplo is live reporting on all sessions from the World Economic Forum 2026 in Davos.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

GPT-5.2 shows how AI can generate real-world cyber exploits

Advanced language models have demonstrated the ability to generate working exploits for previously unknown software vulnerabilities. Security researcher Sean Heelan tested two systems built on GPT-5.2 and Opus 4.5 by challenging them to exploit a zero-day flaw in the QuickJS JavaScript interpreter.

Across multiple scenarios with varying security protections, GPT-5.2 completed every task, while Opus 4.5 failed only 2. The systems produced more than 40 functional exploits, ranging from basic shell access to complex file-writing operations that bypassed modern defences.

Most challenges were solved in under an hour, with standard attempts costing around $30. Even the most complex exploit, which bypassed protections such as address space layout randomisation, non-executable memory, and seccomp sandboxing, was completed in just over three hours for roughly $50.

The most advanced task required GPT-5.2 to write a specific string to a protected file path without access to operating system functions. The model achieved this by chaining seven function calls through the glibc exit handler mechanism, bypassing shadow stack protections.

The findings suggest exploit development may increasingly depend on computational resources rather than human expertise. While QuickJS is less complex than browsers such as Chrome or Firefox, the approach demonstrated could scale to larger and more secure software environments.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

YouTube’s 2026 strategy places AI at the heart of moderation and monetisation

As announced yesterday, YouTube is expanding its response to synthetic media by introducing experimental likeness detection tools that allow creators to identify videos where their face appears altered or generated by AI.

The system, modelled conceptually on Content ID, scans newly uploaded videos for visual matches linked to enrolled creators, enabling them to review content and pursue privacy or copyright complaints when misuse is detected.

Participation requires identity verification through government-issued identification and a biometric reference video, positioning facial data as both a protective and governance mechanism.

While the platform stresses consent and limited scope, the approach reflects a broader shift towards biometric enforcement as platforms attempt to manage deepfakes, impersonation, and unauthorised synthetic content at scale.

Alongside likeness detection, YouTube’s 2026 strategy places AI at the centre of content moderation, creator monetisation, and audience experience.

AI tools already shape recommendation systems, content labelling, and automated enforcement, while new features aim to give creators greater control over how their image, voice, and output are reused in synthetic formats.

The move highlights growing tensions between creative empowerment and platform authority, as safeguards against AI misuse increasingly rely on surveillance, verification, and centralised decision-making.

As regulators debate digital identity, biometric data, and synthetic media governance, YouTube’s model signals how private platforms may effectively set standards ahead of formal legislation.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Snapchat settles social media addiction lawsuit as landmark trial proceeds

Snapchat’s parent company has settled a social media addiction lawsuit in California just days before the first major trial examining platform harms was set to begin.

The agreement removes Snapchat from one of the three bellwether cases consolidating thousands of claims, while Meta, TikTok and YouTube remain defendants.

These lawsuits mark a legal shift away from debates over user content and towards scrutiny of platform design choices, including recommendation systems and engagement mechanics.

A US judge has already ruled that such features may be responsible for harm, opening the door to liability that section 230 protections may not cover.

Legal observers compare the proceedings to historic litigation against tobacco and opioid companies, warning of substantial damages and regulatory consequences.

A ruling against the remaining platforms could force changes in how social media products are designed, particularly in relation to minors and mental health risks.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Why AI systems privilege Western perspectives: ‘The Silicon Gaze’

A new study from the University of Oxford argues that large language models reproduce a distinctly Western hierarchy when asked to evaluate countries, reinforcing long-standing global inequalities through automated judgment.

Analysing more than 20 million English-language responses from ChatGPT’s 4o-mini model, researchers found consistent favouring of wealthy Western nations across subjective comparisons such as intelligence, happiness, creativity, and innovation.

Low-income countries, particularly across Africa, were systematically placed at the bottom of rankings, while Western Europe, the US, and parts of East Asia dominated positive assessments.

According to the study, generative models rely heavily on data availability and dominant narratives, leading to flattened representations that recycle familiar stereotypes instead of reflecting social complexity or cultural diversity.

The researchers describe the phenomenon as the ‘silicon gaze’, a worldview shaped by the priorities of platform owners, developers, and historically uneven training data.

Because large language models are trained on material produced within centuries of structural exclusion, bias emerges not as a malfunction but as an embedded feature of contemporary AI systems.

The findings intensify global debates around AI governance, accountability, and cultural representation, particularly as such systems increasingly influence healthcare, employment screening, education, and public decision-making.

While models are continuously updated, the study underlines the limits of technical mitigation without broader political, regulatory, and epistemic interventions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

How Microsoft is shaping UN reform through digital infrastructure

Microsoft has announced a multi-year pledge to support the United Nations’ UN80 reform initiative, positioning AI and digital infrastructure as central tools for modernising multilateral governance.

The commitment follows agreement among all UN member states on efficiency and financial-stability measures, as the organisation faces growing operational demands amid constrained resources.

The initiative includes a dedicated innovation fund, preferential pricing for digital services, and free AI training for UN staff across agencies and missions.

Rather than focusing on policy direction, Microsoft frames its role as enabling institutional capacity, from procurement and logistics to humanitarian response and development planning, while encouraging other private-sector actors to align behind UN80 priorities.

Microsoft also plans to mobilise partners such as EY to support reform efforts, reinforcing a model where large technology firms contribute expertise, infrastructure, and coordination capacity to global governance systems.

Previous collaborations with UNICEF, UNHCR, ITU, and the ILO are cited as evidence that AI-driven tools can accelerate service delivery at scale.

The pledge highlights how multilateral reform increasingly depends on private technological ecosystems instead of purely intergovernmental solutions.

As AI becomes embedded in the core operations of international institutions, questions around accountability, influence, and long-term dependency are likely to shape debates about the future balance between public authority and corporate power.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

ChatGPT introduces age prediction to strengthen teen safety

New safeguards are being introduced as ChatGPT uses age prediction to identify accounts that may belong to under-18s. Extra protections limit exposure to harmful content while still allowing adults full access.

The age prediction model analyses behavioural and account-level signals, including usage patterns, activity times, account age, and stated age information. OpenAI says these indicators help estimate whether an account belongs to a minor, enabling the platform to apply age-appropriate safeguards.

When an account is flagged as potentially under 18, ChatGPT limits access to graphic violence, sexual role play, viral challenges, self-harm, and unhealthy body image content. The safeguards reflect research on teen development, including differences in risk perception and impulse control.

ChatGPT users who are incorrectly classified can restore full access by confirming their age through a selfie check using Persona, a secure identity verification service. Account holders can review safeguards and begin the verification process at any time via the settings menu.

Parental controls allow further customisation, including quiet hours, feature restrictions, and notifications for signs of distress. OpenAI says the system will continue to evolve, with EU-specific deployment planned in the coming weeks to meet regional regulatory requirements.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

New AI system helps improve cross-neurotype communication

Researchers at Tufts University have developed an AI-based learning tool designed to improve communication between autistic and neurotypical people. The project focuses on helping non-autistic users better understand autistic communication preferences.

The tool, called NeuroBridge, uses large language models to simulate everyday conversations and highlight how wording, tone and clarity can be interpreted differently. Users are guided towards more direct and unambiguous communication styles that reduce misunderstanding.

Unlike many interventions, NeuroBridge does not aim to change how autistic people communicate. The AI system instead trains neurotypical users to adapt their own communication, reflecting principles from the social model of disability.

The research, presented at the ACM SIGACCESS Conference on Computers and Accessibility, received a best student paper award. Early testing showed users gained clearer insight into how everyday language choices can affect cross-neurotype interactions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Amodei warns US AI chip exports to China risk national security

Anthropic chief executive Dario Amodei has criticised the US decision to allow the export of advanced AI chips to China, warning it could undermine national security. Speaking at the World Economic Forum 2026 in Davos, he questioned whether selling US-made hardware abroad strengthens American influence.

Amodei compared the policy to ‘selling nuclear weapons to North Korea‘, arguing that exporting cutting-edge chips risks narrowing the technological gap between the United States and China. He said Washington currently holds a multi-year lead in advanced chipmaking and AI infrastructure.

Sending powerful hardware overseas could accelerate China’s progress faster than expected, Amodei told Bloomberg. He warned that AI development may soon concentrate unprecedented intelligence within data centres controlled by individual states.

Amodei said AI should not be treated like older technologies such as telecoms equipment. While spreading US technology abroad may have made sense in the past, he argued AI carries far greater strategic consequences.

The debate follows recent rule changes allowing some advanced chips, including Nvidia’s H200 and AMD’s MI325X, to be sold to China. The US administration later announced plans for a 25% tariff on AI chip exports, adding uncertainty for US semiconductor firms.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!