UK health sector adopts AI while legacy tech lags

The UK’s healthcare sector has rapidly embraced AI, with adoption rising from 47% in 2024 to 94% in 2025, according to SOTI’s new report ‘Healthcare’s Digital Dilemma’.

AI is no longer confined to administrative tasks, as 52% of healthcare professionals now use it for diagnosis and 57% to personalise treatments. SOTI’s Stefan Spendrup said AI is improving how care is delivered and helping clinicians make more accurate, patient-specific decisions.

However, outdated systems continue to hamper progress. Nearly all UK health IT leaders report challenges from legacy infrastructure, Internet of Things (IoT) tech and telehealth tools.

While connected devices are widely used to support patients remotely, 73% rely on outdated, unintegrated systems, significantly higher than the global average of 65%.

These systems limit interoperability and heighten security risks, with 64% experiencing regular tech failures and 43% citing network vulnerabilities.

The strain on IT teams is evident. Nearly half report being unable to deploy or manage new devices efficiently, and more than half struggle to offer remote support or access detailed diagnostics. Time lost to troubleshooting remains a common frustration.

The UK appears more affected by these challenges than other countries surveyed, indicating a pressing need to modernise infrastructure instead of continuing to patch ageing technology.

While data security remains the top IT concern in UK healthcare, fewer IT teams see it as a priority, falling from 33% in 2024 to 24% in 2025. Despite a sharp increase in data breaches, the number rose from 71% to 84%.

Spendrup warned that innovation risks being undermined unless the sector rebalances priorities, with more focus on securing systems and replacing legacy tools instead of delaying necessary upgrades.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

NSA and allies set AI data security standards

The National Security Agency (NSA), in partnership with cybersecurity agencies from the UK, Australia, New Zealand, and others, has released new guidance aimed at protecting the integrity of data used in AI systems.

The Cybersecurity Information Sheet (CSI), titled AI Data Security: Best Practices for Securing Data Used to Train & Operate AI Systems, outlines emerging threats and sets out 10 recommendations for mitigating them.

The CSI builds on earlier joint guidance from 2024 and signals growing global urgency around safeguarding AI data instead of allowing systems to operate without scrutiny.

The report identifies three core risks across the AI lifecycle: tampered datasets in the supply chain, deliberately poisoned data intended to manipulate models, and data drift—where changes in data over time reduce performance or create new vulnerabilities.

These threats may erode accuracy and trust in AI systems, particularly in sensitive areas like defence, cybersecurity, and critical infrastructure, where even small failures could have far-reaching consequences.

To reduce these risks, the CSI recommends a layered approach—starting with sourcing data from reliable origins and tracking provenance using digital credentials. It advises encrypting data at every stage, verifying integrity with cryptographic tools, and storing data securely in certified systems.

Additional measures include deploying zero trust architecture, using digital signatures for dataset updates, and applying access controls based on data classification instead of relying on broad administrative trust.

The CSI also urges ongoing risk assessments using frameworks like NIST’s AI RMF, encouraging organisations to anticipate emerging challenges such as quantum threats and advanced data manipulation.

Privacy-preserving techniques, secure deletion protocols, and infrastructure controls round out the recommendations.

Rather than treating AI as a standalone tool, the guidance calls for embedding strong data governance and security throughout its lifecycle to prevent compromised systems from shaping critical outcomes.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Fake DeepSeek ads deliver ‘BrowserVenom’ malware to curious AI users

Cybercriminals are exploiting the surge in interest around local AI tools by spreading a new malware strain via Google ads.

According to antivirus firm Kaspersky, attackers use fake ads for DeepSeek’s R1 AI model to deliver ‘BrowserVenom,’ malware designed to intercept and manipulate a user’s internet traffic instead of merely infecting the device.

The attackers purchased ads appearing in Google search results for ‘deep seek r1.’ Users who clicked were redirected to a fake website—deepseek-platform[.]com—which mimicked the official DeepSeek site and offered a file named AI_Launcher_1.21.exe.

Kaspersky’s analysis of the site’s source code uncovered developer notes in Russian, suggesting the campaign is operated by Russian-speaking actors.

Once launched, the fake installer displayed a decoy installation screen for the R1 model, but silently deployed malware that altered browser configurations.

BrowserVenom rerouted web traffic through a proxy server controlled by the hackers, allowing them to decrypt browsing sessions and capture sensitive data, while evading most antivirus tools.

Kaspersky reports confirmed infections across multiple countries, including Brazil, Cuba, India, and South Africa.

The malicious domain has since been taken down. However, the incident highlights the dangers of downloading AI tools from unofficial sources. Open-source models like DeepSeek R1 require technical setup, typically involving multiple configuration steps, instead of a simple Windows installer.

As interest in running local AI grows, users should verify official domains and avoid shortcuts that could lead to malware.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Coreweave expands AI infrastructure with Google tie‑up

CoreWeave has secured a pivotal role in Google Cloud’s new infrastructure partnership with OpenAI. The specialist GPU cloud provider will supply Nvidia‑based compute resources to Google, which will allocate them to OpenAI to support the rising demand for services like ChatGPT.

Already under a $11.9 billion, five‑year contract with OpenAI and backed by a $350 million equity investment, CoreWeave recently expanded the deal by another. 

Adding Google Cloud as a customer helps the company diversify beyond Microsoft, its top client in 2024.

The arrangement positions Google as a neutral provider of AI computing power amid fierce competition with Amazon and Microsoft.

CoreWeave’s stock has surged over 270 percent since its March IPO, illustrating investor confidence in its expanding role in the AI infrastructure boom.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Meta sues AI firm over fake nude images created without consent

Meta has filed a lawsuit against Joy Timeline HK Ltd in Hong Kong, accusing the firm of using its platforms to promote a generative AI app called CrushAI.

The app allows users to digitally strip clothes from images of people, often without consent. Meta said the company repeatedly attempted to bypass ad review systems to push harmful content, advertising phrases like ‘see anyone naked’ on Facebook and Instagram.

The lawsuit follows Meta’s broader investigation into ‘nudity’ apps, which are increasingly being used to create sexualised deepfakes. Despite bans on nonconsensual explicit content, the company said such apps evade detection by disguising ads or rotating domain names after bans.

According to research by Cornell Tech, over 8,000 ads linked to CrushAI appeared on Meta platforms in recent months. Meta responded by updating its detection systems with a broader range of flagged terms and emojis.

While many of the manipulated images target celebrities, concerns are growing about the use of such technology to exploit minors. In one case in Florida, two teenagers used similar AI tools to create sexualised images of classmates.

The issue has sparked legal action in the US, where the Take It Down Act, signed into law earlier this year, criminalises the publication of nonconsensual deepfake imagery and simplifies removal processes for victims.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

New ion trap chip paves way for scalable quantum systems

Researchers at the Quantum Systems Accelerator have announced significant progress in building scalable, stable quantum computers focusing on trapped-ion technology.

Their work marks a series of engineering milestones pushing quantum computing toward practical use.

A new ion trap chip can store up to 200 ions and significantly reduces power loss by redesigning its internal layout.

Developed and tested with collaborators at Duke and Cornell in the US, this design allows for the future creation of far larger qubit systems without overheating or energy waste.

At the University of Maryland, a team achieved parallel quantum gate operations using different spatial directions, overcoming prior interference issues.

However, this innovation boosts processing speed and accuracy, offering more efficient handling of time-sensitive quantum tasks.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

AI cheating crisis leaves teachers in despair

Teachers across the US are growing alarmed by widespread student use of AI for assignments, calling it a crisis that undermines education itself. Some professors report that students now rely on AI for everything from note-taking to essay writing, leaving educators questioning the future of learning.

The fear of false accusations is rising among honest students, with some recording their screens to prove their work is genuine. Detection tools often misfire, further complicating efforts to distinguish real effort from AI assistance.

While some argue for banning tech and returning to traditional classroom methods, others suggest rethinking US education entirely. Rather than fighting AI, some believe it offers a chance to re-engage students by giving them meaningful work they want to do.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Turing Institute urges stronger AI research security

The Alan Turing Institute has warned that urgent action is needed to protect the UK’s AI research from espionage, intellectual property theft and risky international collaborations.

Its Centre for Emerging Technology and Security (CETaS) has published a report calling for a culture shift across academia to better recognise and mitigate these risks.

The report highlights inconsistencies in how security risks are understood within universities and a lack of incentives for researchers to follow government guidelines. Sensitive data, the dual-use potential of AI, and the risk of reverse engineering make the field particularly vulnerable to foreign interference.

Lead author Megan Hughes stressed the need for a coordinated response, urging government and academia to find the right balance between academic freedom and security.

The report outlines 13 recommendations, including expanding support for academic due diligence and issuing clearer guidance on high-risk international partnerships.

Further proposals call for compulsory research security training, better threat communication from national agencies, and standardised risk assessments before publishing AI research.

The aim is to build a more resilient research ecosystem as global interest in UK-led AI innovation continues to grow.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Meta launches AI to teach machines physical reasoning

Meta Platforms has unveiled V-JEPA 2, an open-source AI model designed to help machines understand and interact with the physical world more like humans do.

The technology allows AI agents, including delivery robots and autonomous vehicles, to observe object movement and predict how those objects may behave in response to actions.

The company explained that just as people intuitively understand that a ball tossed into the air will fall due to gravity, AI systems using V-JEPA 2 gain a similar ability to reason about cause and effect in the real world.

Trained using video data, the model recognises patterns in how humans and objects move and interact, helping machines learn to reach, grasp, and reposition items more naturally.

Meta described the tool as a step forward in building AI that can think ahead, plan actions and respond intelligently to dynamic environments. In lab tests, robots powered by V-JEPA 2 performed simple tasks that relied on spatial awareness and object handling.

The company, led by CEO Mark Zuckerberg, is ramping up its AI initiatives to compete with rivals like Microsoft, Google, and OpenAI. By improving machine reasoning through world models such as V-JEPA 2, Meta aims to accelerate its progress toward more advanced AI.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

RIPE NCC and Czech regulator partner to strengthen internet coordination

The RIPE NCC and the Czech Telecommunication Office (CTU) have signed a non-binding Memorandum of Understanding (MoU) to strengthen cooperation in internet coordination and share technical expertise within the Czech Republic. The partnership focuses on training internet operators, collaborating on network measurements, and managing internet number resources.

It is part of a broader initiative by RIPE NCC to establish similar agreements with national regulators in countries like Georgia and Saudi Arabia, reflecting their commitment to closer regional cooperation. The Czech Republic is strategically positioned in European internet infrastructure, hosting several major data centres and internet exchange points (IXPs).

By facilitating collaboration between public and private sectors, the MoU aims to ensure that internet policies are developed with broad input and expertise. The CTU benefits from access to valuable data and technical knowledge that support national digital policy objectives.

Additionally, in a region where geopolitical tensions may affect internet infrastructure, this agreement promotes transparency and cooperation that help stabilise internet operations and build stakeholder trust. Overall, the RIPE NCC continues to evolve as a key technical partner in digital policy discussions across Europe and beyond.

The agreement highlights the need for close cooperation between technical bodies and regulators as digital infrastructure grows more complex, emphasising multistakeholder governance to improve stability and efficiency in Central and Eastern Europe.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!