IBM’s 2025 Cost of a Data Breach Report revealed a sharp gap between rapid AI adoption and the oversight needed to secure it.
Although the global average data breach cost fell slightly to $4.44 million, security incidents involving AI systems remain more severe and disruptive.
Around 13% of organisations reported breaches involving AI models or applications, while 8% were unsure whether they had been compromised.
Alarmingly, nearly all AI-related breaches occurred without access controls, leading to data leaks in 60% of cases and operational disruption in almost one-third. Shadow AI (unsanctioned or unmanaged systems) played a central role, with one in five breaches traced back to it.
Organisations without AI governance policies or detection systems faced significantly higher costs, especially when personally identifiable information or intellectual property was exposed.
Attackers increasingly used AI tools such as deepfakes and phishing, with 16% of studied breaches involving AI-assisted threats.
Healthcare remained the costliest sector, with an average breach price of $7.42 million and the most extended recovery timeline of 279 days.
Despite the risks, fewer organisations plan to invest in post-breach security. Only 49% intend to strengthen defences, down from 63% last year.
Even fewer will prioritise AI-driven security tools. With many organisations also passing costs on to consumers, recovery now often includes long-term financial and reputational fallout, not just restoring systems.
Would you like to learn more aboutAI, tech and digital diplomacy? If so, ask our Diplo chatbot!
The way UK internet users access adult content has undergone a significant change, with new age-verification rules now in force. Under Ofcom’s directive, anyone attempting to visit adult websites must now prove they are over 18, typically by providing credit card or personal ID details.
The move aims to prevent children from encountering harmful content online, but it has raised serious privacy and cybersecurity concerns.
Experts have warned that entering personal and financial information could expose users to cyber threats. Jake Moore from cybersecurity firm ESET pointed out that the lack of clear implementation standards leaves users vulnerable to data misuse and fraud.
There’s growing unease that ID verification systems might inadvertently offer a goldmine to scammers. In response, many have started using VPNs to bypass the restrictions, with providers reporting a surge in UK downloads.
VPNs mask user locations, allowing access to blocked content, but free versions often lack the security features of paid services. As demand rises, cybersecurity specialists are urging users to be cautious.
Free VPNs can compromise user data through weak encryption or selling browsing histories to advertisers. Mozilla and EC-Council have stressed the importance of avoiding no-cost VPNs unless users know the risks.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
The FBI has issued a public warning about a rising scam involving QR codes placed on packages delivered to people who never ordered them.
According to the agency, these codes can lead recipients to malicious websites or prompt them to install harmful software, potentially exposing sensitive personal and financial data.
The scheme is a variation of the so-called brushing scam, in which online sellers send unordered items and use recipients’ names to post fake product reviews. In the new version, QR codes are added to the packaging, increasing the risk of fraud by directing users to deceptive websites.
While not as widespread as other fraud attempts, the FBI urges caution. The agency recommends avoiding QR codes from unknown sources, especially those attached to unrequested deliveries.
It also advised consumers to pay close attention to the web address that appears before tapping on any QR code link.
Authorities have noted broader misuse of QR codes, including cases where criminals place fake codes over legitimate ones in public spaces.
In one recent incident, scammers used QR stickers on parking meters in New York to redirect people to third-party payment pages requesting card details.
Would you like to learn more aboutAI, tech and digital diplomacy? If so, ask our Diplo chatbot!
UK Justice Secretary Shabana Mahmood has rolled out an AI-driven violence prediction tool across prisons and probation services. One system evaluates inmates’ profiles, factoring in age, past behaviour, and gang ties, to flag those likely to become violent. Matching prisoners to tighter supervision or relocation aims to reduce attacks on staff and fellow inmates.
Another feature actively scans content from seized mobile phones. AI algorithms sift through over 33,000 devices and 8.6 million messages, detecting coded language tied to contraband, violence, or escape plans. When suspicious content is flagged, staff receive alerts for preventive action.
Rising prison violence and self-harm underscore the urgency of such interventions. Assaults on staff recently reached over 10,500 a year, the highest on record, while self-harm incidents reached nearly 78,000. Overcrowding and drug infiltration have intensified operational challenges.
Analysts compare the approach to ‘pre‑crime’ models, drawing parallels with sci-fi narratives, raising concerns around civil liberties. Without robust governance, predictive tools may replicate biases or punish potential rather than actual behaviour. Transparency, independent audit, and appeals processes are essential to uphold inmate rights.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Microsoft Research released a comprehensive AI impact assessment, ranking 80 occupations by exposure to generative AI tools such as Copilot and ChatGPT. Roles heavily involved in language, writing, client communication, and routine digital tasks showed the highest AI overlap. Notable examples include translators, historians, customer service agents, political scientists, and data scientists.
By contrast, jobs requiring hands-on work, empathy, real-time physical or emotional engagement, such as nurses, phlebotomists, construction trades, embalmers, and housekeeping staff, were classified as low risk under current AI capabilities. Experts suggest that these kinds of positions remain essential because they involve physical presence, human interaction, and complex real-time decision making.
Although certain professions scored high for AI exposure, Microsoft and independent analysts emphasise that most jobs won’t disappear entirely. Instead, generative AI tools are expected to augment workflows, creating hybrid roles where human judgement and oversight remain critical, especially in sectors such as financial services, healthcare, and creative industries.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
July witnessed a significant surge in cybersecurity mergers and acquisitions (M&A), spearheaded by Palo Alto Networks’ announcement of its definitive agreement to acquire identity security firm CyberArk for an estimated $25 billion.
The transaction, set to be the second-largest cybersecurity acquisition on record, signals Palo Alto’s strategic entry into identity security.
Beyond this significant deal, Palo Alto Networks also completed its purchase of AI security specialist Protect AI. The month saw widespread activity across the sector, including LevelBlue’s acquisition of Trustwave to create the industry’s largest pureplay managed security services provider.
Zurich Insurance Group, Signicat, Limerston Capital, Darktrace, Orange Cyberdefense, SecurityBridge, Commvault, and Axonius all announced or finalised strategic cybersecurity acquisitions.
The deals highlight a strong market focus on AI security, identity management, and expanding service capabilities across various regions.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Amazon is exploring ways to insert ads into conversations with its AI assistant Alexa+, according to CEO Andy Jassy. Speaking during the company’s latest earnings call, he described the feature as a potential tool for product discovery and future revenue.
Alexa+ is Amazon’s upgraded digital assistant designed to support more natural, multi-step conversations using generative AI. It is already available to millions of users through Prime subscriptions or as a standalone service.
Jassy said longer interactions open the door for embedded advertising, although the approach has not yet been fully developed. Industry observers see this as part of a wider trend, with companies like Google and OpenAI also weighing ad-based business models.
Alexa+ has received mixed reviews so far, with delays in feature delivery and technical challenges like hallucinations raising concerns. Privacy advocates have warned that ad targeting within personal conversations may worry users, given the data involved.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
Apple is ramping up its AI efforts, with CEO Tim Cook confirming that the company is significantly increasing its investments in the technology. During the Q3 2025 earnings call, Cook said AI would be embedded across Apple’s devices, platforms and internal operations.
The firm has reallocated staff to focus on AI and continues to acquire smaller companies to accelerate progress, completing seven acquisitions this year alone. Capital expenditure has also risen, partly due to the growing focus on AI.
Despite criticism that Apple has lagged behind in the AI race, the company insists it will not rush features to market. More than 20 Apple Intelligence tools have already been released, with additional features like live translation and an AI fitness assistant expected by year-end.
The updated version of Siri, which promises greater personalisation, has been pushed to 2026. Cook dismissed suggestions that AI-powered hardware, like glasses, would replace the iPhone, instead positioning future devices as complementary.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
OpenAI has removed a feature that allowed users to make their ChatGPT conversations publicly searchable, following backlash over accidental exposure of sensitive content.
Dane Stuckey, OpenAI’s CISO, confirmed the rollback on Thursday, describing it as a short-lived experiment meant to help users find helpful conversations. However, he acknowledged that the feature posed privacy risks.
‘Ultimately, we think this feature introduced too many opportunities for folks to accidentally share things they didn’t intend to,’ Stuckey wrote in a post on X. He added that OpenAI is working to remove any indexed content from search engines.
The move came swiftly after Fast Company and privacy advocate Luiza Jarovsky reported that some shared conversations were appearing in Google search results.
Jarovsky posted examples on X, noting that even though the chats were anonymised, users were unknowingly revealing personal experiences, including harassment and mental health struggles.
To activate the feature, users had to tick a box allowing their chat to be discoverable. While the process required active steps, critics warned that some users might opt in without fully understanding the consequences. Stuckey said the rollback will be complete by Friday morning.
The incident adds to growing concerns around AI and user privacy, particularly as conversational platforms like ChatGPT become more embedded in everyday life.
Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!
UK universities risk losing their competitive edge unless they adopt a clear, forward-looking approach to ΑΙ in teaching. Falling enrolments, limited funding, and outdated digital systems have exposed a lack of AI literacy across many institutions.
As AI skills become essential for today’s workforce, employers increasingly expect graduates to be confident users rather than passive observers.
Many universities continue relying on legacy technology rather than exploring the full potential of modern learning platforms. AI tools can enhance teaching by adapting to individual student needs and helping educators identify learning gaps.
However, few staff have received adequate training, and many universities lack the resources or structure to embed AI into day-to-day teaching effectively.
To close the growing gap between education and the workplace, universities must explore flexible short courses and microcredentials that develop workplace-ready skills.
Introducing ethical standards and data transparency from the start will ensure AI is used responsibly without weakening academic integrity.
Would you like to learn more aboutAI, tech and digital diplomacy? If so, ask our Diplo chatbot!