Cyberattack keeps JLR factories shut, hackers claim responsibility

Jaguar Land Rover (JLR) has confirmed that data was affected in a cyberattack that has kept its UK factories idle for more than a week. The company stated that it is contacting anyone whose data was involved, although it did not clarify whether the breach affected customers, suppliers, or internal systems.

JLR reported the incident to the Information Commissioner’s Office and immediately shut down IT systems to limit damage. Production at Midlands and Merseyside sites has been halted until at least Thursday, with staff instructed not to return before next week.

The disruption has also hit suppliers and retailers, with garages struggling to order spare parts and dealers facing delays registering vehicles. JLR said it is working around the clock to restore operations in a safe and controlled way, though the process is complex.

Responsibility for the hack has been claimed by Scattered Lapsus$ Hunters, a group linked to previous attacks on Marks & Spencer, the Co-op, and Las Vegas casinos in the UK and the US. The hackers posted alleged screenshots from JLR’s internal systems on Telegram last week.

Cybersecurity experts say the group’s claim that ransomware was deployed raises questions, as it appears to have severed ties with Russian ransomware gangs. Analysts suggest the hackers may have only stolen data or are building their own ransomware infrastructure.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Claude AI gains powerful file editing tools for documents and spreadsheets

Anthropic’s Claude has expanded its role as a leading AI assistant by adding advanced tools for creating and editing files. Instead of manually working with different programs, users can now describe their needs in plain language and let the AI produce or update Word, Excel, PowerPoint, and PDF files.

A feature that supports uploads of CSV and TSV data and can generate charts, graphs, or images where needed, with a 30MB size limit applying to uploads and downloads.

The real breakthrough lies in editing. Instead of opening a document or spreadsheet, users can simply type instructions such as replacing text, changing currencies, or updating job titles. Claude processes the prompt and makes all the changes in one pass, preserving the original formatting.

It positions the AI as more efficient than rivals, as Gemini can only export reports but not directly modify existing files.

The feature preview is available on web and desktop for subscribers on Max, Team, or Enterprise plans. Analysts suggest the update could reshape productivity tools, especially after reports that Microsoft has partnered with Anthropic to explore using Claude for Office 365 functions.

By removing repetitive tasks and making file handling conversational, Claude is pushing productivity software into a new phase of automation.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Small business revival could hinge on AI-driven tools

If AI is to matter in the economy, it must first matter to small businesses. Firms employ over 61 million people, nearly half the private workforce, yet most run on outdated technology. While smartphones update monthly, many small businesses still use systems built a decade ago.

Search fund entrepreneurs bridge this gap by upgrading established firms with modern tech. One deal turned a 50-person roadside assistance firm into Asurion, now a global tech-care provider. Others have scaled compliance firms into nationwide SaaS platforms.

Generative AI now accelerates these transformations, cutting work times by over 60% across supply chains, compliance, and document processing functions. Complex tasks can now be completed in hours, unlocking double-digit productivity gains and allowing small businesses to focus on growth.

Search funds are not the only path forward. AI consulting firms, tech studios, and AI-powered roll-up strategies bring enterprise-grade tools to family-run firms. For communities that have relied on traditional playbooks, decades of growth can be compressed into months.

The cost of AI has never been lower, and the opportunity is wide open. Once deployed at scale, AI could power a wave of productivity on Main Street, helping small businesses compete and strengthening the economy for half of their workforce.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

NotebookLM turns notes into flashcards podcasts and quizzes

Google’s learning-focused AI tool NotebookLM has gained a major update, making studying and teaching more interactive.

Instead of offering only static summaries, it now generates flashcards that condense key information into easy-to-remember notes, helping users recall knowledge more effectively.

Reports can also be transformed into quizzes with customisable topics and difficulty, which can then be shared with friends or colleagues through a simple link.

The update extends to audio learning, where NotebookLM’s podcast-style Audio Overviews are evolving with new formats. Instead of a single style, users can now create Brief, Debate, or Critique episodes, giving greater flexibility in how material is explained or discussed.

Google is also strengthening its teaching tools. A new Blog Post format offers contextual suggestions such as strategy papers or explainers, while the ability to create custom report formats allows users to design study resources tailored to their needs.

The most significant addition, however, is the Learning Guide. Acting like a personal tutor, it promotes deeper understanding by asking open-ended questions, breaking problems into smaller steps, and adapting explanations to suit each learner.

With these features, NotebookLM is moving closer to becoming a comprehensive learning assistant, offering a mix of interactive study aids and adaptable teaching methods that go beyond simple note-taking.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Photonic chips open the path to sustainable AI by training with light

A team of international researchers has shown how training neural networks directly with light on photonic chips could make AI faster and more sustainable.

A breakthrough study, published in Nature, involved collaboration between the Politecnico di Milano, EPFL Lausanne, Stanford University, the University of Cambridge, and the Max Planck Institute.

The research highlights how physical neural networks, which use analogue circuits that exploit the laws of physics, can process information in new ways.

Photonic chips developed at the Politecnico di Milano perform mathematical operations such as addition and multiplication through light interference on silicon microchips only a few millimetres in size.

By eliminating the need to digitise information, these chips dramatically cut both processing time and energy use. Researchers have also pioneered an ‘in-situ’ training technique that enables photonic neural networks to learn tasks entirely through light signals, instead of relying on digital models.

The result is a training process that is faster, more efficient and more robust.

Such advances could lead to more powerful AI models capable of running directly on devices instead of being dependent on energy-hungry data centres.

An approach that paves the way for technologies such as autonomous vehicles, portable intelligent sensors and real-time data processing systems that are both greener and quicker.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

New iPhone security ups pressure on spyware

Apple is rolling out Memory Integrity Enforcement on the iPhone 17 line and iPhone Air, an always-on set of protections aimed at blocking memory-safety exploits used by mercenary spyware.

MIE builds on ARM’s Enhanced Memory Tagging Extension in Apple’s A19 chips, alongside secure allocators and tag-confidentiality measures.

Older devices without the new tagging hardware also receive memory-safety upgrades. Apple says new Spectre V1 leak mitigations arrive with virtually no CPU penalty.

Comparable ideas exist elsewhere, such as Windows 11’s memory integrity (HVCI) and Android’s MTE support on Pixel 8, but Apple’s approach is enabled by default across key attack surfaces. Security reporters note the move significantly complicates spyware operations.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

AI chatbot mixes up Dutch political party policies

Dutch voters have been warned not to rely on AI chatbots for political advice after Google’s NotebookLM mixed up VVD and PVV policies.

When asked about Ukrainian refugees, the tool attributed a PVV proposal to send men back to Ukraine to the VVD programme. Similar confusions reportedly occurred when others used the system.

Google acknowledged the mistake and said it would investigate whether the error was a hallucination, the term for incorrect AI-generated output.

Experts caution that language models predict patterns rather than facts, making errors unavoidable. Voting guide StemWijzer stressed that reliable political advice requires up-to-date and verified information.

Professor Claes de Vreese said chatbots might be helpful to supplementary tools but should never replace reading actual party programmes. He also urged stricter regulation to avoid undue influence on election choices.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Canadian news publishers clash with OpenAI in landmark copyright case

OpenAI is set to argue in an Ontario court that a copyright lawsuit by Canadian news publishers should be heard in the United States. The case, the first of its kind in Canada, alleges that OpenAI scraped Canadian news content to train ChatGPT without permission or payment.

The coalition of publishers, including CBC/Radio-Canada, The Globe and Mail, and Postmedia, says the material was created and hosted in Ontario, making the province the proper venue. They warn that accepting OpenAI’s stance would undermine Canadian sovereignty in the digital economy.

OpenAI, however, says the training of its models and web crawling occurred outside Canada and that the Copyright Act cannot apply extraterritorially. It argues the publishers are politicising the case by framing it as a matter of sovereignty rather than jurisdiction.

The dispute reflects a broader global clash over how generative AI systems use copyrighted works. US courts are already handling several similar cases, though no clear precedent has been established on whether such use qualifies as fair use.

Publishers argue Canadian courts must decide the matter domestically, while OpenAI insists it belongs in US courts. The outcome could shape how copyright laws apply to AI training and digital content across borders.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Ransomware 3.0 raises alarm over AI-generated cyber threats

Researchers at NYU’s Tandon School of Engineering have demonstrated how large language models can be utilised to execute ransomware campaigns autonomously. Their prototype, dubbed Ransomware 3.0, simulated every stage of an attack, from intrusion to the generation of a ransom note.

The system briefly raised an alarm after cybersecurity firm ESET discovered its files on VirusTotal, mistakenly identifying them as live malware. The proof-of-concept was designed only for controlled laboratory use and posed no risk outside testing environments.

Instead of pre-written code, the prototype embedded text instructions that triggered AI models to generate tailored attack scripts. Each execution created unique code, evading traditional detection methods and running across Windows, Linux, and Raspberry Pi systems.

The researchers found that the system identified up to 96% of sensitive files and could generate personalised extortion notes, raising psychological pressure on victims. With costs as low as $0.70 per attack using commercial AI services, such methods could lower barriers for criminals.

The team stressed that the work was conducted ethically and aims to help defenders prepare countermeasures. They recommend monitoring file access patterns, limiting outbound AI connections, and developing defences against AI-generated attack behaviours.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Pressure mounts as Apple prepares AI search push with Google ties

Apple’s struggles in the AI race have been hard to miss. Its Apple Intelligence launch was disappointing, and its reliance on ChatGPT appeared to be a concession to rivals.

Bloomberg’s Mark Gurman now reports that Apple plans to introduce its AI-powered web search tool in spring 2026. The move would position it against OpenAI and Perplexity, while renewing pressure on Google.

The speculation comes after news that Google may integrate its Gemini AI into Apple devices. During an antitrust trial in April, Google CEO Sundar Pichai confirmed plans to roll out updates later this year.

According to Gurman, Apple and Google finalised an agreement for Apple to test a Google-developed AI model to boost its voice assistant. The partnership reflects Apple’s mixed strategy of dependence and rivalry with Google.

With a strong record for accurate Apple forecasts, Gurman suggests the company hopes the move will narrow its competitive gap. Whether it can outpace Google, especially given Pixel’s strong AI features, remains an open question.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!