Netherlands and China in talks to resolve Nexperia dispute

The Dutch Economy Minister has spoken with his Chinese counterpart to ease tensions following the Netherlands’ recent seizure of Nexperia, a major Dutch semiconductor firm.

China, where most of Nexperia’s chips are produced and sold, reacted by blocking exports, creating concern among European carmakers reliant on its components.

Vincent Karremans said he had discussed ‘further steps towards reaching a solution’ with Chinese Minister of Commerce Wang Wentao.

Both sides emphasised the importance of finding an outcome that benefits Nexperia, as well as the Chinese and European economies.

Meanwhile, Nexperia’s China division has begun asserting its independence, telling employees they may reject ‘external instructions’.

The firm remains a subsidiary of Shanghai-listed Wingtech, which has faced growing scrutiny from European regulators over national security and strategic technology supply chains.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Meta strengthens protection for older adults against online scams

The US giant, Meta, has intensified its campaign against online scams targeting older adults, marking Cybersecurity Awareness Month with new safety tools and global partnerships.

Additionally, Meta said it had detected and disrupted nearly eight million fraudulent accounts on Facebook and Instagram since January, many linked to organised scam centres operating across Asia and the Middle East.

The social media giant is joining the National Elder Fraud Coordination Center in the US, alongside partners including Google, Microsoft and Walmart, to strengthen investigations into large-scale fraud operations.

It is also collaborating with law enforcement and research groups such as Graphika to identify scams involving fake customer service pages, fraudulent financial recovery services and deceptive home renovation schemes.

Meta continues to roll out product updates to improve online safety. WhatsApp now warns users when they share screens with unknown contacts, while Messenger is testing AI-powered scam detection that alerts users to suspicious messages.

Across Facebook, Instagram and WhatsApp, users can activate passkeys and complete a Security Checkup to reinforce account protection.

The company has also partnered with organisations worldwide to raise scam awareness among older adults, from digital literacy workshops in Bangkok to influencer-led safety campaigns across Europe and India.

These efforts form part of Meta’s ongoing drive to protect users through a mix of education, advanced technology and cross-industry cooperation.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Most EU workers now rely on digital tools and AI

A new EU study finds that 90% of workers rely on digital tools, while nearly a third use AI-powered chatbots in their daily work. The JRC and European Commission surveyed over 70,000 workers across all EU Member States between 2024 and 2025.

The findings show that AI is most commonly used for writing and translation tasks, followed by data processing and image generation. Adoption rates are particularly high in Northern and Central Europe, especially in office-based sectors.

Alongside this digital transformation, workplace monitoring is becoming increasingly widespread, with 37% of EU workers reporting that their working hours are tracked and 36% that their entry and exit times are monitored.

Algorithmic management, where digital systems allocate tasks or assess performance automatically, now affects about a quarter of EU workers. The study also identifies a growing ‘platformisation’ trend, categorising employees based on their exposure to digital monitoring and algorithmic control.

Workers facing full or physical platformisation often report higher stress levels and reduced autonomy, while informational platformisation appears to have milder effects, particularly for remote workers.

Researchers urge EU policymakers to curb digital oversight risks while promoting fair and responsible innovation. The findings support EU initiatives like the Quality Jobs Roadmap and efforts to regulate algorithmic management.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Dutch watchdog warns AI chatbots threaten election integrity

The Dutch data authority warns AI chatbots are biased and unreliable for voting advice ahead of national elections. An AP investigation found chatbots often steered users to the same two parties, ignoring their actual preferences.

In over half of the tests, the bots suggested either Geert Wilders’ far-right Freedom Party (PVV) or the leftwing GroenLinks-PvdA led by Frans Timmermans. Other parties, such as the centre-right CDA, were rarely mentioned even when users’ answers closely matched their platforms.

AP deputy head Monique Verdier said that voters were being steered towards parties that did not necessarily reflect their political views, warning that this undermines the integrity of free and fair elections.

The report comes ahead of the 29 October election, where the PVV currently leads the polls. However, the race remains tight, with GroenLinks-PvdA and CDA still in contention and many voters undecided.

Although the AP noted that the bias was not intentional, it attributed the problem to the way AI chatbots function, highlighting the risks of relying on opaque systems for democratic decisions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Maserati’s new campaign pits human emotion against machine precision

Luxury automobile manufacturer Maserati has launched a new campaign titled ‘Do AIs Dream of Driving?’, created in collaboration with creative agency DUDE and production studio Studio FM.

The 60-second film imagines an advanced AI named ‘Leonardo’ that knows every specification and blueprint of Maserati’s own Nettuno engine, yet, despite this knowledge, cannot experience the emotion, thrill and sensory feel of driving.

The film opens in a futuristic environment where Leonardo narrates the engine’s performance, its 630 hp output, its combustion architecture and Italian engineering heritage. Yet, alongside sleek imagery of Maseratis in motion, Leonardo glitches subtly, illustrating the distinction between technical perfection and human experience.

From a creative standpoint, the campaign intentionally integrates AI in the production process, not as a cost-cutting measure, but as a storytelling device. The protagonist was cast in the traditional human way; his digital twin was generated via AI.

The setting and styling drew from AI-assisted visual research but retained human oversight. Studio FM asserts all stages from casting to post-production adhered to an ethical protocol to protect rights and ensure transparency.

The campaign will run globally across Maserati’s social media channels (Instagram, Facebook, YouTube) and highlights a new phase in how automotive brands use generative AI, not just to showcase specs, but to evoke emotion and identity.

Creative Director Chiara Monticelli (DUDE Milan) noted that the project represented ‘an important step forward … where AI supports and enhances human creativity without replacing it.’

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Nexos.ai raises €30 m to ease enterprise AI adoption

The European startup Nexos.ai, headquartered in Vilnius, Lithuania, has closed a €30 million Series A funding round, co-led by Index Ventures and Evantic Capital, valuing the company at about €300 million (~US $350 million).

Founded by the duo behind cybersecurity unicorn Nord Security (Tomas Okmanas and Eimantas Sabaliauskas), Nexos.ai aims to solve what they describe as the ‘enterprise AI adoption crisis’. In their view, many organisations struggle with governance, cost control, fragmentation and security risks when using large-language models (LLMs).

Nexos.ai’s platform comprises two main components: an AI Workspace for employees and an AI Gateway for developers.

The Gateway offers orchestration across 200+ models, unified access, guardrails, cost monitoring and compliance oversight. The Workspace enables staff to work across formats, compare models and collaborate in a secure interface.

The company’s positioning as a neutral intermediary, likened to ‘Switzerland for LLMs’, underscores its mission to allow enterprises to gain productivity with AI without giving up data control or security.

The new funds will be used to extend support for private model deployment, expand into regulated sectors (finance, public institutions), grow across Europe and North America, and deepen product capabilities in routing, model fallback, and observability.

It’s an illustration of how investors are backing infrastructure plays in the enterprise-AI space: not just building new models, but creating the scaffolding for how organisations adopt, govern and deploy them safely.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Fenland business to close as AI reshapes media work

A Fenland videographer says the rise of AI has forced him to close his business. David Johnson, who runs DMJ-Imagery in Chatteris, will wind up operations in April after client demand collapsed.

He believes companies are turning to AI tools for projects once requiring human filmmakers and editors. Work such as promotional videos, adverts, and scripting has increasingly been replaced by automated content generation.

Johnson said his workload ‘plummeted’ over the past year despite surviving the pandemic. He described AI-made work as lacking ‘passion or emotion’, arguing that human creativity remains an essential component to storytelling.

Despite this, the UK government says AI has vast economic potential, industry groups urge fairer protections for creatives. They argue that existing copyright laws do not adequately safeguard work used to train AI models.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Meta’s ‘Vibes’ feed lets users scroll and remix entirely AI-generated videos

Meta Platforms has introduced Vibes, a new short-form video feed built entirely around AI-generated content, available within its Meta AI app and on the meta.ai website.

The feed allows users to browse videos generated by creators and communities, start videos from scratch via text prompts or upload visual elements, and remix existing videos by adding music or changing styles. Users can then publish these clips to the Vibes feed or cross-post to Instagram Stories, Facebook, and Reels.

Meta says the goal is to make the Meta AI app a hub for creative video generation: ‘You can bring your ideas to life … or remix a video from the feed to make it your own.’ While Meta noted the feature is launching as a preview, it also points to broader ambitions in generative video as part of its AI strategy.

However, media commentary is already acknowledging scepticism. Early feedback has labelled some of the feed’s output as ‘AI slop’, mass-produced synthetic videos that lack authentic human creativity, fueling questions about quality and user demand.

Meta’s timing comes amid heavy investment in its AI efforts and a drive to monetise generative video content and new creator tools. The company sees this as more than experiment, potentially a new vector for engagement and distribution inside its social ecosystem.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Is the world ready for AI to rule justice?

AI is creeping into almost every corner of our lives, and it seems the justice system’s turn has finally come. As technology reshapes the way we work, communicate, and make decisions, its potential to transform legal processes is becoming increasingly difficult to ignore. The justice system, however, is one of the most ethically sensitive and morally demanding fields in existence. 

For AI to play a meaningful role in it, it must go beyond algorithms and data. It needs to understand the principles of fairness, context, and morality that guide every legal judgement. And perhaps more challengingly, it must do so within a system that has long been deeply traditional and conservative, one that values precedent and human reasoning above all else. Jet, from courts to prosecutors to lawyers, AI promises speed, efficiency, and smarter decision-making, but can it ever truly replace the human touch? 

AI is reshaping the justice system with unprecedented efficiency, but true progress depends on whether humanity is ready to balance innovation with responsibility and ethical judgement.

AI in courts: Smarter administration, not robot judges… yet

Courts across the world are drowning in paperwork, delays, and endless procedural tasks, challenges that are well within AI’s capacity to solve efficiently. From classifying cases and managing documentation to identifying urgent filings and analysing precedents, AI systems are beginning to serve as silent assistants within courtrooms. 

The German judiciary, for example, has already shown what this looks like in practice. AI tools such as OLGA and Frauke have helped categorise thousands of cases, extract key facts, and even draft standardised judgments in air passenger rights claims, cutting processing times by more than half. For a system long burdened by backlogs, such efficiency is revolutionary.

Still, the conversation goes far beyond convenience. Justice is not a production line; it is built on fairness, empathy, and the capacity to interpret human intent. Even the most advanced algorithm cannot grasp the nuance of remorse, the context of equality, or the moral complexity behind each ruling. The question is whether societies are ready to trust machine intelligence to participate in moral reasoning.

The final, almost utopian scenario would be a world where AI itself serves as a judge who is unbiased, tireless, and immune to human error or emotion. Yet even as this vision fascinates technologists, legal experts across Europe, including the EU Commission and the OECD, stress that such a future must remain purely theoretical. Human judges, they argue, must always stay at the heart of justice- AI may assist in the process, but it must never be the one to decide it. The idea is not to replace judges but to help them navigate the overwhelming sea of information that modern justice generates.

Courts may soon become smarter, but true justice still depends on something no algorithm can replicate: the human conscience. 

AI is reshaping the justice system with unprecedented efficiency, but true progress depends on whether humanity is ready to balance innovation with responsibility and ethical judgement.

AI for prosecutors: Investigating with superhuman efficiency

Prosecutors today are also sifting through thousands of documents, recordings, and messages for every major case. AI can act as a powerful investigative partner, highlighting connections, spotting anomalies, and bringing clarity to complex cases that would take humans weeks to unravel. 

Especially in criminal law, cases can involve terabytes of documents, evidence that humans can hardly process within tight legal deadlines or between hearings, yet must be reviewed thoroughly. AI tools can sift through this massive data, flag inconsistencies, detect hidden links between suspects, and reveal patterns that might otherwise remain buried. Subtle details that might escape the human eye can be detected by AI, making it an invaluable ally in uncovering the full picture of a case. By handling these tasks at superhuman speed, AI could also help accelerate the notoriously slow pace of legal proceedings, giving prosecutors more time to focus on strategy and courtroom preparation. 

More advanced systems are already being tested in Europe and the US, capable of generating detailed case summaries and predicting which evidence is most likely to hold up in court. Some experimental tools can even evaluate witness credibility based on linguistic cues and inconsistencies in testimony. In this sense, AI becomes a strategic partner, guiding prosecutors toward stronger, more coherent arguments. 

AI is reshaping the justice system with unprecedented efficiency, but true progress depends on whether humanity is ready to balance innovation with responsibility and ethical judgement.

AI for lawyers: Turning routine into opportunity

The adoption of AI and its capabilities might reach their maximum when it comes to the work of lawyers, where transforming information into insight and strategy is at the core of the profession. AI can take over repetitive tasks: reviewing contracts, drafting documents, or scanning case files, freeing lawyers to focus on the work that AI cannot replace, such as strategic thinking, creative problem-solving, and providing personalised client support. 

AI can be incredibly useful for analysing publicly available cases, helping lawyers see how similar situations have been handled, identify potential legal opportunities, and craft stronger, more informed arguments. By recognising patterns across multiple cases, it can suggest creative questions for witnesses and suspects, highlight gaps in the evidence, and even propose potential defence strategies. 

AI also transforms client communication. Chatbots and virtual assistants can manage routine queries, schedule meetings, and provide concise updates, giving lawyers more time to understand clients’ needs and build stronger relationships. By handling the mundane, AI allows lawyers to spend their energy on reasoning, negotiation, and advocacy.

AI is reshaping the justice system with unprecedented efficiency, but true progress depends on whether humanity is ready to balance innovation with responsibility and ethical judgement.

Balancing promise with responsibility

AI is transforming the way courts, prosecutors, and lawyers operate, but its adoption is far from straightforward. While it can make work significantly easier, the technology also carries risks that legal professionals cannot ignore. Historical bias in data can shape AI outputs, potentially reinforcing unfair patterns if humans fail to oversee its use. Similarly, sensitive client information must be protected at all costs, making data privacy a non-negotiable responsibility. 

Training and education are therefore crucial. It is essential to understand not only what AI can do but also its limits- how to interpret suggestions, check for hidden biases, and decide when human judgement must prevail. Without this understanding, AI risks being a tool that misleads rather than empowers. 

The promise of AI lies in its ability to free humans from repetitive work, allowing professionals to focus on higher-value tasks. But its power is conditional: efficiency and insight mean little without the ethical compass of the human professionals guiding it.

Ultimately, the justice system is more than a process. It is about fairness, empathy, and moral reasoning. AI can assist, streamline, and illuminate, but the responsibility for decisions, for justice itself, remains squarely with humans. In the end, the true measure of AI’s success in law will be how it enhances human judgement, not how it replaces it.

So, is the world ready for AI to rule justice? The answer remains clear. While AI can transform how justice is delivered, the human mind, heart, and ethical responsibility must remain at the centre. AI may guide the way, but it cannot and should not hold the gavel.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Adobe unveils AI Foundry for enterprise model building

Adobe has launched a new enterprise service allowing firms to build custom AI models. The platform, called Adobe AI Foundry, lets companies train generative AI on their branding and intellectual property.

Based on Adobe’s Firefly models, the service can produce text, images, video, and 3D content. Pricing depends on usage, offering greater flexibility than Adobe’s traditional subscription model.

Adobe’s Firefly technology, first introduced in 2023, has already helped clients create over 25 billion assets. Foundry’s tailored models are expected to speed up campaign production while maintaining consistent brand identity across markets.

Hannah Elsakr, Adobe’s vice president for generative AI ventures, said the tools aim to enhance, not replace, human creativity. She emphasised that Adobe’s mission remains centred on supporting artists and marketers in telling powerful stories through technology.

The company believes its ethical approach to AI training and licensing could set a standard for enterprise-grade creative tools. Analysts say it also positions Adobe strongly against rivals offering generic AI solutions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot