EU calls on US tech firms to respect rules on handling staff data

Concerns over data protection have intensified as the European Commission calls on major technology companies to apply the EU standards when handling sensitive staff information linked to digital regulation.

Pressure follows requests from the US House Judiciary Committee seeking access to communications between US firms and the EU officials involved in enforcing laws such as the Digital Services Act and Digital Markets Act.

The EU officials emphasise that formal exchanges with companies take place through official channels, including documented correspondence, rather than informal messaging platforms. Internal communication practices may involve encrypted tools, reflecting growing concerns about data security and external scrutiny.

Debate surrounding the issue reflects wider tensions between the EU and the US over digital governance, privacy protections and regulatory authority. Questions over jurisdiction and access to sensitive communications are likely to remain central as transatlantic tech policy evolves.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!  

GDPR changes debated as EU seeks balance on data protection rules

Debate over potential updates to the GDPR is intensifying, as Marina Kaljurand advocates a focused ‘fitness check’ rather than sweeping legislative changes in an omnibus package.

Concerns raised in the European Parliament highlight risks associated with altering foundational elements of the regulation, particularly its definitions to personal data. Preserving these core principles is seen as essential to maintaining the integrity of the EU’s data protection framework.

Ongoing discussions reflect broader policy tensions within the EU, where efforts to reduce regulatory complexity must be balanced against the need to uphold strong privacy safeguards. Proposals for simplification are therefore facing scrutiny from lawmakers prioritising stability and legal clarity.

Future developments are likely to shape how the EU adapts its data protection rules to evolving digital markets, while ensuring that existing protections remain effective in a rapidly changing technological environment.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!  

EU moves to strengthen digital resilience with subsea cable funding

Efforts to improve the security of Europe’s digital infrastructure have advanced as the European Commission opens a €180 million funding call to support backup systems for subsea internet cables.

Investment by the EU will focus on developing alternative routes and redundancy mechanisms, ensuring continuity of connectivity in the event of disruptions affecting critical undersea networks that carry global data traffic.

Growing concerns around infrastructure vulnerability have increased attention on subsea cables, which play a central role in international communications. Strengthening resilience is therefore becoming a priority within broader European strategies on technological sovereignty and security.

Planned projects are expected to enhance reliability across the region, reducing risks associated with outages or potential external threats to essential telecommunications infrastructure.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!  

US senators question Meta facial recognition in smart glasses

Three Democratic senators have raised concerns about Meta’s reported exploration of facial recognition in its smart glasses, warning that it could normalise public surveillance. In a letter to CEO Mark Zuckerberg, Senators Edward Markey, Ron Wyden, and Jeff Merkley asked about consent, biometric data, and the risks of misuse.

The lawmakers said the proposed feature ‘risks normalising mass surveillance at a moment when the federal government is using similar tools to intimidate protesters and chill speech. Although facial recognition may offer real benefits for blind and visually impaired users, Meta’s history of failing to protect user privacy raises serious questions about its plan to deploy this technology in its smart glasses.’

‘Americans do not consent to biometric data collection simply by walking down a public street, entering a café, or standing in a crowd,’ the senators added. ‘Yet, the deployment of this technology would appear to do exactly that – subjecting countless individuals to covert identification without notice, without consent, and without any meaningful opportunity to opt out.’ They warned that such practices would erode longstanding expectations of privacy in public spaces, effectively eliminating public anonymity.’

Concerns grew after reports of US Border Patrol and ICE agents using Meta smart glasses. While there is no evidence of facial recognition use, senators argue that adding identification tools to eyewear could expand undetectable surveillance. The letter questions if Meta might link facial data with information from its platforms, enabling real-time identification tied to profiles. Lawmakers warn that this could increase the risks of harassment and targeting.

Meta had previously discontinued facial recognition on Facebook in 2021, citing societal concerns. The senators argue that reintroducing similar technology in wearable devices suggests a shift rather than a retreat. ‘Five years later, Meta appears less worried about those societal concerns and is reportedly planning to deploy facial recognition technology in one of the most dangerous possible settings,’ they wrote.

‘Moreover,’ they continued, ‘Meta is apparently aware of the risks with this technology,’ noting that ‘an internal memo recommended launching the product ‘during a dynamic political environment where many civil society groups that we would expect to attack us would have their resources focused on other concerns.’

‘In other words,’ the senators added, ‘Meta appears to recognise the serious privacy and civil liberties risks of facial recognition but thinks it can avoid attention by slipping the once-abandoned, ethically fraught product back onto the market while the world is distracted by the Trump administration’s daily chaos.’

The senators have asked Meta to clarify how it would obtain consent from both users and bystanders, how long it would retain biometric data, whether it would use it to train AI models, and whether it could share it with law enforcement, including the Department of Homeland Security. The company has been given until 6 April to respond.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

NSA warns of AI supply chain risks in new cybersecurity guidance

The National Security Agency has released new guidance on managing risks across the AI supply chain, highlighting growing cybersecurity concerns tied to AI and machine learning systems. The joint information sheet outlines how organisations can better assess vulnerabilities when deploying or sourcing AI technologies.

The document defines the AI and machine learning supply chain as a combination of key components, including training data, models, software, infrastructure, hardware, and third-party services. Each element can introduce risks affecting confidentiality, integrity, or availability, particularly as advanced tools such as large language models and AI agents become more widely adopted.

Security risks associated with data include bias, poisoning attacks, and exposure via techniques such as model inversion and data extraction. For models, the guidance warns of hidden backdoors, malware, evasion attacks, and model manipulation. Organisations are advised to use trusted sources, perform integrity checks, and maintain verified model registries to mitigate such threats.

The paper also highlights software and infrastructure vulnerabilities, noting that AI systems often rely on complex dependencies that expand the attack surface. Recommended measures include malware scanning, testing, patching, and maintaining software bills of materials. Additional risks arise from third-party services, which may introduce weaknesses through their own supply chains or shared environments.

To manage these risks, organisations are urged to improve visibility across their AI ecosystems, identify suppliers and subcontractors, and require documentation such as AI and software bills of materials. The guidance aligns with frameworks from the National Institute of Standards and Technology and MITRE, reinforcing the need for coordinated approaches to AI supply chain security.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Exchange Online outage affecting Outlook access resolved by Microsoft

Microsoft has addressed an Exchange Online outage that disrupted access to email and calendar services for users worldwide. The issue affected multiple connection methods, including Outlook on the web, Outlook desktop, and Exchange ActiveSync.

The company first acknowledged the problem early in the day, saying it was investigating reports of users being unable to access their mailboxes. According to a Microsoft 365 admin centre update, several Exchange Online connection protocols were impacted during the outage.

Although Microsoft later reported that telemetry indicated the issue was no longer occurring for most users, some customers continued to experience access problems. At one point, the Office.com portal also displayed an error message, preventing users from logging in.

Microsoft linked the disruption to an issue within its supporting network infrastructure, which affected how traffic was processed. Engineers implemented configuration changes to restore normal service and continue monitoring the platform to ensure stability.

In a later update, Microsoft confirmed that the Exchange Online outage had been mitigated and that services had been restored. The company said it is still investigating the root cause and will provide further details in a post-incident report, while a separate issue affecting Microsoft 365 Copilot web access remains under review.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Vera CPU by NVIDIA accelerates large-scale AI workloads

NVIDIA has unveiled the Vera CPU, designed specifically for agentic AI and reinforcement learning. It delivers 50% faster performance and double the energy efficiency, already adopted by Alibaba, Meta, ByteDance, Oracle Cloud, CoreWeave, and Lambda.

Vera features 88 Olympus cores, high-bandwidth memory, and advanced multithreading, supporting large-scale AI deployments. Liquid-cooled racks can host over 22,500 concurrent CPU environments, allowing enterprises and research labs to scale agentic AI efficiently.

The CPU integrates with NVIDIA GPUs via NVLink-C2C and includes ConnectX SuperNIC and BlueField-4 DPUs to enhance networking, storage, and security. Early users like Cursor and Redpanda report major gains in AI agent throughput and real-time data processing.

High performance, energy efficiency, and GPU integration make Vera a new standard for faster, scalable, and responsive AI systems. The platform supports coding assistants, reinforcement learning, and large-scale data, making it suitable for enterprise and scientific use.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Data centre security evolves with rise of robot dog patrols

Rising demand for AI and cloud computing is driving a surge in data centre construction, pushing operators to adopt new security solutions. Companies are increasingly deploying robotic dogs to patrol sites and monitor operations.

These four-legged machines can inspect equipment, detect anomalies and alert staff before issues escalate. Merry Frayne, senior director of product management at Boston Dynamics, noted a sharp increase in interest as investment in data infrastructure continues to grow.

Developed by firms such as Boston Dynamics and Ghost Robotics, the robots are designed to support rather than replace human guards. Their use can reduce costs by requiring fewer personnel while maintaining continuous monitoring.

The machines can travel long distances on a single charge and operate across both external and internal environments. Some facilities already use them on pre-programmed patrols to collect data and flag unusual activity.

At the same time, competition in robotics is intensifying globally, with companies exploring humanoid and AI-powered systems. Advances from firms like Nvidia and Tesla highlight how automation is expanding beyond security into broader industrial use.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

UK Government commits up to £2 billion to quantum technologies

The UK Government has announced up to £2 billion in funding for quantum technologies, including more than £1 billion over the next four years, confirmed by UKRI in December 2025, and a new procurement programme called ProQure designed to support the scaling of quantum computing across the UK. 

The announcement is being billed as the country’s ‘Quantum Leap’, positioning the UK as a first mover in quantum commercialisation.

The funding is distributed across several areas: over £500 million for quantum computing to help companies scale and develop applications in pharmaceuticals, financial services, and energy; £125 million for quantum networking; and £205 million for quantum sensing and navigation, with dedicated applications in medical diagnostics, greenhouse gas monitoring, and ultra-secure communications.

A further £13.8 million will be injected into the UK’s five National Quantum Research Hubs, with an additional £90 million for quantum infrastructure and £20 million for skills and commercialisation programmes. 

techUK welcomed the announcement, noting that the UK is already home to 11% of the world’s quantum startups and has attracted 12% of global quantum private equity investment.

The trade association highlighted the ProQure procurement programme as a step in the right direction, but cautioned that sustained, long-term private investment will be essential to support deep-tech companies through lengthy development cycles. 

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Advanced AI education unlocks powerful opportunities across Africa

Advanced AI education is expanding across Africa. Google DeepMind has launched new courses to support the next generation of technical learners and reduce the gap between AI talent and opportunities on the continent.

At the same time, the initiative is supported by targeted funding. Google.org is providing $4 million to train lecturers and develop educational toolkits, aiming to strengthen local capacity and scale AI education.

Moreover, the curriculum focuses on practical and technical skills. Learners gain hands-on experience with generative AI models and transformers, including building and fine-tuning language models, moving beyond basic AI literacy.

In addition, the programme is adapted to African contexts. Developed with input from local experts and institutions, such as the African Institute for Mathematical Sciences, the courses include real-world use cases relevant to the continent.

Furthermore, the initiative aims to address Africa’s underrepresentation in AI research. By expanding access to advanced training, it seeks to increase participation and ensure more inclusive global AI development.

Finally, the programme is designed to scale through educators and institutions. Universities and NGOs can integrate the curriculum, supported by training programmes that equip educators to deliver AI courses effectively.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!