Broadcasters and advertisers seek clarity before the EU’s political advertising rules become fully applicable on 10 October. The European Commission has promised further guidance, but details on what qualifies as political advertising remain vague.
Meta and Google will block the EU’s political, election, and social issue ads when the rules take effect, citing operational challenges and legal uncertainty. The regulation, aimed at curbing disinformation and foreign interference, requires ads to display labels with sponsors, payments, and targeting.
Publishers fear they lack the technical means to comply or block non-compliant programmatic ads, risking legal exposure. They call for clear sponsor identification procedures, standardised declaration formats, and robust verification processes to ensure authenticity.
Advertisers warn that the rules’ broad definition of political actors may be hard to implement. At the same time, broadcasters fear issue-based campaigns – such as environmental awareness drives – could unintentionally fall under the scope of political advertising.
The Dutch parliamentary election on 29 October will be the first to take place under the fully applicable rules, making clarity from Brussels urgent for media and advertisers across the bloc.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Scientists from Australian universities and The George Institute for Global Health have developed an AI tool that analyses mammograms and a woman’s age to predict her risk of heart-related hospitalisation or death within 10 years.
Published in Heart on 17 September, the study highlights the lack of routine heart disease screening for women, despite cardiovascular conditions causing 35% of female deaths. The tool delivers a two-in-one health check by integrating heart risk prediction into breast cancer screening.
The model was trained on data from over 49,000 women and performs as accurately as traditional models that require blood pressure and cholesterol data. Researchers emphasise its low-resource nature, making it viable for broad deployment in rural or underserved areas.
Study co-author Dr Jennifer Barraclough said mobile mammography services could adopt the tool to deliver breast cancer and heart health screenings in one visit. Such integration could help overcome healthcare access barriers in remote regions.
Next, before a broader rollout, the researchers plan to validate the tool in more diverse populations and study practical challenges, such as technical requirements and regulatory approvals.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Meta has failed to overturn a jury verdict that found it illegally collected sensitive reproductive health data from users of the Flo period tracking app. US District Judge James Donato rejected Meta’s claim that the data was ‘secondhand’ and not protected under California’s wiretapping law.
The court found that Meta directly intercepted real-time communications between users and the app, such as when users indicated they wanted to track their menstrual cycle or pregnancy.
Judge Donato also dismissed Meta’s argument that Flo users had consented to the data sharing, calling the claim “rank speculation” unsupported by evidence.
The jury’s August verdict marked one of the first major legal decisions involving big tech’s handling of sensitive health information. Legal experts say it could open the door to more lawsuits and greater scrutiny of tech companies’ data practices. Meta has not responded to requests for comment.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Mexico’s competition watchdog has accused Amazon and Mercado Libre of erecting barriers that limit the mobility of sellers in the country’s e-commerce market. The two platforms reportedly account for 85% of the seller market.
The Federal Economic Competition Commission (COFECE) stated that the companies provide preferential treatment to sellers who utilise their logistics services and fail to disclose how featured offers are selected, thereby restricting fair competition.
Despite finding evidence of these practices, COFECE stopped short of imposing corrective measures, citing a lack of consensus among stakeholders. Amazon welcomed the decision, saying it demonstrates the competitiveness of the retail market in Mexico.
The watchdog aims to promote a more dynamic e-commerce sector, benefiting buyers and sellers. Its February report had recommended measures to improve transparency, separate loyalty programme services, and allow fairer access to third-party delivery options.
Trade associations praised COFECE for avoiding sanctions, warning that penalties could harm consumers and shield traditional retailers. Mercado Libre has not yet commented on the findings.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Scientists have unveiled an AI tool capable of predicting the risk of developing over 1,000 medical conditions. Published in Nature, the model can forecast certain cancers, heart attacks, and other diseases more than a decade in advance.
Developed by the German Cancer Research Centre (DKFZ), the European Molecular Biology Laboratory (EMBL), and the University of Copenhagen, the model utilises anonymised health data from the UK and Denmark. It tracks the order and timing of medical events to spot patterns that lead to serious illness.
Researchers said the tool is exceptionally accurate for diseases with consistent progression, including some cancers, diabetes, heart attacks, and septicaemia. Its predictions work like a weather forecast, indicating higher risk rather than certainty.
The model is less reliable for unpredictable conditions such as mental health disorders, infectious diseases, or pregnancy complications. It is more accurate for near-term forecasts than for those decades ahead.
Though not yet ready for clinical use, the system could help doctors identify high-risk patients earlier and enable more personalised, preventive healthcare strategies. Researchers say more work is needed to ensure the tool works for diverse populations.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Ireland has designated 15 authorities to monitor compliance with the EU’s AI Act, making it one of the first EU countries fully ready to enforce the new rules. The AI Act regulates AI systems according to their risk to society and began phasing in last year.
Governments had until 2 August to notify the European Commission of their appointed market surveillance authorities. In Ireland, these include the Central Bank, Coimisiún na Meán, the Data Protection Commission, the Competition and Consumer Protection Commission, and the Health and Safety Authority.
The country will also establish a National AI Office as the central coordinator for AI Act enforcement and liaise with EU institutions. A single point of contact must be designated where multiple authorities are involved to ensure clear communication.
Ireland joins Cyprus, Latvia, Lithuania, Luxembourg, Slovenia, and Spain as countries that have appointed their contact points. The Commission has not yet published the complete list of authorities notified by member states.
Former Italian Prime Minister Mario Draghi has called for a pause in the rollout of the AI Act, citing risks and a lack of technical standards. The Commission has launched a consultation as part of its digital simplification package, which will be implemented in December.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Hammersmith and Fulham Council has approved a £3m upgrade to its CCTV system to see facial recognition and AI integrated across the west London borough.
With over 2,000 cameras, the council intends to install live facial recognition technology at crime hotspots and link it with police databases for real-time identification.
Alongside the new cameras, 500 units will be equipped with AI tools to speed up video analysis, track vehicles, and provide retrospective searches. The plans also include the possible use of drones, pending approval from the Civil Aviation Authority.
Council leader Stephen Cowan said the technology will provide more substantial evidence in a criminal justice system he described as broken, arguing it will help secure convictions instead of leaving cases unresolved.
Civil liberties group Big Brother Watch condemned the project as mass surveillance without safeguards, warning of constant identity checks and retrospective monitoring of residents’ movements.
Some locals also voiced concern, saying the cameras address crime after it happens instead of preventing it. Others welcomed the move, believing it would deter offenders and reassure those who feel unsafe on the streets.
The Metropolitan Police currently operates one pilot site in Croydon, with findings expected later in the year, and the council says its rollout depends on continued police cooperation.
Would you like to learn more aboutAI, tech and digital diplomacy? If so, ask our Diplo chatbot!
The World Economic Forum (WEF) has published an article on using trade policy to build a fairer digital economy. Digital services now make up over half of global exports, with AI investment projected at $252 billion in 2024. Countries from Kenya to the UAE are positioning as digital hubs, but job quality still lags.
Millions of platform workers face volatile pay, lack of contracts, and no access to social protections. In Kenya alone, 1.9 million people rely on digital work yet face algorithm-driven pay systems and sudden account deactivations. India and the Philippines show similar patterns.
AI threatens to automate lower-skilled tasks such as data annotation and moderation, deepening insecurity in sectors where many developing countries have found a competitive edge. Ethical standards exist but have little impact without enforcement or supportive regulation.
Countries are experimenting with reforms: Singapore now mandates injury compensation and retirement savings for platform workers, while the Rider Law in Spain reclassifies food couriers as employees. Yet overly strict regulation risks eroding the flexibility that attracts youth and caregivers to gig work.
Trade agreements, such as the AfCFTA and the Kenya–EU pact, could embed labour protections in digital markets. Coordinated policies and tripartite dialogue are essential to ensure the digital economy delivers growth, fairness, and dignity for workers.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Jaguar Land Rover (JLR) has confirmed that its production halt will continue until at least Wednesday, 24 September, as it works to recover from a major cyberattack that disrupted its IT systems and paralysed production at the end of August.
JLR stated that the extension was necessary because forensic investigations were ongoing and the controlled restart of operations was taking longer than anticipated. The company stressed that it was prioritising a safe and stable restart and pledged to keep staff, suppliers, and partners regularly updated.
Reports suggest recovery could take weeks, impacting production and sales channels for an extended period. Approximately 33,000 employees remain at home as factory and sales processes are not fully operational, resulting in estimated losses of £1 billion in revenue and £70 million in profits.
The shutdown also poses risks to the wider UK economy, as JLR represents roughly four percent of British exports. The incident has renewed calls for the Cyber Security and Resilience Bill, which aims to strengthen defenses against digital threats to critical industries.
No official attribution has been made, but a group calling itself Scattered Lapsus$ Hunters has claimed responsibility. The group claims to have deployed ransomware and published screenshots of JLR’s internal SAP system, linking itself to extortion groups, including Scattered Spider, Lapsus$, and ShinyHunters.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!
Bracknell and Wokingham College has confirmed a cyberattack that compromised data collected for Disclosure and Barring Service (DBS) checks. The breach affects data used by Activate Learning and other institutions, including names, dates of birth, National Insurance numbers, and passport details.
Access Personal Checking Services (APCS) was alerted by supplier Intradev on August 17 that its systems had been accessed without authorisation. While payment card details and criminal conviction records were not compromised, data submitted between December 2024 and May 8, 2025, was copied.
APCS stated that its own networks and those of Activate Learning were not breached. The organisation is contacting only those data controllers where confirmed breaches have occurred and has advised that its services can continue to be used safely.
Activate Learning reported the incident to the Information Commissioner’s Office following a risk assessment. APCS is still investigating the full scope of the breach and has pledged to keep affected institutions and individuals informed as more information becomes available.
Individuals have been advised to closely monitor their financial statements, exercise caution when opening phishing emails, and regularly update security measures, including passwords and two-factor authentication. Activate Learning emphasised the importance of staying vigilant to minimise risks.
Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!