Meta faces backlash over Llama 4 release

Over the weekend, Meta unveiled two new Llama 4 models—Scout, a smaller version, and Maverick, a mid-sized variant it claims outperforms OpenAI’s GPT-4o and Google’s Gemini 2.0 Flash across multiple benchmarks.

Maverick quickly climbed to second place on LMArena, an AI benchmarking platform where human evaluators compare and vote on model outputs. Meta proudly pointed to Maverick’s ELO score of 1417, placing it just beneath Gemini 2.5 Pro, instead of trailing behind the usual leaders.

However, AI researchers noticed a critical detail buried in Meta’s documentation: the version of Maverick that ranked so highly wasn’t the one released to the public. Instead of using the standard model, Meta had submitted an ‘experimental’ version specifically optimised for conversations.

LMArena later criticised this move, saying Meta failed to clearly indicate the model was customised, prompting the platform to update its policies to ensure future evaluations remain fair and reproducible.

Meta’s spokesperson acknowledged the use of experimental variants, insisting the company frequently tests different configurations.

While this wasn’t a violation of LMArena’s existing rules, the episode raised concerns about the credibility of benchmark rankings when companies submit fine-tuned models instead of the ones accessible to the wider community.

Independent AI researcher Simon Willison expressed frustration, saying the impressive ranking lost all meaning once it became clear the public couldn’t even use the same version.

The controversy unfolded against a backdrop of mounting competition in open-weight AI, with Meta under pressure following high-profile releases like China’s DeepSeek model.

Instead of offering a smooth rollout, Meta released Llama 4 on a Saturday—an unusual move—which CEO Mark Zuckerberg explained simply as ‘that’s when it was ready.’ But for many in the AI space, the launch has only deepened confusion around what these models can genuinely deliver.

For more information on these topics, visit diplomacy.edu.

Southampton Airport launches AI assistant to support passengers

Southampton Airport has launched an advanced AI-powered digital assistant to enhance passenger experience and accessibility throughout its terminal. The technology, developed in collaboration with Hello Lamp Post, offers real-time flight updates, personalised navigation assistance, and tailored support, especially for those requiring special assistance.

Following a successful trial at Glasgow Airport with Connected Places Catapult, the AI platform demonstrated a 50% reduction in customer service queries and supported over 12,000 additional passengers annually. Passenger satisfaction during the pilot reached 86%, prompting Southampton to expand the tool for all travellers. The assistant is accessible via QR codes placed throughout the terminal, effectively acting as a virtual concierge.

The initiative forms part of the airport’s broader commitment to inclusive and efficient travel. Southampton Airport recently received the Civil Aviation Authority’s top ‘Very Good’ rating for accessibility. Airport Managing Director Gavin Williams praised the new tool’s ability to enhance customer journeys, while Hello Lamp Post’s CEO, Tiernan Mines, highlighted the value in easing pressure on staff by handling routine queries.

For more information on these topics, visit diplomacy.edu.

New Jersey criminalises AI-generated nude deepfakes of minors

New Jersey has become the first US state to criminalise the creation and sharing of AI-generated nude images of minors, following a high-profile campaign led by 14-year-old Francesca Mani. The US legislation, signed into law on 2 April by Governor Phil Murphy, allows victims to sue perpetrators for up to $1,000 per image and includes criminal penalties of up to five years in prison and fines of up to $30,000.

Mani launched her campaign after discovering that boys at her school had used an AI “nudify” website to target her and other girls. Refusing to accept the school’s minimal disciplinary response, she called for lawmakers to take decisive action against such deepfake abuses. Her efforts gained national attention, including a feature on 60 Minutes, and helped drive the new legal protections.

The law defines deepfakes as media that convincingly depicts someone doing something they never actually did. It also prohibits the use of such technology for election interference or defamation. Although the law’s focus is on malicious misuse, questions remain about whether exemptions will be made for legitimate uses in film, tech, or education sectors.

For more information on these topics, visit diplomacy.edu.

AI tool boosts accuracy of cancer treatment predictions

A Slovenian-US biotech company, Genialis, is harnessing AI to revolutionise cancer treatment by tackling a major obstacle: the lack of reliable biomarkers to predict how patients will respond to therapy. Using an AI-driven model developed from over a million global samples, the company aims to personalise treatment with far greater accuracy.

Founded nine years ago as a spin-off from the University of Ljubljana, Genialis is now headquartered in Boston but maintains strong ties to Slovenia, employing 22 local experts. Initially focused on tools for biologists, the firm shifted towards personalised medicine six years ago, now offering diagnostic insights that predict whether a patient is likely to respond to a specific cancer drug or treatment.

Genialis’ proprietary “Supermodel” analyses RNA data from a diverse range of patients using machine learning, boosting the likelihood of treatment success from 20–30% to as high as 65% when paired with their biomarkers. While the software is already used in research settings, the ultimate goal is to integrate it into routine clinical care. Despite the promise, challenges remain, including securing quality data and investment. Co-founders Rafael Rosengarten and Miha Štajdohar remain optimistic, believing AI-powered precision medicine is the future of effective cancer therapy.

For more information on these topics, visit diplomacy.edu.

Trump administration pushes for pro-AI shift in US federal agencies

The White House announced on Monday a shift in how US federal agencies will approach AI, prioritising innovation over the stricter regulatory framework previously established under President Biden. 

A new memorandum from the Office of Management and Budget instructs agencies to appoint chief AI officers and craft policies to expand the use of AI technologies across government operations.

This pivot includes repealing two Biden-era directives emphasising transparency and safeguards against AI misuse. 

The earlier rules required federal agencies to implement protective measures for civil rights and limit unchecked acquisition of AI tools. 

These protections have now been replaced with a call for a more ‘forward-leaning and pro-innovation’ stance, removing what the current administration views as excessive bureaucratic constraints.

Federal agencies are now expected to develop AI strategies within six months. These plans must identify barriers to responsible AI implementation and improve how the technology is used enterprise-wide. 

The administration also encouraged the development of specific policies for generative AI, emphasising maximising the use of American-made solutions and enhancing interoperability between systems.

The policy change is part of President Trump’s broader rollback of previous AI governance, including his earlier revocation of a 2023 executive order signed by Biden that required developers to disclose sensitive training data. 

The new framework aims to streamline AI procurement processes and eliminate what the administration labels unnecessary reporting burdens while still maintaining basic privacy protections.

Federal agencies have already begun integrating AI into their operations. The Federal Aviation Administration, for example, has applied machine learning to analyse safety reports and identify emerging aviation risks. 

Under the new guidelines, such initiatives are expected to accelerate, signalling a broader federal embrace of AI across sectors.

For more information on these topics, visit diplomacy.edu.

Osney Capital invests in the UK’s cybersecurity innovation

Osney Capital has launched the UK’s first specialist cybersecurity seed fund, focused on investing in promising cybersecurity startups at the Pre-Seed and Seed stages.

The fund, which raised more than its initial £50 million target, will write cheques between £250k and £2.5 million and has the capacity for follow-on investments in Series A rounds.

Led by Adam Cragg, Josh Walter, and Paul Wilkes, the Osney Capital team brings decades of experience in cybersecurity and early-stage investing. Instead of relying on generalist investors, the fund will offer tailored support to early-stage companies, addressing the unique challenges in the cybersecurity sector.

The UK cybersecurity industry has grown to £13.2 billion in 2025, driven by complex cyber threats, regulatory pressures, and the rapid adoption of AI. The fund aims to capitalise on this growth, tapping into the strong talent pipeline boosted by UK universities and specialised cybersecurity programs.

Supported by cornerstone investments from the British Business Bank and accredited by the UK’s National Security Strategic Investment Fund, Osney Capital’s mission is to back the next generation of cybersecurity founders and help them scale globally competitive businesses.

For more information on these topics, visit diplomacy.edu.

Thailand strengthens cybersecurity with Google Cloud

Thailand’s National Cyber Security Agency (NCSA) has joined forces with Google Cloud to strengthen the country’s cyber resilience, using AI-based tools and shared threat intelligence instead of relying solely on traditional defences.

The collaboration aims to better protect public agencies and citizens against increasingly sophisticated cyber threats.

A key part of the initiative involves deploying Google Cloud Cybershield for centralised monitoring of security events across government bodies. Instead of having fragmented monitoring systems, this unified approach will help streamline incident detection and response.

The partnership also brings advanced training for cybersecurity personnel in the public sector, alongside regular threat intelligence sharing.

Google Cloud Web Risk will be integrated into government operations to automatically block websites hosting malware and phishing content, instead of relying on manual checks.

Google further noted the impact of its anti-scam technology in Google Play Protect, which has prevented over 6.6 million high-risk app installation attempts in Thailand since its 2024 launch—enhancing mobile safety for millions of users.

For more information on these topics, visit diplomacy.edu.

Earthquake disrupts Myanmar internet and surveillance ops

A powerful earthquake in Myanmar has significantly disrupted the junta’s cyber operations, severely damaging the National Cyber Security Center and a major e-Government data hub in Naypyitaw.

The Ministry of Transport and Communications confirmed that repairs are underway, though many internet services in quake-affected areas remain unstable.

According to the opposition National Unity Government (NUG), the damage could weaken the military regime’s ability to monitor and suppress dissent online.

The National Cyber Security Center, which played a central role in tracking online activity and orchestrating arrests via surveillance tools, was among the worst-hit sites. Thousands of communication stations were also damaged, with less than half restored so far.

The quake has also taken down several junta ministry websites and left staff in Naypyitaw facing harsh living conditions. With Myanmar recording the highest number of internet shutdowns globally in 2024, activists fear the regime will attempt to regain control, but the full extent of the digital disruption may be greater than reported.

For more information on these topics, visit diplomacy.edu.

GPT-4.5 outperforms humans in updated Turing Test

Two leading AI systems, OpenAI’s GPT-4.5 and Meta’s Llama-3.1, have passed a key milestone by outperforming humans in a modern version of the Turing Test.

The experiment, conducted by researchers at the University of California San Diego, found that GPT-4.5 was mistaken for a human 73% of the time, surpassing the human identification rate. Meta’s Llama-3.1 followed closely, with a 56% success rate.

The study used a three-party test where participants held simultaneous five-minute conversations with both a human and an AI, and then tried to determine which was which.

These trials were conducted across two independent groups: university undergraduates and prolific online workers. The results provide the first substantial evidence that AI can convincingly mimic human responses in spontaneous conversations.

Earlier language models such as ELIZA and GPT-4o were correctly identified as non-human in over 75% of cases.

The success of newer models in passing this benchmark points to how rapidly conversational AI is evolving, raising fresh questions about the ethical and societal implications of indistinguishable AI interactions.

For more information on these topics, visit diplomacy.edu.

Track missing bags with Air India and Apple AirTags

Air India has launched a new partnership with Apple to enhance baggage tracking by integrating Apple’s AirTag technology into its existing systems.

The collaboration connects AirTag’s location-sharing capabilities with Air India’s mobile app and website, allowing passengers to track misplaced luggage more efficiently.

Travellers using Apple devices can generate a location link for their AirTag through the Find My app and submit it to Air India along with a lost baggage report.

Once shared, airline staff can view an interactive map of the item’s location, updated in real time, until the luggage is recovered or the tracking expires after seven days.

This move makes Air India the first Indian airline to adopt Apple’s ‘Share Item Location’ feature, joining a global group that includes British Airways, Lufthansa, and Virgin Atlantic.

The new system is compatible with iPhones, iPads, and Macs running the latest software versions, offering Indian users a smarter way to stay connected with their belongings throughout their journey.

For more information on these topics, visit diplomacy.edu.