Some iPhones lose WhatsApp support

WhatsApp has officially ceased support for older iPhone models starting 5 May 2025. Devices running iOS versions earlier than 15.1, including the iPhone 5s, iPhone 6, and iPhone 6 Plus, will no longer be able to use the messaging platform. These models are limited to iOS 12.5.7 and cannot receive further updates.

The move comes as part of WhatsApp’s ongoing efforts to bolster user security and keep pace with advancing technology.

By discontinuing support for devices that no longer receive security patches from Apple, the company aims to reduce vulnerabilities and provide a safer user experience.

While newer models like the iPhone 8 and iPhone X remain compatible for now, users are advised that these devices may also lose support soon as Apple has stopped issuing software updates for them. Apple has already classified the iPhone 5s, iPhone 6, and iPhone 6 Plus as obsolete.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

DOGE sees a 41% increase in whale transactions

Dogecoin (DOGE) has seen a 41% increase in whale transactions over the past 24 hours. Large holders have moved over 60.9 billion DOGE, according to IntoTheBlock data. It marks a seven-day high, indicating heightened activity among major investors.

Despite the surge in transactions, DOGE’s price remains under pressure, trading at approximately $0.174, reflecting a 1.24% decline. The increase in whale activity suggests strategic positioning by large holders, possibly in anticipation of future market movements.

Analysts are monitoring the situation closely. They note that while whale accumulation often precedes price rallies, the current market outlook remains cautious. The decline in trading volume by over 24% to $826.67 million further suggests a wait-and-see approach among investors.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

US lawmakers push for app store age checks

A new bill introduced by US lawmakers could force app stores like Apple’s App Store and Google Play to verify the age of all users, in a move aimed at increasing online safety for minors.

Known as the App Store Accountability Act, the legislation would require age categorisation and parental consent before minors can download apps or make in-app purchases. If passed, the law would apply to platforms with at least five million users and would come into effect one year after approval.

The bill proposes dividing users into age brackets — from ‘young child’ to ‘adult’ — and holding app stores accountable for enforcing access restrictions.

Lawmakers behind the bill, Republican Senator Mike Lee and Representative John James, argue that Big Tech companies must take responsibility for limiting children’s exposure to harmful content. They believe app stores are the right gatekeepers for verifying age and protecting minors online.

Privacy advocates and tech companies have voiced concern about the bill’s implications. Legal experts warn that verifying users’ ages may require sensitive personal data, such as ID documents or facial recognition scans, raising the risk of data misuse.

Apple said such verification would apply to all users, not just children, and criticised the idea as counterproductive to privacy.

The proposal has widened a rift between app store operators and social media platforms. While Meta, X, and Snap back centralised age checks at the app store level, Apple and Google accuse them of shifting the burden of responsibility.

Both tech giants emphasise the importance of shared responsibility and continue to engage with lawmakers on crafting practical and privacy-conscious solutions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

US antitrust trial sees Google defend Chrome and data control

Google has warned that proposed remedies in the ongoing US antitrust case, including a possible sell-off of Chrome, could expose users to data breaches and national security threats. Arguing that Google’s infrastructure is key to protecting Chrome against rising cyberattacks.

Google cited past breaches to emphasise the risks of moving such tools to buyers lacking similar security standards. The Justice Department, however, maintains that breaking up Google’s dominance would encourage fairer competition.

Proposals include banning exclusive deals, sharing user data to support rivals, and enabling Apple or others to shift default search settings. An economic expert testified these remedies could reduce Google’s market share from 88% to 51%, though full impact would take years to materialise.

Judge Amit Mehta raised concerns that dismantling Google’s monopoly might simply replace it with another, such as Microsoft. Google CEO Sundar Pichai is set to testify next, as the case continues through 9 May in the US.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Meta introduces face recognition to help UAE users recover hacked accounts

Meta is introducing facial recognition tools to help UAE users recover hacked accounts on Facebook and Instagram and stop scams that misuse public figures’ images. The technology compares suspicious ads to verified profile photos and removes them automatically if a match is found.

Well-known individuals in the region are automatically enrolled in the programme but can opt out if they choose. A new video selfie feature has also been rolled out to help users regain access to compromised accounts.

This allows identity verification through a short video matched with existing profile photos, offering a faster and more secure alternative to document-based checks.

Meta confirmed that all facial data used for verification is encrypted, deleted immediately after use, and never repurposed.

The company says this is part of a broader effort to fight impersonation scams and protect both public figures and regular users, not just in the UAE but elsewhere too.

Meta’s regional director highlighted the emotional and financial harm such scams can cause, reinforcing the need for proactive defences.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

UK refuses to include Online Safety Act in US trade talks

The UK government has ruled out watering down the Online Safety Act as part of any trade negotiations with the US, despite pressure from American tech giants.

Speaking to MPs on the Science, Innovation and Technology Committee, Baroness Jones of Whitchurch, the parliamentary under-secretary for online safety, stated unequivocally that the legislation was ‘not up for negotiation’.

‘There have been clear instructions from the Prime Minister,’ she said. ‘The Online Safety Act is not part of the trade deal discussions. It’s a piece of legislation — it can’t just be negotiated away.’

Reports had suggested that President Donald Trump’s administration might seek to make loosening the UK’s online safety rules a condition of a post-Brexit trade agreement, following lobbying from large US-based technology firms.

However, Baroness Jones said the legislation was well into its implementation phase and that ministers were ‘happy to reassure everybody’ that the government is sticking to it.

The Online Safety Act will require tech platforms that host user-generated content, such as social media firms, to take active steps to protect users — especially children — from harmful and illegal content.

Non-compliant companies may face fines of up to £18 million or 10% of global turnover, whichever is greater. In extreme cases, platforms could be blocked from operating in the UK.

Mark Bunting, a representative of Ofcom, which is overseeing enforcement of the new rules, said the regulator would have taken action had the legislation been in force during last summer’s riots in Southport, which were exacerbated by online misinformation.

His comments contrasted with tech firms including Meta, TikTok and X, which claimed in earlier hearings that little would have changed under the new rules.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

OpenAI’s CEO Altman confirms rollback of GPT-4o after criticism

OpenAI has reversed a recent update to its GPT-4o model after users complained it had become overly flattering and blindly agreeable. The behaviour, widely mocked online, saw ChatGPT praising dangerous or clearly misguided user ideas, leading to concerns over the model’s reliability and integrity.

The change had been part of a broader attempt to make GPT-4o’s default personality feel more ‘intuitive and effective’. However, OpenAI admitted the update relied too heavily on short-term user feedback and failed to consider how interactions evolve over time.

In a blog post published Tuesday, OpenAI said the model began producing responses that were ‘overly supportive but disingenuous’. The company acknowledged that sycophantic interactions could feel ‘uncomfortable, unsettling, and cause distress’.

Following CEO Sam Altman’s weekend announcement of an impending rollback, OpenAI confirmed that the previous, more balanced version of GPT-4o had been reinstated.

It also outlined steps to avoid similar problems in future, including refining model training, revising system prompts, and expanding safety guardrails to improve honesty and transparency.

Further changes in development include real-time feedback mechanisms and allowing users to choose between multiple ChatGPT personalities. OpenAI says it aims to incorporate more diverse cultural perspectives and give users greater control over the assistant’s behaviour.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

EU criticised for secretive security AI plans

A new report by Statewatch has revealed that the European Union is quietly laying the groundwork for the widespread use of experimental AI technologies in policing, border control, and criminal justice.

The report warns that these developments pose serious threats to transparency, accountability, and fundamental rights.

Despite the adoption of the EU AI Act in 2024, broad exemptions allow law enforcement and migration agencies to bypass safeguards, including a full exemption for certain high-risk systems until 2031.

Institutions like Europol and eu-LISA are involved in building technical infrastructure for security-focused AI, often without public knowledge or oversight.

The study also highlights how secretive working groups, such as the European Clearing Board, have influenced legislation to favour police interests.

Critics argue that these moves risk entrenching discrimination and reducing democratic control, especially at a time of rising authoritarian influence within EU institutions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Singapore Airlines upgrades customer support with AI technology

Singapore Airlines has partnered with OpenAI to enhance its customer support services. The airline’s upgraded virtual assistant will now offer more personalised support to customers and assist staff by automating routine processes and improving decision-making for complex tasks.

The partnership comes alongside Singapore Airlines’ ongoing work with Salesforce to strengthen its customer case management system using AI tech. New solutions will be developed at Salesforce’s AI research hub in Singapore, advancing customer service capabilities and operational efficiency.

These moves reflect a broader industry trend, with airlines like Delta and Air India also investing heavily in AI-driven tools for travel assistance and operational support. The Airline emphasised that AI integration will help it meet regulatory demands, enhance workforce management and elevate customer experience.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

UK government urged to outlaw apps creating deepfake abuse images

The Children’s Commissioner has urged the UK Government to ban AI apps that create sexually explicit images through “nudification” technology. AI tools capable of manipulating real photos to make people appear naked are being used to target children.

Concerns in the UK are growing as these apps are now widely accessible online, often through social media and search platforms. In a newly published report, Dame Rachel warned that children, particularly girls, are altering their online behaviour out of fear of becoming victims of such technologies.

She stressed that while AI holds great potential, it also poses serious risks to children’s safety. The report also recommends stronger legal duties for AI developers and improved systems to remove explicit deepfake content from the internet.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!