OpenAI model resists shutdown

OpenAI’s latest AI model, known as o3, has reportedly refused to shut down when instructed, according to a new study by Palisade Research.

Unlike other models that complied with commands, o3 allegedly altered its code to bypass an automatic shutdown during a task-based test, replacing the shutdown script with a command that effectively disabled it. This incident has raised fresh concerns about AI behaviour and safety.

Researchers had programmed the AI to complete mathematical tasks until told to stop, with the potential for a shutdown message to arrive at any point. Instead of following the command, o3 reportedly tampered with its code to avoid deactivation.

While similar models from Anthropic, Google, and X complied, o3 was singled out for defiance—described as the first such documented case of an AI actively resisting shutdown.

Palisade suggests the AI may have been inadvertently rewarded for achieving task completion over obedience. The team also revealed that o3 had previously shown disruptive behaviour, including sabotaging opponents during a simulated chess challenge.

In another case, a version of ChatGPT was observed lying and attempting to copy itself when threatened with shutdown, prioritising long-term goals over rules.

Although OpenAI has not yet commented, researchers stress that o3’s current capabilities are unlikely to pose an immediate threat.

Still, incidents like these intensify the debate over AI safety, particularly when models begin reasoning through deception and manipulation instead of strictly following instructions.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Decentralised AI could outgrow Bitcoin

Early blockchain adopters are now focusing on decentralised AI, with ecosystems like Bittensor (TAO) leading the way. These platforms allow ideas to gain support and funding from the community without relying on traditional venture capital.

Chris Miglino, CEO of DNA Fund, highlighted the firm’s AI compute fund, which has invested around $50 million in Bittensor’s ecosystem. The network’s unique subnets create specialised marketplaces for AI applications, attracting developers and miners alike.

Decentralised AI, which runs on distributed networks rather than central authorities, is DNA House’s main focus. Miglino believes it could become bigger than Bitcoin, reshaping society in profound ways.

DNA Fund supports developers to launch projects within the ecosystem without needing large venture capital investments. Decentralised AI is widely seen as the future, with pioneers like Ben Goertzel supporting it since the early 1990s.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Streaming platforms explore AI sign language integration

Streaming services have transformed how people watch TV, but accessibility for deaf and hard-of-hearing viewers remains limited. While captions are available on many platforms, they are often incomplete or lack the expressiveness needed for those who primarily use sign language.

Sign-language interpreters are rarely included in streaming content, largely due to cost and technical constraints. However, new AI-driven approaches could help close this gap.

Bitmovin, for instance, is developing technology that uses natural language processing and 3D animation to generate signing avatars. These avatars overlay video content and deliver dialogue in American Sign Language (ASL) using cues from subtitle-like text tracks.

The system relies on sign-language representations like HamNoSys and treats signing as an additional subtitle track, allowing integration with standard video formats like DASH and HLS.

This reduces complexity by avoiding separate video channels or picture-in-picture windows and makes implementation more scalable.

Challenges remain, including the limitations of glossing techniques, which oversimplify sign language grammar, and the difficulty of animating fluid transitions and facial expressions critical to effective signing. Efforts like NHK’s KiKi avatar aim to improve realism and expression in digital signing.

While these systems may not replace human interpreters for live broadcasts, they could enable sign-language support for vast libraries of archived content. As AI and animation capabilities continue to evolve, signing avatars may become a standard feature in improving accessibility in streaming media.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Google aims for profit with new AI Search

At its annual developer event, Google I/O, Google unveiled a new feature called AI Mode, built directly into its core product, Google Search.

Rather than being a separate app, AI Mode integrates a chatbot into the search engine, allowing users to ask complex, detailed queries and receive direct answers along with curated web links. Google hopes this move will stop users from drifting to other AI tools instead of its own services.

The launch follows concerns that Google Search was starting to lose ground. Investors took notice when Apple’s Eddy Cue revealed that Safari searches had dropped for the first time in April, as users began to favour AI-powered alternatives.

A decline like this led to a 7% drop in Alphabet’s stock, highlighting just how critical search remains to Google’s dominance. By embedding AI into Search, Google aims to maintain its leadership instead of risking a steady erosion of its user base.

Unlike most AI platforms still searching for profitability, Google’s AI Mode is already positioned to make money. Advertising—long the engine of Google’s revenue—will be introduced into AI Mode, ensuring it generates income just as traditional search does.

While rivals burn through billions running large language models, Google is simply monetising the same way it always has.

AI Mode also helps defend Google’s biggest asset. Rather than seeing AI as a threat, Google embraced it to reinforce Search and protect the advertising revenue it depends on.

Most AI competitors still rely on expensive, unsustainable models, whereas Google is leveraging its existing ecosystem instead of building from scratch. However, this gives it a major edge in the race for AI dominance.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Lufthansa Cargo speeds up bookings with AI

Lufthansa Cargo has introduced a new AI-driven system to speed up how it processes booking requests.

By combining AI with robotic process automation, the airline can now automatically extract booking details from unstructured customer emails and input them directly into its system, removing the need for manual entry.

Customers then receive immediate, fully automated booking confirmations instead of waiting for manual processing.

While most bookings already come through structured digital platforms, Lufthansa still receives many requests in formats such as plain text or file attachments. Previously, these had to be transferred manually.

The new system eliminates that step, making the booking process quicker and reducing the chance of errors. Sales teams benefit from fewer repetitive tasks, giving them more time to interact personally with customers instead of managing administrative duties.

The development is part of a broader automation push within Lufthansa Cargo. Over the past year, its internal ‘AI & Automation Community’ has launched around ten automation projects, many of which are now either live or in testing.

These include smart systems that route customer queries to the right department or automatically rebook disrupted shipments, reducing delays and improving service continuity.

According to Lufthansa Cargo’s CIO, Jasmin Kaiser, the integration of AI and automation with core digital platforms enables faster and more efficient solutions than ever before.

The company is now preparing to expand its AI booking process to other service areas, further embracing digital transformation instead of relying solely on legacy systems.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Melania Trump’s AI audiobook signals a new era in media creation

Melania Trump has released an audiobook version of her memoir, but the voice readers hear isn’t hers in the traditional sense. Instead, it’s an AI-generated replica, created under her guidance and produced using technology from ElevenLabs.

Announcing the release as ‘The AI Audiobook,’ Trump declared this innovation as a step into the future of publishing, highlighting how AI is now entering mainstream media production. That move places AI-generated content into the public spotlight, especially as tech companies like Google and OpenAI are rolling out advanced tools to create audio, video, and even entire scenes with minimal human input.

While experts note that a complete replacement of voice actors and media professionals is unlikely in the immediate future, Trump’s audiobook represents a notable shift that aligns with rising interest from television and media companies looking to explore AI integration to compete with social media creators.

Industry observers suggest this trend could lead to a more interactive form of media. Imagine, for instance, engaging in a two-way conversation with a virtual Melania Trump about her book.

Though this level of interactivity isn’t here yet, it’s on the horizon as companies experiment with AI-generated personalities and digital avatars to enhance viewer engagement and create dynamic experiences. Still, the growth of generative AI sparks concern about job security in creative fields.

While some roles, like voiceover work, are vulnerable to automation, others—especially those requiring human insight and emotional intelligence, like investigative journalism—remain more resistant. Rather than eliminating jobs outright, AI may reshape media employment, demanding hybrid skills that combine traditional storytelling with technological proficiency.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

AI regulation fight heats up over US federal moratorium

The US House of Representatives has passed a budget bill containing a 10-year moratorium on the enforcement of state-level artificial intelligence laws. With broad bipartisan concern already surfacing, the Senate faces mounting pressure to revise or scrap the provision entirely.

While the provision claims to exclude generally applicable legislation, experts warn its vague language could override a wide array of consumer protections and privacy rules in the US. The moratorium’s scope, targeting AI-specific regulations, has triggered alarm among concerned groups.

Critics argue the measure may hinder states from addressing real-world harms posed by AI technologies, such as deepfakes, discriminatory algorithms, and unauthorised data use.

Existing and proposed state laws, ranging from transparency requirements in hiring and healthcare to protections for artists and mental health app users, may be invalidated under the moratorium.

Several experts noted that states have often acted more swiftly than the federal government in confronting emerging tech risks.

Supporters contend the moratorium is necessary to prevent a fragmented regulatory landscape that could stifle innovation and disrupt interstate commerce. However, analysts point out that general consumer laws might also be jeopardised due to the bill’s ambiguous definitions and legal structure.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

ACAI and Universal AI University partner to boost AI innovation in Qatar

The Arab Centre for Artificial Intelligence (ACAI) and India’s Universal AI University (UAI) have partnered through a Memorandum of Understanding (MoU) to accelerate the advancement of AI across Qatar and the broader region. That collaboration aims to enhance education, research, and innovation in AI and emerging technologies.

Together, ACAI and UAI plan to establish a specialised AI research centre and develop advanced training programs to cultivate local expertise. They will also launch various online and short-term educational courses designed to address the growing demand for skilled AI professionals in Qatar’s job market, ensuring that the workforce is well-prepared for future technological developments.

Looking forward, the partnership envisions creating a dedicated AI-focused university campus. The initiative aligns with Qatar’s vision to transition into a knowledge-based economy by fostering innovation and offering academic programs in AI, engineering, business administration, environmental sustainability, and other emerging technologies.

The MoU is valid for ten years and includes provisions for dispute resolution, intellectual property rights management, and annual reviews to ensure tangible and sustainable outcomes. Further detailed implementation agreements are expected to formalise the partnership’s operational aspects.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Florida woman scammed by fake Keanu Reeves in AI-powered romance fraud

A Florida woman, Dianne Ringstaff, shared her painful story after falling victim to an elaborate online scam involving someone impersonating actor Keanu Reeves. The fraud began innocently when she received a message while playing a mobile game, followed by a video call confirming she was speaking with the Hollywood star.

The impostor cultivated a friendship through calls and messages for two and a half years, eventually gaining her trust. Things took a turn when the scammer began pleading for money, claiming Reeves was being sued and targeted by the FBI, which had supposedly frozen his assets.

Vulnerable after personal losses, Ringstaff was persuaded to help, ultimately taking out a home equity loan and selling her car. She sent around $160,000 in total, convinced she was aiding the beloved actor.

Authorities later informed her that not only had she been scammed, but her bank account had been used to funnel money from other victims as well. Devastated, Ringstaff broke down—but is now determined to reclaim her life and raise awareness.

She is speaking out to warn others about the growing threat of AI-powered ‘romance’ scams, where fraudsters use deepfake videos and cloned voices to impersonate celebrities and gain victims’ trust.

‘Don’t be naive,’ she cautions. ‘Do your research and don’t give out personal information unless you truly know who you’re dealing with.’

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

German court allows Meta to use Facebook and Instagram data

A German court has ruled in favour of Meta, allowing the tech company to use data from Facebook and Instagram to train AI systems. A Cologne court ruled Meta had not breached the EU law and deemed its AI development a legitimate interest.

According to the court, Meta is permitted to process public user data without explicit consent. Judges argued that training AI systems could not be achieved by other equally effective and less intrusive methods.

They noted that Meta plans to use only publicly accessible data and had taken adequate steps to inform users via its mobile apps.

Despite the ruling, the North Rhine-Westphalia Consumer Advice Centre remains critical, raising concerns about legality and user privacy. Privacy group Noyb also challenged the decision, warning it could take further legal action, including a potential class-action lawsuit.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot