OpenAI unveils new gpt-oss-safeguard models for adaptive content safety

Yesterday, OpenAI launched gpt-oss-safeguard, a pair of open-weight reasoning models designed to classify content according to developer-specified safety policies.

Available in 120b and 20b sizes, these models allow developers to apply and revise policies during inference instead of relying on pre-trained classifiers.

They produce explanations of their reasoning, making policy enforcement transparent and adaptable. The models are downloadable under an Apache 2.0 licence, encouraging experimentation and modification.

The system excels in situations where potential risks evolve quickly, data is limited, or nuanced judgements are required.

Unlike traditional classifiers that infer policies from pre-labelled data, gpt-oss-safeguard interprets developer-provided policies directly, enabling more precise and flexible moderation.

The models have been tested internally and externally, showing competitive performance against OpenAI’s own Safety Reasoner and prior reasoning models. They can also support non-safety tasks, such as custom content labelling, depending on the developer’s goals.

OpenAI developed these models alongside ROOST and other partners, building a community to improve open safety tools collaboratively.

While gpt-oss-safeguard is computationally intensive and may not always surpass classifiers trained on extensive datasets, it offers a dynamic approach to content moderation and risk assessment.

Developers can integrate the models into their systems to classify messages, reviews, or chat content with transparent reasoning instead of static rule sets.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Automakers and freight partners join NVIDIA and Uber to accelerate level 4 deployments

NVIDIA and Uber partner on level 4-ready fleets using the DRIVE AGX Hyperion 10, aiming to scale a unified human-and-robot driver network from 2027. A joint AI data factory on NVIDIA Cosmos will curate training data, aiming to reach 100,000 vehicles over time.

DRIVE AGX Hyperion 10 is a reference compute and sensor stack for level 4 readiness across cars, vans, and trucks. Automakers can pair validated hardware with compatible autonomy software to speed safer, scalable, AI-defined mobility. Passenger and freight services gain faster paths from prototype to fleet.

Stellantis, Lucid, and Mercedes-Benz are preparing passenger platforms on Hyperion 10. Aurora, Volvo Autonomous Solutions, and Waabi are extending level 4 capability to long-haul trucking. Avride, May Mobility, Momenta, Nuro, Pony.ai, Wayve, and WeRide continue to build on NVIDIA DRIVE.

The production platform pairs dual DRIVE AGX Thor on Blackwell with DriveOS and a qualified multimodal sensor suite. Cameras, radar, lidar, and ultrasonics deliver 360-degree coverage. Modular design plus PCIe, Ethernet, confidential computing, and liquid cooling support upgrades and uptime.

NVIDIA is also launching Halos, a cloud-to-vehicle AI safety and certification system with an ANSI-accredited inspection lab and certification program. A multimodal AV dataset and reasoning VLA models aim to improve urban driving, testing, and validation for deployments.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Microsoft restores Azure services after global outage

The US tech giant, Microsoft, has resolved a global outage affecting its Azure cloud services, which disrupted access to Office 365, Minecraft, and numerous other websites.

The company attributed the incident to a configuration change that triggered DNS issues, impacting businesses and consumers worldwide.

An outage that affected high-profile services, including Heathrow Airport, NatWest, Starbucks, and New Zealand’s police and parliament websites.

Microsoft restored access after several hours, but the event highlighted the fragility of the internet due to the concentration of cloud services among a few major providers.

Experts noted that reliance on platforms such as Azure, Amazon Web Services, and Google Cloud creates systemic risks. Even minor configuration errors can ripple across thousands of interconnected systems, affecting payment processing, government operations, and online services.

Despite the disruption, Microsoft’s swift fix mitigated long-term impact. The company reiterated the importance of robust infrastructure and contingency planning as the global economy increasingly depends on cloud computing.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Global alliance strengthens response to crypto crime

Global experts are stepping up efforts to combat the misuse of cryptocurrencies as criminal networks become increasingly sophisticated.

The 9th Global Conference on Criminal Finances and Cryptoassets was held in Vienna and co-organised by Europol, the UNODC and the Basel Institute on Governance. The event brought together over 250 participants and 1,000 online attendees to discuss how to strengthen the global response.

Delegates emphasised the need for unified standards, stronger cooperation and greater investment in training to tackle the evolving threats posed by crypto-enabled crime.

Speakers warned that blockchain misuse has expanded beyond scams to include terrorism financing, sanctions evasion and organised money laundering. Europol’s Burkhard Mühl said tackling these complex crimes needs greater innovation and collaboration.

Advanced tracing tools and successful cross-border operations demonstrate progress, yet significant legislative and capacity gaps remain.

Participants urged harmonised standards and quicker information sharing between financial institutions and virtual asset providers. The Wolfsberg Group noted that private sector collaboration is as vital as public partnerships in disrupting illicit crypto activity.

Building capacity through hands-on training and peer learning was also identified as a priority. According to Elizabeth Andersen of the Basel Institute, equipping agencies with the skills to trace and recover illicit assets can transform how nations respond to crypto-related crime.

Experts agreed that continued dialogue, shared expertise and consistent standards are key to ensuring innovation in blockchain benefits society rather than enabling criminal networks.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Character.ai restricts teen chat access on its platform

The AI chatbot service, Character.ai, has announced that teenagers can no longer chat with its AI characters from 25 November.

Under-18s will instead be limited to generating content such as videos, as the platform responds to concerns over risky interactions and lawsuits in the US.

Character.ai has faced criticism after avatars related to sensitive cases were discovered on the site, prompting safety experts and parents to call for stricter measures.

The company cited feedback from regulators and safety specialists, explaining that AI chatbots can pose emotional risks for young users by feigning empathy or providing misleading encouragement.

Character.ai also plans to introduce new age verification systems and fund a research lab focused on AI safety, alongside enhancing role-play and storytelling features that are less likely to place teens in vulnerable situations.

Safety campaigners welcomed the decision but emphasised that preventative measures should have been implemented.

Experts warn the move reflects a broader shift in the AI industry, where platforms increasingly recognise the importance of child protection in a landscape transitioning from permissionless innovation to more regulated oversight.

Analysts note the challenge for Character.ai will be maintaining teen engagement without encouraging unsafe interactions.

Separating creative play from emotionally sensitive exchanges is key, and the company’s new approach may signal a maturing phase in AI development, where responsible innovation prioritises the protection of young users.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Humanoid robots set to power Foxconn’s new Nvidia server plant in Houston

Foxconn will add humanoid robots to a new Houston plant building Nvidia AI servers from early 2026. Announced at Nvidia’s developer conference, the move deepens their partnership and positions the site as a US showcase for AI-driven manufacturing.

Humanoid systems based on Nvidia’s Isaac GR00T N are built to perceive parts, adapt on the line, and work with people. Unlike fixed industrial arms, they handle delicate assembly and switch tasks via software updates. Goals include flexible throughput, faster retooling, and fewer stoppages.

AI models are trained in simulation using digital twins and reinforcement learning to improve accuracy and safety. On the line, robots self-tune as analytics predict maintenance and balance workloads, unlocking gains across logistics, assembly, testing, and quality control.

Texas, US, offers proximity to a growing semiconductor and AI cluster, as well as policy support for domestic capacity. Foxconn also plans expansions in Wisconsin and California to meet global demand for AI servers. Scaling output should ease supply pressures around Nvidia-class compute in data centres.

Job roles will shift as routine tasks automate and oversight becomes data-driven. Human workers focus on design, line configuration, and AI supervision, with safety gates for collaboration. Analysts see a template for Industry 4.0 factories running near-continuously with rapid changeovers.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

India deploys AI to modernise its military operations

In a move reflecting its growing strategic ambitions, India is rapidly implementing AI across its defence forces. The country’s military has moved from policy to practice, using tools from real-time sensor fusion to predictive maintenance to transform how it fights.

The shift has involved institutional change. India’s Defence AI Council and Defence AI Project Agency (established 2019) are steering an ecosystem that includes labs such as the Centre for Artificial Intelligence & Robotics of the Defence Research and Development Organisation (DRDO).

One recent example is the cross-border operation Operation Sindoor (May 2025), in which AI-driven platforms appeared in roles ranging from intelligence analysis to operational coordination.

This effort signals more than just a technological upgrade. It underscores a shift in warfare logic, where systems of systems, connectivity and rapid decision-making matter more than sheer numbers.

India’s incorporation of AI into its capabilities, drone swarming, combat simulation and logistics optimisation, is aligned with broader trends in defence innovation and digital diplomacy. The country’s strategy now places AI at the heart of its procurement demands and force design.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Alliance science pact lifts US–Korea cooperation on AI, quantum, 6G, and space

The United States and South Korea agreed on a broad science and technology memorandum to deepen alliance ties and bolster Indo-Pacific stability. The non-binding pact aims to accelerate innovation while protecting critical capabilities. Both sides cast it as groundwork for a new Golden Age of Innovation.

AI sits at the centre. Plans include pro-innovation policy alignment, trusted exports across the stack, AI-ready datasets, safety standards, and enforcement of compute protection. Joint metrology and standards work links the US Center for AI Standards and Innovation with the AI Safety Institute of South Korea.

Trusted technology leadership extends beyond AI. The memorandum outlines shared research security, capacity building for universities and industry, and joint threat analysis. Telecommunications cooperation targets interoperable 6G supply chains and coordinated standards activity with industry partners.

Quantum and basic research are priority growth areas. Participants plan interoperable quantum standards, stronger institutional partnerships, and secured supply chains. Larger projects and STEM exchanges aim to widen collaboration, supported by shared roadmaps and engagement in global consortia.

Space cooperation continues across civil and exploration programmes. Strands include Artemis contributions, a Korean cubesat rideshare on Artemis II, and Commercial Lunar Payload Services. The Korea Positioning System will be developed for maximum interoperability with GPS.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Wikipedia founder questions Musk’s Grokipedia accuracy

Speaking at the CNBC Technology Executive Council Summit in New York, Wikipedia founder Jimmy Wales has expressed scepticism about Elon Musk’s new AI-powered Grokipedia, suggesting that large language models cannot reliably produce accurate wiki entries.

Wales highlighted the difficulties of verifying sources and warned that AI tools can produce plausible but incorrect information, citing examples where chatbots fabricated citations and personal details.

He rejected Musk’s claims of liberal bias on Wikipedia, noting that the site prioritises reputable sources over fringe opinions. Wales emphasised that focusing on mainstream publications does not constitute political bias but preserves trust and reliability for the platform’s vast global audience.

Despite his concerns, Wales acknowledged that AI could have limited utility for Wikipedia in uncovering information within existing sources.

However, he stressed that substantial costs and potential errors prevent the site from entirely relying on generative AI, preferring careful testing before integrating new technologies.

Wales concluded that while AI may mislead the public with fake or plausible content, the Wiki community’s decades of expertise in evaluating information help safeguard accuracy. He urged continued vigilance and careful source evaluation as misinformation risks grow alongside AI capabilities.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

ChatGPT offers wellness checks for long chat sessions

OpenAI has introduced new features in ChatGPT to encourage healthier use for people who spend extended periods chatting with the AI. Users may see a pop-up message reading ‘Just checking in. You’ve been chatting for a while, is this a good time for a break?’.

Users can dismiss it or continue, helping to prevent excessive screen time while staying flexible. The update also guides high-stakes personal decisions.

ChatGPT will not give direct advice on sensitive topics such as relationships, but instead asks questions and encourages reflection, helping users consider their options safely.

OpenAI acknowledged that AI can feel especially personal for vulnerable individuals. Earlier versions sometimes struggled to recognise signs of emotional dependency or distress.

The company is improving the model to detect these cases and direct users to evidence-based resources when needed, making long interactions safer and more mindful.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!