Anthropic releases new constitution shaping Claude’s AI behaviour

Anthropic has published a new constitution for its AI model Claude, outlining the values, priorities, and behavioural principles designed to guide its development. Released under a Creative Commons licence, the document aims to boost transparency while shaping Claude’s learning and reasoning.

The constitution plays a central role in training, guiding how Claude balances safety, ethics, compliance, and helpfulness. Rather than rigid rules, the framework explains core principles, enabling AI systems to generalise and apply nuanced judgment.

Anthropic says this approach supports more responsible decision-making while improving adaptability.

The updated framework also enables Claude to refine its own training through synthetic data generation and self-evaluation. Using the constitution in training helps future Claude models align behaviour with human values while maintaining safety and oversight.

Anthropic described the constitution as a living document that will evolve alongside AI capabilities. External feedback and ongoing evaluation will guide updates to strengthen alignment, transparency, and responsible AI development.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

EU urged to accelerate AI deployment under new Apply AI strategy

European policymakers are calling for urgent action to accelerate AI deployment across the EU, particularly among SMEs and scale-ups, as the bloc seeks to strengthen its position in the global AI race.

Backing the European Commission’s Apply AI Strategy, the European Economic and Social Committee said Europe must prioritise trust, reliability, and human-centric design as its core competitive advantages.

The Committee warned that slow implementation, fragmented national approaches, and limited private investment are hampering progress. While the strategy promotes an ‘AI first’ mindset, policymakers stressed the need to balance innovation with strong safeguards for rights and freedoms.

Calls were also made for simpler access to funding, lighter administrative requirements, and stronger regional AI ecosystems. Investment in skills, inclusive governance, and strategic procurement were identified as key pillars for scaling trustworthy AI and strengthening Europe’s digital sovereignty.

Support for frontier AI development was highlighted as essential for reducing reliance on foreign models. Officials argued that building advanced, sovereign AI systems aligned with European values could enable competitive growth across sectors such as healthcare, finance, and industry.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

From chips to jobs: Huang’s vision for AI at Davos 2026

AI is evolving into a foundational economic system rather than a standalone technology, according to NVIDIA chief executive Jensen Huang, who described AI as a five-layer infrastructure spanning energy, hardware, data centres, models and applications.

Speaking at the World Economic Forum in Davos, Huang argued that building and operating each layer is triggering what he called the most significant infrastructure expansion in human history, with job creation stretching from power generation and construction to cloud operations and software development.

Investment patterns suggest a structural shift instead of a speculative cycle. Venture capital funding in 2025 reached record levels, largely flowing into AI-native firms across healthcare, manufacturing, robotics and financial services.

Huang stressed that the application layer will deliver the most significant economic return as AI moves from experimentation to core operational use across industries.

Concerns around job displacement were framed as misplaced. AI automates tasks rather than replacing professional judgement, enabling workers to focus on higher-value activities.

In healthcare, productivity gains from AI-assisted diagnostics and documentation are already increasing demand for radiologists and nurses rather than reducing headcount, as improved efficiency enables institutions to treat more patients.

Huang positioned AI as critical national infrastructure, urging governments to develop domestic capabilities aligned with local language, culture and industrial strengths.

He described AI literacy as an essential skill, comparable to leadership or management, while arguing that accessible AI tools could narrow global technology divides rather than widen them.

Diplo is live reporting on all sessions from the World Economic Forum 2026 in Davos.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

South Korea sets the global standard for frontier AI regulation

South Korea will begin enforcing its Artificial Intelligence Act on Thursday, becoming the first country to introduce formal safety requirements for high-performance, or frontier, AI systems, reshaping the global regulatory landscape.

The law establishes a national AI governance framework, led by the Presidential Council on National Artificial Intelligence Strategy, and creates an AI Safety Institute to oversee safety and trust assessments.

Alongside regulatory measures, the government is rolling out broad support for research, data infrastructure, talent development, startups, and overseas expansion, signalling a growth-oriented policy stance.

To minimise early disruption, authorities will introduce a minimum one-year grace period centred on guidance, consultation, and education rather than enforcement.

Obligations cover three areas: high-impact AI in critical sectors, safety rules for frontier models, and transparency requirements for generative AI, including disclosure of realistic synthetic content.

Enforcement remains light-touch, prioritising corrective orders over penalties, with fines capped at 30 million won for persistent noncompliance. Officials said the framework aims to build public trust while supporting innovation, serving as a foundation for ongoing policy development.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

GPT-5.2 shows how AI can generate real-world cyber exploits

Advanced language models have demonstrated the ability to generate working exploits for previously unknown software vulnerabilities. Security researcher Sean Heelan tested two systems built on GPT-5.2 and Opus 4.5 by challenging them to exploit a zero-day flaw in the QuickJS JavaScript interpreter.

Across multiple scenarios with varying security protections, GPT-5.2 completed every task, while Opus 4.5 failed only 2. The systems produced more than 40 functional exploits, ranging from basic shell access to complex file-writing operations that bypassed modern defences.

Most challenges were solved in under an hour, with standard attempts costing around $30. Even the most complex exploit, which bypassed protections such as address space layout randomisation, non-executable memory, and seccomp sandboxing, was completed in just over three hours for roughly $50.

The most advanced task required GPT-5.2 to write a specific string to a protected file path without access to operating system functions. The model achieved this by chaining seven function calls through the glibc exit handler mechanism, bypassing shadow stack protections.

The findings suggest exploit development may increasingly depend on computational resources rather than human expertise. While QuickJS is less complex than browsers such as Chrome or Firefox, the approach demonstrated could scale to larger and more secure software environments.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

AI model maps how humans form emotions

Researchers in Japan have developed an AI framework designed to model how humans form emotional experiences by integrating bodily signals, sensory input and language. The work was led by scientists at Nara Institute of Science and Technology in collaboration with Osaka University.

The AI model draws on the theory of constructed emotion, which suggests emotions are built by the brain rather than hard-wired responses. Physiological data, visual cues and spoken descriptions were analysed together to replicate how people experience feelings in real situations.

Using unlabeled data from volunteers exposed to emotion-evoking images and videos, the system identified emotional patterns without predefined categories. Results showed about 75 percent alignment with participants’ own emotional assessments, well above chance levels.

The Japanese researchers say the approach could support emotion-aware AI applications in healthcare, robotics and mental health support. Findings were published in IEEE Transactions on Affective Computing, with potential benefits for understanding emotions that are difficult to express verbally.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

How Microsoft is shaping UN reform through digital infrastructure

Microsoft has announced a multi-year pledge to support the United Nations’ UN80 reform initiative, positioning AI and digital infrastructure as central tools for modernising multilateral governance.

The commitment follows agreement among all UN member states on efficiency and financial-stability measures, as the organisation faces growing operational demands amid constrained resources.

The initiative includes a dedicated innovation fund, preferential pricing for digital services, and free AI training for UN staff across agencies and missions.

Rather than focusing on policy direction, Microsoft frames its role as enabling institutional capacity, from procurement and logistics to humanitarian response and development planning, while encouraging other private-sector actors to align behind UN80 priorities.

Microsoft also plans to mobilise partners such as EY to support reform efforts, reinforcing a model where large technology firms contribute expertise, infrastructure, and coordination capacity to global governance systems.

Previous collaborations with UNICEF, UNHCR, ITU, and the ILO are cited as evidence that AI-driven tools can accelerate service delivery at scale.

The pledge highlights how multilateral reform increasingly depends on private technological ecosystems instead of purely intergovernmental solutions.

As AI becomes embedded in the core operations of international institutions, questions around accountability, influence, and long-term dependency are likely to shape debates about the future balance between public authority and corporate power.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Hong Kong crypto licensing overhaul draws industry concern

Hong Kong’s proposed crypto licensing overhaul has drawn criticism from industry leaders, who warn it could disrupt compliant firms and deter blockchain exposure.

Under the proposals, the existing allowance enabling firms to allocate up to 10% of fund assets to crypto without additional licensing would be removed. Even minimal exposure would require a full licence, a move the association called disproportionate and harmful to market experimentation.

Concerns also focused on the absence of transitional arrangements. Without a grace period, firms may be forced to suspend operations while licence applications are reviewed.

The association proposed a six- to 12-month transitional window to allow continued activity during regulatory processing.

Further criticism focused on custody rules restricting client assets to SFC-licensed custodians. Industry representatives warned the measure could limit access to early-stage tokens, restrict Web3 investment, and impose unnecessary geographic constraints.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Youth express higher concern over AI job disruption, global survey indicates

A Randstad survey of global workers shows that younger employees are significantly more worried than older counterparts about the impact of artificial intelligence (AI) on jobs and career prospects.

Respondents aged under 35 reported higher levels of concern that AI could disrupt employment, reshape skill demands and erode job security.

The survey highlights a generational divide in attitudes toward AI adoption: while many older workers express cautious optimism, younger workers often point to risks related to hiring, task automation and future opportunities.

Employers are urged to address these concerns by investing in skills development, training programmes and transparent communication about how AI will be deployed in the workplace.

The findings come amid broader global discussions on the future of work, with AI increasingly integrated into business processes such as recruitment, project planning and administrative automation.

Analysts suggest that targeted education and retraining initiatives could help younger employees adapt and benefit from AI-driven changes rather than feel displaced by them.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

ChatGPT introduces age prediction to strengthen teen safety

New safeguards are being introduced as ChatGPT uses age prediction to identify accounts that may belong to under-18s. Extra protections limit exposure to harmful content while still allowing adults full access.

The age prediction model analyses behavioural and account-level signals, including usage patterns, activity times, account age, and stated age information. OpenAI says these indicators help estimate whether an account belongs to a minor, enabling the platform to apply age-appropriate safeguards.

When an account is flagged as potentially under 18, ChatGPT limits access to graphic violence, sexual role play, viral challenges, self-harm, and unhealthy body image content. The safeguards reflect research on teen development, including differences in risk perception and impulse control.

ChatGPT users who are incorrectly classified can restore full access by confirming their age through a selfie check using Persona, a secure identity verification service. Account holders can review safeguards and begin the verification process at any time via the settings menu.

Parental controls allow further customisation, including quiet hours, feature restrictions, and notifications for signs of distress. OpenAI says the system will continue to evolve, with EU-specific deployment planned in the coming weeks to meet regional regulatory requirements.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!