EU faces pressure to boost action on health disinformation

A global health organisation is urging the EU to make fuller use of its digital rules to curb health disinformation as concerns grow over the impact of deepfakes on public confidence.

Warnings point to a rising risk that manipulated content could reduce vaccine uptake instead of supporting informed public debate.

Experts argue that the Digital Services Act already provides the framework needed to limit harmful misinformation, yet enforcement remains uneven. Stronger oversight could improve platforms’ ability to detect manipulated content and remove inaccurate claims that jeopardise public health.

Campaigners emphasise that deepfake technology is now accessible enough to spread false narratives rapidly. The trend threatens vaccination campaigns at a time when several member states are attempting to address declining trust in health authorities.

The EU officials continue to examine how digital regulation can reinforce public health strategies. The call for stricter enforcement highlights the pressure on Brussels to ensure that digital platforms act responsibly rather than allowing misleading material to circulate unchecked.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!  

Writing as thinking in the age of AI

In his article, Richard Gunderman argues that writing is not merely a way to present ideas but a core human activity through which people think, reflect and form meaning.

He contends that when AI systems generate text on behalf of users, they risk replacing this cognitive process with automated output, weakening the connection between thought and expression.

According to the piece, writing serves as a tool for reasoning, emotional processing and moral judgment. Offloading it to AI can diminish originality, flatten individual voice and encourage passive consumption of machine-produced ideas.

Gunderman warns that this shift could lead to intellectual dependency, where people rely on AI to structure arguments and articulate positions rather than developing those skills themselves.

The article also raises ethical concerns about authenticity and responsibility. If AI produces large portions of written work, it becomes unclear who is accountable for the ideas expressed. Gunderman suggests that overreliance on AI writing tools may undermine trust in communication and blur the line between human and machine authorship.

Overall, the piece calls for a balanced approach: AI may assist with editing or idea generation, but the act of writing itself should remain fundamentally human, as it is central to critical thinking, identity and social responsibility.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Learnovate launches community of practice on AI for learning

The Learnovate Centre, a global innovation hub focused on the future of work and learning at Trinity College Dublin, is spearheading a community of practice on responsible AI in learning, bringing together educators, policymakers, institutional leaders and sector specialists to discuss safe, effective and compliant uses of AI in educational settings.

This initiative aims to help practitioners interpret emerging policy frameworks, including EU AI Act requirements, share practical insights and align AI implementation with ethical and pedagogical principles.

One of the community’s early activities includes virtual meetings designed to build consensus around AI norms in teaching, compliance strategies and knowledge exchange on real-world implementation.

Participants come from diverse education domains, including schools, higher and vocational education and training, as well as representatives from government and unions, reflecting a broader push to coordinate AI adoption across the sector.

Learnovate plays a wider role in AI and education innovation, supporting research, summits and collaborative programmes that explore AI-powered tools for personalised learning, upskilling and ethical use cases.

It also partners with start-ups and projects (such as AI platforms for teachers and learners) to advance practical solutions that balance innovation with safeguards.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

AI in education reveals a critical evidence gap

Universities are increasingly reorganising around AI, treating AI-based instruction as a proven solution for delivering education more efficiently. This shift reflects a broader belief that AI can reliably replace or reduce human-led teaching, despite growing uncertainty about its actual impact on learning.

Recent research challenges this assumption by re-examining the evidence used to justify AI-driven reforms. A comprehensive re-analysis of AI and learning studies reveals severe publication bias, with positive results published far more frequently than negative or null findings. Once corrected, reported learning gains from AI shrink substantially and may be negligible.

More critically, the research exposes deep inconsistency across studies. Outcomes vary so widely that the evidence cannot predict whether AI will help or harm learning in a given context, and no educational level, discipline, or AI application shows consistent benefits.

By contrast, human-mediated teaching remains a well-established foundation of learning. Decades of research demonstrate that understanding develops through interaction, adaptation, and shared meaning-making, leading the article to conclude that AI in education remains an open question, while human instruction remains the known constant.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Study questions reliability of AI medical guidance

AI chatbots are not yet capable of providing reliable health advice, according to new research published in the journal Nature Medicine. Findings show users gain no greater diagnostic accuracy from chatbots than from traditional internet searches.

Researchers tested nearly 1,300 UK participants using ten medical scenarios, ranging from minor symptoms to conditions requiring urgent care. Participants were assigned to use either OpenAI’s GPT-4o, Meta’s Llama 3, Command R+, or a standard search engine to assess symptoms and determine next steps.

Chatbot users identified their condition about one-third of the time, with only 45 percent selecting the correct medical response. Performance levels matched those relying solely on search engines, despite AI systems scoring highly on medical licensing benchmarks.

Experts attributed the gap to communication failures. Users often provided incomplete information or misinterpreted chatbot guidance.

Researchers and bioethicists warned that growing reliance on AI for medical queries could pose public health risks without professional oversight.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Czechia weighs under-15 social media ban as government debate intensifies

A ban on social media use for under-15s is being weighed in Czechia, with government officials suggesting the measure could be introduced before the end of the year.

Prime Minister Andrej Babiš has voiced strong support and argues that experts point to potential harm linked to early social media exposure.

France recently enacted an under-15 restriction, and a growing number of European countries are exploring similar limits rather than relying solely on parental guidance.

The discussion is part of a broader debate about children’s digital habits, with Czech officials also considering a ban on mobile phones in schools. Slovakia has already adopted comparable rules, giving Czech ministers another model to study as they work on their own proposals.

Not all political voices agree on the direction of travel. Some warn that strict limits could undermine privacy rights or diminish online anonymity, while others argue that educational initiatives would be more effective than outright prohibition.

UNICEF has cautioned that removing access entirely may harm children who rely on online platforms for learning or social connection instead of traditional offline networks.

Implementing a nationwide age restriction poses practical and political challenges. The government of Czechia heavily uses social media to reach citizens, complicating attempts to restrict access for younger users.

Age verification, fair oversight and consistent enforcement remain open questions as ministers continue consultations with experts and service providers.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Dubai hosts launch of AI tools for university students

The UAE Ministry of Higher Education and Scientific Research has partnered with Microsoft to develop AI agents to help university students find jobs. The initiative was announced in Dubai during a major policy gathering in the UAE.

The collaboration in the UAE will use Microsoft Azure to build prototype AI agents supporting personalised learning and career navigation. Dubai-based officials said the tools are designed to align higher education with labour market needs in the UAE.

Four AI agents are being developed in the UAE, covering lifelong skills planning, personalised learning, course co creation and research alignment. Dubai remains central to the project as a hub for higher education innovation in the UAE.

Officials in the UAE said the partnership reflects national priorities around innovation and a knowledge based economy. Microsoft said Dubai offers an ideal environment to scale AI driven education tools across the UAE.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

EU split widens over ban on AI nudification apps

European lawmakers remain divided over whether AI tools that generate non-consensual sexual images should face an explicit ban in the EU legislation.

The split emerged as debate intensified over the AI simplification package, which is moving through Parliament and the Council rather than remaining confined to earlier negotiations.

Concerns escalated after Grok was used to create images that digitally undressed women and children.

The EU regulators responded by launching an investigation under the Digital Services Act, and the Commission described the behaviour as illegal under existing European rules. Several lawmakers argue that the AI Act should name pornification apps directly instead of relying on broader legal provisions.

Lead MEPs did not include a ban in their initial draft of the Parliament’s position, prompting other groups to consider adding amendments. Negotiations continue as parties explore how such a restriction could be framed without creating inconsistencies within the broader AI framework.

The Commission appears open to strengthening the law and has hinted that the AI omnibus could be an appropriate moment to act. Lawmakers now have a limited time to decide whether an explicit prohibition can secure political agreement before the amendment deadline passes.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Spain faces escalating battle with Telegram founder

The confrontation between Spain and Telegram founder Pavel Durov has intensified after he claimed that Pedro Sánchez endangered online freedoms.

Government officials responded that the tech executive spread lies rather than engage with the proposed rules in good faith. Sánchez argued that democracy would not be silenced by what he called the techno-oligarchs of the algorithm.

The dispute followed the unveiling of new measures aimed at major technology companies. The plan introduces a ban on social media use for under-16s and holds corporate leaders legally responsible when unlawful or hateful content remains online rather than being removed.

Platforms would also need to adopt age-verification tools such as ID checks or biometric systems, which Durov argued could turn Spain into a surveillance state by allowing large-scale data collection.

Tensions widened as Sánchez clashed with prominent US tech figures. Sumar urged all bodies linked to the central administration to leave X, a move that followed Elon Musk’s accusation that the Spanish leader was acting like a tyrant.

The row highlighted how Spain’s attempt to regulate digital platforms has placed its government in open conflict with influential technology executives.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Claude Opus 4.6 sets new benchmark for enterprise AI

Anthropic has released Claude Opus 4.6, its most advanced AI model to date, introducing significant improvements in coding performance, reasoning depth, and long-context comprehension.

Engineering workflows stand to benefit from stronger debugging, code review, and better large-scale repository management, while agentic task execution now runs for more extended periods with greater reliability.

The model’s 1M token context window, now in beta, enables sustained reasoning across vast datasets and extended conversations. Performance gains span multiple benchmarks, leading in agentic coding, multidisciplinary reasoning, and high-value knowledge work.

Information retrieval in long documents has also improved, addressing persistent industry concerns around context degradation.

Operational capabilities extend beyond software development into enterprise productivity. Financial analysis, research, and document workflows gain direct support, with spreadsheet and presentation integrations enhancing daily business use.

Within Claude Code, newly introduced agent teams allow multiple AI agents to collaborate autonomously on complex workloads.

Safety remains central, with expanded evaluations showing low misalignment risk backed by interpretability research and cybersecurity safeguards. Additional tools- adaptive thinking, effort scaling, and context compaction- add flexibility for deploying long-running AI systems at scale.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!