Rights before risks: Rethinking quantum innovation at WSIS+20

At the WSIS+20 High-Level Event in Geneva, a powerful call was made to ensure the development of quantum technologies remains rooted in human rights and inclusive governance. A UNESCO-led session titled ‘Human Rights-Centred Global Governance of Quantum Technologies’ presented key findings from a new issue brief co-authored with Sciences Po and the European University Institute.

It outlined major risks—such as quantum’s dual-use nature threatening encryption, a widening technological divide, and severe gender imbalances in the field—and urged immediate global action to build safeguards before quantum capabilities mature.

UNESCO’s Guilherme Canela emphasised that innovation and human rights are not mutually exclusive but fundamentally interlinked, warning against a ‘false dichotomy’ between the two. Lead author Shamira Ahmed highlighted the need for proactive frameworks to ensure quantum benefits are equitably distributed and not used to deepen global inequalities or erode rights.

With 79% of quantum firms lacking female leadership and a mere 1 in 54 job applicants being women, the gender gap was called ‘staggering.’ Ahmed proposed infrastructure investment, policy reforms, capacity development, and leveraging the UN’s International Year of Quantum to accelerate global discussions.

Panellists echoed the urgency. Constance Bommelaer de Leusse from Sciences Po advocated for embedding multistakeholder participation into governance processes and warned of a looming ‘quantum arms race.’ Professor Pieter Vermaas of Delft University urged moving from talk to international collaboration, suggesting the creation of global quantum research centres.

Journalist Elodie Vialle raised alarms about quantum’s potential to supercharge surveillance, endangering press freedom and digital privacy, and underscored the need to close the cultural gap between technologists and civil society.

Overall, the session championed a future where quantum technology is developed transparently, governed globally, and serves as a digital public good, bridging divides rather than deepening them. Speakers agreed that the time to act is now, before today’s opportunities become tomorrow’s crises.

Track all key events from the WSIS+20 High-Level Event 2025 on our dedicated page.

Preserving languages in a digital world: A call for inclusive action

At the WSIS+20 High-Level Event in Geneva, UNESCO convened a powerful session on the critical need to protect multilingualism in the digital age. With over 8,000 languages spoken globally but fewer than 120 represented online, the panel warned of a growing digital divide that excludes billions and marginalises thousands of cultures.

Dr Tawfik Jelassi of UNESCO painted a vivid metaphor of the internet as a vast library where most languages have no books on the shelves, calling for urgent action to safeguard humanity’s linguistic and cultural diversity.

Speakers underscored that bridging this divide goes beyond creating language tools—it requires systemic change rooted in policy, education, and community empowerment. Guilherme Canela of UNESCO highlighted ongoing initiatives like the 2003 Recommendation on Multilingualism and the UN Decade of Indigenous Languages, which has already inspired 15 national action plans.

Panellists like Valts Ernstreits and Sofiya Zahova emphasised community-led efforts, citing examples from Latvia, Iceland, and Sámi institutions that show how native speakers and local institutions must lead digital inclusion efforts.

Africa’s case brought the urgency into sharp focus. David Waweru noted that despite hosting a third of the world’s languages, less than 0.1% of websites feature African language content. Yet, promising efforts like the African Storybook project and AI language models show how local storytelling and education can thrive in digital spaces.

Elena Plexida of ICANN revealed that only 26% of email servers accept non-Latin addresses, a stark reminder of the structural barriers to full digital participation.

The session concluded with a strong call for multistakeholder collaboration. Governments, tech companies, indigenous communities, and civil society must work together to make multilingualism the default, not the exception, in digital spaces. As Jelassi put it, ensuring every language has a place online is not just a technical challenge but a matter of cultural survival and digital justice.

Track all key events from the WSIS+20 High-Level Event 2025 on our dedicated page.

Ari Aster warns of AI’s creeping normality ahead of Eddington release

Ari Aster, the director behind Hereditary and Midsommar, is sounding the alarm on AI. In a recent Letterboxd interview promoting his upcoming A24 film Eddington, Aster described his growing unease with AI.

He framed it as a quasi-religious force reshaping reality in ways that are already irreversible. ‘If you talk to these engineers… they talk about AI as a god,’ said Aster. ‘They’re very worshipful of this thing. Whatever space there was between our lived reality and this imaginal reality — that’s disappearing.’

Aster’s comments suggest concern not just about the technology, but about the mindset surrounding its development. Eddington, set during the COVID-19 pandemic, is a neo-Western dark comedy.
It stars Joaquin Phoenix and Pedro Pascal as a sheriff and a mayor locked in a bitter digital feud.

The film reflects Aster’s fears about the dehumanising impact of modern technology. He drew from the ideas of media theorist Marshall McLuhan, referencing his phrase: ‘Man is the sex organ of the machine world.’ Aster asked, ‘Is this technology an extension of us, are we extensions of this technology, or are we here to usher it into being?’

The implication is clear: AI may not simply assist humanity—it might define it. Aster’s films often explore existential dread and loss of control. His perspective on AI taps into similar fears, but in real life. ‘The most uncanny thing about it is that it’s less uncanny than I want it to be,’ he said.

‘I see AI-generated videos, and they look like life. The longer we live in them, the more normal they become.’ The normalisation of artificial content strikes at the core of Aster’s unease. It also mirrors recent tensions in Hollywood over AI’s role in creative industries.

In 2023, WGA and SAG-AFTRA fought for protections against AI-generated scripts and likenesses. Their strike shut down the industry for months, but won language limiting AI use.

The battles highlighted the same issue Aster warns of—losing artistic agency to machines. ‘What happens when content becomes so seamless, it replaces real creativity?’ he seems to ask.

‘Something huge is happening right now, and we have no say in it,’ he said. ‘I can’t believe we’re actually going to live through this and see what happens. Holy cow.’ Eddington is scheduled for release in the United States on 18 July 2025.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Digital rights under threat: Global majority communities call for inclusive solutions at IGF 2025

At the Internet Governance Forum 2025 in Lillestrøm, Norway, a pivotal session hosted by Oxfam’s RECIPE Project shed light on the escalating digital rights challenges facing communities across the Global majority. Representatives from Vietnam, Bolivia, Cambodia, Somalia, and Palestine presented sobering findings based on research with over 1,000 respondents across nine countries.

Despite the diversity of regions, speakers echoed similar concerns: digital literacy is dangerously low, access to safe and inclusive online spaces remains unequal, and legal protections for digital rights are often absent or underdeveloped.

The human cost of digital inequality was made clear from Bolivia to Palestine. In Bolivia, over three-quarters of respondents had experienced digital security incidents, and many reported targeted violence linked to their roles as human rights defenders.

In Somalia, where internet penetration is high, only a fraction understands how to protect their personal data. Palestine, meanwhile, faces systematic digital discrimination, marked by unequal infrastructure access and advanced surveillance technologies used against its population, exacerbated by ongoing occupation and political instability.

Yet amidst these challenges, the forum underscored a strong sense of resilience and innovation. Civil society organisations from Cambodia and Bolivia showcased bottom-up approaches, such as peer-led digital security training and feminist digital safety networks, which help communities protect themselves and influence policy.

Vietnam emphasised the need for genuine participation in policymaking, rather than formalistic consultations, as a path to more equitable digital governance. The session concluded with a shared call to action: digital governance must prioritise human rights and meaningful participation from the ground up.

Speakers and audience members highlighted the urgent need for multistakeholder cooperation—spanning civil society, government, and the tech industry—to counter misinformation and protect freedom of expression, especially in the face of expanding surveillance and online harm. As one participant from Zambia noted, digital safety must not come at the expense of digital freedom; the two must evolve together.

Track all key moments from the Internet Governance Forum 2025 on our dedicated IGF page.

IGF 2025: Africa charts a sovereign path for AI governance

African leaders at the Internet Governance Forum (IGF) 2025 in Oslo called for urgent action to build sovereign and ethical AI systems tailored to local needs. Hosted by the German Federal Ministry for Economic Cooperation and Development (BMZ), the session brought together voices from government, civil society, and private enterprises.

Moderated by Ashana Kalemera, Programmes Manager at CIPESA, the discussion focused on ensuring AI supports democratic governance in Africa. ‘We must ensure AI reflects our realities,’ Kalemera said, emphasising fairness, transparency, and inclusion as guiding principles.

Executive Director of Policy Neema Iyer warned that AI harms governance through surveillance, disinformation, and political manipulation. ‘Civil society must act as watchdogs and storytellers,’ she said, urging public interest impact assessments and grassroots education.

Representing South Africa, Mlindi Mashologu stressed the need for transparent governance frameworks rooted in constitutional values. ‘Policies must be inclusive,’ he said, highlighting explainability, data bias removal, and citizen oversight as essential components of trustworthy AI.

Lacina Koné, CEO of Smart Africa, called for urgent action to avoid digital dependency. ‘We cannot be passively optimistic. Africa must be intentional,’ he stated. Over 1,000 African startups rely on foreign AI models, creating sovereignty risks.

Koné emphasised that Africa should focus on beneficial AI, not the most powerful. He highlighted agriculture, healthcare, and education sectors where local AI could transform. ‘It’s about opportunity for the many, not just the few,’ he said.

From Mauritania, Matchiane Soueid Ahmed shared her country’s experience developing a national AI strategy. Challenges include poor rural infrastructure, technical capacity gaps, and lack of institutional coordination. ‘Sovereignty is not just territorial—it’s digital too,’ she noted.

Shikoh Gitau, CEO of KALA in Kenya, brought a private sector perspective. ‘We must move from paper to pavement,’ she said. Her team runs an AI literacy campaign across six countries, training teachers directly through their communities.

Gitau stressed the importance of enabling environments and blended financing. ‘Governments should provide space, and private firms must raise awareness,’ she said. She also questioned imported frameworks: ‘What definition of democracy are we applying?’

Audience members from Gambia, Ghana, and Liberia raised key questions about harmonisation, youth fears over job loss and AI readiness. Koné responded that Smart Africa is benchmarking national strategies and promoting convergence without erasing national sovereignty.

Though 19 African countries have published AI strategies, speakers noted that implementation remains slow. Practical action—such as infrastructure upgrades, talent development, and public-private collaboration—is vital to bring these frameworks to life.

The panel underscored the need to build AI systems prioritising inclusion, utility, and human rights. Investments in digital literacy, ethics boards, and regulatory sandboxes were cited as key tools for democratic AI governance.

Kalemera concluded, ‘It’s not yet Uhuru for AI in Africa—but with the right investments and partnerships, the future is promising.’ The session reflected cautious optimism and a strong desire for Africa to shape its AI destiny.

Track all key moments from the Internet Governance Forum 2025 on our dedicated IGF page.

Taiwan leads in AI election defence efforts

Taiwan has been chosen to lead a new coalition formed by the International Foundation for Electoral Systems to strengthen democratic resilience against AI-driven disinformation. The AI Advisory Group on Elections will unite policymakers and experts to address AI’s role in protecting fair elections.

The island’s experience has made it a key voice in global AI governance as it counters sophisticated disinformation campaigns linked to authoritarian regimes. Taiwan’s Cyber Ambassador, Audrey Tang, stressed that AI must serve the greater good and help build accountable digital societies.

Taiwan has developed rapid response and civic fact-checking tools that many democracies now look to adopt. These measures helped ensure the integrity of its recent elections despite unprecedented levels of AI-generated disinformation and cyberattacks.

Global democracies are urged to learn from Taiwan’s playbook as threats evolve, and the influence of AI on elections grows. Taiwan’s success shows that resilience can be achieved without sacrificing civil liberties.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

AI governance efforts centre on human rights

At the Internet Governance Forum 2025 in Lillestrøm, Norway, a key session spotlighted the launch of the Freedom Online Coalition’s (FOC) updated Joint Statement on Artificial Intelligence and Human Rights. Backed by 21 countries and counting, the statement outlines a vision for human-centric AI governance rooted in international human rights law.

Representatives from governments, civil society, and the tech industry—most notably the Netherlands, Germany, Ghana, Estonia, and Microsoft—gathered to emphasise the urgent need for a collective, multistakeholder approach to tackle the real and present risks AI poses to rights such as privacy, freedom of expression, and democratic participation.

Ambassador Ernst Noorman of the Netherlands warned that human rights and security must be viewed as interconnected, stressing that unregulated AI use can destabilise societies rather than protect them. His remarks echoed the Netherlands’ own hard lessons from biassed welfare algorithms.

Other panellists, including Germany’s Cyber Ambassador Maria Adebahr, underlined how AI is being weaponised for transnational repression and emphasised Germany’s commitment by doubling funding for the FOC. Ghana’s cybersecurity chief, Divine Salese Agbeti, added that AI misuse is not exclusive to governments—citizens, too, have exploited the technology for manipulation and deception.

From the private sector, Microsoft’s Dr Erika Moret showcased the company’s multi-layered approach to embedding human rights in AI, from ethical design and impact assessments to rejecting high-risk applications like facial recognition in authoritarian contexts. She stressed the company’s alignment with UN guiding principles and the need for transparency, fairness, and inclusivity.

The discussion also highlighted binding global frameworks like the EU AI Act and the Council of Europe’s Framework Convention, calling for their widespread adoption as vital tools in managing AI’s global impact. The session concluded with a shared call to action: governments must use regulatory tools and procurement power to enforce human rights standards in AI, while the private sector and civil society must push for accountability and inclusion.

The FOC’s statement remains open for new endorsements, standing as a foundational text in the ongoing effort to align the future of AI with the fundamental rights of all people.

Track all key moments from the Internet Governance Forum 2025 on our dedicated IGF page.

Parliamentarians call for stronger platform accountability and human rights protections at IGF 2025

At the 2025 Internet Governance Forum in Lillestrøm, Norway, parliamentarians from around the world gathered to share perspectives on how to regulate harmful online content without infringing on freedom of expression and democratic values. The session, moderated by Sorina Teleanu, Diplo’s Director of Knowledge, highlighted the increasing urgency for social media platforms to respond more swiftly and responsibly to harmful content, particularly content generated by AI that can lead to real-world consequences such as harassment, mental health issues, and even suicide.

Pakistan’s Anusha Rahman Ahmad Khan delivered a powerful appeal, pointing to cultural insensitivity and profit-driven resistance by platforms that often ignore urgent content removal requests. Representatives from Argentina, Nepal, Bulgaria, and South Africa echoed the need for effective legal frameworks that uphold safety and fundamental rights.

Argentina’s Franco Metaza, Member of Parliament of Mercosur, cited disturbing content that promotes eating disorders among young girls and detailed the tangible danger of disinformation, including an assassination attempt linked to online hate. Nepal’s MP Yogesh Bhattarai advocated for regulation without authoritarian control, underscoring the importance of constitutional safeguards for speech.

Member of European Parliament, Tsvetelina Penkova from Bulgaria, outlined the EU’s multifaceted digital laws, like the Digital Services Act and GDPR, which aim to protect users while grappling with implementation challenges across 27 diverse member states.

Youth engagement and digital literacy emerged as key themes, with several speakers emphasising that involving young people in policymaking leads to better, more inclusive policies. Panellists also stressed that education is essential for equipping users with the tools to navigate online spaces safely and critically.

Calls for multistakeholder cooperation rang throughout the session, with consensus on the need for collaboration between governments, tech companies, civil society, and international organisations. A thought-provoking proposal from a Congolese parliamentarian suggested that digital rights be recognised as a new, fourth generation of human rights—akin to civil, economic, and environmental rights already codified in international frameworks.

Other attendees welcomed the idea and agreed that without such recognition, the enforcement of digital protections would remain fragmented. The session concluded on a collaborative and urgent note, with calls for shared responsibility, joint strategies, and stronger international frameworks to create a safer, more just digital future.

Track all key moments from the Internet Governance Forum 2025 on our dedicated IGF page.

World gathers in Norway to shape digital future

The Internet Governance Forum (IGF) 2025 opened in Lillestrøm, Norway, marking its 20th anniversary and coinciding with the World Summit on the Information Society Plus 20 (WSIS+20) review.

UN Secretary-General António Guterres, in a video message, underscored that digital cooperation has shifted from aspiration to necessity. He highlighted global challenges such as the digital divide, online hate speech, and concentrated tech power, calling for immediate action to ensure a more equitable digital future.

https://twitter.com/intgovforum/status/1937473277695246428

Norwegian leaders, including Prime Minister Jonas Gahr Støre and Digitisation Minister Karianne Tung, reaffirmed their country’s commitment to democratic digital governance and human rights, echoing broader forum themes of openness, transparency, and multilateral cooperation. They emphasised the importance of protecting the internet as a public good in an era marked by fragmentation, misinformation, and increasing geopolitical tension.

https://twitter.com/intgovforum/status/1937461829891915844

The ceremony brought together diverse voices—from small island states and the EU to civil society and the private sector. Mauritius’ President Dharambeer Gokhool advocated for a citizen-centered digital transformation, while European Commission Vice President Henna Virkkunen introduced a new EU international digital strategy rooted in human rights and sustainability.

Actor and digital rights activist Joseph Gordon-Levitt cautioned against unregulated AI development, arguing for governance frameworks that protect human agency and economic fairness.

Why does it matter?

Echoing across speeches was a shared call to action: to strengthen the multistakeholder model of internet governance, bridge the still-massive digital divide, and develop ethical, inclusive digital policies. As stakeholders prepare to delve into deeper dialogues during the forum, the opening ceremony made clear that the next chapter of digital governance must be collaborative, human-centered, and urgently enacted.

Track all key moments from the Internet Governance Forum 2025 on our dedicated IGF page.

Big Tech’s grip on information sparks urgent debate at IGF 2025 in Norway

At the Internet Governance Forum 2025 in Lillestrøm, Norway, global leaders, tech executives, civil society figures, and academics converged for a high-level session to confront one of the digital age’s most pressing dilemmas: how to protect democratic discourse and human rights amid big tech’s tightening control over the global information space. The session, titled ‘Losing the Information Space?’, tackled the rising threat of disinformation, algorithmic opacity, and the erosion of public trust, all amplified by powerful AI technologies.

Norwegian Minister Lubna Jaffery sounded the alarm, referencing the annulled Romanian presidential election as a stark reminder of how influence operations and AI-driven disinformation campaigns can destabilise democracies. She warned that while platforms have democratised access to expression, they’ve also created fragmented echo chambers and supercharged the spread of propaganda.

Estonia’s Minister of Justice and Digital Affairs Liisa Ly Pakosta echoed the concern, describing how her country faces persistent information warfare—often backed by state actors—and announced Estonia’s rollout of AI-based education to equip youth with digital resilience. The debate revealed deep divides over how to achieve transparency and accountability in tech.

TikTok’s Lisa Hayes defended the company’s moderation efforts and partnerships with fact-checkers, advocating for what she called ‘meaningful transparency’ through accessible tools and reporting. But others, like Reporters Without Borders’ Thibaut Bruttin, demanded structural reform.

He argued platforms should be treated as public utilities, legally obliged to give visibility to trustworthy journalism, and rejected the idea that digital space should remain under the control of private interests. Despite conflicting views on the role of regulation versus collaboration, panellists agreed that the threat of disinformation is real and growing—and no single entity can tackle it alone.

The session closed with calls for stronger international legal frameworks, cross-sector cooperation, and bold action to defend truth, freedom of expression, and democratic integrity in an era where technology’s influence is pervasive and, if unchecked, potentially perilous.

Track all key moments from the Internet Governance Forum 2025 on our dedicated IGF page.