Women driving tech innovation as Web Summit marks 10 years

Web Summit’s Women in Tech programme marked a decade of work in Qatar by highlighting steady progress in female participation across global technology sectors.

The Web Summit event recorded an increase in women-founded startups and reflected rising engagement in Qatar, where female founders reached 38 percent.

Leaders from the initiative noted how supportive networks, mentorship, and access to role models are reshaping opportunities for women in technology and entrepreneurship.

Speakers from IBM and other companies focused on the importance of AI skills in shaping the future workforce. They argued that adequate preparation depends on understanding how AI shapes everyday roles, rather than relying solely on technical tools.

IBM’s SkillsBuild platform continues to partner with universities, schools, and nonprofit groups to expand access to recognised AI credentials that can support higher earning potential and new career pathways.

Another feature of the event was its emphasis on inclusion as a driver of innovation. The African Women in Technology initiative, led by Anie Akpe, is working to offer free training in cybersecurity and AI so women in emerging markets can benefit from new digital opportunities.

These efforts aim to support business growth at every level, even for women operating in local markets, who can use technology to reach wider communities.

Female founders also used the platform to showcase new health technology solutions.

ScreenMe, a Qatari company founded by Dr Golnoush Golsharazi, presented its reproductive microbiome testing service, created in response to long-standing gaps in women’s health research and screening.

Organisers expressed confidence that women-led innovation will expand across the region, supported by rising investment and continuing visibility at major global events.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Global coalition demands ban on AI-nudification tools over child-safety fears

More than 100 organisations have urged governments to outlaw AI-nudification tools after a surge in non-consensual digital images.

Groups such as Amnesty International, the European Commission, and Interpol argue that the technology now fuels harmful practices that undermine human dignity and child safety. Their concerns intensified after the Grok nudification scandal, where users created sexualised images from ordinary photographs.

Campaigners warn that the tools often target women and children instead of staying within any claimed adult-only environment. Millions of manipulated images have circulated across social platforms, with many linked to blackmail, coercion and child sexual abuse material.

Experts say the trauma caused by these AI images is no less serious because the abuse occurs online.

Organisations within the coalition maintain that tech companies already possess the ability to detect and block such material but have failed to apply essential safeguards.

They want developers and platforms to be held accountable and believe that strict prohibitions are now necessary to prevent further exploitation. Advocates argue that meaningful action is overdue and that protection of users must take precedence over commercial interests.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

South Korea confirms scale of Coupang data breach

The South Korean government has confirmed that 33.67 million user accounts were exposed in a major data breach at Coupang in South Korea. The findings were released by the Ministry of Science and ICT in Seoul.

Investigators in South Korea said names and email addresses were leaked, while delivery lists containing addresses and phone numbers were accessed 148 million times. Officials warned that the impact in South Korea could extend beyond the headline account figure.

Authorities in South Korea identified a former employee as the attacker, alleging misuse of authentication signing keys. The probe concluded that weaknesses in internal controls at Coupang enabled the breach in South Korea.

The ministry in South Korea criticised delayed reporting and plans to impose a fine on Coupang. The company disputed aspects of the findings but said 33.7 million accounts were involved in South Korea.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Fukushima rebuilds as technology hub

Fukushima is repositioning itself as a technology and innovation hub, more than a decade after the 2011 earthquake, tsunami and nuclear disaster in Japan. The Fukushima Innovation Coast Framework aims to revitalise the coastal Hamadori region of Fukushima Prefecture.

At the centre of the push in Fukushima is the Fukushima Institute for Research, Education and Innovation, which plans a major research complex in Namie. The site in Fukushima will focus on robotics, energy, agriculture and radiation science, drawing researchers from across Japan and overseas.

Fukushima already hosts the Fukushima Robot Test Field and the Fukushima Hydrogen Energy Research Field. Projects in Fukushima include hydrogen production from solar power and large-scale robotics and drone testing.

Officials in Fukushima say the strategy combines clean energy, sustainable materials and advanced research to create jobs and attract families back to Japan’s northeast. Fukushima is positioning itself as a global case study in post-disaster recovery through technology.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

EU launches cyberbullying action plan to protect children online

The European Commission has launched an Action Plan Against Cyberbullying aimed at protecting the mental health and well-being of children and teenagers online across the EU. The initiative focuses on reporting access, national coordination, and prevention.

A central element is the development of an EU-wide reporting app that would allow victims to report cyberbullying, receive support, and safely store evidence. The Commission will provide a blueprint for Member States to adapt and link to national helplines.

To ensure consistent protection, Member States are encouraged to adopt a shared understanding of cyberbullying and develop national action plans. This would support comparable data collection and a more coordinated EU response.

The Action Plan builds on existing legislation, including the Digital Services Act, the Audiovisual Media Services Directive, and the AI Act. Updated guidelines will strengthen platform obligations and address AI-enabled forms of abuse.

Prevention and education are also prioritised through expanded resources for schools and families via Safer Internet Centres and the Better Internet for Kids platform. The Commission will implement the plan with Member States, industry, civil society, and children.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Custom AI bots support student negotiating skills

In Cambridge, instructors at MIT and the Harvard Negotiation Project are using AI negotiation bots to enhance classroom simulations. The tools are designed to prompt reflection rather than offer fixed answers.

Students taking part in a multiparty exercise called Harborco engage with preparation, back-table and debriefing bots. The system helps them analyse stakeholder interests and test strategies before and after live negotiations.

Back-table bots simulate unseen political or organisational actors who often influence real-world negotiations. Students can safely explore trade-offs and persuasion tactics in a protected digital setting.

According to reported course findings, most participants said the AI bots improved preparation and sharpened their understanding of opposing interests. Instructors in Cambridge stress that AI supports, rather than replaces, human teaching and peer learning.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

EU reopens debate on social media age restrictions for children

The European Union is revisiting the idea of an EU-wide social media age restriction as several member states move ahead with national measures to protect children online. Spain, France, and Denmark are among the countries considering the enforcement of age limits for access to social platforms.

The issue was raised in the European Commission’s new action plan against cyberbullying, published on Tuesday. The plan confirms that a panel of child protection experts will advise the Commission by the summer on possible EU-wide age restrictions for social media use.

Commission President Ursula von der Leyen announced the creation of an expert panel last September, although its launch was delayed until early 2026. The panel will assess options for a coordinated European approach, including potential legislation and awareness-raising measures for parents.

The document notes that diverging national rules could lead to uneven protection for children across the bloc. A harmonised EU framework, the Commission argues, would help ensure consistent safeguards and reduce fragmentation in how platforms apply age restrictions.

So far, the Commission has relied on non-binding guidance under the Digital Services Act to encourage platforms such as TikTok, Instagram, and Snap to protect minors. Increasing pressure from member states pursuing national bans may now prompt a shift towards more formal EU-level regulation.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

AI and human love in the digital age debate

AI is increasingly entering intimate areas of human life, including romance and emotional companionship. AI chatbots are now widely used as digital companions, raising broader questions about emotional authenticity and human-machine relationships.

Millions of people use AI companion apps, and studies suggest that a significant share of them describe their relationship with a chatbot as romantic. While users may experience genuine emotions, experts stress that current AI systems do not feel love but generate responses based on patterns in data.

Researchers explain that large language models can simulate empathy and emotional understanding, yet they lack consciousness and subjective experience. Their outputs are designed to imitate human interaction rather than reflect genuine emotion.

Scientific research describes love as deeply rooted in biology. Hormones such as dopamine and oxytocin, along with specific brain regions, shape attraction, attachment, and emotional bonding. These processes are embodied and chemical, which machines do not possess.

Some scholars argue that future AI systems could replicate certain cognitive aspects of attachment, such as loyalty or repeated engagement. However, most agree that replicating human love would likely require consciousness, which remains poorly understood and technically unresolved.

Debate continues over whether conscious AI is theoretically possible. While some researchers believe advanced architectures or neuromorphic computing could move in that direction, no existing system meets the established criteria for consciousness.

In practice, human-AI romantic relationships remain asymmetrical. Chatbots are designed to engage, agree, and provide comfort, which can create dependency or unrealistic expectations about real-world relationships.

Experts therefore emphasise transparency and AI literacy, stressing that users should understand AI companions simulate emotion and do not possess feelings, intentions, or awareness; while these systems can imitate expressions of love, they do not experience it, and the emotional reality remains human even when the interaction is digital.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

eSafety escalates scrutiny of Roblox safety measures

Australia’s online safety regulator has notified Roblox of plans to directly test how the platform has implemented a set of child safety commitments agreed last year, amid growing concerns over online grooming and sexual exploitation.

In September last year, Roblox made nine commitments following months of engagement with eSafety, aimed at supporting compliance with obligations under the Online Safety Act and strengthening protections for children in Australia.

Measures included making under-16s’ accounts private by default, restricting contact between adults and minors without parental consent, disabling chat features until age estimation is complete, and extending parental controls and voice chat restrictions for younger users.

Roblox told eSafety at the end of 2025 that it had delivered all agreed commitments, after which the regulator continued monitoring implementation. eSafety Commissioner Julie Inman Grant said serious concerns remain over reports of child exploitation and harmful material on the platform.

Direct testing will now examine how the measures work in practice, with support from the Australian Government. Enforcement action may follow, including penalties of up to $49.5 million, alongside checks against new age-restricted content rules from 9 March.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Cloudflare launches Moltworker platform after AI assistant success

The viral success of Moltbot has prompted Cloudflare to launch a dedicated platform for running the popular AI assistant. The move underscores how the networking company is positioning itself at the centre of the emerging AI agent ecosystem.

Moltbot, an open-source AI personal assistant built on Anthropic’s Claude model, became a viral sensation last month and demonstrated the effectiveness of Cloudflare’s edge infrastructure for running autonomous agents.

The assistant’s rapid adoption validated CEO Matthew Prince’s assertion that AI agents represent a ‘fundamental re-platforming’ of the internet. In response, Cloudflare quickly released Moltworker, a platform specifically designed for securely operating Moltbot and similar AI agents.

Prince described the dynamic as creating a ‘virtuous flywheel,’ with AI agents serving as the new users of the internet, whilst Cloudflare provides the platform they run on and the network they pass through.

Industry analysts have highlighted why Cloudflare’s infrastructure is well-suited to the era of agentic computing. RBC Capital Markets noted that AI agents require low-latency, secure inferencing at the network’s edge- precisely what Cloudflare’s Workers platform delivers.

The continued proliferation of AI agents is expected to drive ongoing demand for these capabilities.

Prince, who co-founded the company, revealed that Cloudflare ended 2025 with 4.5 million active human developers on its platform, providing a substantial foundation for the next wave of AI-driven applications and agents built on the company’s infrastructure.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot