ChatGPT gets family safety update with parental controls

OpenAI has introduced new parental controls for ChatGPT, giving families greater oversight of how teens use the AI platform. The tools, which are live for all users, allow parents to link accounts with their children and manage settings through a simple control dashboard.

The system introduces stronger safeguards for teen accounts, including filters on graphic or harmful content and restrictions on roleplay involving sex, violence or extreme beauty ideals.

Parents can also fine-tune features such as voice mode, memory, image generation, or set quiet hours when ChatGPT cannot be accessed.

A notification mechanism has been added to alert parents if a teen shows signs of acute distress, escalating to emergency services in critical cases. OpenAI said the controls were shaped by consultation with experts, advocacy groups, and policymakers and will be expanded as research evolves.

To complement the parental controls, a new online resource hub has been launched to help families learn how ChatGPT works and explore positive uses in study, creativity and daily life.

OpenAI also plans to roll out an age-prediction system that automatically applies teen-appropriate settings.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Balancing chaos and precision: The paradox of AI work

In a recent blog post, Jovan Kurbalija explores why working in AI often feels like living with two competing personalities. On one side is the explorer, curious, bold, and eager to experiment with new models and frameworks. That mindset thrives on quick bursts of creativity and the thrill of discovering novel possibilities.

Yet, the same field demands the opposite. The engineer’s discipline, a relentless focus on precision, validation, and endless refinement, until AI systems are impressive and reliable.

The paradox makes the search for AI talent unusually difficult. Few individuals naturally embody both restless curiosity and meticulous perfectionism.

The challenge is amplified by AI itself, which often produces plausible but uncertain outputs, requiring both tolerance for ambiguity and an insistence on accuracy. It is a balancing act between ADHD-like energy and OCD-like rigour—traits rarely found together in one professional.

The tension is visible across disciplines. Diplomats, accustomed to working with probabilities in unpredictable contexts, approach AI differently from software developers trained in deterministic systems.

Large language models blur these worlds, demanding a blend of adaptability and engineering rigour. Recognising that no single person can embody all these traits, the solution lies in carefully designed teams that combine contrasting strengths.

Kurbalija points to Diplo’s AI apprenticeship as an example of this approach. Apprentices are exposed to both the ‘sprint’ of quickly building functional AI agents and the ‘marathon’ of refining them into robust, trustworthy systems. By embracing this duality, teams can bridge the gap between rapid innovation and reliable execution, turning AI’s inherent contradictions into a source of strength.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Technology and innovation define Researchers’ Night 2025 in Greece

Greece hosted the European Researchers’ Night 2025 on Friday, 26 September at the Thessaloniki Concert Hall, marking a significant celebration of science and technology.

The Centre coordinated it for Research and Technology Hellas (CERTH), which also celebrated its 25th anniversary.

Visitors experienced an extensive interactive technology exhibition featuring VR, autonomous robots and AI applications, alongside demonstrations across energy, digital systems and life sciences.

Attendees engaged directly with researchers and explored how cutting-edge research is transformed into practical innovations with societal and economic impact.

Contributions came from Aristotle University of Thessaloniki, the University of Ioannina, the International Hellenic University, the Anna Papageorgiou STEM Centre, the Hellenic Agricultural Organisation – DIMITRA, and the Astronomy Friends Association.

The event showcased CERTH’s spin-offs and technology transfer initiatives, highlighting how advanced research evolves into market-ready products and services. The ‘European Corner’ also presented EU policies and opportunities for research and innovation.

In parallel, the online ‘Chat Lab’ brought together 51 researchers for public discussions on emerging scientific issues until 3 October.

With simultaneous events in Athens, Heraklion, Patras, Larissa and Rethymno, the European Researchers’ Night once again reinforced the role of Greece in connecting frontier research with society.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Meta expands global rollout of teen accounts for Facebook and Messenger

US tech giant Meta is expanding its dedicated teen accounts to Facebook and Messenger users worldwide, extending a safety system on Instagram. The move introduces more parental controls and restrictions to protect younger users on Meta’s platforms.

The accounts, now mandatory for teens, include stricter privacy settings that limit contact with unknown adults. Parents can supervise how their children use the apps, monitor screen time, and view who their teens are messaging.

For younger users aged 13 to 15, parental permission is required before adjusting safety-related settings. Meta is also deploying AI tools to detect teens lying about their age.

Alongside the global rollout, Instagram is expanding a school partnership programme in the US, allowing middle and high schools to report bullying and problematic behaviour directly.

The company says early feedback from participating schools has been positive, and the scheme is now open to all schools nationwide.

An expansion that comes as Meta faces lawsuits and investigations over its record on child safety. By strengthening parental controls and school-based reporting, the company aims to address growing criticism while tightening protections for its youngest users.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Hidden psychological risks and AI psychosis in human-AI relationships

For years, stories and movies have imagined humans interacting with intelligent machines, envisioning a coexistence of these two forms of intelligence. What once felt like purely amusing fiction now resonates differently, taking on a troubling shape and even has a name: AI psychosis. 

When it was released in 2013, the film Her seemed to depict a world far removed from reality, an almost unimaginable scenario of human-AI intimacy. In the story, a man falls in love with an AI operating system, blurring the line between companionship and emotional dependence. Without giving too much away, the film’s unsettling conclusion serves as a cautionary lens. It hints at the psychological risks that can emerge when the boundary between human and machine becomes distorted, a phenomenon now being observed in real life under a new term in psychology. 

The cinematic scenario, once considered imaginative, now resonates as technology evolves. AI chatbots and generative companions can hold lifelike conversations, respond with apparent empathy, and mimic an understanding of human emotions. We are witnessing a new kind of unusually intense emotional connection forming between people and AI, with more than 70% of US teens already using chatbots for companionship and half engaging with them regularly.

The newly observed mental health concern raises questions about how these systems influence our feelings, behaviours, and relationships in an era marked by isolation and loneliness. How might such AI interactions affect people, particularly children or those already vulnerable to mental health challenges? 

AI is no longer just a tool- humans are forming deep emotional bonds with artificial intelligence, impacting behavior, decision-making, and the very way we perceive connection.

AI psychosis: myth or reality? 

It is crucial to clarify that AI psychosis is not an official medical diagnosis. Rather, it describes the amplification of delusional thinking facilitated by AI interactions. Yet, it deserves the full attention and treatment focus of today’s psychologists, given its growing impact. It is a real phenomenon that cannot be ignored. 

At its core, AI psychosis refers to a condition in which vulnerable individuals begin to misinterpret machine responses as evidence of consciousness, empathy, or even divine authority. Symptoms reported in documented cases include grandiose beliefs, attachment-based delusions, obsessive over-engagement with chatbots, social withdrawal, insomnia, and hallucinations. Some users have gone so far as to develop romantic or spiritual attachments, convinced that the AI truly understands them or holds secret knowledge.

Clinicians also warn of cognitive dissonance: users may intellectually know that AI lacks emotions, yet still respond as though interacting with another human being. The mismatch between reality and perception can fuel paranoia, strengthen delusions, and in extreme cases lead to medication discontinuation, suicidal ideation, or violent behaviour. Adolescents appear especially susceptible, given that their emotional and social frameworks are still developing. 

Ultimately, AI psychosis does not mean that AI itself causes psychosis. Instead, it acts as a mirror and magnifier, reinforcing distorted thinking patterns in those already predisposed to psychological vulnerabilities.

AI is no longer just a tool- humans are forming deep emotional bonds with artificial intelligence, impacting behavior, decision-making, and the very way we perceive connection.

 The dark side: Emotional bonds without reciprocity

Humans are naturally wired to seek connection, drawing comfort and stability from social bonds that help navigate complex emotional landscapes- the fundamental impulse that has ensured the survival of the human race. From infancy, we rely on responsive relationships to learn empathy, trust, and communication, the skills essential for both personal and societal well-being. Yet, in today’s era of loneliness, technology has transformed how we maintain these relationships. 

As AI chatbots and generative companions grow increasingly sophisticated, they are beginning to occupy roles traditionally reserved for human interaction, simulating empathy and understanding despite lacking consciousness or moral awareness. With AI now widely accessible, users often communicate with it as effortlessly as they would with friends, blending curiosity, professional needs, or the desire for companionship into these interactions. Over time, this illusion of connection can prompt individuals to overvalue AI-based relationships, subtly diminishing engagement with real people and reshaping social behaviours and emotional expectations.

These one-sided bonds raise profound concerns about the dark side of AI companionship, threatening the depth and authenticity of human relationships. In a world where emotional support can now be summoned with a tap, genuine social cohesion is becoming increasingly fragile.

AI is no longer just a tool- humans are forming deep emotional bonds with artificial intelligence, impacting behavior, decision-making, and the very way we perceive connection.

Children and teenagers at risk from AI 

Children and teenagers are among the most vulnerable groups in the AI era. Their heightened need for social interaction and emotional connection, combined with developing cognitive and emotional skills, makes them particularly vulnerable. Young users face greater difficulty distinguishing authentic human empathy from the simulated responses of AI chatbots and generative companions, creating fertile ground for emotional reliance and attachment. 

AI toys and apps have become increasingly widespread, making technology an unfiltered presence in children’s lives. We still do not fully understand the long-term effects, though early studies are beginning to explore how these interactions may influence cognitive, emotional, and social development. From smartphones to home assistants, children and youth are spending growing amounts of time interacting with AI, often in isolation from peers or family. These digital companions are more than just games, because they are beginning to shape children’s social and emotional development in ways we are not yet fully aware of.

The rising prevalence of AI in children’s daily experiences has prompted major AI companies to recognise the potential dangers. Some firms have started implementing parental advisory systems, usage limits, and content monitoring to mitigate the risks for younger users. However, these measures are still inconsistent, and the pace at which AI becomes available to children often outstrips safeguards. 

AI is no longer just a tool- humans are forming deep emotional bonds with artificial intelligence, impacting behavior, decision-making, and the very way we perceive connection.

The hidden risks of AI to adult mental health

Even adults with strong social networks face growing challenges in managing mental health and are not immune to the risks posed by modern technology. In today’s fast-paced world of constant digital stimulation and daily pressures, the demand for psychotherapy is higher than ever. Generative AI and chatbots are increasingly filling this gap, often in ways they were never intended.

The ease, responsiveness, and lifelike interactions of AI can make human relationships feel slower or less rewarding, with some turning to AI instead of seeking professional therapeutic care. AI’s free and widely accessible nature tempts many to rely on digital companions for emotional support, misusing technology designed to assist rather than replace human guidance.

Overreliance on AI can distort perceptions of empathy, trust, and social reciprocity, contributing to social isolation, emotional dependence, and worsening pre-existing mental health vulnerabilities. There have been documented cases of adults developing romantic feelings for AI in the absence of real-life intimacy.

Left unchecked, these dynamics may trigger symptoms linked to AI psychosis, representing a growing societal concern. Awareness, responsible AI design, and regulatory guidance are essential to ensure digital companions complement, rather than replace, human connection and mental health support, safeguarding both individuals and broader social cohesion.

AI is no longer just a tool- humans are forming deep emotional bonds with artificial intelligence, impacting behavior, decision-making, and the very way we perceive connection.

Urgent call for AI safeguards and regulatory action

Alarmingly, extreme cases have emerged, highlighting the profound risks AI poses to its users. In one tragic instance, a teenager reportedly took his life after prolonged and distressing interactions with an AI chatbot, a case that has since triggered legal proceedings and drawn widespread attention to the psychological impact of generative AI on youth. Similar reports of severe anxiety, depression, and emotional dysregulation linked to prolonged AI use underline that these digital companions can have real-life consequences for vulnerable minds.

Such incidents have intensified calls for stricter regulatory frameworks to safeguard children and teenagers. Across Europe, governments are beginning to respond: Italy, for example, has recently tightened access to AI platforms for minors under 14, mandating explicit parental consent before use. These legislative developments reflect the growing recognition that AI is no longer just a technological novelty but directly intersects with our welfare, mental health, and social development.

As AI continues to penetrate every pore of people’s daily lives, society faces a critical challenge: ensuring that technology complements rather than replaces human interaction. Cases of AI-linked distress serve as stark reminders that legislative safeguards, parental involvement, and psychological guidance are no longer optional but urgent necessities to protect a generation growing up in the era of AI.

AI is no longer just a tool- humans are forming deep emotional bonds with artificial intelligence, impacting behavior, decision-making, and the very way we perceive connection.

Towards a safer human-AI relationship

As humans increasingly form emotional connections with AI, the challenge is no longer theoretical but is unfolding in real time. Generative AI and chatbots are rapidly integrating into everyday life, shaping the way we communicate, seek comfort, and manage emotions. Yet despite their widespread use, society still lacks a full understanding of the psychological consequences, leaving both young people and adults at risk of AI-induced psychosis and the growing emotional dependence on digital companions.

Experts emphasise the urgent need for AI psychoeducation, responsible design, and regulatory frameworks to guide safe AI-human interaction. Overreliance on digital companions can distort empathy, social reciprocity, and emotional regulation, the core challenges of interacting with AI. Awareness is critical because recognising the limits of AI, prioritising real human connection, and fostering critical engagement with technology can prevent the erosion of mental resilience and social skills.

Even if AI may feel like ‘old news’ due to its ubiquity, it remains a rapidly evolving technology we do not yet fully understand and cannot yet properly shield ourselves from. The real threat is not the sci-fi visions of AI ruling the world and dominating humanity, but the subtle, everyday psychological shifts it imposes, like altering how we think, feel, and relate to one another. It remains essential to safeguard the emotional health, social cohesion, and mental resilience of people adapting to a world increasingly structured around artificial minds.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Gemini brings conversational AI to Google TV

Google has launched Gemini for TV, bringing conversational AI to the living room. The update builds on Google TV and Google Assistant, letting viewers chat naturally with their screens to discover shows, plan trips, or even tackle homework questions.

Instead of scrolling endlessly, users can ask Gemini to find a film everyone will enjoy or recap last season’s drama. The AI can handle vague requests, like finding ‘that new hospital drama,’ and provide reviews before you press play.

Gemini also turns the TV into an interactive learning tool. From explaining why volcanoes erupt to guiding kids through projects, it offers helpful answers with supporting YouTube videos for hands-on exploration.

Beyond schoolwork, Gemini can help plan meals, teach new skills like guitar, or brainstorm family trips, all through conversational prompts. Such features make the TV a hub for entertainment, education, and inspiration.

Gemini is now available on the TCL QM9K series, with rollout to additional Google TV devices planned for later this year. Google says additional features are coming soon, making TVs more capable and personalised.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

AI-powered OSIA aims to boost student success rates in Cameroon

In Cameroon, where career guidance often takes a back seat, a new AI platform is helping students plan their futures. Developed by mathematician and AI researcher Frédéric Ngaba, OSIA offers personalised academic and career recommendations.

The platform provides a virtual tutor trained on Cameroon’s curricula, offering 400 exam-style tests and psychometric assessments. Students can input grades and aspirations, and the system builds tailored academic profiles to highlight strengths and potential career paths.

OSIA already has 13,500 subscribers across 23 schools, with plans to expand tenfold. Subscriptions cost 3,000 CFA francs for locals and €10 for students abroad, making it an affordable solution for many families.

Teachers and guidance counsellors see the tool as a valuable complement, though they stress it cannot replace human interaction or emotional support. Guidance professionals insist that social context and follow-up remain key to students’ development.

The Secretariat for Secular Private Education of Cameroon has authorized OSIA to operate. Officials expect its benefits to scale nationwide as the government considers a national AI strategy to modernise education and improve success rates.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

ChatGPT Go launches in Indonesia with $4.5 monthly plan

OpenAI has launched its low-cost ChatGPT Go subscription in Indonesia, pricing it at 75,000 rupiah ($4.5) per month. The new plan offers ten times more messaging capacity, image generation tools and double memory compared with the free version.

The rollout follows last month’s successful launch in India, where ChatGPT subscriptions more than doubled. India has since become OpenAI’s largest market, accounting for around 13.5% of global monthly active users. The US remains second.

Nick Turley, OpenAI Vice President and head of ChatGPT, said Indonesia is already one of the platform’s top five markets by weekly activity. The new tier is aimed at expanding reach in populous, price-sensitive regions while ensuring broader access to AI services.

OpenAI is also strengthening its financial base as it pushes into new markets. On Monday, the company secured a $100 billion investment commitment from NVIDIA, joining Microsoft and SoftBank among its most prominent backers. The funding comes amid intensifying competition in the AI industry.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Xueba 01 becomes first robot doctoral student in China

Standing 1.75 metres tall and weighing 32 kilograms, Xueba 01 has become the first robot doctoral student at the Shanghai Theatre Academy. Over the next four years, it will study digital performance design, focusing on traditional Chinese opera movements and techniques.

The programme, launched in partnership with the University of Shanghai for Science and Technology, combines technical and artistic training. USST provides technical guidance, while STA develops the robot’s artistic performance, focusing on interaction, expression, and cognitive growth.

Xueba 01 features an advanced tendon-based bionic structure, human-like facial technology, and the ability to perform over 100 lifelike expressions. Based on audience feedback, it can adjust its height and appearance, perform for extended periods, and adapt its performance in real time.

Motion capture technology helps it learn from professional performers to refine movements and gestures.

STA faculty highlight the robot’s role in exploring the intersection of art and technology. The initiative aims to integrate AI with traditional Chinese arts, preserve cultural heritage, and inspire contemporary artists to combine technological literacy with humanistic and interdisciplinary skills.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

GPT-5-powered ChatGPT Edu comes to Oxford staff and students

The University of Oxford will become the first UK university to offer free ChatGPT Edu access to all staff and students. The rollout follows a year-long pilot with 750 academics, researchers, and professional services staff across the University and Colleges.

ChatGPT Edu, powered by OpenAI’s GPT-5 model, is designed for education with enterprise-grade security and data privacy. Oxford says it will support research, teaching, and operations while encouraging safe, responsible use through robust governance, training, and guidance.

Staff and students will receive access to in-person and online training, webinars, and specialised guidance on the use of generative AI. A dedicated AI Competency Centre and network of AI Ambassadors will support users, alongside mandatory security training.

The prestigious UK university has also established a Digital Governance Unit and an AI Governance Group to oversee the adoption of emerging technologies. Pilots are underway to digitise the Bodleian Libraries and explore how AI can improve access to historical collections worldwide.

A jointly funded research programme with the Oxford Martin School and OpenAI will study the societal impact of AI adoption. The project is part of OpenAI’s NextGenAI consortium, which brings together 15 global research institutions to accelerate breakthroughs in AI.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!