Teens launch High Court bid to stop Australia’s under-16 social media ban

Two teenagers in Australia have taken the federal government to the High Court in an effort to stop the country’s under-16 social media ban, which is due to begin on 10 December. The case was filed by the Digital Freedom Project with two 15-year-olds, Noah Jones and Macy Neyland, listed as plaintiffs. The group says the law strips young people of their implied constitutional right to political communication.

The ban will lead to the deactivation of more than one million accounts held by users under 16 across platforms such as YouTube, TikTok, Snapchat, Twitch, Facebook and Instagram. The Digital Freedom Project argues that removing young people from these platforms blocks them from engaging in public debate. Neyland said the rules silence teens who want to share their views on issues that affect them.

The Digital Freedom Project’s president, John Ruddick, is a Libertarian Party politician in New South Wales. After the lawsuit became public, Communications Minister Anika Wells told Parliament the government would not shift its position in the face of legal threats. She said the government’s priority is supporting parents rather than platform operators.

The law, passed in November 2024, is supported by most Australians according to polling. The government says research links heavy social media use among young teens to bullying, misinformation and harmful body-image content.

Companies that fail to comply with the ban risk penalties of up to A$49.5 million. Lawmakers and tech firms abroad are watching how the rollout unfolds, as Australia’s approach is among the toughest efforts globally to restrict minors’ access to social platforms.

Would you like to learn more aboutAI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Teenagers still face harmful content despite new protections

In the UK and other countries, teenagers continue to encounter harmful social media content, including posts about bullying, suicide and weapons, despite the Online Safety Act coming into effect in July.

A BBC investigation using test profiles revealed that some platforms continue to expose young users to concerning material, particularly on TikTok and YouTube.

The experiment, conducted with six fictional accounts aged 13 to 15, revealed differences in exposure between boys and girls.

While Instagram showed marked improvement, with no harmful content displayed during the latest test, TikTok users were repeatedly served posts about self-harm and abuse, and one YouTube profile encountered videos featuring weapons and animal harm.

Experts warned that changes will take time and urged parents to monitor their children’s online activity actively. They also recommended open conversations about content, the use of parental controls, and vigilance rather than relying solely on the new regulatory codes.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

YouTube launches likeness detection to protect creators from AI misuse

YouTube has expanded its AI safeguards with a new likeness detection system that identifies AI-generated videos imitating creators’ faces or voices. The tool is now available to eligible members of the YouTube Partner Program after a limited pilot phase.

Creators can review detected videos and request their removal under YouTube’s privacy rules or submit copyright claims.

YouTube said the feature aims to protect users from having their image used to promote products or spread misinformation without consent.

The onboarding process requires identity verification through a short selfie video and photo ID. Creators can opt out at any time, with scanning ending within a day of deactivation.

YouTube has backed recent legislative efforts, such as the NO FAKES Act in the US, which targets deceptive AI replicas. The move highlights growing industry concern over deepfake misuse and the protection of digital identity.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

New YouTube tools provide trusted health advice for teens

YouTube is introducing a new shelf of mental health and wellbeing content designed specifically for teenagers. The feature will provide age-appropriate, evidence-based videos covering topics such as depression, anxiety, ADHD, and eating disorders.

Content is created in collaboration with trusted organisations and creators, including Black Dog Institute, ReachOut Australia, and Dr Syl, to ensure it is both reliable and engaging.

The initiative will initially launch in Australia, with plans to expand to the US, the UK, and Canada. Videos are tailored to teens’ developmental stage, offering practical advice, coping strategies, and medically-informed guidance.

By providing credible information on a familiar platform, YouTube hopes to improve mental health literacy and reduce stigma among young users.

YouTube has implemented teen-specific safeguards for recommendations, content visibility, and advertising eligibility, making it easier for adolescents to explore their interests safely.

The company emphasises that the platform is committed to helping teens access trustworthy resources, while supporting their wellbeing in a digital environment increasingly filled with misinformation.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Google cautions Australia on youth social media ban proposal

The US tech giant, Google (also owner of YouTube), has reiterated its commitment to children’s online safety while cautioning against Australia’s proposed ban on social media use for those under 16.

Speaking before the Senate Environment and Communications References Committee, Google’s Public Policy Senior Manager Rachel Lord said the legislation, though well-intentioned, may be difficult to enforce and could have unintended effects.

Lord highlighted the 23-year presence of Google in Australia, contributing over $53 billion to the economy in 2024, while YouTube’s creative ecosystem added $970 million to GDP and supported more than 16,000 jobs.

She said the company’s investments, including the $1 billion Digital Future Initiative, reflect its long-term commitment to Australia’s digital development and infrastructure.

According to Lord, YouTube already provides age-appropriate products and parental controls designed to help families manage their children’s experiences online.

Requiring children to access YouTube without accounts, she argued, would remove these protections and risk undermining safe access to educational and creative content used widely in classrooms, music, and sport.

She emphasised that YouTube functions primarily as a video streaming platform rather than a social media network, serving as a learning resource for millions of Australian children.

Lord called for legislation that strengthens safety mechanisms instead of restricting access, saying the focus should be on effective safeguards and parental empowerment rather than outright bans.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Google fights to bundle Gemini AI with Maps and YouTube

Google is seeking permission to bundle its Gemini AI application with long-standing services such as YouTube and Maps, even as US regulators press for restrictions to curb its dominance in search.

At a recent court hearing, Google’s lawyer John Schmidtlein told Judge Amit Mehta that tying Gemini to its core apps is vital to delivering a consistent AI experience across its ecosystem.

He insisted the courts should not treat the AI market as a settled domain subject to old rules, and claimed that neither Maps nor YouTube is a monopoly product justifying special constraints.

The government’s position is more cautious. During the hearing, Judge Mehta questioned whether allowing Google to require its AI app to be installed to access Maps or YouTube would give it unfair leverage over competitors, mirroring past practices that regulators found harmful in search and browser markets.

This moment frames a broader tension: how antitrust frameworks will adapt (or not) when dominant platforms seek to integrate generative AI across many services. The outcome could shape the future of bundling practices and interoperability in AI ecosystems.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

YouTube settles Donald Trump lawsuit over account suspension for $24.5 million

YouTube has agreed to a $24.5 million settlement to resolve a lawsuit filed by President Donald Trump, stemming from the platform’s decision to suspend his account after the 6 January 2021 Capitol riot.

The lawsuit was part of a broader legal push by Trump against major tech companies over what he calls politically motivated censorship.

As part of the deal, YouTube will donate $22 million to the Trust for the National Mall on Trump’s behalf, funding a new $200 million White House ballroom project. Another $2.5 million will go to co-plaintiffs, including the American Conservative Union and author Naomi Wolf.

The settlement includes no admission of wrongdoing by YouTube and was intended to avoid further legal costs. The move follows similar multimillion-dollar settlements by Meta and X, which also suspended Trump’s accounts post-January 6.

Critics argue the settlement signals a retreat from consistent content moderation. Media scholar Timothy Koskie warned it sets a troubling precedent for global digital governance and selective enforcement.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Google tests AI hosts for YouTube Music

Google is testing AI-generated hosts for YouTube Music through its new YouTube Labs programme. The AI hosts will appear while users listen to mixes and radio stations, providing commentary, fan trivia, and stories to enrich the listening experience.

The feature is designed to resemble a radio jockey but relies on AI, so there is a risk of occasional inaccuracies.

YouTube Labs, similar to Google Labs, allows the company to trial new AI features and gather user feedback before wider release. The AI hosts are currently available to a limited group of US testers, who can sign up via YouTube Labs and snooze commentary for an hour or all day.

The rollout follows Google’s Audio Overviews in NotebookLM, which turns research papers and documents into podcast-style summaries. Past AI experiments on YouTube, such as automatic dubbing, faced criticism as viewers had limited control over translations.

The AI hosts experiment shows Google’s push to integrate AI across its apps, enhancing engagement while monitoring feedback before wider rollout.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

YouTube rolls back rules on Covid-19 and 2020 election misinformation

Google’s YouTube has announced it will reinstate accounts previously banned for repeatedly posting misinformation about Covid-19 and the 2020 US presidential election. The decision marks another rollback of moderation rules that once targeted health and political falsehoods.

The platform said the move reflects a broader commitment to free expression and follows similar changes at Meta and Elon Musk’s X.

YouTube had already scrapped policies barring repeat claims about Covid-19 and election outcomes, rules that had led to actions against figures such as Robert F. Kennedy Jr.’s Children’s Health Defense Fund and Senator Ron Johnson.

An announcement that came in a letter to House Judiciary Committee Chair Jim Jordan, amid a Republican-led investigation into whether the Biden administration pressured tech firms to remove certain content.

YouTube claimed the White House created a political climate aimed at shaping its moderation, though it insisted its policies were enforced independently.

The company said that US conservative creators have a significant role in civic discourse and will be allowed to return under the revised rules. The move highlights Silicon Valley’s broader trend of loosening restrictions on speech, especially under pressure from right-leaning critics.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Gemini brings conversational AI to Google TV

Google has launched Gemini for TV, bringing conversational AI to the living room. The update builds on Google TV and Google Assistant, letting viewers chat naturally with their screens to discover shows, plan trips, or even tackle homework questions.

Instead of scrolling endlessly, users can ask Gemini to find a film everyone will enjoy or recap last season’s drama. The AI can handle vague requests, like finding ‘that new hospital drama,’ and provide reviews before you press play.

Gemini also turns the TV into an interactive learning tool. From explaining why volcanoes erupt to guiding kids through projects, it offers helpful answers with supporting YouTube videos for hands-on exploration.

Beyond schoolwork, Gemini can help plan meals, teach new skills like guitar, or brainstorm family trips, all through conversational prompts. Such features make the TV a hub for entertainment, education, and inspiration.

Gemini is now available on the TCL QM9K series, with rollout to additional Google TV devices planned for later this year. Google says additional features are coming soon, making TVs more capable and personalised.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!