AI-powered study rooms are revolutionising online education in China by offering personalised, tech-driven learning experiences. These spaces cater to students aged 8 to 18, using advanced software to provide interactive lessons and real-time feedback. The AI systems analyse mistakes, adjust course materials, and generate detailed progress reports for parents, who can track their child’s improvement remotely. By leveraging technology, these study rooms aim to make education more engaging and tailored to individual learning needs.
These AI rooms are marketed as self-study spaces rather than traditional tutoring centres, allowing them to navigate China’s strict private tutoring regulations by framing their services as facility rentals or membership plans. This creative positioning allows them to operate within a regulatory grey area, avoiding restrictions on off-campus tutoring for students in grades one through nine. Membership fees range from 1,000 to 3,000 yuan monthly, making them a more affordable long-term alternative to expensive one-on-one tutoring sessions.
Despite their growing popularity, education experts remain sceptical of their educational value. Critics argue that many of these systems lack proper AI functionality, relying instead on preloaded prompts and automated responses. Furthermore, there are concerns that their heavy emphasis on drilling questions to improve test scores may neglect critical thinking and deeper comprehension. However, proponents believe these AI-powered study rooms represent an essential step toward integrating technology into education and expanding access to personalised learning.
Lina Khan, a prominent advocate of strong antitrust enforcement, has announced her resignation as chair of the US Federal Trade Commission (FTC) in a memo to staff. Her departure, set to occur in the coming weeks, marks the end of a tenure that challenged numerous corporate mergers and pushed for greater accountability among powerful companies.
During her leadership, Khan spearheaded high-profile lawsuits against Amazon, launched investigations into Microsoft, and blocked major deals, including Kroger’s planned $25 billion acquisition of Albertsons. Her efforts often focused on protecting consumers and workers from potential harms posed by dominant corporations.
Khan, the youngest person to lead the FTC, first gained recognition in 2017 for her work criticising Amazon’s market practices. She argued that tech giants exploited outdated antitrust laws, allowing them to sidestep scrutiny. Her aggressive approach divided opinion, with courts striking down some of her policies, including a proposed ban on noncompete clauses.
Following Khan’s exit, the FTC faces a temporary deadlock with two Republican and two Democratic commissioners. Republican Andrew Ferguson has assumed the role of chair, and a Republican majority is expected once the Senate approves Mark Meador, a pro-enforcement nominee, to complete the five-member commission.
Younger members of Generation Z are turning to ChatGPT for schoolwork, with a new Pew Research Centre survey revealing that 26% of US teens aged 13 to 17 have used the AI-powered chatbot for homework. This figure has doubled since 2023, highlighting the growing reliance on AI tools in education. The survey also showed mixed views among teens about its use, with 54% finding it acceptable for research, while smaller proportions endorsed its use for solving maths problems (29%) or writing essays (18%).
Experts have raised concerns about the limitations of ChatGPT in academic contexts. Studies indicate the chatbot struggles with accuracy in maths and certain subject areas, such as social mobility and African geopolitics. Research also shows varying impacts on learning outcomes, with Turkish students who used ChatGPT performing worse on a maths test than peers who didn’t. German students, while finding research materials more easily, synthesised information less effectively when using the tool.
Educators remain cautious about integrating AI into classrooms. A quarter of public K-12 teachers surveyed by Pew believed AI tools like ChatGPT caused more harm than good in education. Another study by the Rand Corporation found only 18% of K-12 teachers actively use AI in their teaching practices. The disparities in effectiveness and the tool’s limitations underscore the need for careful consideration of its role in learning environments.
The United States Federal Trade Commission (FTC) has referred a complaint about Snap Inc’s AI-powered chatbot, My AI, to the Department of Justice (DOJ) for further investigation. The FTC alleges the chatbot caused harm to young users, though specific details about the alleged harm remain undisclosed.
Snap Inc defended its chatbot, asserting that My AI operates under rigorous safety and privacy measures and criticised the FTC for lacking concrete evidence to support its claims. Despite the company’s reassurances, the FTC stated it had uncovered indications of potential legal violations.
The announcement impacted Snap’s stock performance, with shares dropping by 5.2% to close at $11.22 on Thursday. The US FTC noted that publicising the complaint’s transfer to the DOJ was in the public interest, underscoring the gravity of the allegations.
Apple has halted AI-powered notification summaries for news and entertainment apps after backlash over misleading news alerts. A BBC complaint followed a summary that misrepresented an article about a murder case involving UnitedHealthcare’s CEO.
The latest developer previews for iOS 18.3, iPadOS 18.3, and macOS Sequoia 15.3 disable notification summaries for such apps, with Apple planning to reintroduce them after improvements. Notification summaries will now appear in italics to help users distinguish them from standard alerts.
Users will also gain the ability to turn off notification summaries for individual apps directly from the Lock Screen. Apple will notify users in the Settings app that the feature remains in beta and may contain errors.
A public beta is expected next week, but the general release date for iOS 18.3 remains unclear. Apple had already announced plans to clarify that summary texts are generated by Apple Intelligence.
The US Supreme Court will hear a challenge on Wednesday regarding a Texas law that mandates adult websites verify the age of users before granting access to potentially harmful material. The law, which is part of a broader trend across Republican-led states, requires users to submit personal information proving they are at least 18 years old to access pornographic content. The case raises significant First Amendment concerns, as adult entertainment industry groups argue that the law unlawfully restricts free speech and exposes users to risks such as identity theft and data breaches.
The challengers, including the American Civil Liberties Union and the Free Speech Coalition, contend that alternative methods like content-filtering software could better protect minors without infringing on adults’ rights to access non-obscene material. Texas, however, defends the law, citing concerns over the ease with which minors can access explicit content online.
This case is significant because it will test the balance between state efforts to protect minors from explicit content and the constitutional rights of adults to access protected expression. If the Supreme Court upholds the law, it could set a precedent for similar age-verification measures across the US.
Indonesia plans to implement interim guidelines to protect children on social media as it works toward creating a law to establish a minimum age for users, a senior communications ministry official announced on Wednesday. The move follows discussions between Communications Minister Meutya Hafid and President Prabowo Subianto, aiming to address concerns about online safety for children.
The proposed law will mirror recent regulations in Australia, which banned children under 16 from accessing social media platforms like Instagram, Facebook, and TikTok, penalising tech companies that fail to comply. In the meantime, Indonesia will issue regulations requiring platforms to follow child protection guidelines, focusing on shielding children from harmful content while still allowing access to some degree.
Public opinion on the initiative is divided. While parents like Nurmayanti support stricter controls to reduce exposure to harmful material, human rights advocates, including Anis Hidayah, urge caution to ensure children’s access to information is not unduly restricted. A recent survey revealed nearly half of Indonesian children under 12 use the internet, with many accessing social media platforms such as Facebook, Instagram, and TikTok.
This regulatory push reflects Indonesia’s broader efforts to balance digital innovation with safeguarding younger users in its rapidly growing online landscape
Indonesia is preparing to introduce regulations setting a minimum age for social media users, aiming to shield children from potential online risks, according to Communications Minister Meutya Hafid. The announcement follows Australia’s recent ban on social media access for children under 16, which imposes penalties on platforms like Meta’s Facebook and Instagram, as well as TikTok, for non-compliance.
While the specific age limit for Indonesia remains undecided, Minister Hafid stated that President Prabowo Subianto supports the initiative, emphasising the importance of child protection in the digital space. The move highlights concerns about young users’ exposure to inappropriate content and data privacy risks.
Indonesia, with a population of approximately 280 million, has significant internet usage. A recent survey found internet penetration at 79.5%, with nearly half of children under 12 accessing the web, often using platforms like Facebook, Instagram, and TikTok. Among “Gen Z” users aged 12 to 27, internet penetration reached 87%. The proposed regulation reflects growing global efforts to prioritise child safety online.
Ian Russell, father of Molly Russell, has called on the UK government to take stronger action on online safety, warning that delays in regulation are putting children at risk. In a letter to Prime Minister Sir Keir Starmer, he criticised Ofcom’s approach to enforcing the Online Safety Act, describing it as a “disaster.” Russell accused tech firms, including Meta and X, of prioritising profits over safety and moving towards a more dangerous, unregulated online environment.
Campaigners argue that Ofcom’s guidelines contain major loopholes, particularly in addressing harmful content such as live-streamed material that promotes self-harm and suicide. While the government insists that tech companies must act responsibly, the slow progress of new regulations has raised concerns. Ministers acknowledge that additional legislation may be required as AI technology evolves, introducing new risks that could further undermine online safety.
Russell has been a prominent campaigner for stricter online regulations since his daughter’s death in 2017. Despite the Online Safety Act granting Ofcom the power to fine tech firms, critics believe enforcement remains weak. With concerns growing over the effectiveness of current safeguards, pressure is mounting on the government to act decisively and ensure platforms take greater responsibility in protecting children from harmful content.
A new app designed to help children aged seven to twelve manage anxiety through gaming is being launched in Lincolnshire, UK. The app, called Lumi Nova, combines cognitive behavioural therapy (CBT) techniques with personalised quests to gently expose children to their fears in a safe and interactive way.
The digital game has been created by BFB Labs, a social enterprise focused on digital therapy, in collaboration with children, parents, and mental health experts. The app aims to make mental health support more accessible, particularly in rural areas, where traditional services may be harder to reach.
Families in Lincolnshire can download the app for free without needing a prescription or referral. Councillor Patricia Bradwell from Lincolnshire County Council highlighted the importance of flexible mental health services, saying: ‘We want to ensure children and young people have easy access to support that suits their needs.’
By using immersive videos and creative tasks, Lumi Nova allows children to confront their worries at their own pace from the comfort of home, making mental health care more engaging and approachable. The year-long pilot aims to assess the app’s impact on childhood anxiety in the region.