US tech firms warn Vietnam’s draft law could limit growth

US tech companies have raised concerns over a proposed data protection law in Vietnam, warning it could restrict their ability to grow in one of Asia’s largest digital markets. The draft law, which is under discussion in Vietnam’s parliament, aims to tighten controls on data protection, limit data transfers abroad, and give authorities easier access to information. Major industry players, represented by the Information Technology Industry Council, argue that these restrictions could hinder companies like Meta, Google, and Equinix from effectively reaching their large Vietnamese user base and building new data centres.

Vietnam, home to 100 million people, is an attractive market for tech and social media companies and has ambitions to expand its data centre industry through foreign investments. However, the new law would require companies to obtain prior authorisation before transferring “core” or “important” data abroad—terms that critics say are vaguely defined. In addition, companies may be required to share data with the government in cases broadly categorised as being in the “public interest.”

The US tech sector has voiced opposition, citing an “undue expansion of government access” that could create significant compliance challenges. The American Chamber of Commerce in Hanoi has joined the call, urging lawmakers to reconsider the legislation’s quick adoption, which is scheduled for a vote on November 30. Industry analysts are watching closely, as the law could impact foreign investment plans, including Google’s potential new data centre in southern Vietnam.

UK court sides with Google in YouTube Shorts trademark case

Google has won a trademark lawsuit brought by Shorts International, a British company specialising in short films, over the use of the word ‘shorts’ in YouTube‘s short video platform, YouTube Shorts. London’s High Court found no risk of consumer confusion between Shorts International’s brand and YouTube’s platform, which launched in 2020 as a response to TikTok‘s popularity.

Shorts International, known for its short film television channel, argued that YouTube Shorts infringed on its established trademark. However, Google’s lawyer, Lindsay Lane, countered that it was clear the ‘Shorts’ platform belonged to YouTube, removing any chance of brand confusion.

Judge Michael Tappin ruled in favour of Google, stating that the use of ‘shorts’ by YouTube would not affect the distinctiveness or reputation of Shorts International’s trademark. The court’s decision brings the legal challenge to a close, dismissing all claims of infringement.

Mozambique and Mauritius to block social media

Mozambique and Mauritius are facing criticism for recent social media shutdowns amid political crises, with many arguing these actions infringe on digital rights. In Mozambique, platforms like Facebook and WhatsApp were blocked following protests over disputed election results.

Meanwhile, in Mauritius, the government imposed a similar blackout until after the 10 November election, following a wiretapping scandal involving leaked conversations of high-profile figures. Furthermore, digital rights groups such as Access Now and the #KeepItOn coalition have condemned these actions, arguing that they violate international human rights standards, including the African Commission on Human and Peoples’ Rights (ACHPR) resolution 580 and the International Covenant on Civil and Political Rights (ICCPR), as well as national constitutions.

In response, digital rights advocates are calling on telecommunications providers, including Emtel and Mauritius Telecom, to resist government orders to enforce the shutdowns. By maintaining internet connectivity, these companies could help preserve citizens’ access to information and uphold democratic principles in politically sensitive times.

Additionally, rights organisations argue that internet service providers have a unique role in supporting transparency and accountability, which are vital to democratic societies.

The US federal agency investigates how Meta uses consumer financial data for targeted advertising

The Consumer Financial Protection Bureau (CFPB) has informed Meta of its intention to consider ‘legal action’ concerning allegations that the tech giant improperly acquired consumer financial data from third parties for its targeted advertising operations. This federal investigation was revealed in a recent filing that Meta submitted to the Securities and Exchange Commission (SEC).

The filing indicates that the CFPB notified Meta on 18 September that it evaluated whether the company’s actions violate the Consumer Financial Protection Act, designed to protect consumers from unfair and deceptive financial practices. The status of the investigation remains uncertain, with the filing noting that the CFPB could initiate a lawsuit soon, seeking financial penalties and equitable relief.

Meta, the parent company of Instagram and Facebook, is facing increased scrutiny from regulators and state attorneys general regarding various concerns, including its privacy practices.

In the SEC filing, Meta disclosed that the CFPB has formally notified the company about an investigation focusing on the alleged receipt and use for advertising of financial information from third parties through specific advertising tools. The inquiry targets explicitly advertising related to ‘financial products and services,’ although it remains to be seen whether the scrutiny pertains to Facebook, Instagram, or both platforms.

While a Meta spokesperson refrained from commenting on the matter, the company stated in the filing that it disputes the allegations and believes any enforcement action would be unjustified. The CFPB also opted not to provide additional comments.

Amid this scrutiny, Meta recently reported $41 billion in revenue for the third quarter, a 19 percent increase from the previous year. A significant portion of this revenue is generated from its targeted advertising business, which has faced criticism from the Federal Trade Commission (FTC) and European regulators for allegedly mishandling user data and violating privacy rights.

In 2019, Meta settled privacy allegations related to the Cambridge Analytica scandal by paying the FTC $5 billion after it was revealed that the company had improperly shared Facebook user data with the firm for voter profiling. Last year, the European Union fined Meta $1.3 billion for improperly transferring user data from Europe to the United States.

Kremlin seeks end to YouTube ban on Russian state media

The Kremlin has called on Google to lift its restrictions on Russian broadcasters on YouTube, highlighting mounting legal claims against the tech giant as potential leverage. Google blocked more than a thousand Russian channels and over 5.5 million videos, including state-funded media, after halting ad services in Russia following its invasion of Ukraine in 2022.

Russia’s legal actions against Google, initiated by 17 Russian TV channels, have led to compound fines based on the company’s revenue in Russia, accumulating to a staggering figure reportedly in the “undecillions,” according to Russian media. Kremlin spokesperson Dmitry Peskov described this enormous number as symbolic but urged Google to take these legal pressures seriously and reconsider its restrictions.

In response, Google has not commented on these demands. Russian officials argue that such restrictions infringe on the country’s broadcasters and hope the significant financial claims will compel Google to restore access to Russian media content on YouTube.

TikTok faces lawsuit in France after teen suicides linked to platform

Seven families in France are suing TikTok, alleging that the platform’s algorithm exposed their teenage children to harmful content, leading to tragic consequences, including the suicides of two 15-year-olds. Filed at the Créteil judicial court, this grouped case seeks to hold TikTok accountable for what the families describe as dangerous content promoting self-harm, eating disorders, and suicide.

The families’ lawyer, Laure Boutron-Marmion, argues that TikTok, as a company offering its services to minors, must address its platform’s risks and shortcomings. She emphasised the need for TikTok’s legal liability to be recognised, especially given that its algorithm is often blamed for pushing disturbing content. TikTok, like Meta’s Facebook and Instagram, faces multiple lawsuits worldwide accusing these platforms of targeting minors in ways that harm their mental health.

TikTok has previously stated it is committed to protecting young users’ mental well-being and has invested in safety measures, according to CEO Shou Zi Chew’s remarks to US lawmakers earlier this year.

OpenAI adds search capabilities to ChatGPT

OpenAI has introduced new search functions to its popular ChatGPT, making it a direct competitor with Google, Microsoft’s Bing, and other emerging AI-driven search tools. Instead of launching a separate search engine, OpenAI chose to integrate search capabilities directly into ChatGPT, which will pull information from the web and relevant sources based on user questions.

Initially, ChatGPT’s search feature will be available to Plus and Team users, with plans to expand access to enterprise and educational users, as well as free users, in the coming months. OpenAI’s partnerships with major publishers like Condé Nast, Time, and the Financial Times aim to provide a rich pool of content for ChatGPT’s search.

This launch follows OpenAI’s selective testing of SearchGPT, an AI-based search prototype, earlier in the year. With its recent funding round boosting its valuation to an estimated $157 billion, OpenAI continues to strengthen its standing as a leading private AI company.

Indonesia bans Google and Apple smartphone sales

Indonesia has banned sales of Google’s Pixel smartphones due to regulations requiring a minimum of 40% locally manufactured components in devices sold within the country. This decision follows a similar ban on Apple’s iPhone 16 for failing to meet these content standards. According to Febri Hendri Antoni Arief, a spokesperson for Indonesia’s industry ministry, the rules aim to ensure fairness among investors by promoting local sourcing and partnerships.

Google stated that its Pixel phones are not officially distributed in Indonesia, though consumers can still import them independently if they pay applicable taxes. Officials are also considering measures to deactivate unauthorised imports to enforce compliance.

Despite Google and Apple not being leading brands in Indonesia, the market holds significant potential for global tech firms due to its large, tech-savvy population. However, Bhima Yudhistira from the Centre of Economic and Law Studies warned that these restrictions may deter foreign investment, creating what he calls ‘pseudo protectionism’ that could dampen investor sentiment in the region.

EU moves to formalise disinformation code under DSA

The EU‘s voluntary code of practice on disinformation will soon become a formal set of rules under the Digital Services Act (DSA). According to Paul Gordon, assistant director at Ireland’s media regulator Coimisiúin na Meán, efforts are underway to finalise the transition by January. He emphasised that the new regulations should lead to more meaningful engagement from platforms, moving beyond mere compliance.

Originally established in 2022 and signed by 44 companies, including Google, Meta, and TikTok, the code outlines commitments to combat online disinformation, such as increasing transparency in political advertising and enhancing cooperation during elections. A spokesperson for the European Commission confirmed that the code aims to be recognised as a ‘Code of Conduct’ under the DSA, which already mandates content moderation measures for online platforms.

The DSA, which applies to all platforms since February, imposes strict rules on the largest online services, requiring them to mitigate risks associated with disinformation. The new code will help these platforms demonstrate compliance with the DSA’s obligations, as assessed by the Commission and the European Board of Digital Services. However, no specific timeline has been provided for the code’s formal implementation.

AI robocall threats loom over US election

Election officials across the US are intensifying efforts to counter deepfake robocalls as the 2024 election nears, worried about AI-driven disinformation campaigns. Unlike visible manipulated images or videos, fake audio calls targeting voters are harder to detect, leaving officials bracing for the impact on public trust. A recent incident in New Hampshire, where a robocall falsely claimed to be from President Biden urging people to skip voting, highlighted how disruptive these AI-generated calls can be.

Election leaders have developed low-tech methods to counter this high-tech threat, such as unique code words to verify identities in sensitive phone interactions. In states like Colorado, officials have been trained to respond quickly to suspicious calls, including hanging up and verifying information directly with their offices. Colorado’s Secretary of State Jena Griswold and other leaders are urging election directors to rely on trusted contacts to avoid being misled by convincing deepfake messages.

To counter misinformation, some states are also enlisting local leaders and community figures to help debunk false claims. Officials in states like Minnesota and Illinois have collaborated with media outlets and launched public awareness campaigns, warning voters about potential disinformation in the lead-up to the election. These campaigns, broadcasted widely on television and radio, aim to preempt misinformation by providing accurate, timely information.

While no confirmed cases show that robocalls have swayed voters, election officials regard the potential impact as severe. Local efforts to counteract these messages, such as public statements and community outreach, serve as a reminder of the new and evolving risks AI technology brings to election security.