Digital Watch newsletter – Issue 59 – April 2021

Home | Newsletters & Shorts | Digital Watch newsletter – Issue 59 – April 2021

Download your copy

EN FR

TRENDS

The top 3 trends in April: Regulating AI, data scraping, and the aftermath of SolarWinds

 Each month we analyse hundreds of unfolding developments to identify key trends and underlying issues which impact the work of policy practitioners working in these fields. Here’s what mattered in April.

 

1. Regulating AI

The EU’s proposed regulation could become a global standard, but there’s a long journey – and some tough decisions – ahead 

Governments’ interest in regulating AI systems is now more manifest than ever, especially after the EU’s long-awaited draft proposal published on 21 April. The proposal takes a risk-based regulatory approach: If a system poses exceptional risk, it will be banned; if it’s considered high-risk, it will be heavily regulated; if the risk is limited, it’s then a matter of being transparent about it. Read more on page 10.

Experts say there’s a striking similarity with the GDPR: Both take a long-hand approach, meaning that the proposed rules regulate anyone – anywhere – whose system is used within the EU or affects EU citizens. This normative approach and jurisdictional reach  could help the proposal become a global gold standard for AI regulation and the development of AI systems. 

But the EU needs more than a claim to a global standard. It will have to assess the potential impact of the new legal framework, or risk losing out (again) to the USA and China: Tight regulation is likely to discourage the private sector from innovating and developing AI systems in the EU; but a laxer approach would simply delay finding solutions to important issues, including deeply ethical ones, maybe until it’s too late. Experts dealing with antitrust issues around the globe know this all too well.

Essentially, European policymakers and politicians will need to agree on how far and deep the regulation should go. They will need to decide on which systems are too risky for society, which can be tolerated (albeit regulated) in the interest of innovation, and which to let off the hook. These decisions will ultimately impact the level of trust which society will have in AI systems and those who run them – not just in the EU, but around the world. 

 

2. Data scraping

Facebook and LinkedIn’s latest breaches were a data scraping incident, and not a hack... there’s more to this than just technical lingo

We’re quite accustomed to learning about massive data breaches. Facebook believes this activity happens – and will continue to happen – regularly

What stood out this month, however, is that both Facebook and LinkedIn responded to the massive data breaches – affecting 500+ million users on each platform – in a similar way. According to the social media giants, it wasn’t a hack, but rather a data scraping incident from years ago, with the databases of personal information being posted online only now. 

At face value, both terms – hacking and scraping – conjure elements of technical mischief. However, both may lead to the same dire consequences: Personal data that is meant to be private or protected is now in the public domain, and can be exploited by bad actors. Breach or no breach, users are at risk of falling victim to phishing or other fraudulent practices, as has already been the case. Both are an unwelcome invitation for more criminal activity, which is broadly regulated and punishable by law around the globe. 

However, this reveals deeper societal issues, and is a reflection of the price we’re willing to pay for more free services, social connections, innovative products, and convenience. Ultimately, whether we want to – or how we go about – holding companies to higher standards is a reflection of the value we attach to our personal data and privacy, and that of future generations.  

We take a deep dive at what data scraping means for regulators and users further down.

 

3. The aftermath of SolarWinds

The espionage was carefully planned and executed – what happens next is even more critical

The USA has formally attributed last year’s SolarWinds attack to the Russian Foreign Intelligence Service (SVR). At least nine US federal agencies were breached and several other countries affected.

What followed since this attribution is tit-for-tat: In an executive order, US President Joe Biden sanctioned around 40 Russian individuals and companies, after which Russia expelled 10 US diplomats. NATO, the EU, the UK, and Australia immediately expressed support in favour of the USA – an indication that the Biden administration worked closely with its allies to amplify the message and show a united front. 

What happens next will be much more critical. Although the attack was labelled as espionage (meaning information was taken, but nothing was destroyed or disrupted), it doesn’t mean that the perpetrators won’t use their newly acquired intelligence for a cyberattack. Plus, it’s still unclear to the USA whether it has managed to wipe its breached systems clean from the intrusion. 

The main question is whether the sanctions will be enough to dissuade Russia (and others) from carrying out similar hostile activities, and whether the USA will change its deterrence and ‘defend forward’ strategies. The White House has already signalled that the sanctions were only a part of the steps the USA would be taking, and that ‘there will be elements of our responses to these actions that will remain unseen.’ 

The USA will need to weigh the implications of its responses, and make sure it does not cross the threshold of what’s considered an armed attack. While deterrence may be ineffective, experts have suggested that the USA should focus more on beefing up its defences.

On a more positive note, experts also say both sides have left the door open for dialogue: Biden said it was now time for tensions to de-escalate through dialogue and diplomacy; Russia did not shoot down Biden’s idea for a summit. Time will tell.

 

SolarWinds timeframe
The SolarWinds timeframe. Source: Microsoft

BAROMETER

The digital policy developments that made headlines in April

The digital policy landscape changes on a daily basis. Our aim is to save practitioners time: We decode, contextualise, and analyse ongoing developments, offering a bite-sized authoritative update.

 

Global IG architecture

G7 digital and technology ministers outlined plans to cooperate on a wide range of issues, including supply chains, digital standards, data flows, and Internet safety.

The 2020–2021 Group of Governmental Experts on lethal autonomous weapons systems concluded its work.

The call for session proposals for the 16th Internet Governance Forum meeting is open until 26 May.

 

Sustainable development

Egypt launched the Decent Life project to connect 1300 villages to optical fibre.

Bangladesh, Brazil, China, Egypt, India, Indonesia, Mexico, Nigeria, and Pakistan initiated the digital learning initiative E9.

The Economic Commission for Africa established the Learning Girls in ICT Initiative.

 

Security

The USA formally attributed the SolarWinds cyberattack to Russia and imposed sanctions. Russia responded, while the EU and NATO expressed support for the USA. Read the commentary further up.

The European Parliament adopted a regulation requiring online platforms to remove or disable access to terrorist content within one hour of notification.

Several EU institutions were affected by an IT security incident. Apple supplier Quanta Computer Inc. was hit by a ransomware attack.

Interpol expressed support for the Coalition Against Stalkerware in the fight against online abuse, stalking, and harassment.

 

E-commerce and Internet economy

China’s competition authority fined Alibaba US$2.8 billion over anticompetitive practices, and launched an antitrust investigation into food delivery company Meituan. It also ordered 34 tech companies to correct anticompetitive practices within 1 month. Russia’s competition authority launched investigations into Yandex and YouTube. A report by Australia’s antitrust regulator argues that measures are needed to address the market dominance of Apple and Google’s app stores.

Google was fined more than US$36 million in Turkey for abusing its dominant position in search engine services. Associated Newspapers sued Google for anticompetitive practices in the advertisement business.

German industry associations filed an antitrust complain against Apple’s new AppTrackingTransparency requirements. UK software company ValueLincensing sued Microsoft over alleged anticompetitive practices.

The USA proposed a global minimum tax for companies, including the tech sector. The Organisation for Economic Co-operation and Development (OECD) reaffirmed plans to reach a global solution for taxing the digital economy by July 2021.

US cryptocurrency exchange Coinbase joined the stock market.

China’s digital yuan will allow anonymous transactions in small amounts. The central banks in the UK and Japan launched exploratory work into digital currencies.

 

Digital rights

The data of 533 million Facebook users was leaked to an online hacking forum, triggering investigations in several countries. Additionally, the data of 500 million LinkedIn users was posted for saleRead our analysis throughout this newsletter.

A German data protection regulator opened proceedings ‘to prevent an illegal mass exchange of data’ between Facebook and WhatsApp. Privacy group noyb filed a complaint in France against Google’s use of its Android Advertising ID.

TikTok was sued in the UK over the processing of children’s data. Consumer and child protection organisations asked Facebook to cancel plans to create an Instagram for children.

The UK National Health System’s COVID-19 app was blocked on Apple and Google’s app stores for privacy violations.

 

Content policy

Facebook expanded the scope of its Oversight Board to also decide in cases in which the company chose not to take down content. The Board postponed its decision on Trump’s ban from Facebook and Instagram.

Livestream platform Twitch announced it would ban users for severe misconduct that occurs outside of the platform.

Twitter was criticised for acting upon a request from the Indian government to remove tweets seen as critical of the government’s handling of the COVID-19 pandemic.

Instagram released new features to address hate speech. Twitter launched an initiative to analyse the harmful impact of its algorithms.

 

Jurisdiction and legal issues

The US Supreme Court issued its opinion in the Google LL.C v. Oracle America InC. case, ruling that copying a portion of the Java SE computer program by Google is protected as ‘fair use’. 

A high court in Pakistan lifted the TikTok ban imposed by the country’s telecom regulator. A Russian court fined Twitter for not removing content encouraging minors to take part in unauthorised protests.

A federal court ruled that Google misled Android users regarding the personal location data it collected.

An Amsterdam court ordered Uber to reintegrate five drivers excluded from the platform through an automated process.

 

Infrastructure

At an online summit on semiconductor and supply chain resilience, US President Joe Biden reiterated plans to strengthen the country’s semiconductor industry and secure its supply chain.

The USA added seven Chinese supercomputing entities to the export control list.

The UK opened an investigation into Nvidia’s proposed acquisition of chip designer Arm.

 

New technologies (IoT, AI, etc.)

The European Commission launched a draft regulation for AI systems. Read more further down. The US Federal Trade Commission warned it would take action against discriminatory AI systems.

Brazil adopted an AI strategy. Canada outlined plans for updating its AI strategy and launching a national quantum strategy.Finland established the Quantum Institute.

The UK outlined plans to regulate the cybersecurity of internet of things devices.

Like what you read? We have more...

Our weekly digests break up the month’s developments into shorter updates. We publish the digest every Friday, allowing you to wrap up the week with an overview of what has happened, or start the following Monday with a re-cap.

Read an issue from our archives, and sign up for the digest.

DW weekly

IN FOCUS

Data scraping and the problems regulators and users face

On 3 April, Business Insider revealed that a database containing the personal data of 553 million Facebook users was uploaded to a hacking forum.  A few days later, an archive with the personal information of 500 million LinkedIn users was also posted for sale.

Facebook and LinkedIn’s immediate response was similar: The breaches were large-scale scraping incidents, rather than hacks. This drew the attention of data protection regulators and users around the world.

Facebook said the incident happened in August 2019, when malicious actors used a vulnerability in the ‘How people can find me’ feature, which was later fixed. It later emerged that the data was scraped on several occasions from 2018 to 2019 (we’ll get to the issue of timelines further down). In LinkedIn’s case, it is unclear how old this information is and whether it was aggregated solely from the platform.

What’s the difference between data scraping, a data breach, and a data leak?

Data scraping is an automatic extraction of large amounts of data from websites, databases, or apps. In web scraping, a program ‘crawls’ through content on a website or platform and lifts, or copies, publicly available information automatically and creates a document (excel sheet, accounting ledger, website) for scrapers’ use. It is a common practice in business; for instance, this is what search engines do to provide us with the information we’re searching for. It’s also how price comparison websites work, such as those comparing air ticket prices or insurance costs.

Data scraping is not unlawful, but it needs to follow certain rules – such as requiring people’s consent to scrape, and a legal reason to use data scraping. The problems arising out of data scraping are related to how the data is scraped (with or without consent), what it was used for (data analysis or posting on hackers website), and who shoulders responsibility for any risks and threats.

A data breach is when a computer system or data is unlawfully accessed or weakened by malicious actors without the knowledge or authorisation of the owner, leading to exposure of confidential, sensitive, or protected information. Data breaches (or hacks) are unlawful. In many cases, they are considered a crime.

A data leak is when private data (or a private database, even if made up of various pieces of publicly available data) is made available to the public. A data leak does not require an active breach into the computer system. It usually is a result of poor data security practices or accidental human error. The responsibility for data leaks is tied to the obligation to keep data secure and to undertake necessary measures to do so.

Here are at least five main issues which regulators must resolve.

Issue #1: How platforms classify personal data as private or public

Both companies said the data was public. In Facebook’s case, the company added that it’s the users’ responsibility to make sure ‘their settings align with what they want to be sharing publicly’. In other words, if a user chooses certain settings on the platform, Facebook takes this as the consent which certain data protection regulations require. The problem here is that Facebook is not properly explaining to the users the full extent of the consequences of certain settings for users to make an informed decision.

Plus, if we look at the dedicated resource page, the platform says that public information refers to ‘some of the information you give us when you fill out your profile…’ or that ‘your Public Profile includes your name, gender, username…’  The fact that Facebook talks in terms of ‘some’ or ‘including’ leaves users in the dark. As a result, users could be providing personal data to Facebook under the mistaken idea that Facebook will keep it private, and hence more secure.

Issue #2: The platforms’ responsibility for the personal data they hold, public or not

In some jurisdictions, platforms are indeed responsible for personal data, even if it's already out there and visible to others. For instance, the EU’s GDPR says that companies need to have certain technical measures in place to ensure that personal data is processed safely, securely, and according to law.

One can therefore argue that if the data was lifted off the platforms – whether through a vulnerability in the system or by failing to detect scraping activity – then the platforms have a degree of responsibility.

The Irish data protection authority, for one, is arguing along these lines. Digital Rights Ireland, a digital rights advocacy group, is also gearing up to file a mass lawsuit against Facebook on behalf of users who want to be compensated for having their personal data exposed, in breach of the platform’s terms and conditions. In the USA, the incident could be considered a violation of Facebook’s US$5 billion privacy settlement with the Federal Trade Commission (FTC) to resolve the Cambridge Analytica data scandal, among others. Other user lawsuits are expected to hit Facebook in India and in other countries.

Issue #3: The platforms’ attempt to normalise recurrent incidents

Facebook has downplayed the incident not only to shrug off liability, but to normalise this as a chronic occurrence. Facebook is telling users and regulators that the world needs to get used to such incidents and that it does not intend to notify affected users every time this happens.

This insight comes from internal emails that were circulated: ‘We expect more scraping incidents and think it's important to both frame this as a broad industry issue and normalise the fact that this activity happens regularly.’

Issue #4: The platforms’ obligation to report a data leak

The two platforms did not disclose or notify users or authorities of these incidents at the time. In countries like EU member states and Brazil, which have comprehensive data protection regulations, Facebook and LinkedIn face steep fines for failure to disclose and notify breaches and violations of data protection regulations.

Under the GDPR, and similarly under its Brazilian counterpart the LGPD, scraping of personal data requires consent and a lawful basis. Both the Irish Data Protection Commissioner and the Brazilian Protection and Consumer Defence Foundation of the State of São Paulo  – plus others worldwide – are investigating Facebook’s practices. Similar steps are being taken in the LinkedIn incident, with the Italian data protection agency opening a probe. The crucial issue for the regulators is to establish, in legal terms, how these breaches fall under the type of incidents that would trigger the platforms’ responsibility to report them.

Here is where the timeline comes in: The date when the breach (or leak) took place has a direct impact on the platforms’ responsibility. In Europe, for instance, reporting a breach became a legal obligation when the GDPR came into effect in May 2018.

In the USA, Facebook is to some extent shielded from liability for any breaches occurring before June 2019, as part of the Facebook/Cambridge Analytica settlement, except in the state of California, where the Consumer Privacy Act came into effect mid-2018. Yet, the regulators will still need to evaluate the compound data sets and identify timing of the data they include, to then determine when the platforms learned about the scraping incidents, and what responsibilities Facebook and LinkedIn were under at those times. 

Issue #5: Users need more awareness of and protection from the risks of exposed personal data

Personal data in the public domain can be used to harm people. Whether the responsibility is the users’ (when sharing personal details on social media without understanding, or bothering to understand, the implications), or the platforms’ (for not doing enough to protect people’s data), bad actors will exploit it.

In these two cases, for instance, the value of the leaks is in the ability to associate mobile numbers and other personal data with an individual. Fraudsters who call someone on a mobile phone will sound more genuine and legitimate if they also know the date of birth, hometown, and place of work. One would think only the authorities and their bank hold such information, right?

The companies themselves, being victims of malicious actors who exploited vulnerabilities on their platforms, will be expected to file legal action, as they have done in the past.  But this should also serve as a wake-up call for users. As a consumer watchdog put it: ‘People are going to be a lot more skeptical and a lot more careful about providing this whole trove of information to social media platforms.’

DATA ANALYSIS

The world needs more chips: What this means for the big players

A global shortage: How did we come to this?

Semiconductors – or chips – are at the core of virtually every electronic device. In turn, the production of chips relies on a very complex global supply chain. Producing a single computer chip can involve over 1 000 steps and 70 separate border crossings.

It involves mining and processing the raw materials (such as silicon and boron), designing the chip, and developing the manufacturing equipment. All of this needs to happen before the tiny chip is actually manufactured and assembled into products.

The USA dominates the architecture and design components of the supply chain. US companies, with Intel in the lead, account for almost 50% of the world’s chip sales. But more than 80% of the global chip manufacturing happens in Asia, with Taiwan Semiconductor Manufacturing Company (TSMC) and Samsung taking the lead.

Right now, there’s a global shortage of chips, which is affecting companies, and indirectly, consumers. Samsung, a chipmaker itself, warned about a ‘serious imbalance in the supply and demand of chips in the IT sector globally’.

Apple reportedly had to postpone the production of some laptops and tablets due to the tight supply of chip and display components. Sony has been facing difficulties in supplying its new PlayStation 5.The shortage is expected to last until 2022 or 2023.

 

Global microchip consumers
Global microchip consumers. Source: Semiconductor Industry Association

 

One of the reasons is that chip manufacturers – and that’s basically TSMC, Samsung and, to a lesser extent, Intel – are sensitive to spikes in demand. The pandemic generated a higher-than-usual demand for smartphones, laptops, and other similar devices, prompting producers to increase their demand. The demand is also coming from other industries, such as carmakers and producers of internet of things devices.

The handful or so companies producing chips have found it difficult to meet this demand. As digital devices become more and more sophisticated, so do chip designs. In turn, this has increased costs for chip manufacturers, which creates a barrier for new players wanting to enter the market. The result is a highly concentrated market in which only a handful of companies can afford to compete and respond to the global demand.

Global market share in 2020
Global market share in 2020. Source: Semiconductor Industry Association

The geopolitical tensions, including export controls and licencing policies imposed by the US government on Chinese companies in recent years, have not helped. For instance, US and foreign firms are now barred from using US intellectual property and technology to produce chips for Huawei (including Huawei’s own designed Kirin chips).

Before these sanctions took effect, Huawei was forced to stockpile chips, and reports argue that large producers such as TSMC prioritised Huawei's orders to the detriment of other clients, affecting the supply chain with long-term effects.Chinese chipmaker Semiconductor Manufacturing International Corporation (SMIC) can only acquire American technology if its suppliers receive a license from the US government.

Natural disasters and incidents are also affecting the production and supply of chips. Chip factories have had to close down or reduce production. Among them were factories in Texas (closed down in February 2021 due to power failures caused by low temperatures); in Japan (after a fire broke out in March 2021); and in Taiwan (due to the ongoing drought).

Geopolitics: How the world’s main players have reacted

China and the USA, which have for long competed in the technological sector, have big ambitions. They both want to become self-reliant in the chip market. This would ensure that their home-grown tech industries do not depend on external, possibly unstable, global supply mechanisms.

It is estimated that a country requires a minimum of US$1 trillion up front to achieve a self-sufficient local supply chain; such an investment, in turn, would result in a 35–65% increase in chip prices. That’s steep.

China has, for the past few years, been a net importer of chips. Only 30% of its chips have been manufactured domestically. The trade sanctions and export controls imposed by the USA and their effects on Chinese companies like Huawei and SMIC have fuelled China’s ambitions for becoming self-reliant.

As a result, China is investing heavily in its own chip manufacturing capabilities, as part of its multibillion-dollar five-year plan. Released in March 2021, the plan envisions considerable support for the local semiconductor industry. This is not completely new, as the Chinese chip manufacturers have already benefited from substantial government support in recent years.

The USA is also not producing enough chips on its own, making its large tech industry heavily dependent on Asian manufacturers. During April’s chip summit, President Joe Biden pledged to strengthen the country’s semiconductor industry. At the same meeting, key companies affected by the chip supply shortage called on the president to help increase the country’s capacity to produce chips.

Biden also called on Congress to establish a US$50 billion fund for semiconductor manufacturing and research, in line with the draft CHIPS for America Act being debated in the Congress (H.R. 7178/ S. 3933).

There are expectations that a portion of such a fund would go to support the development of advanced chip plants in the USA by TSMC, Samsung, and Intel.

If complete chip self-reliance is almost an impossible task, the next best thing is to source chips from elsewhere. Both China and the USA are now relying on Taiwan’s TSMC, which has so far managed to navigate the geopolitical complexities between the two countries by making itself indispensable to both.

Unsurprisingly, TSMC is now facing increased pressure to pick a side between access to the world largest chip market in the USA and the fastest growing one in China.

The chip shortage and complex supply chain also offers an opportunity for new tech partnerships. The EU, for instance, has interesting options ahead, as it tries to achieve its goal to produce one-fifth of advanced chips globally by 2030.

On the one hand, it could join forces with the USA to counterbalance the regional concentration of chip production in Asia. On the other, it could also partner with TSMC and/or Samsung to build or expand factories in Europe. Of course, the bloc could also choose to invest more in European players.

Another potential partnership is between the USA, Japan, Korea, and India (known as the Quad), which have recently agreed to create a Critical and Emerging Technologies Working Group. Among the Quad’s aims is to cooperate on diversifying critical technology supply chains.

The adage ‘if you can’t beat them, join them’ couldn’t be more true.

LEGAL

The EU’s proposed rules for regulating AI systems: The good, the bad, and the ugly

AI is all around us, from our search engines to the smart devices we’re using in our homes. The industry continues to push the boundaries of what AI systems can do. New and improved systems make our lives more comfortable, our decisions easier to make, and our surroundings safer.   

This all sounds blissful… until it’s not. The EU thinks there should be cut-off lines depending on the level of risk to people or infrastructure. Beyond these lines, AI systems should be outright banned (or heavily regulated). Here’s the risk-based approach the new European Commission proposal takes:

  • Systems that pose an enormous risk will be considered outright unacceptable. The cases are listed specifically in the regulation, and include types of facial recognition, and algorithms used to manipulate how we think or what we do. 
  • Systems that pose a high-risk will be heavily scrutinised. These include critical infrastructure which could prove risky to people’s lives, and some of the safety components in products (think robot-assisted surgery). The regulation includes both a predefined list of cases and a set of criteria to determine how other systems could be classified as high risk. If a system is deemed high-risk, it needs to be assessed before being put on the market, and then registered. There’s actually a host of other obligations involved, and of course, penalties for not observing them.    
  • Systems that pose a limited risk will carry only minor obligations. Among these are chatbots: The rules will require that we’re at least made aware we’re talking with a machine. 
  • Systems that pose minimal risk can be developed and used freely. Essentially, that’s anything else not falling under the three above. Most of the AI systems currently in use fall under this category. 
The European Commission's proposed AI regulation
The European Commission's proposed AI regulation. Source: European Union

The next step is for the European Parliament and member states to consider the proposal. It sounds straightforward, but this will actually take years of readings, amendments, and lobbying before the regulations are adopted.

 

Proposal gets mixed reactions

The proposed rules have been criticised for their vague and broad language, and for not going far enough. They’re also riddled with loopholes, especially when it comes to facial recognition technologies and their oft-times discriminatory practices. 

For instance, systems that identify people through real-time biometrics are generally banned. Yet, there are some exceptions that apply to law enforcement, which critics believe can be abused

Plus, the ban does not apply to other public entities or companies. Who knows whether they are already using real-time biometrics? Activists say they are, in fact.

Real-time biometrics may be banned, but systems that work on pre-captured images are not. These can still be used to identify people according to race, gender, and sexuality, so although they may well be heavily regulated, companies could still get a green light to use them.

And then, there’s the issue of balancing innovation and regulation. Companies think there are too many compliance obligations in place. This argument will not hold enough water for Margrethe Vestager, the European Commission’s executive vice-president, who is adamant that Europe ‘can only reap the full benefits of AI's societal and economic potential if we trust we can mitigate the associated risks’.

However, it gets more complex when companies say the law will slow down the development of advanced AI systems in the EU. Although big companies may find it difficult – but not impossible – smaller companies or start-ups could be discouraged altogether

The Commission certainly wants Europe to trust the development of AI systems, but it also wants to encourage a home-grown industry. As we said in our commentary (page 2), the EU will be walking a tightrope, and will have some tough calls to make.

GENEVA

Policy updates from International Geneva

Many policy discussions take place in Geneva every month. In this space, we update you with all that’s been happening in April. For other event reports, visit the Past Events section.

Protection of Critical Water-related Infrastructure (Part II) – 13 April 2021 

This event, organised by the Permanent Missions of Slovenia and Israel to Geneva, the World Meteorological Organization (WMO), and Microsoft, was a follow-up to the November 2020 roundtable. This second part explored how some of the lessons that emerged from the roundtable could be applied in developing countries, especially in the African region. The event focused on national experiences, and the frameworks and standards that have been put in place, such as the Malabo Convention.

Promoting Circular Economy and Sustainable Use of Natural Resources in the UNECE Region – 20–21 April 2021

The United Nations Economic Commission on Europe (UNECE) focused its 69th session on the circular economy and the use of natural resources in achieving the SDGs. One of the roundtables dealt with circular energy, mobility, and digital transformation, and the corresponding new policies and industry best practices. While energy is aspiring to be low carbon or carbon negative, there’s also a major push for e-mobility (electric-powered vehicles) in cities and for long-distance transport. Technology is at the core of the fourth industrial revolution, which is embracing all aspects of life and work. The digital transformation has significant bearing on how natural resources are used, moving away from past linear models to new sustainable, service-focused, and circular economy business models. 

Digital Economy: Trade and Finance Tour – 28 April 2021 

The roundtable discussion, organised by the GIP as part of the series 12 Tours to Navigate Digital Geneva, debated the ongoing e-commerce discussions taking place at the World Trade Organization (WTO), with close attention having been paid to the Joint Statement Initiative (JSI) and regional trade agreements. The speakers also looked at the impacts of the COVID-19 pandemic on e-commerce and the role that the latter can play in the post-pandemic world. 

12 Tours - Digital Economy

UPCOMING

Upcoming global policy events: Are you tuned in?

Learn more about upcoming digital policy events around the world and use DeadlineR to remind you about important dates and deadlines.