Artificial Intelligence & Emerging Tech

9 Oct 2023 06:15h - 07:45h UTC

Event report

Speakers and Moderators

  • Pamela Chogo,Tanzania IGF
  • Jörn Erbguth
  • Kamesh Shekar, Youth Ambassador at The Internet Society
  • Tanara Lauschner
  • Umut Pajaro Velasquez
  • Victor Lopez Cabrera

Moderators

  • Jennifer Chung

Table of contents

Disclaimer: This is not an official record of the IGF session. The DiploAI system automatically generates these resources from the audiovisual recording. Resources are presented in their original format, as provided by the AI (e.g. including any spelling mistakes). The accuracy of these resources cannot be guaranteed. The official record of the session can be found on the IGF's official website.

Knowledge Graph of Debate

Session report

Audience

The analysis explores multiple aspects of the relationship between artificial intelligence (AI) and developing countries. One significant challenge faced by developing countries is the limited internet connectivity and lack of electronic devices. This hampers their ability to fully harness the benefits of AI. Most people in these countries are far from internet connectivity and lack electronic devices, making it difficult to access AI services. This is seen as a negative aspect of the situation.

However, despite these resource limitations, a question is raised about how developing countries can still benefit from AI services, despite their limited resources. This suggests that creative solutions and strategies can be explored to ensure that developing countries can leverage AI. This neutral stance highlights the importance of finding alternative ways for these countries to benefit from the advancements in AI.

Furthermore, AI is viewed as an opportunity for up-skilling and re-skilling youth in developing countries. This positive argument suggests that AI can provide educational opportunities and empower young people in these regions. Equipping youths with AI skills can better prepare them for the future job market and contribute to the economic growth of their countries.

Connectivity issues in developing countries for leveraging AI are also highlighted. This negative sentiment emphasizes that without adequate infrastructure and connectivity, the full potential of AI cannot be realized. It underscores the importance of investing in resilient infrastructure and promoting sustainable industrialization in these regions.

On a more positive note, there is support for the use of generative AI and data governance at a local level. This viewpoint suggests that AI can be a valuable tool for societal progress and development, and local communities should take advantage of it. The positive sentiment towards generative AI and local data governance indicates a belief in their ability to contribute to the achievement of global goals such as industry innovation and strong institutions.

Regulation in the field of AI is recognized as a necessary measure. It is understood that each country will eventually establish its own regulations for AI. Professionals in the field are urged to consider the implications and importance of regulation. Similarly, policymakers are encouraged to consider the need for regulations in the artificial intelligence industry. This neutral stance highlights the importance of ensuring ethical considerations, privacy protection, and responsible use of AI technologies.

In summary, the analysis sheds light on the challenges faced by developing countries in relation to AI, such as inadequate connectivity and limited resources. It also highlights the potential benefits of AI, such as up-skilling and re-skilling opportunities for youth. The need for investing in infrastructure and connectivity is emphasized, along with the importance of local data governance and regulation in the field of AI. These insights provide valuable considerations for policymakers, professionals, and stakeholders involved in the integration of AI in developing countries.

Kamesh Shekar

The use of AI technology needs to be more responsible, considering both the intervention and ecosystem perspectives. Various stakeholders, such as technology companies, AI developers, and AI deployers, have unique responsibilities in ensuring the responsible use of AI. Fine-tuning in the operationalization of AI principles is crucial. Certain principles, like “human in the loop,” can have different interpretations at different stages of AI deployment.

A balanced approach is required for the regulation of emerging technologies to prevent the creation of problems while solving traditional ones. The implementation of regulatory interventions in the context of emerging technologies is important, but regulations should not disrupt beneficial technological innovations, especially in developing countries.

Balancing AI utilization with innovation can unlock maximum benefits. India’s Digital Personal Data Protection Act serves as an example of finding this balance. The act focuses on the usage of data value as well as the protection of privacy. This approach demonstrates how AI can be used to drive economic growth and innovation while also protecting individual rights and data.

Regulatory frameworks for AI should include both government compliance and market-enabling mechanisms. India, as the chair for the Global Partnership on Artificial Intelligence, is working towards creating a well-thought-through regulatory framework. Consensus-building at the global level for AI regulation is being suggested. It is important for legislation to evolve to keep up with the rapid advancements in AI and emerging technologies.

The ongoing debate over the role of consent in data processing and utilization is a significant concern. While consent has traditionally played a critical role, its effectiveness in new technological contexts such as AI and the Internet of Things (IoT) is debatable. There is a growing need for new mechanisms to safeguard data, in addition to consent, to protect individual privacy and ensure responsible AI use.

Advocates call for data protection regulations that are in tune with technological advancements. As AI and emerging technologies continue to evolve, regulations must adapt to keep up. Innovative regulations are necessary to work in tandem with technological advancements and ensure responsible AI use.

Data governance in relation to AI is a crucial concern for the future. Emerging technologies raise important questions about how publicly available personal information is used. With new data protection regulations, the use of AI technologies with this information needs to be carefully considered to avoid potential misuse.

Regulatory considerations for AI should take into account the positive aspects and innovations that the technology brings. While regulations are necessary to address any potential risks, they should not hinder or disrupt the beneficial technological innovations associated with AI, especially in developing countries.

Global coordination is necessary for the progressive development of AI. Multiple entities with an interest in AI need to come together for discussions and collaborations. This will aid in addressing common challenges and ensuring responsible and ethical use of AI on a global scale.

The role of government and private entities is crucial in advancing AI. Governments and private sectors can work together to drive the progression of AI by implementing AI governance frameworks and utilizing market mechanisms to encourage adoption.

Trust and compliance play significant roles in AI systems. Trust can serve as a competitive advantage for AI adopters, and compliance, while burdensome, can bring trust in AI systems. It is important for regulations and frameworks to be seen as beneficial rather than burdensome to foster trust in AI technology.

In conclusion, the responsible use of AI technology requires a balanced approach that considers intervention, ecosystem perspectives, fine-tuning of AI principles, and regulatory frameworks. Legislation needs to evolve to keep up with the pace of AI and emerging technologies. Global coordination and collaboration are essential for the progression of AI, and the role of government and private entities is crucial in advancing responsible AI use. Trust and compliance are key factors in building confidence in AI systems. By addressing these various aspects, AI can be utilized in a responsible and beneficial manner.

Pamela Chogo

The importance of increasing awareness of AI in Tanzania was highlighted due to the low level of understanding of AI in the country. Many people are using AI or contributing to it without fully comprehending its implications. There is a misconception that AI is solely a technical aspect and not a social-technical aspect. To address this, it is argued that there needs to be a raise in awareness and education about AI, including its benefits and challenges.

The importance of ensuring fairness, explainability, robustness, transparency, and data privacy in the AI development process was emphasized. It is crucial to examine how data is gathered and the ethical considerations involved. The models and frameworks used in AI should also be scrutinized. By adopting these principles, AI systems can be developed in a more responsible and accountable manner.

An AI Convention is proposed as a means to create a safe and regulated environment for the use of AI. Drawing parallels to the importance of environmental conventions, it is argued that an AI Convention should be established to ensure all stakeholders adhere to ethical standards and guidelines. This convention would provide a framework for the responsible use and development of AI, allowing individuals and communities to benefit from its potential while minimizing risks and harm.

Concerns were raised about the digital gap in Tanzania and the lack of clear guidelines and standards in AI development. Technology is advancing rapidly while the country is still grappling with bridging the digital divide. To address this issue, it is suggested that comprehensive guidelines and standards be put in place to ensure AI development is inclusive and accessible to all.

The necessity of AI advocacy and awareness creation for policymakers, developers, users, and the community at large is emphasized. There is a great need for advocacy initiatives to promote AI awareness and educate various stakeholders about its potential and impact. Additionally, ethics in data collection should be ensured, and the accuracy and validity of data used in AI systems should be upheld.

AI is seen as beneficial in sectors like health and agriculture. AI can assist in predicting outcomes in agriculture, and AI-based devices can be used in hospitals to address the shortage of experts and resources. Furthermore, community-based AI solutions are identified as a way to mitigate challenges related to resource constraints and access to information and knowledge.

Despite resource constraints, it is maintained that Africa can still benefit from AI with the right approach. Many African countries are facing resource challenges, but the use of AI can provide community-level solutions and fill knowledge gaps.

The global nature of AI is emphasized, with AI being seen as a global asset. Cooperation and partnerships at an international level are crucial to fully harness the potential of AI and address the challenges it presents.

There is a call for more discussion forums and sharing of work on AI. By fostering collaboration and knowledge exchange, the development and understanding of AI can be accelerated.

Finally, the importance of an AI Code of Conduct or AI Convention is underscored. Such a code or convention is essential in establishing ethical standards, ensuring transparency, and promoting responsible and accountable use of AI.

Jörn Erbguth

During the EuroDIG session, the topic of AI regulation was extensively discussed. The importance of AI literacy and capacity building was emphasized as crucial for effectively utilizing AI. The session recognized that everyone, including children, needs to understand AI and its potential applications.

Transparency and explainability in AI regulation were explored. Although it was acknowledged that achieving full transparency and explainability with current technology presents challenges, it was considered a critical aspect of AI regulation. The session also highlighted the need to address issues such as discrimination, data governance, and other ethical considerations.

A global set of core principles and a multi-stakeholder approach were proposed as essential for effective AI regulation. The session emphasized that humans should ultimately remain in control of AI, and different regions may need to apply these principles in different ways, considering their societal contexts.

Data governance’s significance in AI development was emphasized, notably through the implementation of data governance policies like the EU Data Governance Act. This is seen as crucial to prevent monopolies on training data and ensure equal access to AI technology. Access to necessary data for training AI systems was considered fundamental for emerging nations and small-medium enterprises.

The EU’s risk-based approach to regulating AI, where systems with higher risks face more stringent regulations, was discussed. However, doubts were raised about the effectiveness of this approach. Some participants argued that regulations should focus on the applications using AI rather than the technology itself, due to the diverse nature of AI and its various use cases.

Flexibility was identified as a key factor in addressing potential risks and applications of technology. Given the uncertainty surrounding future developments, a flexible regulatory framework would enable effective adaptation and response.

Regulating technology before its development was recognized as having limitations. While the EU positions itself as a technology regulation leader, the specific form of regulation for the next decade remains uncertain.

In conclusion, the EuroDIG session emphasized the importance of AI literacy, global principles, data governance, and equal access in AI regulation. Challenges in achieving transparency and explainability were acknowledged, advocating for a flexible approach to address potential future risks. The session highlighted a multi-stakeholder approach to ensure effective and responsible use of AI technologies.

Tanara Lauschner

The analysis provides a comprehensive overview of different perspectives on artificial intelligence (AI) and its impact on social dynamics. It highlights both the positive and negative aspects of AI and emphasizes the need for responsible development and governance.

One viewpoint acknowledges the transformative potential of AI and its ability to generate novel content and integrate diverse ideas. This is exemplified by the outstanding capabilities demonstrated by large language models. These models can merge and connect ideas, leading to the creation of innovative concepts. The argument here is that AI holds great promise for various social dynamics.

On the other hand, challenges in AI are also identified. One such challenge is the manifestation of unintended behaviors and biases in large language models. Their decisions often lack transparent explanations. The argument presented here is that there is a need for interpretability and control in AI systems.

The importance of multi-stakeholder discussions in AI policy and governance is emphasized. The establishment of the Brazilian Artificial Intelligence Strategy under the Ministry of Science, Technology, and Innovations is highlighted as an example. The argument made is that involving all stakeholders in collaborative discussions allows for the sharing of ideas and the formulation of consensus policies.

Preserving and leveraging the internet and digital ecosystem as a catalyst for innovation is recognized as crucial. However, the analysis does not provide specific evidence or supporting facts for this viewpoint.

Another important aspect highlighted in the analysis is the ethical considerations in AI development. It is argued that AI should focus on responsibility, fairness, equality, and creating opportunities for all. Unfortunately, no supporting facts are provided to substantiate this perspective.

Offline AI applications are identified as a valuable solution for areas with sporadic internet availability. Language translation tools, health diagnostic apps, and AI-driven services hosted by community centers in remote areas are cited as examples. This viewpoint highlights the positive impact of AI in bridging the digital divide.

However, financial and accessibility barriers are acknowledged as limiting factors for the utilization of AI technologies. The argument presented is that people without internet access or financial means may not be able to benefit from AI solutions.

The analysis also emphasizes the necessity of community-driven governance for safe AI. The Brazilian Internet Governance Forum (IGF) and Lusophone IGF are highlighted as platforms where discussions on AI topics have taken place. The argument made is that community-driven governance ensures the safety and responsible implementation of AI technologies.

International cooperation is identified as a critical requirement for ensuring the inclusivity of AI. The fostering of debates and actions within the Brazilian IGF community and among National and Regional Internet Governance Forums (NRIs) is seen as essential in progressing towards this goal.

The need for trustworthiness and understandability of AI systems is emphasized. The argument put forward is that for AI to be trusted, it is necessary to understand how these systems work, what they do, what they don’t do, and what we don’t want them to do.

In conclusion, the analysis presents a balanced understanding of the impact of AI on social dynamics. While acknowledging its transformative potential, it also highlights challenges such as bias and lack of transparency. The analysis advocates for responsible development, multi-stakeholder discussions, and community-driven governance to ensure the ethical and inclusive implementation of AI technologies. Trustworthiness, understandability, and human well-being are identified as crucial considerations in the use of AI.

Jennifer Chung

During the meeting, the importance and potential of artificial intelligence (AI) were extensively discussed by the speakers. It was acknowledged that AI plays a crucial role in societal development and offers transformative opportunities across various sectors. The Prime Minister of Japan, Kishida-san, specifically highlighted the significance of AI in his speech. The importance of effective governance and regulations to ensure accountability, transparency, and fairness in the development and use of AI systems was emphasized.

To address this, policies, regulations, and ethical frameworks were deemed necessary to guide the development of AI and ensure its responsible deployment. The aim is to establish guidelines that prioritize human values and rights while avoiding any negative consequences. Secretary General Antonio Guterres established a high-level advisory body on artificial intelligence to further support these efforts.

An interesting aspect discussed was the involvement of various regions in AI discussions and policy recommendations. Representatives from regional IGFs in Tanzania, Germany, Panama, Colombia, Brazil, and India actively participated in the discussions. Each regional IGF highlighted different topics and priorities that are important in their respective jurisdictions and home regions, reflecting the diverse perspectives and challenges faced globally.

Furthermore, the potential of AI to address societal challenges was extensively emphasized. The speakers highlighted that AI has the capacity to drive economic growth and can tackle various challenges ranging from healthcare and education to transportation and energy. This highlights the potential of AI to contribute to the achievement of Sustainable Development Goals (SDGs) such as SDG 3 (Good Health and Well-being), SDG 4 (Quality Education), and SDG 7 (Affordable and Clean Energy).

The human-centric approach to AI was another significant point of discussion. It was stressed that AI should aim to benefit society as a whole and avoid increasing the digital divide, especially for rural and indigenous populations. It was also highlighted that careful consideration of language barriers and cultural sensitivities is crucial in the design and implementation of AI technologies.

The importance of multi-stakeholder collaboration was emphasized in the development of AI regulations. The speakers recognized that addressing the complexities of AI requires input from various stakeholders, including government, industry, academia, and civil society. Collaboration and dialogue among these stakeholders are crucial for creating robust and inclusive regulatory frameworks.

Efforts to coordinate in the development of AI regulatory frameworks were deemed essential. It was suggested that instead of reinventing the wheel, existing good practices and frameworks should be utilized and built upon. This highlights the importance of avoiding duplication and ensuring efforts are channeled in the most effective manner.

The meeting also brought attention to the significance of awareness and capacity building around AI. Speakers stressed the need to educate and build knowledge around AI, as it is a tool that can greatly improve human societies. Digital literacy and AI literacy were identified as crucial components in the successful implementation and adoption of AI.

In conclusion, the meeting underscored the importance of AI in societal development and how it can address various challenges. It highlighted the need for effective governance, regulations, and ethical frameworks to guide the development and deployment of AI. Collaboration among different regions and stakeholders is essential for creating inclusive and comprehensive policies. Additionally, the meeting emphasized the significance of awareness and capacity building to ensure the successful integration of AI into society. The European Union was commended for its advanced approach to AI regulation and risk management. Overall, the discussions emphasized that a responsible and human-centric approach is vital to harness the full potential of AI for the benefit of all.

Victor Lopez Cabrera

According to the information provided, Latin America is predicted to become an ageing society by 2053, with the number of individuals aged 60 and above surpassing other age groups. This demographic shift highlights the necessity for enhanced healthcare provision for the elderly. AI-driven automation in healthcare is identified as a potential ally in addressing this challenge, especially in regions where human resources are limited. The use of AI technologies has the potential to improve healthcare services and contribute to better outcomes for the elderly population.

Furthermore, AI is also seen as instrumental in the field of education, particularly in promoting intergenerational learning. One example mentioned is a pilot project in Panama, where seniors and younger tech aides worked together, encouraging joint exploration of technology usage. AI has the capacity to facilitate education and bridge the gap between different generations.

While AI presents numerous opportunities, it should not replace genuine human connections. Interactions with AI, regardless of their sophistication, cannot fully substitute the importance of human interaction and connection. This perspective emphasizes the need to balance the use of AI technology with maintaining genuine human relationships.

Another significant consideration is the protection of data privacy. In an age characterised by concerns about data privacy, it is crucial for citizens to have the necessary AI literacy to discern when to share personal data and when to abstain. The responsible usage of AI should not compromise individual privacy, highlighting the importance of a balanced approach in AI implementation.

The responsible management of AI technology is deemed essential. As AI applications and methodologies continue to evolve rapidly, ensuring trust and maintaining ethical practices in the development, deployment, and use of AI systems becomes crucial. The implementation of a trust-certified system is proposed, as trust lies at the heart of the data sharing dynamic.

The analysis also suggests that the Global South should contribute more data to AI algorithms. It is stated that the biological markers of elderly people vary according to geographical and cultural context, and such diversity is not adequately represented in current datasets. The inclusion of locally specific data is deemed necessary to address complications in diagnostic tools and improve the efficacy of AI applications in these regions.

Efforts to address the digital divide are viewed as a shared responsibility. In Panama, for instance, despite being a small country, significant gaps in digital device ownership and internet access persist. Organisations such as the Internet Society are working towards establishing community networks for indigenous people, aiming to bridge the digital divide and promote equal access to digital resources.

Upskilling and reskilling in AI are highlighted as crucial for individuals and communities. However, this process should not solely focus on developing technical capabilities but also on the development of soft skills and humanity. The dynamic interaction between freshman students and elderly individuals, as observed in the speaker’s experience, was found to offer valuable opportunities for both technical and personal growth.

The importance of explainability in AI systems is also emphasized. The speaker suggests that if a computer cannot explain its behaviour, people will not trust it. Therefore, achieving explainability in AI applications is crucial to enhance trust and adoption of AI technologies.

Regarding Large Language Models (LLMs), they are viewed with some degree of scepticism due to their complex mathematical and computational nature. They are often considered as black boxes, lacking transparency in how they arrive at their outputs. Thus, research organisations and academic institutions are called upon to play a role in helping citizens understand AI and navigate its complexities.

In conclusion, while AI presents numerous opportunities in various sectors, it is important to approach its implementation with caution and responsibility. Genuine human connections should not be replaced by AI, data privacy should be safeguarded, and AI technology should be managed responsibly. Furthermore, the inclusion of diverse data, efforts to bridge the digital divide, and the development of holistic skills in AI education are essential for a balanced and equitable AI-driven future.

Umut Pajaro Velasquez

The discussions on AI governance stressed the importance of adopting a transversal approach that includes perspectives from the global south and promotes meaningful partnerships. It was argued that the majority of the world’s population resides in the global south, and therefore, their insights and contributions are crucial to creating a resilient governance framework. By incorporating diverse perspectives, AI governance can avoid being monopolised by the global north and instead reflect the needs and aspirations of a global community.

Another key point discussed was the integration of a human rights approach into each phase of the AI life cycle. It was emphasised that AI systems should prioritise and safeguard the rights of under-represented communities, such as black, indigenous people, people of colour, and LGBTQI+ communities. By embedding human rights principles into the design, development, and implementation of AI, it is possible to mitigate the perpetuation of existing inequalities and ensure equal access and opportunities for all.

The participants in the discussions also recognised that AI regulation should be a collective responsibility of all stakeholders. True progress in AI development and implementation requires collaboration and cooperation among governments, the private sector, academia, and other relevant actors. The involvement of multiple stakeholders ensures that a wide range of perspectives and expertise are considered, allowing for more comprehensive and effective regulation.

Concerns were also raised regarding the environmental impact of AI and emerging technologies. It was highlighted that as AI becomes more widely implemented, it is necessary to address the consequences it may have on our environment. Finding ways to minimise AI’s carbon footprint and ecological impact will be essential for achieving sustainable development and combating climate change.

Another important aspect that was discussed is the need to review and strengthen data protection rules for AI technologies and future developments such as quantum computing. Data plays a central role on the internet and in internet governance. Therefore, ensuring robust data protection rules, privacy measures, and accountability mechanisms are in place is crucial to build trust and maintain the integrity of AI technologies.

The discussions underscored the significance of providing access to AI tools and technologies for youth. It was noted that AI can be a valuable educational resource that engages young people and provides them with hands-on experiences in real-life applications. By offering the youth access to AI tools, we can nurture their skills and empower them to contribute actively to the future of AI and technology.

Furthermore, community efforts were deemed fundamental in AI learning and development. Acknowledging that AI is a collaborative field, suggestions were made to organise events such as hackathons where diverse individuals can come together to understand, improve, and democratise AI tools. By fostering a sense of community, AI can be developed in a more inclusive and equitable manner.

The discussions also emphasised the importance of continuing the dialogue and sharing insights from AI developments in different regions and nations. These discussions provide opportunities to better understand the real-world applications of AI and facilitate the formulation of improved governmental policies that address societal challenges effectively.

It was argued that to ensure the ethical and responsible use of AI, there should be a commitment to sharing AI discussions with governments and the final users. Open discussions and involvement of AI users can lead to more informed decision-making processes and a more human-centric approach. By prioritising the needs and values of users, AI systems can better serve society and complement human well-being.

In conclusion, the discussions on AI governance highlighted the need for a transversal approach that incorporates perspectives from the global south and promotes meaningful partnerships. The integration of a human rights approach into the AI life cycle, the collective responsibility of all stakeholders in AI regulation, and the consideration of the environmental impact of AI were also emphasised. Strengthening data protection rules, providing access to AI tools for youth, fostering community efforts, sharing insights, involving governments and final users, and prioritising human-centred AI were all identified as essential components for the responsible and beneficial development and use of AI technologies.

Speakers

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more

Speech speed

0 words per minute

Speech length

words

Speech time

0 secs

Click for more