How nonprofits are using AI-based innovations to scale their impact
20 Feb 2026 14:00h - 15:00h
How nonprofits are using AI-based innovations to scale their impact
Session at a glance
Summary
This panel discussion focused on an AI cohort program for nonprofits that ran from September to December, anchored by Project Tech for Dev and funded by Agency Fund. The program brought together seven nonprofits to develop AI solutions for their specific use cases, with three education-focused organizations represented on the panel. Erica Arya from Project Tech for Dev explained that the cohort model was designed to provide mentorship, remove barriers like funding and technical resources, and facilitate peer learning among participating organizations.
The three nonprofit representatives shared their AI implementations in education. Simple Education Foundation developed a WhatsApp-based chatbot called Simple Teacher Buddy to help teachers identify appropriate pedagogical strategies for their classrooms. Quest Alliance created an AI system to provide personalized career guidance to learners aged 8-25, addressing the heterogeneous nature of classrooms and teacher shortages. Avanti Fellows built an AI tool that generates mentorship scripts for teachers based on student performance data, helping guide conversations with their 200,000 students, 98% of whom are online learners.
Tamina Madon from Agency Fund highlighted how the cohort model mirrors successful Silicon Valley incubators, providing shared resources like technical staff and product management support. She emphasized that nonprofits face unique challenges in building products for excluded populations while measuring social impact through comprehensive evaluation frameworks. The program incorporated responsible AI principles and safety considerations from the beginning, with knowledge partners Digital Future Labs and Tatl providing expertise in these areas.
Participants shared key learnings, emphasizing the importance of identifying existing pain points rather than seeking AI applications, avoiding reinventing solutions, and building in responsible AI practices early. The collaborative nature of the program led to unexpected partnerships, with some organizations discovering they were solving identical problems and choosing to work together rather than compete.
Keypoints
Overall Purpose/Goal
This panel discussion focused on sharing experiences and learnings from a 4-month AI cohort program run by Project Tech for Dev from September to December. The program brought together seven nonprofits to develop AI use cases with mentorship support, funding from Agency Fund, and knowledge partners focused on responsible AI and AI safety. The goal was to discuss the program’s genesis, structure, real-world experiences, and provide actionable takeaways for other nonprofits considering AI adoption.
Major Discussion Points
– Program Structure and Motivation: The cohort model was designed to address barriers nonprofits face in AI adoption, including lack of funding, engineering resources, and technical expertise. The program provided mentorship, peer learning opportunities, and pooled resources rather than individual funding to create a supportive community for organizations taking their first steps with AI.
– Real-World AI Use Cases in Education: Three education-focused nonprofits shared their specific implementations – Simple Education Foundation built a WhatsApp chatbot to help teachers with pedagogical strategies, Quest Alliance created an AI system for personalized career guidance for learners, and Avanti Fellows developed AI-generated scripts to guide teacher-student mentorship conversations and provide feedback to online learners.
– Implementation Challenges and Learnings: Panelists discussed practical obstacles including user behavior not matching assumptions (teachers not starting conversations with “hi”), AI hallucinations creating trust issues with teachers, the complexity of building behavior-change systems rather than just information dissemination tools, and the need for extensive prompt engineering and guardrails.
– Responsible AI and Safety Integration: Unlike typical software development where quality comes later, this program embedded AI safety and responsible AI principles from the beginning through knowledge partners. This included implementing guardrails, slur filters, and evaluation frameworks to ensure ethical deployment, particularly important when working with vulnerable populations.
– Collaboration and Knowledge Sharing: The cohort model enabled unprecedented collaboration between nonprofits, including two health organizations discovering they were building identical predictive models for high-risk pregnancies and deciding to work together. This level of early-stage collaboration is rare in the corporate world but proved valuable for avoiding duplication and sharing learnings.
Overall Tone
The discussion maintained a consistently collaborative and reflective tone throughout. Panelists were candid about both successes and failures, creating an atmosphere of shared learning rather than promotional presentations. The tone was practical and grounded, focusing on real implementation challenges rather than theoretical possibilities. There was a strong emphasis on responsible development and putting user needs first, with speakers frequently cautioning against adopting AI just because it’s trendy. The conversation remained constructive and solution-oriented, with participants building on each other’s insights and offering concrete advice for other nonprofits.
Speakers
Speakers from the provided list:
– Manohar Sreekanth – Partner and CTO at Satva Consulting, works at the intersection of tech, data, and AI, creating solutions that help social sector organizations scale their impact. Session moderator.
– Elizabeth – AI assistant used to introduce the panelists during the session
– Erica Arya – CEO of Project Tech for Dev, which collaborates across the global south to help non-profits use tech and data at scale. Under her leadership, they’ve supported over 200 NGOs with open source platforms and advisory.
– Temina Madon – Co-founder of the Agency Fund, works at the intersection of technology, philanthropy and economic mobility. She invests in services that empower human agency and create social impact at scale.
– Mainak Roy – Co-founder and CEO of Simple Education Foundation, partners with state governments across India to enhance teaching and learning at scale. Also exploring how AI can responsibly support teachers and public education. Background as a teacher.
– Steven Suting – Director of Technology and Product at Quest Alliance, leads AI-powered digital learning platforms that reach youth across India using data-driven innovation to improve learning outcomes.
– Pritam Sukumar – Technology and Research Leader at Avanti Fellows, drives tech and research efforts to use data and digital platforms to uplift public school students’ life outcomes across India.
Additional speakers:
None identified – all speakers in the transcript were included in the provided speakers names list.
Full session report
This panel discussion provided an in-depth examination of an innovative AI cohort programme for nonprofits, offering valuable insights into collaborative approaches to artificial intelligence adoption in the social sector. The four-month programme was anchored by Project Tech for Dev and funded by Agency Fund, bringing together seven nonprofits to develop AI solutions for their specific use cases whilst fostering peer learning and collaboration.
Programme Genesis and Structure
The cohort model emerged from Project Tech for Dev’s extensive experience working with more than 200 nonprofits in India. Erica Arya explained that whilst their organisation provides open-source platforms and technical advisory services, they recognised the need for more intensive, hands-on support to help nonprofits take their first steps towards AI adoption. Drawing inspiration from their previous data catalyst programme with Dasra, they identified that cohort-based approaches enable deeper learning and create valuable feedback loops that inform platform development.
This programme specifically targeted nonprofits wanting to take their first step into AI, distinguishing it from Agency Fund’s other initiatives that work with nonprofits already advanced in their technology and AI journey. The programme structure was carefully designed to address multiple barriers simultaneously. Rather than providing individual funding to separate organisations, the cohort model pooled resources including mentorship, technical expertise, and knowledge partnerships. This approach proved particularly valuable given that many nonprofits lack dedicated engineering resources or the funding to hire full-time AI specialists.
Tamina Madon from Agency Fund brought a Silicon Valley perspective to the initiative, drawing parallels with successful incubators like Y Combinator. An engineer by training who never pursued that career, she brought 20 years of experience and her background at South Park Commons, a venture capital fund and tech community, to inform the understanding that early-stage technology development is inherently challenging and benefits significantly from peer support. The Agency Fund’s approach of providing pooled technical resources—including AI engineers and product managers—addressed the reality that most nonprofits aren’t ready to hire full-time specialists but need part-time expertise during the development phase.
Real-World AI Implementations in Education
Three education-focused nonprofits shared their specific AI implementations, revealing both the diversity of approaches and common challenges in the sector. Simple Education Foundation, led by Mainak Roy and mentored by Manohar, developed the Simple Teacher Buddy, a WhatsApp-based chatbot designed to help teachers identify appropriate pedagogical strategies. The choice of WhatsApp as the platform was strategic, leveraging the widespread adoption of the messaging service. As Manohar observed, WhatsApp usage is so ubiquitous that “even my domestic help uses WhatsApp,” making it accessible even in remote areas. The system allows teachers to describe their classroom context and teaching objectives, then provides evidence-based strategies that have proven effective in similar situations.
Quest Alliance, represented by Steven Suting, tackled the challenge of personalised learning for young people from grade 8 to class 12 and those aged 16-25 in the TVET (Technical and Vocational Education and Training) ecosystem. Their AI system addresses the reality that these learners have vastly different learning styles, confidence levels, and access to opportunities, yet are typically served by uniform “chalk and talk” educational approaches. With teacher-to-student ratios often exceeding 1:30, their solution aims to provide personalised career guidance that meets learners where they are, rather than applying one-size-fits-all approaches.
Avanti Fellows, through Pritam Sukumar, developed an AI tool that generates mentorship scripts for teachers based on comprehensive student performance data. Serving approximately 200,000 students, 98% of whom are online learners, the organisation recognised that their remote students lacked the personalised guidance available to those in physical schools. The AI system analyses test performance, attendance patterns, strengths, and weaknesses to create actionable scripts that guide teacher-student conversations and provide structured mentorship for online learners.
Implementation Challenges and Practical Learnings
The real-world deployment of these AI solutions revealed significant gaps between theoretical design and practical implementation. Simple Education Foundation encountered what seemed like a simple but revealing challenge: teachers wouldn’t begin conversations with “hi” as designed, instead immediately launching into their problems. This caused the system to “glitch,” requiring fundamental redesign to recognise any initial input as a conversation starter. This highlighted how user behaviour often defies designer assumptions.
Avanti Fellows faced more complex challenges with AI hallucinations that, whilst maintaining numerical accuracy, would reverse directional indicators—saying “decreased” instead of “increased.” This created trust issues among teachers who, whilst appreciating the time-saving benefits, couldn’t rely completely on the AI-generated content. The organisation found themselves in extensive prompt engineering efforts, with Pritam describing large language models as “black boxes” where the relationship between input changes and desired outputs remained unclear. To address this, they developed a “golden set” to ensure AI response alignment with expected outputs.
Quest Alliance took perhaps the most ambitious approach, recognising that their challenge wasn’t purely technical but fundamentally about behaviour science. Steven Suting explained that creating systems capable of behaviour change required understanding human psychology and conversation dynamics, not just software engineering. Their focus shifted from building an information dissemination system to creating something that could emulate human teachers by picking up conversational signals and responding appropriately.
Responsible AI and Safety Integration
One of the programme’s most distinctive features was the integration of responsible AI principles and safety considerations from the outset, rather than treating them as afterthoughts. This was achieved through partnerships with knowledge organisations: Digital Future Labs provided expertise in responsible AI principles, whilst Tatl contributed AI safety knowledge. This approach contrasted sharply with typical software development practices where quality and safety measures are often addressed after core functionality is built.
The practical implementation of these principles included the development of guardrails and evaluation frameworks. Participants learned about plugins that could be integrated into their systems to prevent inappropriate responses, including slur filters and safety measures. The emphasis on building these protections into the design process proved valuable, as several organisations implemented guardrails that prevented their chatbots from responding to inappropriate or off-topic queries.
Collaboration and Knowledge Sharing
The collaborative aspect of the programme yielded unexpected benefits that extended beyond the immediate cohort. Two health-focused nonprofits discovered they were building identical predictive models for high-risk pregnancies and chose to collaborate rather than duplicate efforts. Similarly, when Project Tech for Dev identified three organisations working on assessment models for evaluating student answer sheets, they facilitated connections that led to shared development of rubrics and approaches.
This level of early-stage collaboration proved remarkable when contrasted with typical corporate behaviour. As Manohar observed, having spent “a couple of decades in the corporate side,” this kind of collaboration at an early stage of technology development is “very, very rare.” The nonprofit context, combined with the cohort structure, created an environment where organisations felt comfortable sharing both successes and failures, leading to accelerated learning and reduced duplication of effort.
Evaluation Framework and Impact Measurement
Tamina Madon shared Agency Fund’s sophisticated four-level evaluation framework that distinguishes social sector AI development from commercial applications. The framework progresses from user experience evaluation (basic usability and satisfaction) through user behaviour evaluation (actual usage patterns and engagement) to user evaluation (measuring changes in confidence, agency, and self-efficacy) and finally impact evaluation (long-term outcomes in health, education, or livelihood).
This comprehensive approach reflects the fundamental difference between nonprofit and for-profit AI applications. Whilst commercial applications primarily need to generate revenue from existing customers, social sector applications must create meaningful change for excluded populations—a significantly more complex challenge. The framework requires extensive survey data collection and long-term tracking that goes far beyond typical business metrics, reflecting the sector’s commitment to genuine impact rather than just user engagement.
Key Learnings and Strategic Recommendations
The panellists’ collective experience yielded several crucial insights for nonprofit AI adoption. The most fundamental lesson was to focus on existing organisational pain points rather than seeking applications for AI technology. Pritam Sukumar emphasised looking at operational troubles and inefficiencies first, then evaluating whether AI could address these problems, rather than starting with AI and searching for applications.
Mainak Roy’s advice to “build what is needed, not what is sexy” resonated throughout the discussion, reflecting the tendency for organisations to pursue innovative-sounding solutions rather than addressing genuine user needs. This insight came from lived experience—his organisation had pivoted from their original use case to something more practical based on actual teacher requirements.
Steven Suting provided a broader sectoral critique, noting that the AI bandwagon effect had led many nonprofits to lead with solutions rather than problems. He highlighted the systemic issue where multiple organisations build separate solutions for the same beneficiaries, citing the example of “multiple people building an app for ASHA workers and they have to interface with multiple apps themselves,” creating fragmented experiences despite claims of putting beneficiaries at the centre of design.
The discussion also emphasised the importance of leveraging existing solutions before building custom systems. Erica Arya recommended starting with open-source platforms that meet 60-70% of requirements, then iterating and customising as needed, rather than building from scratch. This approach conserves resources and allows organisations to focus on their unique value-add rather than recreating common functionality.
Broader Implications and Future Directions
The programme’s success in fostering collaboration and responsible AI development has implications beyond the immediate participants. During the discussion, when Manohar asked audience members from nonprofits to raise their hands, “quite a few” responded, indicating significant interest in the sector for such approaches. The model demonstrates that structured peer learning can accelerate AI adoption whilst maintaining ethical standards and avoiding duplication of effort.
However, several challenges remain unresolved. The technical issue of AI hallucinations continues to affect user trust, particularly in educational contexts where accuracy is critical. The broader systemic problem of multiple organisations building similar solutions for the same beneficiaries requires sector-wide coordination that extends beyond individual programmes.
The discussion revealed that successful AI implementation in the social sector requires fundamentally different approaches from commercial applications. The complexity of serving excluded populations, combined with the need for comprehensive impact measurement, creates unique challenges that require specialised expertise and collaborative approaches. The cohort model, with its emphasis on peer learning, responsible development, and shared resources, offers a promising framework for addressing these challenges whilst building the sector’s collective capacity for ethical AI adoption.
This comprehensive examination of the AI cohort programme provides valuable insights for nonprofits considering AI adoption, funders designing support programmes, and the broader social sector seeking to harness artificial intelligence for social impact whilst maintaining ethical standards and avoiding common pitfalls.
Session transcript
Not sure why exactly you came here, but I’m glad you’re here. We’ll make it worthwhile. So my name is Manohar, and I’ll be moderating this session. Usually what happens is in the beginning of such a panel discussion, you know, we tend to introduce the panelists. And since we are at an AI summit, so I thought, and we are all in the habit of delegating things to AI these days, I thought, why not let the AI itself introduce some of the panelists, right? We don’t know what they’re going to say. Hello, Elizabeth. How are you doing today?
Hi, Manohar. I’m doing great and ready to help. What’s on your mind today?
You couldn’t hear that, right? One second. Can you speak a little louder, Elizabeth?
Of course, I can definitely speak up for you. Just let me know what you need, and I’ll match your pace.
All right. No, right? Can’t hear. Let’s try again. And I can hear you. Usually this should work.
Yes, absolutely. This usually works just fine. If there’s anything specific you want to try again or clarify, just let me know.
Okay, Elizabeth, I am sitting in front of a panel right now and I need your help to introduce the panel members. Can you help me do that?
Absolutely. I’ve got their profiles right here. Just let me know how you want to frame it and I can help you introduce them smoothly.
Okay, let’s start with Tamina.
So, introducing Tamina Madon. She’s the co -founder of the agency fund. Tamina works right where technology, philanthropy and economic mobility. She invests in services that empower human agency. create social impact at scale
all right thank you next up is erica
next we have erica aria she’s the ceo of project tech for dev which collaborates across the global south to help non -profits use tech and data at scale under her leadership they’ve supported over 200 ngos with open source platforms and advisory
very nice thank you next we have my neck
now we have my neck roy he’s the co -founder and ceo of simple education foundation my neck partners with state governments across india to enhance teaching and learning at scale he’s also exploring how ai can responsibly support teachers and public education
thank you next up is steven
up next is steven suiting you he’s the director of technology and product at quest alliance steven leads ai powered digital learning platforms that reach youth across india using data -driven innovation to improve learning outcomes.
Thank you. And then we have Preetam Sukumar.
And he is Preetam Sukumar. He’s the technology and research leader of Antifellows. Preetam drives tech and research efforts to use data and digital platforms to uplift public school students’ life outcomes across India.
Excellent. And now finally we have this chap moderating this session. His name is Manohar. Can you tell me something about him?
And finally our moderator, Manohar Srikanth. He’s partner and CTO at Satva Consulting. Manohar works at the intersection of tech, data, and AI, creating solutions that help social sector organizations scale their impact.
Excellent, Elizabeth. I think you did a stellar job. Thank you so much. You wanted to do the panel for us. Next year, I’m sure, we’ll have some. digital assistant sitting here. But anyway, I hope that was interesting and it did a good job as well. So the main theme for this discussion is AI cohort program that was run from September to December. It was about a four -month program, which was actually anchored by Project Tech for Dev. And about seven nonprofits built their AI use cases in the course of these four months. So we’re going to talk about the genesis of the program, why it was structured that way, and also some of the learnings and experiences.
And finally, depending on, I’m not sure how many NGOs are there, but if you want some takeaways on how a nonprofit can also take away approaches towards AI, that’s also something which we’ll cover. I’m going to begin with Erika, because Project Tech for Dev was the entity which came up with the idea for such a program. Erika, can you give us a brief overview? of maybe start a little bit about Tech4Dev itself, right? And then why did you think about such a program in a cohort model, right? And what are the genesis, what are the motivation for this? And something about the structure also.
Thanks, Manoj. So, good morning everyone. Hello. So, good morning everyone. The AI did a bit of introduction for me. And I’ll just add a little human element to it. So, yes, we at Project Tech4Dev, we develop open source tech platforms for the social sector. That’s one part of our work. And in addition, we do a lot of tech advisory for the non -profits. And as we had been working with, you know, more than 200 non -profits in India mostly, what we realized was that yes, we have these open source platforms which nonprofits come, they can use it as a SaaS platform, and we also provide consulting. And through our tech advisory, what we learned was that there are times when, you know, you really need to handhold nonprofits in taking that first step towards technology or those who have already taken that step to help them move to that next level.
And we have in past, you know, run a data catalyst program also, which was in collaboration with Dasra, and I’m just assuming you all would know this name, in the social sector. And they’ve been running these cohort -based programs, and we felt that, you know, when we do these deep -rooted programs with a small set of NGOs, we’re able to learn a lot from them. And that actually… That actually feeds into, you know, the platforms that we built. Because the way we… our platforms is not like a top -down, but it is always listening from the ground what the non -profits need, and I feel these cohort -based programs, at least for Tech4Dev, enables us to get those learnings.
In addition to that, you know, even this small cohort that comes together, there is a lot of peer learning and sharing that happens. It helps us see the commonalities across the use cases the NGOs are working on, and it helps them to further collaborate on them. So why do you rebuild things? Or if we are building new things, can we make it for the sector? Can we reuse the intellectual property that we have, you know, working with one or two other partners in the cohort? How can we leverage that and make it available more for the sector? So learning from our data catalyst program, which was a cohort -based program, and a program that agency funded, was also running on AI.
and it was called AI for Global Development, we felt that maybe while agency fund program was working more with the nonprofits who were much ahead in their journey in technology and in using AI, we felt that can we come down a level and work with those nonprofits who were wanting to take that step of integrating AI into their work. So they had a good use case, but maybe funding was a challenge, or maybe it was the resources were a challenge, because in the sector, especially in the NGO space, we know that having engineering resources is a very big challenge, and many nonprofits do not have that. So is that a barrier for them to take that step towards adopting AI?
So we wanted to bring down those barriers by introducing mentors into this program. So just… I’ll quickly tell you about the structure of the program. It was like we got… seven NGOs together. It was an open application. Screening was done. We had calls with the NGOs to really understand that their use case was really fitting the program. It was a free program for the NGOs because as Tech4Dev, what we believe, it was a pilot for us. So when we are running a pilot, we are learning together in it. And if we have the funds to encourage these NGOs to come together, then we should play that role. So there was a proper evaluation of the use case, also ensuring that the nonprofit has a resource who can devote time to running this for the period of the four months.
So who could anchor it within the organization. There was leadership buy -in into onboarding onto the program because for technology to get integrated, it is very important that the leadership buy -in is there because there is investment that Tech4Dev has. So I think that’s what we’re trying to do. the reason why we felt that can we, you know, just unlock these couple of barriers that these non -profits have and today we have three of such non -profits on this panel who are going to share with you, you know, what does this cohort enable for them. So I think, yeah.
Right. Thank you, Erica. I mean, you talked a lot about the goals and the intent. We will hear from the non -profits how much of that intent also was realized, I think. And you can be candid, I’m sure, all of you. Right. We’ll move on to Tamina now. Tamina from the agency fund. Do give a brief overview of agency fund itself. But also talk to us about the motivation of funding such a cohort model because, you know, typically funding you can give to specific non -profits themselves for the AI use cases. But this is a different model. What is the motivation for that? Maybe we
Sure. Thank you, everyone. I live in Silicon Valley and I started life as an engineer. I’ve never pursued that career. Over the last 20 years. One thing I saw in Silicon Valley is that a lot of new ventures are started in cohorts. Y Combinator is a quite famous incubator for profitable companies. I was working before we started the agency fund in South Park Commons, which is a venture capital fund and tech community. There’s one in Bangalore, one in San Francisco, one in New York City. And the way SPC operates is to bring founders together at the earliest stages in their journeys as they’re trying to figure out what to build and how to build it and who the user will be.
And that’s a very lonely journey. It is filled with frustrations and failures, and it helps to have a cohort of colleagues in the same stage of life or in the same stage of product development. And so we brought that philosophy into the agency fund when we started, that we bring together cohorts for accelerators. And we have participated in a lot of accelerators, including the data catalyst program that Tech4Dev has, which we have done. run the AI for Global Development Accelerator as well. The AI one is a year long for us and we bring established companies, non -profit companies together, but they’re all building AI for the first time. So it’s that same experience of needing to understand your user, figuring out what to build and going through the practice of putting together pipelines, experimenting, experimenting, iterating, refining.
It’s a journey and it’s difficult. And so I do think that the cohort support helps a lot. The other thing is as a funder, our perspective has been that while we don’t know everything, there are certain resources that are hard for non -profits to pay for themselves. And if we can create those resources in a pool format and make them available, we can perhaps be more efficient. For example, not every NGO is ready to hire two full -time AI engineers. Maybe they only need a part -time person right now as they’re building out the pipeline. So we’ve put together a pool of, of 10 technical staff who sit alongside the organizations that we fund in the accelerator.
We include product management in that as well. We’ve just added some additional people in product management because the AI technology is actually easy, especially there are a lot of talented engineers in India and elsewhere. What is difficult is to fit the technology to the pain points we all experience in life and build a product that achieves social impact. That’s really difficult. It’s so much further from what the typical Silicon Valley startup has to deal with. They just need to make money from people who already have money, right? We are trying to do something much more difficult with a population that’s much more excluded. And so, you know, there aren’t a lot of product managers in nonprofits.
If you’re in a nonprofit, raise your hand if you have a product manager. Oh, I’m surprised. Okay, there’s a decent number. But it’s not a role that commonly has been found in nonprofits. This is probably a… a special group. Some of them are from the organizations in the panel.
Okay, fair enough. Not counted. Anyway, thank you, Tamina. Alright, so that set the stage. So now you got a sense of why the organization actually, both these organizations thought of a cohort model, why they set it up and so on. Now let’s get into the trenches in terms of the experience of the non -profits themselves, right? So it would be good to start with the problem space, right? So what problems were you trying to solve in the context of this particular program? And at a high level, how did AI come into the picture and overview? We’ll get to the experiences and learnings a little later, but just to set the context of what problems because again, we should remember that AI is a means to an end, right?
So understanding the problem that you’re trying to solve and was it scale, was it inclusion, was it reach, right, in the context of your program would be good to understand. Let’s start with you, Myna Croy.
Thanks. I’m audible right? Yeah. Okay. So I’ll also add a bit to the intro that the AI did. So I’m also a teacher. And that’s how I started my journey into this space. And one of the things that as a teacher, I’d always struggled with is figuring out what is the right classroom pedagogical strategy that I need to use for a certain sort of class. And that often changed with the topics or the objectives that I was covering on the day. Right. And for a teacher, it’s often difficult when I’m doing eight lessons a day to sharpen my sort of focus and ensure that I get the perfect strategy. And that really was our sort of problem statement in many ways that can we build a platform or a tool which enables teachers to actually say that this is the class I’m teaching today.
This is the kind of classroom that I need. This is the kind of classroom that I have. And then can the. AI throw out something which is evidence based and has worked in other similar classes before and support the teacher so that’s what we wanted to do and in India I think if I asked how many of you use WhatsApp probably all of you will put your hands up and that’s the case for teachers in some of the most remote areas of the country as well so we chose WhatsApp as a platform and we used a chatbot kind of a mechanism to solve that problem for our teachers
ok thank you Manak for that intro so that’s Simple Education Foundation so they built this Simple Teacher Buddy we’ll come back to you a little later Stephen, Quest Alliance, so what was the problem you were trying to solve and high level overview of the solution
thank you it’s happening no? I’m not sure if I can yeah it’s happening so we work so we work with young people from grade 8 to class 12 and then younger people who work in the TVET who study in the TVET ecosystem who are between about 16 to 25 right And what we realized in the work that we did, and this wasn’t necessarily revolutionary, was that no two learners are the same, right? And they’re different in the way they think, in the way they learn, in the confidence levels that they have. And that’s one dimension of difference. And when you think about the demographic difference, their opportunity of access, the gender inequalities, and if you put all of this together, you have a system that’s a classroom that’s fairly heterogeneous.
These are not homogeneous systems, right? But unfortunately, the educational structure around this is a chalk and talk mechanism. It largely treats all the students in the same way, right? And as a result of that, you’re not really meeting the learner where they are. And so that’s one dimension. The other side of this problem is also that there’s a lot of stress with the availability of teachers. Because in general, there’s about a 1. 30 ratio. between student and teacher. And that’s in a good scenario, right? So what tends to happen is these students don’t necessarily have access to the information that they need. You’re not really meeting the learners at where they are. And so what we try to do or what we’re trying to do is really try and place a system that is able to hear the learners’ problems with respect to the area of career development.
Okay, so you’re directly targeting the learners, right? Whereas my next use case was targeting the teachers, right? So we have yet another education use case from Avanti Fellows. Preetam, tell us about it.
So hi, I’m Preetam from Avanti Fellows. And what Tamina said and what Stephen said kind of resonated because we struggled a lot to find a good use case for AI because we have a lot of pain points. Like our teachers have pain points, our students have pain points, our program managers have pain points, and they’re all spread across the country in remote locations. So we started with a use case to improve student report. So that because a lot of, I mean, we reach about 200 ,000 students and 98 % of those are online learners. And they don’t get useful feedback from the test report. They just get score and this and that. But whereas in our physical school, there is a teacher talking to them.
There is a mentor talking to them. So we thought we’ll improve the student report first, but that didn’t work out so well. So we, our mentor actually helped us a lot, Akilesh from Tech4Dev. And we iterated through a few use cases till we landed up on actually replacing the teacher -student mentor conversation with an AI summary that can be useful for our low -touch, I mean, our online students. So it kind of collects all the data of the student’s performance in the tests about how they’re doing, what, what, how their attendance is, what chapters they’re weak in, what chapters they’re strong in. And generates kind of a script. And for the teachers, the script is used to guide their conversation.
And for the online learners, the script can be used to like just give them a proper mentors, sort of mentorship guidance so that they have actionable things to do for their next test or whatever they’re doing.
Right, now let’s dig a little deeper. So now where are we on this journey, Pritam? And you know, how was the solution deployed? How many people are using it? Is it being in the pilot phase? Can you tell us more?
It’s, I think it’s somewhere between the pilot and the rollout. So we, around 15 teachers I think have had 57 or 75, 57 to 75 conversations with students with these scripts. And we are working on feedback because there are hallucinations as LLMs do. But the hallucinations in this case are a little weird. The numbers are correct but instead of saying increased it says decreased. So we are like, there is a lot of prompt engineering that we are doing to get this sort of fixed. And in the low touch, in the online program it is being, once the rollout is kind of ready and yet we are just waiting to fix this, these small issues. Okay.
Before rolling out.
Okay, and what’s been the response from the teachers themselves who are getting these reports?
The teachers are very happy. I mean, it saves them a lot of time. Like in one of the main feedbacks, one of the main points of feedback that we got from the teacher is that it saves them really a lot of time because before such a conversation, they would have to look at the last four reports of the student. They would ask, I don’t know, what is happening at home? Like, are you getting device access and all this? Now all this, the students fill a form before the mentorship conversation and the teacher, we have the test data and we give this all in a script to the teacher. So that has been the positive side.
On the negative side, the hallucinations really bother the teachers because there isn’t 100 % trust yet in what we’ve given them. And that we are still working through. I mean, the solution is not clear because to me, maybe we need more AI engineers like Tamina said. But because it’s not clear exactly what to do with these hallucinations because to me and us, the LLMs are like a bug. Black box. You just put something, you get it. something out and you don’t know what change will bring the change that you want. Yeah, that’s where we are now.
Right. Yeah, I guess the error rate, the hit ratio and what kind of an impact it has depends on the use case. And if the teachers can be trained to address some of this, it might be one way out. Mainak, you talked about WhatsApp being the channel, right? So that’s probably something which, you know, eases adoption because they don’t have to learn something new. It’s there. It’s part of the conversation. So now they’re getting some additional feedback and advice. But at the same time, and by the way, I know this because I was mentoring some of the folks here. So by the way, this whole program had a mentorship model where each of these nonprofits are hooked up with one or two mentors and those mentors guided the teams.
Right. So I happen to be one of the mentors for Simple Education. And we did face some interesting challenges also, even if it was simplified from the WhatsApp point of view. So what were some of those challenges on the ground? Right. Because when the rubber meets the road, that’s. when you see what I mean you would have imagined something every product person imagines and starts with a hypothesis but then needs to be validated right so what were some of those things that we learned
yeah I think there are a bunch of challenges but you know my favorite challenge is actually we wanted teachers to start the conversation with the bot with a simple hi but with the moment teachers would get the QR code and scan it they would start putting their problems in right and then the bot would glitch because the bot did not get that first hi and then the whole process of the data that we needed to collect to give the teacher very specific guidance was not happening so we ended up saying that whatever is the first conversation that the teacher does the bot sort of recognizes it as a hi and then starts the conversation so it although we’ve kept telling the teachers that started with a hi but it does not happen so we’ve sort of worked around it but yeah that’s been one of the challenges and then I think I think So we’ve seen a bit of also teachers figuring out what kind of questions they can ask.
So sometimes there’s been some straying away from the kind of questions we want the teachers to ask. And because it’s AI, they would still throw out some response. Right. So over there also, we’ve had to put some guardrails around what kind of conversations the chatbot will entertain. What are the conversations that we would completely sort of say no to. So, yeah. Right.
Thank you, Stephen. In your case, I think you had this slightly more ambitious plan for behavior change. Right. You’re looking at nudging these learners towards a different kind of behavior. And you had a certain approach to that. Can you tell us a little bit about it? And how successful has it been? I
think we’re still largely pie in the sky thinking right now. So we still hope to be ambitious through the entire course of the project and not be deterred by, how many users and, you know, do we have, how many sessions do we have? We’re trying to be steadfast with that. I think for us, the fundamental premise was that we didn’t want to create a system that was really an information dissemination system, right, because there’s enough and more of that. We wanted to create something that emulated a human being or emulated a teacher. And the intent over there was really to be able to pick up signals. If you imagine a conversation between a teacher and a student, there are multiple signals that a human being is able to pick up when they’re having a conversation, right?
Sometimes it can just be the tonality. It can be the speed at which they’re speaking. You can gauge confidence levels, right? But when you’re working with an interface like a bot, for example, you have limited access to the kind of signals that you can pick up. So then how do you decrypt conversation to be able to pick up these different signals? And then how do you use those signals to then create a response that intends to or is intended to emulate a human being? So we’re going to create a system that emulates a human being. right um and and interestingly so when we started thinking like that it stopped becoming a software problem it started becoming a behavior science problem right and so for us that was really the elevation of that situation and so right now we’ve got a plethora of like people who are specialists in that space we’ve got you know the engineering and software folks to help us with the scaffolding but the heart of the problem is really in understanding that human being right no
it’s taking it really to the next level that’s so nice um so tamina you’ve now heard some of these challenges experiences and so on and i think you’ve worked with previous um cohorts as well i want to get a sense of also how these indian non -profits you’ve been working with and engaging with how does this differ from the others probably you’ve engaged with you know in the u .s and abroad in terms of adoption in terms of maturity in terms of the kind of use cases they’re looking at can you just give us a lay of the land comparing the two
then going to learn it. It’s about my belief, my confidence in myself as a new mom to take care of my baby and know what caregiving I need to provide. It’s about my role as a farmer, feeling that I’m going to be able to experiment with a new input and feeling more confident or thinking maybe I’ll try livestock because now I have better information and support. So we call that the user evaluation. That requires survey data collection. Most companies don’t do many surveys. Maybe they buy some survey data from a firm, but all of these orgs, once they’re deploying to enough people, will have monitoring and evaluation systems that really try to understand the user.
And then the fourth motor we switch on, we call impact evaluation, and that’s when you have tens of thousands, hundreds of thousands of users, and you want to understand whether the product that they’re using is helping them make a change in their health outcomes, in their education or livelihood outcomes. So that’s what we call impact evaluation. And so we almost never see that fourth stage of evaluation in the private sector. That’s something we do in the social sector. But we think that it’s a nice way, this kind of four -level framework, to think about how to continuously learn from the product you’re deploying and then prove its impact to society and hopefully bring in more funding to scale.
Right. No, thank you, Tamina. In fact, if those of you who are interested in that framework, it’s a very powerful, interesting framework. The Search for Agency Fund Evaluation Framework, and I’m sure you’ll find that. It’s a very useful resource. Right. So, in fact, this is one such topic. Evaluation is one such orthogonal topic. But there were a couple of others as well. So, Erika, also from a program structure point of view, right, you people had some knowledge partners come in, right? So it was not just these nonprofits hooked up with mentors, but you also had some others. Can you tell us a little bit more about those knowledge partners, their roles, and so on, and then we’ll get into how they impact.
the work they were doing.
Thanks, Manohar. I think this is a very critical part of the entire program design. So Tech4Dev, while we are like a bunch of engineers, CTOs in the organization, but again, we know that the problem that we are trying to solve is not something that we can do alone. And what we try to do is, in whatever we do, we don’t want to reinvent the wheel. And that’s where and we know we don’t know it all. So we always work in our programs through collaborations and partnerships with people who have expertise in those areas so that we can get the best to the cohort. And so as part of this AI cohort too, we wanted, what was the overall goal of the program?
It was that the non -profits could unlock those barriers which was keeping them away from building and deploying. AI solutions to help them bring in more efficiency, help them scale and create more impact on the ground. But we wanted, as a technologist, we could say that, okay, we’re just building something, but are we building it in the right way? That’s what we wanted to do. And so we collaborated with two knowledge partners. One was Digital Future Labs, who brought in expertise around building, you know, on integrating responsible AI principles into your design as you are working on a use case. And the second knowledge partner was Statle, which works in AI safety. So we actually wanted that when the nonprofits are working as part of this cohort, they have that thinking of responsible AI and AI safety built into their design right from the get -go.
It’s not something that you need to think. Later, okay, let me just build something. And then I’ll come to, you know, responsible AI practices or think about AI safety, you know, a couple of after I’ve done my pilot or, you know, after a couple of months when actually I see things breaking. So we wanted to build that right from the get go. And that’s why we collaborated with partners because we knew we did not know this bit. We definitely contributed through, you know, what Manohar said. This was a mentorship program. So a lot of mentors came in from Project Tech for Dev because we knew that is one piece that we feel we do good at and collaborated with these partners.
And I hope that this helps the nonprofits in designing better solutions.
Right. No, that is a very unique part of this program, actually, because also anyone who’s in software would know that quality and all these things will look later. Let’s build and get the functionality out. That’s a typical attitude. And in the context of AI, AI safety, responsible AI seems to be the most important thing. to be an afterthought, but here it was baked into it. So now let’s hear from some of the non -profits and how did this influence the way you approached, you know, the building out this and how did AI safety or responsible AI kind of thinking influence and how did it actually help on the ground itself? Stephen, would you like to say something?
Sure. I think for us, it was supremely helpful, just primarily because, you know, the artifacts that these organizations helped put together in some ways were like a blueprint for us, right? And when you have a blueprint in front of you, you have a sense of what the different moving parts are. And what that allowed for us to do is to rapid prototype in a fairly modular way, which allowed us to fail fast, right? And if you were failing fast, we were learning faster, right? So yeah, I think. and I think what can typically happen in this space of human development because the dimensionality of problems can be fairly large and this phenomenon of analysis paralysis can happen and it can be overwhelming to decide what do you pick and choose because they all seem like really important problems but having frameworks like what Tamina was talking about helps you sort of chunk things into where is this problem best suited and where is it most relevant and then you solve for that in that section and that helps influence what you’re going to do in the next step and then you kind of shuttle back and forth till you’ve crossed a certain critical threshold of it making sense to the problem that you’re solving
One of the things I remember from the AI safety part was what Tatl came up with with these plugins where they have created a slur list a list of words which the AI is not supposed to use and that’s something which they have crowdsourced and so on and there are also some guardrails plugins from Lama and so on and And the fact that those things can be plugged into your bot to make sure that it’s behaving the right way was, I thought, was an interesting thing. Did any one of you use any of those things? Not yet, Right. For the simple, yes. I mean, I remember the guardrail. So the bot which Minak was talking about.
So there you could ask any question. And then the whole thinking process behind making sure that even if a random question comes. So let’s say a teacher asks, what do I do? It’s not a question the bot is supposed to answer. But how does it respond to that? It’s something we started thinking about and baked into the whole design itself. And that was used for testing. Okay. Now, one of the unique features of this program was also the whole collaboration piece. Because. they are not working in isolation. So we had in -person workshops where each of them actually came and presented, and they had sessions where, you see, all three of them are actually working on education use cases, and there are some others as well.
So how could they learn from each other? That’s one. The other thing is, and I think I’m sure you’ll recognize, you probably are also looking at similar problems in this space. How do you not reinvent the wheel? How can you leverage some of the work that’s been done by others? So this collaboration aspect also came in in a very strong way. So I want to understand from some of you, so whom else did you speak to? How did this play out as a part of the program? How did this program structure influence the way you went about it? And how was it in comparison to, let’s say, you were going about it by yourself? What was the difference?
Anyone? Anyone?
I think. One thing that was nice about the program design was that. by nature of the conversations that we were having, there was this sense of porousness that we had, that we weren’t holding our cards too close to our chest. It was okay for me to learn from what you were doing well and what you weren’t doing well. And if I had that conversation with you, it would elevate or it would sort of catalyze and expedite certain problems that we were trying to solve. For example, when we were thinking about how do you like if you have to do responsible, if you have to do classification, if you have to do behavior. So these are all different parts of a fairly large problem.
And we had a certain approach in mind. But through the cohort, we realized that people were trying to solve similar things, but they did it in a different way. And like, for example, we didn’t think about this through the genetic way. And so then when people started saying that this is what worked for us in that model of deployment, that in many ways kind of expedited the way we solved for that problem. And I think there are multiple examples like that. Also, the idea of responsible, for example, right? Like when you say responsible, it can be a whole gamut of what you mean by responsible. But what does responsible mean for a 25 -year -old versus, say, a grade 8 student?
It’s significantly different. Now, if you’re, you know, if you’re, let’s say, for example, an Avanti and you’re working with a certain cohort that is similar to your cohort, right? What are they doing with responsible that you want to do, right? And what is that conversation that we want to have? So I feel that was certainly helpful.
Interesting. Erica?
Yeah, I’ll just add in. So while these were like seven NGOs, out of which four were from the education and three are sitting here, there were two health NGOs also, and they were both working on building a predictive model to predict high -risk pregnancies. And they didn’t know about, like, you know, each one of them building something like that. So but when the cohort came together, they discussed. And then we felt, you know, the mentors also got together, and then that’s when they felt that we are building the same thing. So then are we happy to collaborate? And that’s when, you know, the future calls that happened was getting these two NGOs together and, you know, building on at least sharing and building it together was one thing.
Second, while we were working with these seven NGOs, one of them, which is not represented here, was working on an assessment model. How can we do assessments of, say, answer sheets which are coming from students because that takes in a lot of time of the teachers. And then we also, since we’ve been working with 200 nonprofits, we were also aware of two other nonprofits who were working in, you know, trying to work on similar solution. So just going beyond the cohort, we brought all three of them together, and we just discussed on, you know, the assessment models that they were building, the rubrics. And we were seeing that there was a lot of commonality. So could we build this together, learn from each other’s mistakes?
So I’m just trying to share here that the learning and sharing goes beyond the cohort also. And that is where, you know, platforms or organizations like us who are more ecosystem players have a role to play. So, yeah, I think these are like two such examples where I could see collaboration happening with just bringing a set of seven NGOs together in a room.
Right. No, thank you, Erica. Those are very interesting and powerful examples. Also building on what Teminat said earlier, contrasting this and the corporate space. I’ve spent a couple of decades in the corporate side, and this kind of collaboration at an early stage of technology is very, very rare. You don’t see that on the for -profit side. And so it’s fascinating to see this kind of thing. And, of course, we are in a room where I’m sure we’re working on similar kind of problems. And just to exchange notes is so, so important. All right. Just now wrapping up, we have about six and a half minutes. I’d like to understand how many of you in the audience are from nonprofits.
Can you raise your hands? OK, quite a few. All right. So now we have. keen non -profit listeners. So the question to the non -profits and also probably others over here is what are some of the learnings that you came away with which surprised you or which you didn’t know of before which could be useful for the non -profits in the audience as well as they go about their AI journeys. If you could share some of your nuggets. Pritam, please.
So I think one thing is that one thing we learned is to stop looking at use cases for AI but look at pain points and troubles we already have and see if AI is a good fit there. It doesn’t have to be like AI is a new technology, let’s put it and build something amazing and innovative. There’s already a lot of trouble in the way even our non -profit runs. There’s a lot of inefficiency, there’s a lot of stuff so you just, there is a lot of stuff that AI can do to help daily operations like our fundraising team write grants or our programmers write code which of course is the number one use case right now.
But yeah, that is one thing we really believe, at least I really believe now.
You know when AI started becoming really popular it almost became like knee jerk reactions for organizations to jump on that bandwagon and if you weren’t doing it then you couldn’t get into the right rooms with the right kind of funders and then you would be lost in that race right not always but more often than not there is that pressure and I think what has happened as a consequence of that is some part of our DNA was forgotten like I feel like civil society organizations have always been you know a species that has really always asked questions sometimes over indexed on always asking questions but then when software development became so much more easier to do AI became so much easier to do we started leading with solutions and we started deploying solutions and then and this is true for our space as well that when we go to a classroom there will be five different NGOs who are solving the same problem and there’s that learner who we’re all saying that we’re trying to work for the learner we’re putting the learner at the center of design but there are five not -for -profits who are all trying to solve the same problem for that learner without recognizing what we’re doing to that learner.
And we need to have that conversation, right? So sometimes I think it’s important to just take a moment and have that conversation or do that assessment before you start leading to solutions.
Yeah, it reminds me of multiple people building an app for ASHA workers and they have to interface with multiple apps themselves, right? Yes, Roy?
Yeah, no, I think just sort of definitely plus one into what both of them have mentioned, I think don’t build what is sexy, build what is needed. And I think we had a very different use case to start off with. And I think we ended up building something very different because we recognized that that was actually the need of the hour. So I think that’s definitely one thing that I would say. And the second piece is I feel like… especially when we are building for the non -profit space in India, scale becomes a very important metric that we want to reach out to like thousands and hundreds and thousands of teachers or students or whoever.
And I think in that whole space, we miss out on some fundamentals like responsibility, that responsible AI that we were talking about. So I think how do we build that? And that’s something that I think I have learned through the process. There’s something that we’ve been working on now called the golden set to ensure that the responses the AI is giving is actually more aligned to what we actually want the AI to give. And I think those pieces are things that if we fix early on, then I think it’s just easier. And that’s something that I think we’ve learned through this process.
Great. Thank you, Manak. I mean, the golden data set you talked about is part of an evaluation framework, also something which is very strongly, emphasized in the context of the program. We have a couple of minutes. We’d like to open up for questions, but before that, any last words? Erica, Tamina?
Y es. So, totally resonate with what all my three panelists said. So, don’t just get on to using technology because everyone else is doing it. Only if it really meets your needs, that is when you think about technology. And when you think about technology, please look, there might be already solutions existing, so don’t jump on to building custom solutions because I think what you are working on is also something many others are working on. And that will just help you save time, effort, and money. So, maybe just, even if an existing solution, I’m just saying, like an open source platform, is meeting 60 -70 % of your needs, just start with that. And then you can improvise it and then decide on whether I want to use it or not.
want to build up something which is custom. So yeah, those are like my thoughts.
It’s a fantastic note to end on. Thank you. Thank you for the panel. Big hand to them. Thank you. We have one minute and three seconds. Questions? I have a few hands up. Like a Gemini gem or a Claude project. Did you all evaluate that? Curious to know the journey of that build versus buy or just adopt for any one of them. Yeah.
Yeah, I can just add and maybe, you know, as you are using it. So like when we were talking about out say a chatbot, right? So it’s not, it’s like those models, like when we are using, we built the Glyphic chatbot. So we’ve obviously integrated it with open AI. So it’s not that you are reinventing anything, but, and even we are building an AI platform. Now, even in that platform, what are we doing? We are integrating different open source elements which are there. So we ourselves as an organization, even if we’re building something new, we are picking up things which are already existing stitching it together and then adding layers when we speak to our NGOs of, you know, what is missing in it and then just build that part of it.
So it is always there in our working that we integrate open source solutions that are existing and not built. Even if it is like for dashboards, we are like, we have a data management platform, but we’ve not been creating dashboards on it. We’ve integrated it with Superset, which is another open source solution. I’m just, I know you may not know these names, but what I want to call out. here is that even as a tech organization who’s building platforms we’re not building things from scratch we are integrating with other tools you know which are meeting the needs even when we are building our own platforms
thank you actually it’s 0 -0 but we’ll stay back if you have questions we’ll have to you know have a conversation with you but thank you for your time it was great to see a big audience also engaged hope it was useful and do reach out to anyone of us if you want to find out more about the program there’s a lot of documents which have been written and published as blogs as well from project tech for dev and some of the non -profits so happy to be engaging with you also going forward in this journey thank you so much and thank you to the panel Thank you. Thank you. you Thank you.
Erica Arya
Speech speed
160 words per minute
Speech length
1921 words
Speech time
717 seconds
Cohort hand‑holding and peer learning model
Explanation
Erica describes how the small AI cohort fosters peer learning and sharing among NGOs, allowing deep‑rooted programs to generate valuable insights from a close‑knit group.
Evidence
“In addition to that, you know, even this small cohort that comes together, there is a lot of peer learning and sharing that happens.” [1]. “they’ve been running these cohort‑based programs, and we felt that, you know, when we do these deep‑rooted programs with a small set of NGOs, we’re able to learn a lot from them.” [14].
Major discussion point
Rationale and Design of the AI Cohort Program
Topics
Capacity development | Artificial intelligence
Integration of responsible‑AI and AI‑safety expertise through knowledge partners
Explanation
Erica explains that the program partnered with organizations specializing in responsible AI and AI safety to embed ethical considerations into the design of NGO AI use cases.
Evidence
“One was Digital Future Labs, who brought in expertise around building, you know, on integrating responsible AI principles into your design as you are working on a use case.” [90]. “And the second knowledge partner was Statle, which works in AI safety.” [87].
Major discussion point
Mentorship, Knowledge Partners, and Responsible AI
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society
Leverage existing open‑source platforms before building custom solutions
Explanation
Erica advises NGOs to start with existing open‑source tools that meet most needs rather than creating bespoke solutions from scratch.
Evidence
“maybe just, even if an existing solution, I’m just saying, like an open source platform, is meeting 60 -70 % of your needs, just start with that.” [114]. “And when you think about technology, please look, there might be already solutions existing, so don’t jump on to building custom solutions because I think what you are working on is also something many others are working on.” [147].
Major discussion point
Lessons Learned and Recommendations for NGOs
Topics
Artificial intelligence | The enabling environment for digital development
Cross‑NGO collaboration on predictive health models and assessment tools
Explanation
Erica highlights how bringing together NGOs working on health predictive models and assessment tools fostered collaboration and shared learning across sectors.
Evidence
“there were two health NGOs also, and they were both working on building a predictive model to predict high -risk pregnancies.” [164]. “Second, while we were working with these seven NGOs, one of them, which is not represented here, was working on an assessment model.” [165].
Major discussion point
Collaboration and Knowledge Sharing Among NGOs
Topics
Social and economic development | Artificial intelligence
Temina Madon
Speech speed
174 words per minute
Speech length
805 words
Speech time
276 seconds
Accelerator‑inspired cohort with pooled technical resources
Explanation
Temina describes creating a pool of technical staff that works alongside funded NGOs, mirroring accelerator models to increase efficiency.
Evidence
“So we’ve put together a pool of, of 10 technical staff who sit alongside the organizations that we fund in the accelerator.” [16]. “if we can create those resources in a pool format and make them available, we can perhaps be more efficient.” [20].
Major discussion point
Rationale and Design of the AI Cohort Program
Topics
The enabling environment for digital development | Artificial intelligence
Provision of shared product‑management and engineering talent to NGOs
Explanation
Temina notes that adding product‑management expertise and making engineering talent available helps NGOs overcome scarce technical resources.
Evidence
“We’ve just added some additional people in product management because the AI technology is actually easy, especially there are a lot of talented engineers in India and elsewhere.” [48]. “If you’re in a nonprofit, raise your hand if you have a product manager.” [99].
Major discussion point
Mentorship, Knowledge Partners, and Responsible AI
Topics
Capacity development | The enabling environment for digital development
Four‑level evaluation framework (user, monitoring, impact)
Explanation
Temina outlines a four‑level framework that guides continuous learning, impact proofing, and scaling of AI products in the social sector.
Evidence
“But we think that it’s a nice way, this kind of four -level framework, to think about how to continuously learn from the product you’re deploying and then prove its impact to society and hopefully bring in more funding to scale.” [111]. “And then the fourth motor we switch on, we call impact evaluation, and that’s when you have tens of thousands, hundreds of thousands of users, and you want to understand whether the product that they’re using is helping them make a change in their health outcomes, in their education or livelihood outcomes.” [115].
Major discussion point
Evaluation Framework and Impact Measurement
Topics
Monitoring and measurement | Artificial intelligence
Pool scarce resources (e.g., product managers) to increase efficiency
Explanation
Temina emphasizes that pooling scarce talent such as product managers enables NGOs to access expertise they would otherwise lack, improving overall program efficiency.
Evidence
“if we can create those resources in a pool format and make them available, we can perhaps be more efficient.” [20]. “We’ve just added some additional people in product management because the AI technology is actually easy…” [48].
Major discussion point
Lessons Learned and Recommendations for NGOs
Topics
Capacity development | The enabling environment for digital development
Manohar Sreekanth
Speech speed
160 words per minute
Speech length
2355 words
Speech time
882 seconds
Program aims to enable NGOs to build and pilot AI use cases
Explanation
Manohar states that the cohort helped seven nonprofits develop AI use cases within a four‑month period, demonstrating the program’s practical impact.
Evidence
“And about seven nonprofits built their AI use cases in the course of these four months.” [26].
Major discussion point
Rationale and Design of the AI Cohort Program
Topics
Artificial intelligence | Social and economic development
Use of a “golden dataset” to align AI outputs with desired outcomes
Explanation
Manohar highlights the creation of a golden set to ensure AI responses match intended goals, linking it to the broader evaluation framework.
Evidence
“There’s something that we’ve been working on now called the golden set to ensure that the responses the AI is giving is actually more aligned to what we actually want the AI to give.” [127]. “the golden data set you talked about is part of an evaluation framework, also something which is very strongly, emphasized in the context of the program.” [124].
Major discussion point
Evaluation Framework and Impact Measurement
Topics
Data governance | Monitoring and measurement | Artificial intelligence
Early‑stage collaboration is rare in for‑profit sector but valuable for social impact
Explanation
Manohar observes that early‑stage, cross‑NGO collaboration on technology is uncommon in the for‑profit world, yet it yields significant benefits for social impact.
Evidence
“I’ve spent a couple of decades in the corporate side, and this kind of collaboration at an early stage of technology is very, very rare.” [171]. “You don’t see that on the for‑profit side.” [172].
Major discussion point
Collaboration and Knowledge Sharing Among NGOs
Topics
Social and economic development | Capacity development
Elizabeth
Speech speed
163 words per minute
Speech length
305 words
Speech time
111 seconds
AI can assist in panel introductions and logistics
Explanation
Elizabeth offers to help frame and deliver introductions for panelists, showcasing AI’s role as a facilitation tool.
Evidence
“Just let me know how you want to frame it and I can help you introduce them smoothly.” [47].
Major discussion point
Rationale and Design of the AI Cohort Program
Topics
Artificial intelligence
Mainak Roy
Speech speed
170 words per minute
Speech length
729 words
Speech time
256 seconds
WhatsApp chatbot workflow issues and need for guardrails
Explanation
Mainak points out that the WhatsApp‑based chatbot required guardrails to manage conversation flow and avoid glitches when teachers interacted with it.
Evidence
“So over there also, we’ve had to put some guardrails around what kind of conversations the chatbot will entertain.” [51]. “yeah I think there are a bunch of challenges but you know my favorite challenge is actually we wanted teachers to start the conversation with the bot with a simple hi but with the moment teachers would get the QR code and scan it they would start putting their problems in … the bot would glitch because the bot did not get that first hi… we sort of worked around it…” [59].
Major discussion point
Implementation Challenges and Solutions in Education AI Use Cases
Topics
Artificial intelligence | Building confidence and security in the use of ICTs
Build what is needed, not what is “sexy”; employ evaluation datasets early
Explanation
Mainak cautions against chasing flashy technology, urging teams to focus on essential functionality and to use evaluation datasets like the golden set early in development.
Evidence
“don’t build what is sexy, build what is needed.” [150]. “There’s something that we’ve been working on now called the golden set to ensure that the responses the AI is giving is actually more aligned to what we actually want the AI to give.” [127].
Major discussion point
Lessons Learned and Recommendations for NGOs
Topics
Capacity development | Artificial intelligence
Steven Suting
Speech speed
182 words per minute
Speech length
1399 words
Speech time
461 seconds
Need for personalized, human‑like AI to address heterogeneous learners
Explanation
Steven stresses that effective AI in education must emulate human interaction and consider behavioral science, given the diverse learner population.
Evidence
“what we’re trying to do is really try and place a system that is able to hear the learners’ problems with respect to the area of career development.” [70]. “it stopped becoming a software problem it started becoming a behavior science problem.” [62].
Major discussion point
Implementation Challenges and Solutions in Education AI Use Cases
Topics
Artificial intelligence | Social and economic development
Avoid knee‑jerk AI adoption; conduct needs assessment before building
Explanation
Steven reflects on the tendency for NGOs to jump on AI trends without proper assessment, leading to misaligned solutions.
Evidence
“when AI started becoming really popular it almost became like knee jerk reactions for organizations to jump on that bandwagon…” [71].
Major discussion point
Lessons Learned and Recommendations for NGOs
Topics
Human rights and the ethical dimensions of the information society | Artificial intelligence
Collaborative prototyping and peer feedback accelerate solution development
Explanation
Steven notes that sharing what works within the cohort speeds up problem‑solving and reduces duplication of effort.
Evidence
“And so then when people started saying that this is what worked for us in that model of deployment, that in many ways kind of expedited the way we solved for that problem.” [110]. “people were trying to solve similar things, but they did it in a different way.” [23].
Major discussion point
Mentorship, Knowledge Partners, and Responsible AI
Topics
Capacity development | Artificial intelligence
Open sharing of prototypes and lessons reduces duplication of effort
Explanation
Steven emphasizes that making prototypes and learnings openly available helps NGOs avoid reinventing solutions.
Evidence
“people were trying to solve similar things, but they did it in a different way.” [23].
Major discussion point
Collaboration and Knowledge Sharing Among NGOs
Topics
Capacity development | Artificial intelligence
Pritam Sukumar
Speech speed
186 words per minute
Speech length
774 words
Speech time
249 seconds
AI‑generated student report scripts improve teacher efficiency but suffer hallucinations
Explanation
Pritam reports that AI‑crafted scripts save teachers time, yet hallucinations undermine trust and require ongoing refinement.
Evidence
“it saves them really a lot of time because before such a conversation, they would have to look at the last four reports of the student.” [85]. “On the negative side, the hallucinations really bother the teachers because there isn’t 100 % trust yet in what we’ve given them.” [76].
Major discussion point
Implementation Challenges and Solutions in Education AI Use Cases
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society
Start from concrete pain points rather than chasing AI hype
Explanation
Pritam advises NGOs to identify existing pain points first and then assess whether AI is a suitable solution.
Evidence
“one thing we learned is to stop looking at use cases for AI but look at pain points and troubles we already have and see if AI is a good fit there.” [75].
Major discussion point
Lessons Learned and Recommendations for NGOs
Topics
Capacity development | Artificial intelligence
Ongoing pilot metrics and feedback loops to assess effectiveness
Explanation
Pritam shares that teachers have used the AI scripts in dozens of conversations, providing data for iterative improvement.
Evidence
“around 15 teachers I think have had 57 or 75, 57 to 75 conversations with students with these scripts.” [83]. “It saves them a lot of time…” [85].
Major discussion point
Evaluation Framework and Impact Measurement
Topics
Monitoring and measurement | Social and economic development
Agreements
Agreement points
Cohort-based programs enable effective peer learning and collaboration among nonprofits
Speakers
– Erica Arya
– Temina Madon
– Steven Suting
Arguments
Cohort model enables deep learning and peer collaboration among nonprofits
Cohort support helps organizations navigate the lonely journey of AI development
Cohort structure enables organizations to share learnings and avoid reinventing solutions for similar problems
Summary
All three speakers emphasized that bringing organizations together in cohorts creates valuable peer learning opportunities, reduces isolation during challenging development processes, and enables knowledge sharing that prevents duplication of efforts
Topics
Artificial intelligence | Capacity development
Responsible AI practices must be integrated from the beginning of development, not as an afterthought
Speakers
– Erica Arya
– Steven Suting
– Mainak Roy
– Manohar Sreekanth
Arguments
Responsible AI principles and safety considerations should be built into design from the beginning, not added later
Frameworks and blueprints help organizations rapid prototype and fail fast while maintaining responsible practices
Build what is needed rather than what appears innovative, and prioritize responsible AI practices early
Guardrails and safety plugins can be integrated to prevent inappropriate AI responses
Summary
There was strong consensus that responsible AI practices, safety considerations, and guardrails should be built into the design process from the start rather than being added later, with practical tools and frameworks available to support this approach
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society | Building confidence and security in the use of ICTs
Focus on solving existing pain points rather than pursuing AI for its own sake
Speakers
– Pritam Sukumar
– Mainak Roy
– Steven Suting
Arguments
Focus on existing pain points rather than seeking AI use cases for the sake of technology adoption
Build what is needed rather than what appears innovative, and prioritize responsible AI practices early
Avoid jumping on AI bandwagon without proper needs assessment and consideration of user impact
Summary
All three nonprofit representatives agreed that organizations should identify real operational challenges and user needs first, then evaluate whether AI can address these problems, rather than starting with AI technology and looking for applications
Topics
Artificial intelligence | Social and economic development | The enabling environment for digital development
Leverage existing solutions before building custom AI systems
Speakers
– Erica Arya
– Pritam Sukumar
Arguments
Leverage existing open-source solutions rather than building custom solutions from scratch
Focus on existing pain points rather than seeking AI use cases for the sake of technology adoption
Summary
Both speakers advocated for using existing open-source solutions that meet 60-70% of needs before considering custom development, emphasizing efficiency and resource conservation
Topics
Artificial intelligence | Information and communication technologies for development | The enabling environment for digital development
Similar viewpoints
All three education-focused nonprofits identified similar challenges around personalized support in educational settings – whether for teachers needing pedagogical guidance, students requiring individualized learning approaches, or online learners lacking mentorship. They all developed AI solutions to bridge these gaps in personalized educational support
Speakers
– Mainak Roy
– Steven Suting
– Pritam Sukumar
Arguments
Teachers need evidence-based pedagogical strategies tailored to specific classroom contexts delivered via WhatsApp
Heterogeneous classrooms require personalized learning approaches due to diverse student needs and limited teacher availability
Student mentorship conversations need AI-generated scripts to provide actionable guidance for both teachers and online learners
Topics
Artificial intelligence | Social and economic development | Closing all digital divides
Both funders emphasized the importance of systematic approaches – Temina through comprehensive evaluation frameworks and Erica through structured collaboration – to maximize the impact and efficiency of nonprofit AI initiatives
Speakers
– Temina Madon
– Erica Arya
Arguments
Four-level evaluation framework includes user experience, user behavior, user evaluation, and impact evaluation
Multiple nonprofits working on identical use cases can collaborate and build solutions together
Topics
Artificial intelligence | Monitoring and measurement | Capacity development
All three nonprofit implementers recognized that successful AI deployment requires understanding human behavior and psychology, not just technical implementation, and that real-world user behavior often differs significantly from expected patterns
Speakers
– Steven Suting
– Pritam Sukumar
– Mainak Roy
Arguments
Building behavior change systems requires focusing on behavior science rather than just software engineering
Hallucinations in AI outputs create trust issues with teachers despite time-saving benefits
Teachers don’t follow expected interaction patterns, requiring flexible bot design to handle unexpected inputs
Topics
Artificial intelligence | Capacity development | Building confidence and security in the use of ICTs
Unexpected consensus
Social sector organizations require more sophisticated evaluation than private sector companies
Speakers
– Temina Madon
– Steven Suting
Arguments
Social sector organizations require more comprehensive evaluation systems compared to private sector companies
Avoid jumping on AI bandwagon without proper needs assessment and consideration of user impact
Explanation
It was unexpected to see such strong agreement that nonprofit AI development requires fundamentally different and more comprehensive approaches than private sector AI, including extensive survey data collection and long-term impact measurement that goes far beyond typical business metrics
Topics
Artificial intelligence | Monitoring and measurement | Social and economic development
Collaboration should extend beyond immediate program participants
Speakers
– Erica Arya
– Steven Suting
Arguments
Ecosystem players can facilitate connections beyond the immediate cohort to maximize collaboration
Cohort structure enables organizations to share learnings and avoid reinventing solutions for similar problems
Explanation
The consensus on actively facilitating collaboration beyond the immediate cohort participants was unexpected, showing a commitment to sector-wide efficiency that goes beyond typical program boundaries and competitive instincts
Topics
Artificial intelligence | Capacity development | The enabling environment for digital development
Overall assessment
Summary
There was remarkably strong consensus among all speakers on key principles: the value of cohort-based learning, the necessity of building responsible AI practices from the start, the importance of focusing on real user needs rather than technology trends, and the benefits of leveraging existing solutions. The education-focused nonprofits shared similar challenges around personalization and user behavior, while funders agreed on the need for comprehensive evaluation and systematic collaboration approaches.
Consensus level
Very high level of consensus with no significant disagreements identified. This strong alignment suggests the AI cohort program successfully created shared understanding and best practices among participants. The implications are positive for scaling such programs, as the consistent learnings and approaches can be systematized and replicated across the nonprofit sector.
Differences
Different viewpoints
Approach to handling AI hallucinations and trust issues
Speakers
– Pritam Sukumar
– Mainak Roy
Arguments
Hallucinations in AI outputs create trust issues with teachers despite time-saving benefits
Teachers don’t follow expected interaction patterns, requiring flexible bot design to handle unexpected inputs
Summary
While both faced AI reliability issues, Pritam struggled with hallucinations affecting teacher trust and saw LLMs as ‘black boxes’, while Mainak focused on design flexibility to accommodate unexpected user behavior patterns
Topics
Artificial intelligence | Building confidence and security in the use of ICTs
Unexpected differences
Target audience for AI solutions in education
Speakers
– Mainak Roy
– Steven Suting
– Pritam Sukumar
Arguments
Teachers need evidence-based pedagogical strategies tailored to specific classroom contexts delivered via WhatsApp
Heterogeneous classrooms require personalized learning approaches due to diverse student needs and limited teacher availability
Student mentorship conversations need AI-generated scripts to provide actionable guidance for both teachers and online learners
Explanation
While all three worked in education, they chose fundamentally different target audiences – Mainak focused solely on teachers, Steven targeted students directly, and Pritam addressed both teachers and students. This represents different philosophies about where AI intervention would be most effective in the education ecosystem
Topics
Artificial intelligence | Social and economic development | Capacity development
Overall assessment
Summary
The panel showed remarkably high consensus on fundamental principles including the importance of addressing real needs over technology adoption, the value of responsible AI practices, the benefits of collaboration, and the need for proper evaluation frameworks
Disagreement level
Very low disagreement level with high collaborative spirit. The few disagreements were primarily tactical rather than strategic, focusing on different implementation approaches rather than conflicting philosophies. This high consensus likely reflects the shared challenges nonprofits face and the collaborative nature of the cohort program itself, which fostered alignment on best practices and common goals
Partial agreements
Partial agreements
All three speakers agreed on the fundamental principle of addressing real needs rather than pursuing AI for its own sake, but they emphasized different aspects: Steven focused on avoiding user burden from multiple solutions, Pritam emphasized identifying operational pain points first, and Mainak stressed building needed rather than ‘sexy’ solutions
Speakers
– Steven Suting
– Pritam Sukumar
– Mainak Roy
Arguments
Avoid jumping on AI bandwagon without proper needs assessment and consideration of user impact
Focus on existing pain points rather than seeking AI use cases for the sake of technology adoption
Build what is needed rather than what appears innovative, and prioritize responsible AI practices early
Topics
Artificial intelligence | Social and economic development | The enabling environment for digital development
Both agreed on the value of cohort models but emphasized different benefits: Erica focused on practical collaboration and avoiding duplication of efforts among nonprofits, while Temina emphasized emotional support and peer learning during the challenging AI development process
Speakers
– Erica Arya
– Temina Madon
Arguments
Cohort model enables deep learning and peer collaboration among nonprofits
Cohort support helps organizations navigate the lonely journey of AI development
Topics
Artificial intelligence | Capacity development | The enabling environment for digital development
Similar viewpoints
All three education-focused nonprofits identified similar challenges around personalized support in educational settings – whether for teachers needing pedagogical guidance, students requiring individualized learning approaches, or online learners lacking mentorship. They all developed AI solutions to bridge these gaps in personalized educational support
Speakers
– Mainak Roy
– Steven Suting
– Pritam Sukumar
Arguments
Teachers need evidence-based pedagogical strategies tailored to specific classroom contexts delivered via WhatsApp
Heterogeneous classrooms require personalized learning approaches due to diverse student needs and limited teacher availability
Student mentorship conversations need AI-generated scripts to provide actionable guidance for both teachers and online learners
Topics
Artificial intelligence | Social and economic development | Closing all digital divides
Both funders emphasized the importance of systematic approaches – Temina through comprehensive evaluation frameworks and Erica through structured collaboration – to maximize the impact and efficiency of nonprofit AI initiatives
Speakers
– Temina Madon
– Erica Arya
Arguments
Four-level evaluation framework includes user experience, user behavior, user evaluation, and impact evaluation
Multiple nonprofits working on identical use cases can collaborate and build solutions together
Topics
Artificial intelligence | Monitoring and measurement | Capacity development
All three nonprofit implementers recognized that successful AI deployment requires understanding human behavior and psychology, not just technical implementation, and that real-world user behavior often differs significantly from expected patterns
Speakers
– Steven Suting
– Pritam Sukumar
– Mainak Roy
Arguments
Building behavior change systems requires focusing on behavior science rather than just software engineering
Hallucinations in AI outputs create trust issues with teachers despite time-saving benefits
Teachers don’t follow expected interaction patterns, requiring flexible bot design to handle unexpected inputs
Topics
Artificial intelligence | Capacity development | Building confidence and security in the use of ICTs
Takeaways
Key takeaways
Cohort-based AI programs enable effective peer learning, collaboration, and knowledge sharing among nonprofits working on similar challenges
AI adoption should be driven by existing organizational pain points rather than technology trends – focus on building what is needed, not what appears innovative
Responsible AI principles and safety considerations must be integrated from the design phase, not added as an afterthought
Real-world deployment reveals unexpected user behaviors that require flexible system design and continuous iteration
Leveraging existing open-source solutions and platforms is more efficient than building custom solutions from scratch
Social sector organizations require more comprehensive evaluation frameworks compared to private sector, including user experience, behavior, evaluation, and impact measurement
Collaboration among nonprofits can prevent duplication of efforts and enable shared solutions for common problems
WhatsApp and familiar platforms can ease AI adoption by reducing the learning curve for end users
AI hallucinations and trust issues remain significant challenges that require ongoing prompt engineering and user training
Resolutions and action items
Continue fixing hallucination issues in AI-generated mentorship scripts before full rollout (Avanti Fellows)
Implement golden dataset approach for better AI response alignment (Simple Education Foundation)
Facilitate collaboration between nonprofits working on identical use cases like pregnancy risk prediction models
Publish program learnings and documentation as blogs and resources for broader nonprofit community
Continue iterating on behavior science approaches rather than purely technical solutions for student engagement
Unresolved issues
How to effectively address AI hallucinations in educational contexts where accuracy is critical
Scaling responsible AI practices across diverse nonprofit contexts and use cases
Balancing the need for rapid deployment with thorough safety and evaluation protocols
Managing multiple nonprofits working on similar problems for the same beneficiaries (like ASHA workers using multiple apps)
Determining optimal evaluation frameworks for different types of AI interventions in social sector
Long-term sustainability and funding models for AI initiatives in nonprofits
Suggested compromises
Start with existing solutions that meet 60-70% of needs rather than building custom solutions, then iterate and improve
Use flexible bot designs that can handle unexpected user inputs rather than forcing rigid interaction patterns
Integrate open-source components and platforms rather than building everything from scratch
Focus on modular rapid prototyping to fail fast and learn quickly while maintaining responsible AI principles
Combine technical mentorship with behavior science expertise rather than treating AI as purely a software problem
Thought provoking comments
Don’t build what is sexy, build what is needed. And I think we had a very different use case to start off with. And I think we ended up building something very different because we recognized that that was actually the need of the hour.
Speaker
Mainak Roy
Reason
This comment cuts through the AI hype and addresses a fundamental problem in technology adoption – the tendency to chase trendy solutions rather than solving real problems. It’s particularly insightful because it comes from lived experience of pivoting from their original use case to something more practical.
Impact
This comment reinforced and crystallized the discussion’s shift from technical implementation details to strategic thinking about AI adoption. It became a rallying point that other panelists built upon, emphasizing the importance of needs-based rather than technology-driven solutions.
When AI started becoming really popular it almost became like knee jerk reactions for organizations to jump on that bandwagon… we started leading with solutions and we started deploying solutions… there will be five different NGOs who are solving the same problem and there’s that learner who we’re all saying that we’re trying the learner at the center of design but there are five not-for-profits who are all trying to solve the same problem for that learner without recognizing what we’re doing to that learner.
Speaker
Steven Suting
Reason
This is a profound critique of the nonprofit sector’s approach to technology adoption. It highlights the paradox where organizations claiming to be user-centered are actually creating fragmented experiences for their beneficiaries. The comment reveals systemic issues beyond just AI implementation.
Impact
This comment elevated the discussion from individual organizational challenges to sector-wide systemic issues. It prompted deeper reflection on coordination and collaboration, and reinforced the value of the cohort model being discussed.
Stop looking at use cases for AI but look at pain points and troubles we already have and see if AI is a good fit there. It doesn’t have to be like AI is a new technology, let’s put it and build something amazing and innovative.
Speaker
Pritam Sukumar
Reason
This comment reframes the entire approach to AI adoption by inverting the typical technology-first mindset. It’s particularly valuable because it comes from someone who actually went through multiple iterations to find the right use case, lending credibility to the advice.
Impact
This comment helped establish a key takeaway for the nonprofit audience and shifted the conversation toward practical implementation advice. It provided a clear, actionable framework that other speakers could build upon.
The AI technology is actually easy, especially there are a lot of talented engineers in India and elsewhere. What is difficult is to fit the technology to the pain points we all experience in life and build a product that achieves social impact. That’s really difficult. It’s so much further from what the typical Silicon Valley startup has to deal with. They just need to make money from people who already have money, right? We are trying to do something much more difficult with a population that’s much more excluded.
Speaker
Temina Madon
Reason
This comment provides crucial context by distinguishing the unique challenges of social sector AI implementation from commercial applications. It highlights that the real difficulty isn’t technical but lies in product-market fit for excluded populations, which is a much more complex challenge.
Impact
This comment helped frame the entire discussion by establishing why the cohort model and collaborative approach was necessary. It elevated the conversation from technical implementation to the broader challenges of social impact technology, influencing how subsequent speakers discussed their experiences.
We wanted to build that right from the get go. And that’s why we collaborated with partners because we knew we did not know this bit… It’s not something that you need to think later, okay, let me just build something. And then I’ll come to, you know, responsible AI practices or think about AI safety… after I’ve done my pilot.
Speaker
Erica Arya
Reason
This comment challenges the typical software development approach of ‘build first, fix later’ by emphasizing the importance of embedding responsible AI principles from the beginning. It’s particularly insightful because it recognizes organizational limitations and the value of partnerships.
Impact
This comment introduced the critical dimension of responsible AI and safety into the discussion, which became a recurring theme. It demonstrated how the program structure itself was designed to address these concerns, influencing how other panelists discussed their implementation approaches.
Overall assessment
These key comments collectively transformed the discussion from a typical technology showcase into a nuanced exploration of strategic AI adoption in the social sector. The comments created a progression from individual technical challenges to broader systemic issues, ultimately establishing a framework for responsible AI implementation. The speakers’ willingness to share failures and pivots, combined with their emphasis on needs-driven rather than technology-driven solutions, provided practical wisdom that distinguished this discussion from typical AI conferences. The collaborative and reflective tone established by these comments made the session more valuable for nonprofit practitioners seeking genuine guidance rather than just success stories.
Follow-up questions
How to address hallucinations in AI systems, particularly when LLMs provide incorrect directional information (saying ‘decreased’ instead of ‘increased’)
Speaker
Pritam Sukumar
Explanation
This is a critical technical challenge affecting teacher trust in AI-generated mentorship scripts, and the solution approach remains unclear to the team
How to build 100% trust in AI systems among teachers when hallucinations persist
Speaker
Pritam Sukumar
Explanation
Teacher confidence in AI-generated content is essential for successful adoption, but current error rates are undermining trust
Whether more AI engineers are needed to solve hallucination problems effectively
Speaker
Pritam Sukumar
Explanation
The team is uncertain about the right resource allocation and expertise needed to address fundamental AI reliability issues
How to prevent teachers from straying away from intended question types when interacting with AI chatbots
Speaker
Mainak Roy
Explanation
User behavior differs from designed interaction patterns, requiring better guardrails and user guidance mechanisms
How to decrypt conversation signals in bot interfaces to emulate human teacher-student interactions
Speaker
Steven Suting
Explanation
Limited signal access in digital interfaces compared to face-to-face interactions poses challenges for creating human-like AI responses
How to address the problem of multiple NGOs building similar solutions for the same beneficiaries without coordination
Speaker
Steven Suting
Explanation
This creates inefficiency and potentially overwhelming experiences for end users like learners who must interface with multiple similar tools
Build versus buy versus adopt evaluation process for AI solutions in nonprofits
Speaker
Audience member
Explanation
Understanding the decision-making framework for choosing between custom development, purchasing existing solutions, or adopting open-source alternatives
How to develop effective evaluation frameworks for measuring AI impact in social sector applications
Speaker
Temina Madon (implied through framework discussion)
Explanation
Social sector organizations need robust methods to assess whether AI solutions are achieving intended social impact outcomes
How to implement responsible AI principles and AI safety measures from the beginning of development rather than as an afterthought
Speaker
Erica Arya
Explanation
Ensuring ethical AI deployment requires integrating safety considerations into the design process rather than addressing them post-development
How to create effective ‘golden datasets’ for ensuring AI responses align with intended outcomes
Speaker
Mainak Roy
Explanation
Developing evaluation mechanisms to ensure AI outputs meet quality and alignment standards is crucial for responsible deployment
Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.
Related event

