OpenAI and Stack Overflow team up for better AI models

OpenAI and developer platform Stack Overflow have joined forces in a new partnership to enhance AI capabilities and provide richer technical information. Under this collaboration, OpenAI gains access to Stack Overflow’s API and will incorporate feedback from the developer community to refine AI models. In return, Stack Overflow will receive attribution in ChatGPT, offering users access to Stack Overflow’s extensive knowledge base when seeking coding or technical advice. Both companies anticipated that this collaboration would deepen user engagement with content.

Stack Overflow plans to leverage OpenAI’s large language models to enhance Overflow AI, introduced last year as its generative AI application. Overflow AI aims to incorporate AI-powered natural language search functionality into Stack Overflow, providing users with more intuitive access to coding solutions. Stack Overflow emphasises that it will integrate feedback from its community and internal testing of OpenAI models to develop additional AI products for its user base.

The initial phase of integrations resulting from this partnership is expected to roll out in the first half of the year, although Stack Overflow has yet to specify the exact features to be released first. The collaboration follows Stack Overflow’s similar arrangement with Google in February, where Gemini for Google Cloud users could access coding suggestions directly from Stack Overflow.

Why does it matter?

For years, developers have relied on Stack Overflow for coding solutions. Still, the company faced challenges in 2022, including a significant hiring push followed by layoffs of 28% of its workforce in October of the same year. While Stack Overflow did not provide a specific reason for the layoffs, they coincided with the growing prominence of AI-assisted coding. Additionally, Stack Overflow briefly prohibited users from sharing ChatGPT responses on its platform in 2022.

Financial Times partners with OpenAI to license content and develop AI tools

The Financial Times has announced a collaboration with OpenAI, allowing the AI company to license its content and utilise it to develop AI tools. Under this partnership, ChatGPT users will encounter summaries, quotes, and article links from the Financial Times, with all information sourced attributed to the publication. In return, OpenAI will collaborate with the Financial Times to innovate and create new AI products, building upon their existing relationship, as the publication already utilises OpenAI’s ChatGPT Enterprise.

John Ridding, CEO of the Financial Times Group, emphasises the importance of maintaining ‘human journalism’ even amidst collaborations with AI platforms. Ridding asserts that AI products must incorporate reliable sources, highlighting the significance of partnerships like the one with OpenAI. Notably, OpenAI has secured similar agreements with other news organisations, including Axel Springer and the Associated Press, to license content for training AI models.

However, OpenAI’s licensing agreements have drawn attention for their comparatively lower payouts to publishers, ranging from $1 million to $5 million, in contrast to offers from companies like Apple. This discrepancy has led to legal disputes, with the New York Times and other news outlets suing OpenAI and Microsoft for alleged copyright infringement related to ChatGPT’s use of their content. These legal battles underscore the complexities and challenges surrounding the integration of AI technology within the news industry.

NOYB files a privacy complaint against OpenAI’s ChatGPT

OpenAI, a startup supported by Microsoft, faces a privacy complaint from the European Center for Digital Rights (NOYB), an advocacy group, for allegedly failing to address incorrect information provided by its AI chatbot, ChatGPT, which could violate the EU privacy regulations. ChatGPT, renowned for its ability to mimic human conversation and perform various tasks, including summarising texts and generating ideas, has been scrutinised after reportedly providing inaccurate responses to queries about a public figure’s birthday.

NOYB claims that despite the complainant’s requests, OpenAI refused to rectify or erase the erroneous data, citing technical limitations. Additionally, the group alleges that OpenAI did not disclose crucial information regarding data processing, sources, or recipients, prompting NOYB to file a complaint with the data protection authority in Austria.

According to NOYB’s data protection lawyer, Maartje de Graaf, the incident underscores the challenge of ensuring compliance with the EU law when processing individuals’ data using chatbots like ChatGPT. She emphasised the necessity for technology to adhere to legal requirements rather than vice versa.

OpenAI has previously acknowledged ChatGPT’s tendency to provide plausible yet incorrect responses, citing it as a complex issue. However, NOYB’s complaint highlights the urgency for companies to ensure the accuracy and transparency of personal data processed by large language models like ChatGPT.

OpenAI hires first employee in India for AI policy

OpenAI, the company behind ChatGPT, has appointed Pragya Misra, its first employee in India, to lead government relations and public policy affairs. This move comes as India prepares for a new administration to influence AI regulations in one of the world’s largest and fastest-growing tech markets. Previously with Truecaller AB and Meta Platforms Inc., Misra brings a wealth of experience navigating policy issues and partnerships within the tech industry.

The hiring reflects OpenAI’s strategic efforts to advocate for favourable regulations amid the global push for AI governance. Given its vast population and expanding economy, India presents a significant growth opportunity for tech giants. However, regulatory complexities in India have posed challenges, with authorities aiming to protect local industries while embracing technological advancements.

Why does it matter?

OpenAI’s engagement in India mirrors competition from other tech giants like Google, which is developing AI models tailored for the Indian market to address linguistic diversity and expand internet access beyond English-speaking urban populations. OpenAI’s CEO, Sam Altman, emphasised the need for AI research to enhance government services like healthcare, underscoring the importance of integrating emerging technologies into public sectors.

During Altman’s visit to India last year, he highlighted the country’s early adoption of OpenAI’s ChatGPT. Altman has advocated for responsible AI development, calling for regulations to mitigate potential harms from AI technologies. While current AI versions may not require major regulatory changes, Altman believes that evolving AI capabilities will soon necessitate comprehensive governance.

OpenAI inaugurates its office in Tokyo

OpenAI, supported by Microsoft, has set its sights on Japan, inaugurating its first Asia office in Tokyo. CEO Sam Altman expressed enthusiasm for a long-term collaboration with Japan, envisioning partnerships with government bodies, businesses, and research institutions. With the success of its ChatGPT AI chatbot, OpenAI seeks to expand its revenue streams globally.

Altman and COO Brad Lightcap have been actively engaging Fortune 500 executives in the US and UK, signalling a concerted effort to attract business. Last year’s meeting with Prime Minister Fumio Kishida laid the groundwork for OpenAI’s expansion into Japan, joining its offices in London and Dublin. Japan, aiming to bolster its competitiveness against China, sees AI as pivotal in its digital transformation and addressing labour shortages.

OpenAI is strategically positioned with a tailored model for the Japanese language, led by Tadao Nagasaki, former president of Amazon Web Services in Japan. Despite Japan’s reputation as a technology follower, companies like SoftBank and NTT are investing in large language models. Notable Japanese clients of OpenAI include Toyota Motor, Daikin Industries, and local government entities.

The move aligns with Microsoft’s recent commitment of $2.9 billion over two years to bolster cloud and AI infrastructure in Japan. The investment surge from US tech giants underscores Japan’s growing importance in the global AI landscape and its alignment to maintain a solid place in the race for cutting-edge technology development.

New OpenAI and Meta AI models close to human-like reasoning

Meta and OpenAI are close to unveiling advanced AI models that can reason and plan, according to a Financial Times report. OpenAI’s COO, Brad Lightcap, hinted at the upcoming release of GPT-5, which will make significant progress in solving ‘hard problems’ of reasoning.

Yann LeCun, Meta’s chief AI scientist, and Joelle Pineau, VP of AI Research, envision AI agents capable of complex, multi-stage operations. The enhanced reasoning should enable the AI models to ‘search over possible answers,’ ‘plan sequences of actions,’ and model out the outcomes and consequences before execution.

Why does it matter?


Meta is getting ready to launch Llama 3 in various model sizes optimized for different apps and devices, including WhatsApp and Ray-Ban smart glasses. OpenAI is less open about its plans for GPT-5, but Lightcap expressed optimism about the model’s potential to reason.


Getting AI models to reason and plan is a critical step towards reaching artificial general intelligence (AGI). Multiple definitions of AGI exist, but it can be simply described as a sort of AI capable of performing at or beyond human levels on a broad range of activities.

Some scientists and experts have expressed concerns about building technology that will outperform human abilities. AI godfathers Yoshua Bengio, and Geoffrey Hinton have even warned us against the threats to humanity posed by AI. Both Meta and OpenAI claim to be aiming for AGI, which could be worth trillions for the company that achieves it.

EU antitrust probe into Microsoft’s OpenAI investment nears conclusion

The EU regulators are swiftly moving to conclude a preliminary investigation into Microsoft’s relationship with OpenAI, according to Margrethe Vestager, the EU’s antitrust chief. The probe, initiated in January, aims to determine whether Microsoft’s substantial investment of $13 billion into OpenAI should undergo scrutiny under the EU merger regulations. Vestager indicated in an interview with Bloomberg TV that a resolution is forthcoming, highlighting ongoing discussions with other regulatory authorities.

Vestager emphasised that the EU authorities closely monitor Microsoft’s investments and the broader trend of large tech companies investing in AI. The scrutiny extends beyond Microsoft to include other significant AI investments from major tech firms like Google, Amazon, and Nvidia. The EU mainly ensures competitiveness and prevents anti-competitive practices in this rapidly evolving AI landscape.

Microsoft’s involvement with OpenAI represents a significant stake, with the tech giant investing in other AI ventures, such as French startup Mistral and acquiring the team from Inflection AI. This investment landscape extends to other major players like Google and Amazon, which have their stakes in AI ventures. Vestager stressed the importance of vigilance in this emerging field, characterising it as a critical area for regulatory oversight to safeguard competition and innovation in the AI sector.

AI giants OpenAI, Google, Meta and Mistral unveil new LLMs in rapid succession

Three major players in the AI field, OpenAI, Google, and Mistral, have unveiled new versions of their cutting-edge AI models within 12 hours, signalling a burst of innovation anticipated for the summer. Meta’s Nick Clegg hinted at the imminent release of Meta’s Llama 3 at an event in London, while Google swiftly followed with the launch of its Gemini Pro 1.5, a sophisticated large language model with a limited free usage tier. Shortly after, OpenAI introduced its milestone model, GPT-4 Turbo, which, like Gemini Pro 1.5, supports multimodal input, including images.

In France, Mistral, a startup founded by former Meta AI team members, debuted Mixtral 8x22B, a frontier AI model released as a 281GB download file, following an open-source philosophy. While this approach is criticised for potential risks due to a lack of oversight, it reflects a trend towards democratising access to AI models beyond the control of tech giants like Meta and Google.

Experts caution that the prevailing approach centred on large language models (LLMs) might be reaching its limitations. Meta’s chief AI scientist, Yann LeCun, challenges the notion of imminent artificial general intelligence (AGI) and emphasises the need for AI systems capable of reasoning and planning beyond language manipulation. LeCun advocates for a shift towards ‘objective-driven’ AI to achieve truly superhuman capabilities, thereby highlighting the ongoing evolution and challenges in the AI landscape.

OpenAI utilised one million hours of YouTube content to train GPT-4

In recent reports by The New York Times, the challenges faced by AI companies in acquiring high-quality training data have come to light. The New York Times elaborates on how companies like OpenAI and Google have navigated this issue, often treading in legally ambiguous territories related to AI copyright law.

OpenAI, for instance, resorted to developing its Whisper audio transcription model by transcribing over a million hours of YouTube videos to train GPT-4, its advanced language model. Although this approach raised legal concerns, OpenAI believed it fell within fair use. The company’s president, Greg Brockman, reportedly played a hands-on role in collecting these videos.

According to a Google spokesperson, there were unconfirmed reports of OpenAI’s activities, and both Google’s terms of service and robots.txt files prohibit unauthorised scraping or downloading of YouTube content. Google also utilised transcripts from YouTube, aligned with its agreements with content creators.

Similarly, Meta encountered challenges with data availability for training its AI models. The company’s AI team discussed using copyrighted works without permission to catch up with OpenAI. Meta explored options like paying for book licenses or acquiring a large publisher to address this issue.

Why does it matter?

AI companies, including Google and OpenAI, are grappling with the dwindling availability of quality training data to improve their models. The future of AI training may involve synthetic data or curriculum learning methods, but these approaches still need to be proven. In the meantime, companies continue to explore various avenues for data acquisition, sometimes straying into legally contentious territories as they navigate this evolving landscape.

OpenAI enhances model performance and customisation options

OpenAI has unveiled new features to enhance model performance and customizability, catering to developers seeking to optimise AI implementations for speed, accuracy, and cost-efficiency. These enhancements include a self-serve fine-tuning API for GPT-3.5, which has already facilitated the training of hundreds of thousands of models across various organisations since its launch in August 2023. Fine-tuning empowers models to grasp content deeply and augment their existing capabilities, resulting in superior outcomes tailored to tasks such as code generation, text summarisation, or personalised content creation.

One standout success story involves a leading job-matching platform that utilised fine-tuning GPT-3.5 Turbo to refine personalised job recommendations for users. By leveraging fine-tuning, it achieved an 80% reduction in token usage, allowing the platform to scale its messaging volume from under a million to roughly 20 million monthly messages, enhancing user engagement and satisfaction.

In addition to the fine-tuning API, OpenAI has introduced new features to provide developers with enhanced control and visibility over their fine-tuning projects. These features include epoch-based checkpoint creation, a comparative playground for model evaluation, third-party integrations with platforms like Weights and Biases, comprehensive validation metrics, and improved dashboard functionality.

Why does it matter?

The ChatGPT’s owner envisions widespread adoption of customised models across industries and use cases. It emphasises the importance of effectively scoping use cases, implementing robust evaluation systems, and leveraging the right techniques to optimise model performance.

OpenAI accommodates organisations seeking fully custom-trained models capable of comprehending complex domain-specific knowledge and behaviours. The company has expanded its Custom Models program, offering assisted fine-tuning services for organisations that require tailored AI solutions beyond the capabilities of standard fine-tuning. Assisted fine-tuning involves collaborative efforts with technical teams to implement advanced techniques, such as parameter-efficient fine-tuning, to maximise model performance across specific use cases or tasks. Thus, for organisations seeking to harness AI capabilities for personalised impact, OpenAI’s offerings provide a pathway towards tailored and effective AI implementations.