The Financial Times has announced a collaboration with OpenAI, allowing the AI company to license its content and utilise it to develop AI tools. Under this partnership, ChatGPT users will encounter summaries, quotes, and article links from the Financial Times, with all information sourced attributed to the publication. In return, OpenAI will collaborate with the Financial Times to innovate and create new AI products, building upon their existing relationship, as the publication already utilises OpenAI’s ChatGPT Enterprise.
John Ridding, CEO of the Financial Times Group, emphasises the importance of maintaining ‘human journalism’ even amidst collaborations with AI platforms. Ridding asserts that AI products must incorporate reliable sources, highlighting the significance of partnerships like the one with OpenAI. Notably, OpenAI has secured similar agreements with other news organisations, including Axel Springer and the Associated Press, to license content for training AI models.
However, OpenAI’s licensing agreements have drawn attention for their comparatively lower payouts to publishers, ranging from $1 million to $5 million, in contrast to offers from companies like Apple. This discrepancy has led to legal disputes, with the New York Times and other news outlets suing OpenAI and Microsoft for alleged copyright infringement related to ChatGPT’s use of their content. These legal battles underscore the complexities and challenges surrounding the integration of AI technology within the news industry.
OpenAI, a startup supported by Microsoft, faces a privacy complaint from the European Center for Digital Rights (NOYB), an advocacy group, for allegedly failing to address incorrect information provided by its AI chatbot, ChatGPT, which could violate the EU privacy regulations. ChatGPT, renowned for its ability to mimic human conversation and perform various tasks, including summarising texts and generating ideas, has been scrutinised after reportedly providing inaccurate responses to queries about a public figure’s birthday.
NOYB claims that despite the complainant’s requests, OpenAI refused to rectify or erase the erroneous data, citing technical limitations. Additionally, the group alleges that OpenAI did not disclose crucial information regarding data processing, sources, or recipients, prompting NOYB to file a complaint with the data protection authority in Austria.
According to NOYB’s data protection lawyer, Maartje de Graaf, the incident underscores the challenge of ensuring compliance with the EU law when processing individuals’ data using chatbots like ChatGPT. She emphasised the necessity for technology to adhere to legal requirements rather than vice versa.
OpenAI has previously acknowledged ChatGPT’s tendency to provide plausible yet incorrect responses, citing it as a complex issue. However, NOYB’s complaint highlights the urgency for companies to ensure the accuracy and transparency of personal data processed by large language models like ChatGPT.
OpenAI, the company behind ChatGPT, has appointed Pragya Misra, its first employee in India, to lead government relations and public policy affairs. This move comes as India prepares for a new administration to influence AI regulations in one of the world’s largest and fastest-growing tech markets. Previously with Truecaller AB and Meta Platforms Inc., Misra brings a wealth of experience navigating policy issues and partnerships within the tech industry.
The hiring reflects OpenAI’s strategic efforts to advocate for favourable regulations amid the global push for AI governance. Given its vast population and expanding economy, India presents a significant growth opportunity for tech giants. However, regulatory complexities in India have posed challenges, with authorities aiming to protect local industries while embracing technological advancements.
Why does it matter?
OpenAI’s engagement in India mirrors competition from other tech giants like Google, which is developing AI models tailored for the Indian market to address linguistic diversity and expand internet access beyond English-speaking urban populations. OpenAI’s CEO, Sam Altman, emphasised the need for AI research to enhance government services like healthcare, underscoring the importance of integrating emerging technologies into public sectors.
During Altman’s visit to India last year, he highlighted the country’s early adoption of OpenAI’s ChatGPT. Altman has advocated for responsible AI development, calling for regulations to mitigate potential harms from AI technologies. While current AI versions may not require major regulatory changes, Altman believes that evolving AI capabilities will soon necessitate comprehensive governance.
OpenAI, supported by Microsoft, has set its sights on Japan, inaugurating its first Asia office in Tokyo. CEO Sam Altman expressed enthusiasm for a long-term collaboration with Japan, envisioning partnerships with government bodies, businesses, and research institutions. With the success of its ChatGPT AI chatbot, OpenAI seeks to expand its revenue streams globally.
Altman and COO Brad Lightcap have been actively engaging Fortune 500 executives in the US and UK, signalling a concerted effort to attract business. Last year’s meeting with Prime Minister Fumio Kishida laid the groundwork for OpenAI’s expansion into Japan, joining its offices in London and Dublin. Japan, aiming to bolster its competitiveness against China, sees AI as pivotal in its digital transformation and addressing labour shortages.
OpenAI is strategically positioned with a tailored model for the Japanese language, led by Tadao Nagasaki, former president of Amazon Web Services in Japan. Despite Japan’s reputation as a technology follower, companies like SoftBank and NTT are investing in large language models. Notable Japanese clients of OpenAI include Toyota Motor, Daikin Industries, and local government entities.
The move aligns with Microsoft’s recent commitment of $2.9 billion over two years to bolster cloud and AI infrastructure in Japan. The investment surge from US tech giants underscores Japan’s growing importance in the global AI landscape and its alignment to maintain a solid place in the race for cutting-edge technology development.
Meta and OpenAI are close to unveiling advanced AI models that can reason and plan, according to a Financial Times report. OpenAI’s COO, Brad Lightcap, hinted at the upcoming release of GPT-5, which will make significant progress in solving ‘hard problems’ of reasoning.
Yann LeCun, Meta’s chief AI scientist, and Joelle Pineau, VP of AI Research, envision AI agents capable of complex, multi-stage operations. The enhanced reasoning should enable the AI models to ‘search over possible answers,’ ‘plan sequences of actions,’ and model out the outcomes and consequences before execution.
Why does it matter?
Meta is getting ready to launch Llama 3 in various model sizes optimized for different apps and devices, including WhatsApp and Ray-Ban smart glasses. OpenAI is less open about its plans for GPT-5, but Lightcap expressed optimism about the model’s potential to reason.
Getting AI models to reason and plan is a critical step towards reaching artificial general intelligence (AGI). Multiple definitions of AGI exist, but it can be simply described as a sort of AI capable of performing at or beyond human levels on a broad range of activities.
Some scientists and experts have expressed concerns about building technology that will outperform human abilities. AI godfathers Yoshua Bengio, and Geoffrey Hinton have even warned us against the threats to humanity posed by AI. Both Meta and OpenAI claim to be aiming for AGI, which could be worth trillions for the company that achieves it.
The EU regulators are swiftly moving to conclude a preliminary investigation into Microsoft’s relationship with OpenAI, according to Margrethe Vestager, the EU’s antitrust chief. The probe, initiated in January, aims to determine whether Microsoft’s substantial investment of $13 billion into OpenAI should undergo scrutiny under the EU merger regulations. Vestager indicated in an interview with Bloomberg TV that a resolution is forthcoming, highlighting ongoing discussions with other regulatory authorities.
Vestager emphasised that the EU authorities closely monitor Microsoft’s investments and the broader trend of large tech companies investing in AI. The scrutiny extends beyond Microsoft to include other significant AI investments from major tech firms like Google, Amazon, and Nvidia. The EU mainly ensures competitiveness and prevents anti-competitive practices in this rapidly evolving AI landscape.
Microsoft’s involvement with OpenAI represents a significant stake, with the tech giant investing in other AI ventures, such as French startup Mistral and acquiring the team from Inflection AI. This investment landscape extends to other major players like Google and Amazon, which have their stakes in AI ventures. Vestager stressed the importance of vigilance in this emerging field, characterising it as a critical area for regulatory oversight to safeguard competition and innovation in the AI sector.
Three major players in the AI field, OpenAI, Google, and Mistral, have unveiled new versions of their cutting-edge AI models within 12 hours, signalling a burst of innovation anticipated for the summer. Meta’s Nick Clegg hinted at the imminent release of Meta’s Llama 3 at an event in London, while Google swiftly followed with the launch of its Gemini Pro 1.5, a sophisticated large language model with a limited free usage tier. Shortly after, OpenAI introduced its milestone model, GPT-4 Turbo, which, like Gemini Pro 1.5, supports multimodal input, including images.
In France, Mistral, a startup founded by former Meta AI team members, debuted Mixtral 8x22B, a frontier AI model released as a 281GB download file, following an open-source philosophy. While this approach is criticised for potential risks due to a lack of oversight, it reflects a trend towards democratising access to AI models beyond the control of tech giants like Meta and Google.
Experts caution that the prevailing approach centred on large language models (LLMs) might be reaching its limitations. Meta’s chief AI scientist, Yann LeCun, challenges the notion of imminent artificial general intelligence (AGI) and emphasises the need for AI systems capable of reasoning and planning beyond language manipulation. LeCun advocates for a shift towards ‘objective-driven’ AI to achieve truly superhuman capabilities, thereby highlighting the ongoing evolution and challenges in the AI landscape.
In recent reports by The New York Times, the challenges faced by AI companies in acquiring high-quality training data have come to light. The New York Times elaborates on how companies like OpenAI and Google have navigated this issue, often treading in legally ambiguous territories related to AI copyright law.
OpenAI, for instance, resorted to developing its Whisper audio transcription model by transcribing over a million hours of YouTube videos to train GPT-4, its advanced language model. Although this approach raised legal concerns, OpenAI believed it fell within fair use. The company’s president, Greg Brockman, reportedly played a hands-on role in collecting these videos.
According to a Google spokesperson, there were unconfirmed reports of OpenAI’s activities, and both Google’s terms of service and robots.txt files prohibit unauthorised scraping or downloading of YouTube content. Google also utilised transcripts from YouTube, aligned with its agreements with content creators.
AI companies, including Google and OpenAI, are grappling with the dwindling availability of quality training data to improve their models. The future of AI training may involve synthetic data or curriculum learning methods, but these approaches still need to be proven. In the meantime, companies continue to explore various avenues for data acquisition, sometimes straying into legally contentious territories as they navigate this evolving landscape.
OpenAI has unveiled new features to enhance model performance and customizability, catering to developers seeking to optimise AI implementations for speed, accuracy, and cost-efficiency. These enhancements include a self-serve fine-tuning API for GPT-3.5, which has already facilitated the training of hundreds of thousands of models across various organisations since its launch in August 2023. Fine-tuning empowers models to grasp content deeply and augment their existing capabilities, resulting in superior outcomes tailored to tasks such as code generation, text summarisation, or personalised content creation.
One standout success story involves a leading job-matching platform that utilised fine-tuning GPT-3.5 Turbo to refine personalised job recommendations for users. By leveraging fine-tuning, it achieved an 80% reduction in token usage, allowing the platform to scale its messaging volume from under a million to roughly 20 million monthly messages, enhancing user engagement and satisfaction.
In addition to the fine-tuning API, OpenAI has introduced new features to provide developers with enhanced control and visibility over their fine-tuning projects. These features include epoch-based checkpoint creation, a comparative playground for model evaluation, third-party integrations with platforms like Weights and Biases, comprehensive validation metrics, and improved dashboard functionality.
Why does it matter?
The ChatGPT’s owner envisions widespread adoption of customised models across industries and use cases. It emphasises the importance of effectively scoping use cases, implementing robust evaluation systems, and leveraging the right techniques to optimise model performance.
OpenAI accommodates organisations seeking fully custom-trained models capable of comprehending complex domain-specific knowledge and behaviours. The company has expanded its Custom Models program, offering assisted fine-tuning services for organisations that require tailored AI solutions beyond the capabilities of standard fine-tuning. Assisted fine-tuning involves collaborative efforts with technical teams to implement advanced techniques, such as parameter-efficient fine-tuning, to maximise model performance across specific use cases or tasks. Thus, for organisations seeking to harness AI capabilities for personalised impact, OpenAI’s offerings provide a pathway towards tailored and effective AI implementations.
OpenAI unveiled a new AI tool, Voice Engine, capable of generating lifelike speech by analysing a mere 15-second audio sample, as announced in OpenAI’s blog post. The tool aims to offer reading assistance, aid translation efforts, and provide a voice for nonverbal individuals with speech conditions. Despite its potential benefits, OpenAI acknowledges the serious risks associated with the technology, especially during an election year.
Voice Engine, developed by OpenAI in late 2022, has undergone private testing with a select group of partners who have agreed to usage policies requiring explicit consent from original speakers and prohibiting unauthorised impersonation. OpenAI stresses the importance of transparency, with partners mandated to disclose that the voices are AI-generated and that all audio produced by Voice Engine includes watermarking for traceability.
OpenAI advocates for responsible deployment of synthetic voices, suggesting measures such as voice authentication and a ‘no-go voice list’ to prevent misuse. The company recommends phasing out voice-based authentication to access sensitive information like bank accounts. However, the widespread release of Voice Engine remains uncertain as OpenAI seeks feedback and evaluates the results of its small-scale tests before deciding on broader deployment.
Why does it matter?
The introduction of Voice Engine comes amid rising concerns over AI-generated deepfakes and their potential to disseminate misinformation, particularly in political contexts. Recent incidents, such as a fake robocall imitating President Biden and an AI-generated video of a Senate candidate, underscore the urgency of addressing advanced AI technologies’ ethical and societal implications.