Three major players in the AI field, OpenAI, Google, and Mistral, have unveiled new versions of their cutting-edge AI models within 12 hours, signalling a burst of innovation anticipated for the summer. Meta’s Nick Clegg hinted at the imminent release of Meta’s Llama 3 at an event in London, while Google swiftly followed with the launch of its Gemini Pro 1.5, a sophisticated large language model with a limited free usage tier. Shortly after, OpenAI introduced its milestone model, GPT-4 Turbo, which, like Gemini Pro 1.5, supports multimodal input, including images.
In France, Mistral, a startup founded by former Meta AI team members, debuted Mixtral 8x22B, a frontier AI model released as a 281GB download file, following an open-source philosophy. While this approach is criticised for potential risks due to a lack of oversight, it reflects a trend towards democratising access to AI models beyond the control of tech giants like Meta and Google.
Experts caution that the prevailing approach centred on large language models (LLMs) might be reaching its limitations. Meta’s chief AI scientist, Yann LeCun, challenges the notion of imminent artificial general intelligence (AGI) and emphasises the need for AI systems capable of reasoning and planning beyond language manipulation. LeCun advocates for a shift towards ‘objective-driven’ AI to achieve truly superhuman capabilities, thereby highlighting the ongoing evolution and challenges in the AI landscape.
In recent reports by The New York Times, the challenges faced by AI companies in acquiring high-quality training data have come to light. The New York Times elaborates on how companies like OpenAI and Google have navigated this issue, often treading in legally ambiguous territories related to AI copyright law.
OpenAI, for instance, resorted to developing its Whisper audio transcription model by transcribing over a million hours of YouTube videos to train GPT-4, its advanced language model. Although this approach raised legal concerns, OpenAI believed it fell within fair use. The company’s president, Greg Brockman, reportedly played a hands-on role in collecting these videos.
According to a Google spokesperson, there were unconfirmed reports of OpenAI’s activities, and both Google’s terms of service and robots.txt files prohibit unauthorised scraping or downloading of YouTube content. Google also utilised transcripts from YouTube, aligned with its agreements with content creators.
AI companies, including Google and OpenAI, are grappling with the dwindling availability of quality training data to improve their models. The future of AI training may involve synthetic data or curriculum learning methods, but these approaches still need to be proven. In the meantime, companies continue to explore various avenues for data acquisition, sometimes straying into legally contentious territories as they navigate this evolving landscape.
OpenAI has unveiled new features to enhance model performance and customizability, catering to developers seeking to optimise AI implementations for speed, accuracy, and cost-efficiency. These enhancements include a self-serve fine-tuning API for GPT-3.5, which has already facilitated the training of hundreds of thousands of models across various organisations since its launch in August 2023. Fine-tuning empowers models to grasp content deeply and augment their existing capabilities, resulting in superior outcomes tailored to tasks such as code generation, text summarisation, or personalised content creation.
One standout success story involves a leading job-matching platform that utilised fine-tuning GPT-3.5 Turbo to refine personalised job recommendations for users. By leveraging fine-tuning, it achieved an 80% reduction in token usage, allowing the platform to scale its messaging volume from under a million to roughly 20 million monthly messages, enhancing user engagement and satisfaction.
In addition to the fine-tuning API, OpenAI has introduced new features to provide developers with enhanced control and visibility over their fine-tuning projects. These features include epoch-based checkpoint creation, a comparative playground for model evaluation, third-party integrations with platforms like Weights and Biases, comprehensive validation metrics, and improved dashboard functionality.
Why does it matter?
The ChatGPT’s owner envisions widespread adoption of customised models across industries and use cases. It emphasises the importance of effectively scoping use cases, implementing robust evaluation systems, and leveraging the right techniques to optimise model performance.
OpenAI accommodates organisations seeking fully custom-trained models capable of comprehending complex domain-specific knowledge and behaviours. The company has expanded its Custom Models program, offering assisted fine-tuning services for organisations that require tailored AI solutions beyond the capabilities of standard fine-tuning. Assisted fine-tuning involves collaborative efforts with technical teams to implement advanced techniques, such as parameter-efficient fine-tuning, to maximise model performance across specific use cases or tasks. Thus, for organisations seeking to harness AI capabilities for personalised impact, OpenAI’s offerings provide a pathway towards tailored and effective AI implementations.
OpenAI unveiled a new AI tool, Voice Engine, capable of generating lifelike speech by analysing a mere 15-second audio sample, as announced in OpenAI’s blog post. The tool aims to offer reading assistance, aid translation efforts, and provide a voice for nonverbal individuals with speech conditions. Despite its potential benefits, OpenAI acknowledges the serious risks associated with the technology, especially during an election year.
Voice Engine, developed by OpenAI in late 2022, has undergone private testing with a select group of partners who have agreed to usage policies requiring explicit consent from original speakers and prohibiting unauthorised impersonation. OpenAI stresses the importance of transparency, with partners mandated to disclose that the voices are AI-generated and that all audio produced by Voice Engine includes watermarking for traceability.
OpenAI advocates for responsible deployment of synthetic voices, suggesting measures such as voice authentication and a ‘no-go voice list’ to prevent misuse. The company recommends phasing out voice-based authentication to access sensitive information like bank accounts. However, the widespread release of Voice Engine remains uncertain as OpenAI seeks feedback and evaluates the results of its small-scale tests before deciding on broader deployment.
Why does it matter?
The introduction of Voice Engine comes amid rising concerns over AI-generated deepfakes and their potential to disseminate misinformation, particularly in political contexts. Recent incidents, such as a fake robocall imitating President Biden and an AI-generated video of a Senate candidate, underscore the urgency of addressing advanced AI technologies’ ethical and societal implications.
OpenAI, the Microsoft-backed company in AI, is gearing up to establish its presence in Tokyo this April, marking its foray into Asia as it expands its global operations. This move comes after the successful establishment of offices in London and Dublin last year, according to a source familiar with the matter who preferred to remain anonymous due to the sensitive nature of the information.
The decision to set up in Japan underscores the growing significance of the Asian market in AI development and adoption. OpenAI’s presence in Tokyo signifies a strategic move to tap into Japan’s burgeoning interest in AI technology, with major players like SoftBank Corp. and Nippon Telegraph and Telephone Corp. actively venturing into AI-driven services tailored for Japanese speakers.
The AI giant’s expansion plans in Japan have been in the pipeline for some time, with discussions intensifying after a meeting between OpenAI’s CEO, Sam Altman, and Prime Minister Fumio Kishida last April. Altman expressed the organisation’s intent to bolster its Japanese language services and collaborate with the government to address potential risks and establish regulatory frameworks.
Why does it matter?
OpenAI’s decision to establish a foothold in Tokyo reflects its commitment to catering to the Japanese market’s evolving needs and its broader strategy to engage with international stakeholders and foster AI innovation on a global scale. As the demand for AI-powered solutions continues to surge worldwide, OpenAI’s move into Tokyo signals a significant milestone in its quest to shape the future of AI.
Microsoft has outlined its commitment to safeguarding customer data privacy as businesses increasingly utilise generative AI tools such as Azure OpenAI Service and Copilot. In a blog post published on 28 March, the tech giant assured that customer organisations leveraging these services are protected under existing privacy policies and contractual agreements. Notably, Microsoft emphasised that organisations’ data is only utilised to train OpenAI models or foundational models if explicitly permitted by the users.
The tech giant clarified that customer data used in its generative AI solutions, including Azure OpenAI Service and Copilot, is not accessible for training open-source AI, addressing concerns raised by data privacy experts in the past. Furthermore, Microsoft affirmed that it does not share customer data with third parties like OpenAI without explicit permission, nor does it use it to train OpenAI’s foundational models. Any fine-tuned AI solutions resulting from organisations using their data will remain exclusive to them and not be shared externally.
The blog post highlights measures to protect organisations from copyright infringement lawsuits related to using Azure OpenAI and Microsoft Copilot services. Through the 2023 Customer Copyright Commitment plan, Microsoft pledged to defend customers and cover settlements in the event of copyright infringement lawsuits, provided customers utilise available guardrails and content filters within the products.
In addition to copyright protection, Microsoft is focused on safeguarding sensitive data associated with AI usage. Chief Privacy Officer Julie Brill detailed how Microsoft Purview enables corporate customers to identify risks linked to AI usage, including sensitive prompts. Azure OpenAI and Copilot users can employ sensitivity labels and classifications to protect their sensitive data, with Copilot summarising content only when authorised by users. This integration ensures that Copilot-generated output inherits sensitivity labels from reference files, maintaining data protection policies and preventing unauthorised access.