Snapchat introduces advanced AI-powered AR features

Snap Inc, the owner of Snapchat, has unveiled a new iteration of its generative AI technology, enabling users to apply more realistic special effects when using their phone cameras. That move aims to keep Snapchat ahead of its social media competitors by enhancing its augmented reality (AR) capabilities, which superimpose digital effects onto real-world photos and videos.

In addition to this AI upgrade, Snap has introduced an enhanced version of its developer program, Lens Studio. The upgrade will significantly reduce the time required to create AR effects, cutting it from weeks to hours. The new Lens Studio also incorporates generative AI tools, including an AI assistant to help developers and a feature that can generate 3D images from text prompts.

Bobby Murphy, Snap’s chief technology officer, highlighted that these tools expand creative possibilities and are user-friendly, allowing even newcomers to create unique AR effects quickly. Plans for Snap include developing full-body AR experiences, such as generating new outfits, which are currently challenging to produce.

SewerAI utilises AI to detect sewer pipe issues

Sewage failures exacerbated by climate change and ageing infrastructure are becoming increasingly costly and common across the United States. The Environmental Protection Agency estimates that nearly $700 billion is required over the next two decades to maintain existing wastewater and stormwater systems. In response to these challenges, Matthew Rosenthal and Billy Gilmartin, veterans of the wastewater treatment industry, founded SewerAI five years ago. Their goal was to leverage AI to improve the inspection and management of sewer infrastructure.

SewerAI’s AI-driven platform offers cloud-based subscription products tailored for municipalities, utilities, and private contractors. Their tools, such as Pioneer and AutoCode, streamline field inspections and data management by enabling inspectors to upload data and automatically tag issues. That approach enhances efficiency and helps project managers plan and prioritise infrastructure repairs based on accurate 3D models generated from inspection videos.

Unlike traditional methods that rely on outdated on-premise software, SewerAI’s technology increases productivity and reduces costs by facilitating more daily inspections. The company has distinguished itself in the competitive AI-assisted pipe inspection market by leveraging a robust dataset derived from 135 million feet of sewer pipe inspections. This data underpins their AI models, enabling precise defect detection and proactive infrastructure management.

Recently, SewerAI secured $15 million in funding from investors like Innovius Capital, bringing their total raised capital to $25 million. This investment will support SewerAI’s expansion efforts, including AI model refinement, hiring initiatives, and diversification of their product offerings beyond inspection tools. The company anticipates continued growth as it meets rising demand and deploys its technology to empower organisations to achieve more with existing infrastructure budgets.

AI award-winning headless flamingo photo found to be real

A controversial AI-generated photo of a headless flamingo has ignited a heated debate over the ethical implications of AI in art and technology. The image, which was honored in the AI category of the 1839 Awards’ Color Photography Contest, has drawn criticism and concern from various sectors, including artists, technologists, and ethicists. 

The photo, titled ‘F L A M I N G O N E,’ depicts a flamingo without its head. It was created by photographer Miles Astray using a sophisticated AI model designed to generate lifelike images. Contrary to initial impressions, the photo wasn’t generated from a text prompt but was instead based on a real — and not at all beheaded — flamingo that Astray captured on the beaches of Aruba two years ago. After the photo won both third place in the category and the People’s Vote award, Astray revealed the truth, leading to his disqualification.

Proponents of AI-generated art assert that such creations push the boundaries of artistic expression, offering new and innovative ways to explore and challenge traditional concepts of art. They argue that the AI’s ability to produce unconventional and provocative images can be seen as a form of artistic evolution, allowing for greater diversity and creativity in the art world. However, detractors highlight the potential risks and ethical dilemmas posed by such technology. The headless flamingo photo, in particular, has been described as unsettling and inappropriate, sparking a broader conversation about the limits of AI-generated content. Concerns have been raised about the potential for AI to produce harmful or distressing images, and the need for guidelines and oversight to ensure responsible use.

The release of the headless flamingo photo has prompted a range of responses from the art and tech communities. Some artists view the image as a provocative statement on the nature of AI and its role in society, while others see it as a troubling example of the technology’s potential to create disturbing content. Tech experts emphasise the importance of developing ethical frameworks and guidelines for AI-generated art. They argue that while AI has the potential to revolutionize creative fields, it is crucial to establish clear boundaries and standards to prevent misuse and ensure that the technology is used responsibly.

‘‘F L A M I N G O N E’ accomplished its mission by sending a poignant message to a world grappling with ever-advancing, powerful technology and the profusion of fake images it brings. My goal was to show that nature is just so fantastic and creative, and I don’t think any machine can beat that. But, on the other hand, AI imagery has advanced to a point where it’s indistinguishable from real photography. So where does that leave us? What are the implications and the pitfalls of that? I think that is a very important conversation that we need to be having right now.”, Miles Astray told The Washington Post.

Why does it matter?

The controversy surrounding the AI-generated headless flamingo photo highlights the broader ethical challenges posed by artificial intelligence in creative fields. As AI technology continues to advance, it is increasingly capable of producing highly realistic and complex images. That raises important questions about the role of AI in art, the responsibilities of creators and developers, and the need for ethical guidelines to navigate these new frontiers.

McDonald’s halts AI ordering test in drive-thrus

McDonald’s has decided to discontinue the use of AI ordering technology that was being tested at over 100 drive-thru locations in the US. The company had collaborated with IBM to develop and test this AI-driven, voice-automated system. Despite this decision, McDonald’s remains committed to exploring AI solutions, noting that IBM will remain a trusted partner in other areas. The discontinuation of this specific technology is set to occur by 26 July 2024.

The partnership between McDonald’s and IBM began in 2021 as part of McDonald’s ‘Accelerating the Arches’ growth plan, which aimed to enhance customer experience through Automated Order Taking (AOT) technology. IBM highlighted the AOT’s capabilities as being among the most advanced in the industry, emphasising its speed and accuracy. Nonetheless, McDonald’s is reassessing its strategy for implementing AOT and intends to find long-term, scalable AI solutions by the end of 2024.

McDonald’s move to pause its AI ordering technology reflects broader challenges within the fast-food industry’s adoption of AI. Other chains like White Castle and Wendy’s have also experimented with similar technologies. However, these initiatives have faced hurdles, including customer complaints about incorrect orders due to the AI’s difficulty in understanding different accents and filtering out background noise. Despite these setbacks, the fast-food sector continues to push forward with AI innovations to improve operational efficiency and customer service.

FCC names Royal Tiger as first official AI robocall scammer gang

The US Federal Communications Commission (FCC) has identified Royal Tiger as the first official AI robocall scammer gang, marking a milestone in efforts to combat sophisticated cyber fraud. Royal Tiger has used advanced techniques like AI voice cloning to impersonate government agencies and financial institutions, deceiving millions of Americans through robocall scams.

These scams involve automated systems that mimic legitimate entities to trick individuals into divulging sensitive information or making fraudulent payments. Despite the FCC’s actions, experts warn that AI-driven scams will likely increase, posing significant challenges in protecting consumers from evolving tactics such as caller ID spoofing and persuasive social engineering.

While the FCC’s move aims to raise awareness and disrupt criminal operations, individuals are urged to remain vigilant. Tips include scepticism towards unsolicited calls, utilisation of call-blocking services, and verification of caller identities by contacting official numbers directly. Avoiding sharing personal information over the phone without confirmation of legitimacy is crucial to mitigating the risks posed by these scams.

Why does it matter?

As technology continues to evolve, coordinated efforts between regulators, companies, and the public are essential in staying ahead of AI-enabled fraud and ensuring robust consumer protection measures are in place. Vigilance and proactive reporting of suspicious activities remain key in safeguarding against the growing threat of AI-driven scams.

AI tools struggle with election questions, raising voter confusion concerns

As the ‘year of global elections’ reaches its midpoint, AI chatbots and voice assistants are still struggling with basic election questions, risking voter confusion. The Washington Post found that Amazon’s Alexa often failed to correctly identify Joe Biden as the 2020 US presidential election winner, sometimes providing irrelevant or incorrect information. Similarly, Microsoft’s Copilot and Google’s Gemini refused to answer such questions, redirecting users to search engines instead.

Tech companies are increasingly investing in AI to provide definitive answers rather than lists of websites. This feature is particularly important as false claims about the 2020 election being stolen persist, even after multiple investigations found no fraud. Trump faced federal charges for attempting to overturn Biden’s victory, who won decisively with over 51% of the popular vote.

OpenAI’s ChatGPT and Apple’s Siri, however, correctly answered election questions. Seven months ago, Amazon claimed to have fixed Alexa’s inaccuracies, and recent tests showed Alexa correctly stating Biden won the 2020 election. Nonetheless, inconsistencies were spotted last week. Microsoft and Google, in return, said they avoid answering election-related questions to reduce risks and prevent misinformation,, a policy also applied in Europe due to a new law requiring safeguards against misinformation.

Why does it matter?

Tech companies are increasingly tasked with distinguishing fact from fiction as it develops AI-enabled assistants. Recently, Apple announced a partnership with OpenAI to enhance Siri with generative AI capabilities. Concurrently, Amazon is set to launch a new AI version of Alexa as a subscription service in September, although it remains unclear how it will handle election queries. An early prototype struggled with accuracy, and internal doubts about its readiness persist. The new AI assistants from Amazon and Apple aim to merge traditional voice commands with conversational capabilities, but experts warn this integration may pose new challenges.

G7 summit underscores ethical AI, digital inclusion, and global solidarity

The G7 leaders met with counterparts from several countries, including Algeria, Argentina, Brazil, and India, along with heads of major international organisations such as the African Development Bank and the UN, to address global challenges impacting the Global South. They emphasised the need for a unified and equitable international response to these issues, underscoring solidarity and shared responsibility to ensure inclusive solutions.

Pope Francis made an unprecedented appearance at the summit, contributing valuable insights on AI. The leaders discussed AI’s potential to enhance industrial productivity while cautioning against its possible negative impacts on the labour market and society. They stressed the importance of developing AI that is ethical, transparent, and respects human rights, advocating for AI to improve services while protecting workers.

The leaders highlighted the necessity of bridging digital divides and promoting digital inclusion, supporting Italy’s proposal for an AI Hub for Sustainable Development. The hub aims to strengthen local AI ecosystems and advance AI’s role in sustainable development.

They also emphasised the importance of education, lifelong learning, and international mobility to equip workers with the necessary skills to work with AI. Finally, the leaders committed to fostering cooperation with developing and emerging economies to close digital gaps, including the gender digital divide, and achieve broader digital inclusion.

AI in news sparks global concerns

A new report from the Reuters Institute for the Study of Journalism highlights growing global concerns about the use of AI in news production and the spread of misinformation. The Digital News Report, based on surveys of nearly 100,000 people across 47 countries, reveals that consumers are particularly uneasy about AI-generated news, especially on sensitive topics like politics. In the US, 52% of respondents expressed discomfort with AI-produced news; this figure was 63% in the UK.

The report underscores the challenges newsrooms face in maintaining revenue and trust. Concerns about the reliability of AI-generated content are significant, with 59% of global respondents worried about false news, which rises to 81% in South Africa and 72% in the US, both of which are holding elections this year. Additionally, the reluctance of audiences to pay for news subscriptions remains a problem, with only 17% of respondents in 20 countries paying for online news, a figure unchanged for three years.

Why does it matter?

A significant trend noted in the report is the growing influence of news personalities on platforms like TikTok. Among 5,600 TikTok users surveyed, 57% said they primarily follow individual personalities for news, compared to 34% who follow journalists or news brands. The report suggests that newsrooms must establish direct relationships with their audiences and strategically use social media to reach younger, more elusive viewers. The shift is illustrated by figures like Vitus ‘V’ Spehar, a TikTok creator known for delivering news uniquely and engagingly.

IOC implements AI for athlete safety at Paris Olympics

The International Olympic Committee (IOC) will deploy AI to combat social media abuse directed at 15,000 athletes and officials during the Paris Olympics next month, IOC President Thomas Bach announced on Friday. With the Games set to begin on 26 July, more than 10,500 athletes will compete across 32 sports, generating over half a billion social media engagements.

The AI system aims to safeguard athletes by monitoring and automatically erasing abusive posts to provide extensive protection against cyber abuse. That initiative comes amid ongoing global conflicts, including the wars in Ukraine and Gaza, which have already led to social media abuse cases.
Russian and Belarusian athletes, who will compete as neutral athletes without their national flags, are included in the protective measures. The IOC did not specify the level of access athletes would need to grant for the AI monitoring.

Despite recent political developments in France, including a snap parliamentary election called by President Emmanuel Macron, Bach assured that preparations for the Olympics remain on track. He emphasised that both the government and opposition are determined to ensure that France presents itself well during the Games.

In the beginning was the word, and the word was with the chatbot, and the word was the chatbot

By introducing the argument to discuss, there is not much need to mention how important the word, respectively, the language and its narrow disciplines, is and what we humans have achieved in time through our enriched communication systems, especially in technological and diplomatic contexts where the word is an essential and powerful instrument

Since linguistics, especially nowadays, is an inseparable element from the realm of technology, it is absolutely legitimate to question the way chatbots, the offshoots of the latest technology, work. In other words, it is legitimate to question the way chatbots learn through digital, that is, algorithmic cognition and the way they accurately and articulately express themselves in response to someone’s most diverse queries or inputs.

What makes the human-like cognitive power of deep learning LLMs?

To understand AI and the epicentre of its evolution, chatbots, which interact with people by responding to most different prompts, we should delve into the branches of linguistics called semantics and syntax, and the process of learning and elaboration of most diverse and articulated info by chatbots. 

The complex understanding of language and how it is being assimilated by humans, (and in this case) by deep learning machines, was explained as far back as in some segments of language studies by Ferdinand de Saussure.

For that reason, we will explore the cognitive mechanisms underlying semantics and syntax in large language models (LLMs) such as ChatGPT, integrating the theoretical perspectives of one of the most renowned linguistic philosophers such as Saussure. By synthesising linguistic theories with contemporary AI methodologies, the aim is to provide a comprehensive understanding of how LLMs process, understand and generate natural language. What follows is a modest examination of the models’ training processes, data integration, and real-time interaction with users, highlighting the interplay between linguistic theories and AI language assimilation systems.

Overview of Saussure’s studies related to synta(x)gmatic relations and semantics 

 Face, Head, Person, Photography, Portrait, Adult, Male, Man, Mustache, Clothing, Coat, Accessories, Formal Wear, Tie, Ferdinand de Saussure

Starting with Ferdinand de Saussure, one of the first linguistic scientists of the 20th century (along with Charles Sanders Peirce and Leonard Bloomfield), and an introduction to syntax and semantics from the reading ‘Course in General Linguistics’, he depicts language as a scientific phenomenon, emphasising the synchronic study of language, focusing on its current state rather than its historical evolution, in a structuralist view, with syntax and semantics as some of the fundamental components of its structure. 

Syntax

Syntax, within this framework, is a grammar discipline which represents and explains the systematic and linear arrangement of words and phrases to form meaningful sentences within a given language. Saussure views syntax as an essential aspect of language, an abstract language system, which encompasses grammar, vocabulary, and rules. He argues that syntax operates according to inherent principles and conventions established within a linguistic community rather than being governed by individual speakers. His structuralist approach to linguistics highlights the interdependence between syntax and other linguistic elements, such as semantics, phonology and morphology, within the overall structure of language.

Semantics

Semantics is a branch of linguistics and philosophy concerned with the study of meaning in language. It explores how words, phrases, sentences, and texts convey meaning and how interpretation is influenced by context, culture, and usage. Semantics covers various aspects, including the meaning of words (lexical semantics), the meaning of sentences (compositional semantics or syntax), and the role of context in understanding language (pragmatics).

However, one of Saussure’s biggest precepts within semantics posits that language is a system of signs composed of the signifier (sound/image) and the signified (concept). This dyadic structure is crucial for understanding how LLMs process the understanding of words and their possible ambiguity. 

 Lighting, Nature, Night, Outdoors, Art, Graphics, Light, Person, Face, Head

How do chatbots cognise semantics and syntax in linguistic processes?

Chatbots’ processing and understanding of language usage involves several key steps: training on vast amounts of textual data from the internet to predict the next word in a sequence; tokenisation to divide the text into smaller units; learning relationships between words and phrases for semantic understanding; using vector representations to recognise similarities and generate contextually relevant responses; and leveraging transformer architecture to efficiently process long contexts and complex linguistic structures. Although it does not learn in real time, the model is periodically updated with new data to improve performance, enabling it to generate coherent and useful responses to user queries.

As explained earlier, in LLMs, words and phrases are tokenised and transformed into vectors within a high-dimensional space. These vectors function similarly to Saussure’s signifiers, with their positions and relationships encoding meaning (the signified). Thus, within the process of ‘Tokenisation and Embedding,’ LLMs tokenise text into discrete units (signifiers) and map them to embeddings that capture their meanings (signified). The model learns these embeddings by processing vast amounts of text, identifying patterns and relationships analogous to Saussure’s linguistic structures.

Chatbots’ ability to understand and generate text relies on their grasp of semantics (meaning) and syntax (structure). It processes semantics through contextual word embeddings that capture meanings based on usage, an attention mechanism that weighs word importance in context, and layered contextual understanding that handles polysemy and synonymy. The model is pre-trained on general language patterns and fine-tuned on specific datasets for enhanced semantic comprehension. For syntax, it uses positional encoding to understand word order, attention mechanisms to maintain syntactic coherence, layered processing to build complex structures, and probabilistic grammar learning from vast text exposure. Tokenisation and sequence modelling help track dependencies and coherence, while the transformer model integrates syntax and semantics at each layer, ensuring that responses are both meaningful and grammatically correct. Training on diverse datasets further enhances its ability to generalise across various language uses, making the chatbot a powerful natural language processing tool.

Interesting invention..

Recently, researchers in the Netherlands developed an AI platform capable of recognising sarcasm, which was presented at the Acoustical Society of America and Canadian Acoustical Association meeting. By training a neural network with the Multimodal Sarcasm Detection Dataset (MUStARD) using video clips and text from sitcoms like ‘Friends’ and ‘The Big Bang Theory,’ the large language model accurately detected sarcasm in about 75% of unlabeled exchanges.

Sarcasm generally takes the form of a, linguistically speaking, layered and ironic remark, often rooted in humour, that is intended to mock or satirise something. When a speaker is being sarcastic, they say something different than what they actually mean, and that’s why it is hard for a large language machine to detect such nuances in someone’s speech.

This process leverages deep learning techniques that analyse both syntax and semantics and the concepts of syntagma and idiom to understand the layered structure and meaning of language and how comprehensive the acquisition of human speech by an LLM is.

By integrating Saussure’s linguistic theories with the cognitive mechanisms of large language models, we gain a deeper understanding of how these models process and generate language. The interplay between structural rules, contextual usage, and fluidity of meaning partially depicts the sophisticated performance of LLMs’ language generation. This synthesis not only illuminates the inner workings of contemporary AI systems but also reinforces the enduring relevance of classical linguistic theories in the age of AI.