-
-
-
-
URL copied!
There is a big difference between NLP (Natural Language Processing) and NLU (Natural Language Understanding).
NLP models, predicts, spots and generates word sequences learned over terabytes of uncurated language data scraped from the internet (hence potentially biased, fake or nonsensical).
On the other hand, NLU abstracts beyond the words and even word sequences and tries to assign some kind of meaning to the word sequences (user intent and goal, semantic category, ontology mapping).
There is a parallel in the Voice AI and Speech Recognition world: we juxtapose "word accuracy" (whether the automated speech recognition [ASR] got all the words right) and "concept accuracy" (whether the ASR got the right "words", i.e. extracted the right user intent or slot value).
A Speech IVR, Voicebot or Chatbot may only get half the words that you speak or type right (50% word accuracy), but may still get at what you actually want to say or do (100% concept accuracy). Very often the opposite can be true too; the ASR may recognise all the words typed in a chat, as they are already in its vocabulary (100% word accuracy), but may not have a clue what you're on about (0% concept accuracy), because this concept, goal, intent, slot, slot value has not yet been included in its domain model.
Statistics + Linguistics = Explainable NLU (xNLU)
This is why language analysis and modelling, whether it's for content extraction, text summarisation or Conversational Experience Design, should be a combination of robust data modelling and Computational Linguistics. Neither, by itself, can get it right:
- ❌ Conversation Design that is not based on a sufficient amount of real-world data (80k utterances is the standard in the Contact Centre world) is bound to give rise to Voicebots and Chatbots that will invariably fail in the face of actual language use and user behaviour, irrespective of how natural, friendly or even endearing the system messages are.
- ❌ A Conversation Designer who hasn't been trained in Linguistics is also bound to get the system messages wrong, by confusing, misleading or even putting the user off with messages that are too vague, ambiguous or chatty.
- ❌ Data Science that does not supplement string manipulation (NLP) with semantic ontologies and Human-Machine interaction patterns (NLU), will, too, result in natural language interfaces that have limited functionality, and are, hence, unhelpful, if not nonsensical.
- ❌ Engineers who design conversations and write system prompts always get the messaging wrong (verbose, full of jargon, too formal, confusing), which trips the user and the conversation.
Enter Google Imagen
More recently, the GOOGLE text-to-image diffusion model IMAGEN wowed everyone by generating incredibly realistic – usually beautiful – pictures. However, once again, it was trained on large, mostly uncurated, web-scraped datasets, meaning it unsurprisingly also replicates social biases and stereotypes, as the researchers themselves admit:
- "an overall bias towards generating images of people with lighter skin tones"
- "portraying different professions to align with Western gender stereotypes"
- "social and cultural biases when generating images of activities, events, and objects".
Good job the code is not in the public domain yet, as, in the wrong hands, Imagen can generate fake, defamatory or even harmful content. The future should be trodden with caution!
TL:DR
In short, don't get overexcited over GPT3 and such LMs, or even text-to-image algorithms such as IMAGEN; they may have impressive (creative, beautiful, funny, even scary) generative power, but next to no explanatory power. That is, despite the name, current large Language Models only represent a tiny fraction of what human language use is about and what human language conveys. The right terminology is "word sequence modelling", which does not equal actual language understanding (or even ... human language modelling)! We should instead strive for explainability: Explainable NLP, NLU, VUI and Conversation Design.
More about the author:
Dr Maria Aretoulaki is Principal Consultant CX Design (Voice & Conversational AI) at GlobalLogic UK&I. She has been designing Voice User Interfaces (VUIs and Speech IVRs) for the past 25+ years, mainly for Contact Centre automation and Customer Self-Service. In 2019, she coined the term "Explainable VUI Design" to promote the principle of data-based Conversation Design and Computational Linguistics-based Language Engineering.
Top Insights
Manchester City Scores Big with GlobalLogic
AI and MLBig Data & AnalyticsCloudDigital TransformationExperience DesignMobilitySecurityMediaTwitter users urged to trigger SARs against energy...
Big Data & AnalyticsDigital TransformationInnovationRetail After COVID-19: How Innovation is Powering the...
Digital TransformationInsightsConsumer and RetailTop Insights Categories
Let’s Work Together
Related Content
Accelerating Digital Transformation with Structured AI Outputs
Enterprises increasingly rely on large language models (LLMs) to derive insights, automate processes, and improve decision-making. However, there are two significant challenges to the use of LLMs: transforming structured and semi-structured data into suitable formats for LLM prompts and converting LLM outputs back into forms that integrate with enterprise systems. OpenAI's recent introduction of structured … Continue reading NLP vs NLU →
Learn More
If You Build Products, You Should Be Using Digital Twins
Digital twin technology is one of the fastest growing concepts of Industry 4.0. In the simplest terms, a digital twin is a virtual replica of a real-world object that is run in a simulation environment to test its performance and efficacy
Learn More
Accelerating Enterprise Value with AI
As many organizations are continuing to navigate the chasm between AI/GenAI pilots and enterprise deployment, Hitachi is already making significant strides. In this article, GlobaLogic discusses the importance of grounding any AI/GenAI initiative in three core principles: 1) thoughtful consideration of enterprise objectives and desired outcomes; 2) the selection and/or development of AI systems that are purpose-built for an organization’s industry, its existing technology, and its data; and 3) an intrinsic commitment to responsible AI. The article will explain how Hitachi is addressing those principles with the Hitachi GenAI Platform-of-Platforms. GlobalLogic has architected this enterprise-grade solution to enable responsible, reliable, and reusable AI that unlocks a high level of operational and technical agility. It's a modular solution that GlobalLogic can use to rapidly build solutions for enterprises across industries as they use AI/GenAI to pursue new revenue streams, greater operational efficiency, and higher workforce productivity.
Learn More
Share this page:
-
-
-
-
URL copied!