- Chatbots Life
- Posts
- Conversational UX Mini-Course
Conversational UX Mini-Course
Discover New CUX Design Stack & What it takes to Design & Build Conversational Agents.
Welcome to the Conversational UX Mini-Course
In just 15 minutes, learn the essentials of designing and building AI Agents with our quick guide to Conversational User Experience (Conversational UX). Dive into practical techniques and the latest technologies that are transforming how conversational agents are developed, especially with the help of NLLP, NLU and Large Language Models (LLMs).”
What You Will Learn:
New Design Stack: Explore the components of the modern conversational AI project stack, including flows, intents, entities, LLM integration, and the art of prompt engineering.
Impact of LLMs: Discover how LLMs streamline the creation of conversational agents, reducing the need for intricate intent flows and enabling rapid, effective communication solutions.
Overcoming Bias & Hallucination Management: Delve into the ethical dimensions of conversational UX, addressing biases inherent in AI technologies, and strategies to mitigate their impact on system outputs.
Time: 15 Min
Table of Contents
Conversational Design Stack
Conversational AI Project consist of the following design stack:
1. Flows, Intents & Entities: These will still be useful in use cases where the business needs very specific answers. Additionally, they can be used as follow up action to interest a user has shown in a particular product or service.
2. Large Language Models (LLMs): Integrating Large Language Models into the your project at multiple stages. LLMs can be used to interpret user intent, entities and sentiments. They can be used to answer questions. Multiple LLMs can be used in a single project, where each LLMs has a specific function.
3. Prompt Engineering: Now the fun begins. Prompt Engineering includes all of the instructions given to an LLM in order to get back high quality, accurate responses. This includes giving the LLM are role, for example "you are a super start customer service agent" and prompt chaining.
4. Knowledge Bases: This entail training the LLMs on your own data. FAQs and website information is a great example of this. The main challenge here will be in designing knowledge bases in such a way so the the LLM can understand the information quickly and accurately. Knowledge Base design is key!
5. Storytelling, Tone & Personality: Basically at all times, we all have a story going in our minds. The story is about what we want, what we believe will happen when we get it and if we are on the right track. Even as you are reading this sentence, you are comparing this information to that story and evaluating its fit. Designers will need to consider the Customers Journey, what story customers have in their minds and create a design, narrative and tone that fit the expectation and goals of the user.
Role of LLMs
LLMs are redefining how we build Conversational Agents. Today, LLMs can do most of the heavy lifting in minutes. The days of building hundreds of flows of every useful intent or FAQ question are gone. Instead, LLMs can be trained on a knowledge base and be able to give great answers.
The role of a Conversation Designer has changed. Designers now have to start thinking differently. Instead of thinking linearly and having a flow-based approach to design, now we need to think more about knowledge, information, data and how all of this leads to mean making.
Through stories, knowledge bases and prompt engineering, Designers will soon be creating experiences we once only dreamed of. The Job has changed and it is now more creative then ever.
Will the Web Become more Conversational?
As Conversational AI Agents evolve and become more and more commonplace in our everyday lives we need to imagine and consider the implications to our businesses.
Right now, Search Generative Experience (SGE), which is Google's AI Search Engine Bot, is decreasing website traffic by 18% to 68% according to latest data from Search Engine Land. SGE is only in its infancy and at some point it will be able to answer most questions much better and more directly than the average person could after reading the corresponding information. In other words, the Web as we know it is changing.
Websites now have to change. To date, websites have been like digital magazines. They are mostly made up of static content.
In the future websites will have to become destinations. A place where people go to meet, have dynamic, personalized experiences that SGE can not give them.
Main Challenges: Accuracy, Bias & Hallucinations
What is an LLM?
When a new technology really wows and gets us excited it becomes a part of us. We make it ours and we anthropomorphize it. We project human like qualities on it and this can hold us back from really understanding what we are actually dealing it.
So let's consider a few questions. Mainly what is an LLM and what are its limitations?
Perhaps these questions and ideas will illuminate our understanding:
Is an LLM a program?
Is an LLM a knowledge base?
Does an LLM know anything?
If an LLM is a program how does it compute its 70-100 Billion parameters in only a few seconds?
If an LLM is a knowledge base, why does it need to predict?
How can an LLM Model with Billions of parameters that has been trained on pretty much the entire internet, fit on a 100GB drive?
What are some simple tasks that LLMs can't do?
Hopefully these questions dispel some of the mystic around LLMs. There are a number of things that most people believe about them that are contradictory and wrong.
First, LLMS are not knowledge bases and they are not really programs either. What they are is a statistical representation of a knowledge bases.
In other words, an LLM like ChatGPT4 has been trained on hundreds of billions of parameters that it has condensed into statistical patterns. It does't have any knowledge but it has the patters of knowledge. When you ask it a question, it predicts the answer based on its statistical model.
A good way to think about this is that it is basically like a filter. Your question becomes the filter and configures the shape of the model so the right content or words drip out.
If you want to learn more about LLMs and how they work check out our LLM Section.
What is Bias?
Bias is an inclination to favor or oppose a person, group, idea, or thing, often in an unfair, inaccurate, or prejudicial way. Biases can be conscious or unconscious, and can be held by individuals, groups, or institutions
Bias has its roots in our preferences. For example, if you prefer coffee over tea, you are more likely to show bias towards coffee. You might believe that more people drink coffee, that coffee increases your mental focus better, even that it is healthier. At the bare minimum, you will have more information about coffee which will skew how you view both coffee and tea.
Why do Model’s Hallucinate?
LLMs are Statistical Representations of Knowledge Bases. They have taken the world's information and knowledge and boiled it down to statistical principles.
These principles are like icons. Icons represent something much more than what they are. They are a low resolution images that represent a much bigger chunk of information. They give you a lot more information than meets the eye.
When we combine both low-resolution models and bias we will have a hallucinations and poor accuracy.
How does this work and when does it happen?
It happens when you ask a detailed question about something specific. In our example, if you ask detailed questions about the icon, the model might make up those details in a way that conforms to its biases.
Exploring Current Solutions
The consequences of unethical AI are profound and multifaceted, affecting every level of society. There would be breakdown at each level of responsibility.
1) Awareness
Awareness is the first step to address both issues: Bias and Accuracy. By understanding what LLMs actually are, how the work, and what their limitations are we can build much better products.
2) Knowledge Bases & RAG
Next, we want to train the Models on our own data. By doing so, we can limit bias and give the model the right information so it can answer questions about our business.
You can also use your proprietary data to build unique and personalized experiences that delight customers.
Here are the variables to consider:
Hierarchical Structure: This relates to how information is organized in your company and the overall hierarchy of data.
Document Context & Semitics: This relates to how information is organized within a document, pdf, or webpage and how it relates to the context and overall meaning.
3) Prompt Engineering
This is the other great way to work around the limitations of LLMs.
Prompt Engineering is all about triggering the model to find the right 'icon' and then to follow up and go deeper into that 'icon' by asking the right contextual questions.
Here are a few variables to consider:
Consider your overall goal
How many steps are in your goal?
What knowledge is required to accomplish the goal?
Who is the right role or person for the job?
What is the context of your goal?
What are the instructions for each step?
Using these parameters, you can break it down into multiple Prompt and give very clear, focused instructions in each one.
Next: Deepen your Understanding of AI or jump into building your own AI Agent.
Congrats! You know know more about Conversational AI then most people. Next you can go deeper into Understanding AI or jump into Building an AI Agent in our Course.
Special Thanks to our Sponsors
Need a personal assistant? We do too, that’s why we use AI.
Ready to embrace a new era of task delegation?
HubSpot’s highly anticipated AI Task Delegation Playbook is your key to supercharging your productivity and saving precious time.
Learn how to integrate AI into your own processes, allowing you to optimize your time and resources, while maximizing your output with ease.
The fastest way to build AI apps
Writer is the full-stack generative AI platform for enterprises. Quickly and easily build and deploy AI apps with Writer AI Studio, a suite of developer tools fully integrated with our LLMs, graph-based RAG, AI guardrails, and more.
Use Writer Framework to build Python AI apps with drag-and-drop UI creation, our API and SDKs to integrate AI into your existing codebase, or intuitive no-code tools for business users.
Reply