Natural Language UI - The Next Big Leap in UX?
Introduction
This is the second part of my "AI application architecture” series. You can read the first one about the overall architecture here:
In this article, I will dive deeper into natural language user interfaces and how they could change the way we interact with computers today.
Why current UIs are broken?
Let’s face it: the UI has not much evolved since Apple popularized touch and haptic user interfaces. We are still mainly communicating with forms, lists/grids, and buttons. While they are good for simple browsing of data, they really suck at:
In the next chapters, I go through the possible solutions for these problems and how AI could help to overcome these.
Build relationships, not single conversations
Imagine a situation when you go to work and you would always need to start from scratch with your colleagues. Even though your work would be saved, all previous discussions would have been lost. You would always need to explain your goals and needs, try to bind relationships, and build co-operations. This is what the current IT systems feel like, it’s a constant groundhog day!
What if the systems would learn and adapt to your needs? They would remember who you are and what is your history with them. Even before you tell, they could remember you and serve you with the data you are after. They would be more of your co-worker than a computer.
But isn’t that something we could already do with current UIs? Just save the typical paths and clicks the user is making and learn from there. Yes, partly that is possible. But anyways those paths and selections are something we have coded into the system and then taught users to use those, when they need something. They don’t necessarily match the actual needs that the user has.
This is why we need to have Natural Language UIs, a natural way for the user to express their actual needs and wishes. An interface that is not limited by the choices made by the designers and developers. And more over it, a system, which can remember the past with you to create meaningful relationships!
This table summarizes the difference between old (HTML) and new (NLP) UIs.
Understanding and serving humans
There is a big difference between collecting data with pre-defined inputs and really understanding the user needs. The former is just data and text that we use to fill the needed data structures in order to perform the functions we want. It does not understand the semantics of the inputs and the data models it feeds.
The nicest thing about the new large language models (LLMs) is that they do understand the semantics of the user language. They don’t actually know what a certain word or sentence is, but they understand its relationship to other words and sentences. That way we can use LLM to map the users’ natural language into needs, intentions, and even computer languages/data models.
Recommended by LinkedIn
There are various techniques to collect user understanding from a conversation. You can basically let the AI do the analysis by feeding in data schemas and asking AI to fill in the data based on some discussion history. With an understanding of semantics, it can quite easily map the topics from discussion to attribute names in schema. By asking from multiple angles, you can quite fast create a unique profile of the user.
Once you have a user profile collected, the possibility to personalize the service is outstanding. The old way of segmenting the users based on a few criteria will feel antique. Personal is really personal, only based on the communication history and the user’s unique profile. Say goodbye to pushy email marketing, which only tries to guess what the user wants!
With Generative AI, the sales and marketing stack will not be the same!
AI facilitating discussion / processes
Let's face it: many times when we have some need or problem, it is not possible to solve it right away in one session. Whether we are looking for a new car or planning a new project, it requires a lot of discussion and questions to be answered. There might be many people involved and sometimes you just need to wait for something to happen. This breaks the process and if you don't manage/poll it continuously, it might be lost forever.
What if we let AI facilitate and manage this process? We start the process with AI by introducing the problem and defining the needed outcome + people involved. After that AI could keep observing the process and polling information real time when something happens.
E.g. in car sales, we could have a conversation with the user to fetch the needs. If there isn't a correct car available, we would involve car salespeople to be aware of this need. Once a car that matches this need comes up, we could just contact the user automatically or by the salesperson.
AI would be aware of the communication channels that the user uses and the discussion would be seamless no matter the channel. It could be sending reminders if the process seems stuck. It could even try to keep the connection and discussion warm by different means like linking with interesting content.
With the NLP UI, the concept of logging into the web app will become less important. Conversation as a UI fits really well in the current messengers, emails, and even phones. Once the speech-to-text interfaces have evolved enough, I believe it will be normal to communicate with speech to users. While text chatting works well with laptops, for mobile the typing process is still typically too slow and difficult.
I believe Speech-to-Text is the way humans and computers communicate in the future!
Summary - “The Good Old Days” is back
I must confess I miss the “good old days” when most stuff was bought at brick&mortar stores. Call me old, but I cannot really get the same feeling and experience in eCommerce sites. It’s not only the human service and interaction but also all the expertise that sales staff had at that time. Being able to tell the staff about my preferences and tastes while discussing products and services.
Now when playing with OpenAI, I finally see a glimpse of that same experience. Finally, the computer is able to understand me and adapt to my personal needs! Though we are still living in the early days, current versions are good enough to beat the cold and impersonal eCommerce sites!
Who am I: a passionate software architect/developer with over 25 years of experience in multiple fields. Having seen the impact of AI is creating in our industry, I decided to quit my job and put all-in at building new world with AI. Just bootstrapping and designing multiple ideas to create a big impact!
Educator | UX expert | SW Engineer
2yIt would be good to define what you mean by UX in this context. It's such a psychological and context dependent concept. A smart, personal, and multimodal interface might be a good solution in some context but not necessarily so good in another. In some cases people might be really concerned about the scope of information collection and feel need to know the exact state of the system and have total control. It might be challenging to provide that with a multimodal system with a generative system involved. Often UX involves anticipated experience, repetitive experience, and retrospective experience in addition to situational experiences. The experience may also have practical and hedonic perspectives also. In this context it makes sense to clearly separate usability from UX as usability is easier to measure. At least if you define it in ISO standard way and measure effectiveness, efficiency, and satisfaction (as ergonomics and suitability to situation). The multimodal and smart interface may even reduce the objective usability of the UI since it may not always be consistent and might forget or lose users favourite settings. The user satisfaction may increase a lot in the beginning (if you are to trust the Kano model) because the novelty acts as excitement property. The satisfaction may reduce over time when the excitement wears of and the usability issues are realized. This may result in a long term UX that is described as a "mixed bag of feelings and opinions".
Hacker | Cyber security nerd
2yAnd what about personalized operating systems? How cool it would be if my phone learned the dimensions of my fingers and adjusted the places and sizes of elements on the screen accordingly.
Helping AIs cure cancer. AI generalist working from Spain. Experience in leadership, AI research, software engineering.
2yWe can surely fix all sorts of existing pain points in UIs using chatbots, and they can be used for many purposes. From personal assistants to customer interfaces, from search engines to coding assistants, from automatic consultants to virtual romantic partners, from automatic summarizers to filling our world with endless songs, from automatic game narratives to intelligent NPCs. And all this is only the very beginning. We can do so much with this tech, but we should be careful not to force people into a discussion where a button is enough.
Tech, Business, Progress
2yGreat post! I would love to see demos of these ideas.
Startup Ghost | Product Designer | UX Expert | Digital Strategist
2yGoof stuff, Pasi Vuorio! Although, personalization could be done without AI, but the way it could be represented/provided in UI is definitely worth to explore more.