Summary: AI is introducing the third user-interface paradigm in computing history, shifting from command-based interaction to intent-based outcome specification, where users tell the computer what they want, not how to do it, reversing the locus of control.
ChatGPT and other AI systems are shaping up to launch the 3rd user-interface paradigm in the history of computing — the first new paradigm in more than 60 years.
The First Two Paradigms
Paradigm 1: Batch Processing
From the birth of computers around 1945, the first UI paradigm was batch processing. In this paradigm, users specified a complete workflow of everything they wanted the computer to do. This batch of instructions was submitted to a data center (often as a deck of punched cards) and was processed at some unspecified later time, often overnight.
Later, often the next morning, the user would pick up the result of his or her batch: usually, this would be a thick fanfold of printouts, but it could also be a new deck of punched cards. If the original batch contained even the slightest error, there would be no output, or the result would be meaningless because it would not reflect the complete input batch.
From a UI perspective, batch processing did not involve any back-and-forth between the user and the computer. The UI was a single point of contact: that batch of punched cards. Usability was horrible, and it was common to need multiple days to finetune the batch to the point where executing it would produce the desired end result.
Paradigm 2: Command-Based Interaction Design
Around 1964, the advent of time-sharing led to the second UI paradigm: command-based interaction. In this paradigm, the user and the computer would take turns, one command at a time. This paradigm is so powerful that it has dominated computing ever since — for more than 60 years. Command-based interactions have been the underlying approach throughout three generations of user-interface technology: command lines (like DOS and Unix), full-screen text-based terminals (common with IBM mainframes), and graphical user interfaces (GUI: Macintosh, Windows, and all current smartphone platforms). Powerful and long-lasting indeed.
The benefit of command-based interactions compared to batch processing is clear: after each command has been executed, the user can reassess the situation and modify future commands to progress toward the desired goal. In fact, users don’t even need to have a fully specified goal in mind because they can adjust their approach to the problem at hand as they get more information from the computer and see the results of their initial commands. At least that’s the case if the design follows the first of the 10 usability heuristics: visibility of system status. Early command-line systems often didn’t do this, with horrible usability as a result. For example, in Unix, no news is good news because you only get any feedback from the computer if your command results in an error message. No errors meant no information from the computer about the new state, which made it harder for users to compose the following command.
The beauty of GUIs is that they do show the status after each command, at least when designed well. The graphical user interface has dominated the UX world since the launch of the Macintosh in 1984: about 40 years of supremacy until it possibly is replaced by the next generation of UI technology and, more importantly, the next UI paradigm in the form of artificial intelligence.
(See follow-up article on why we won't continue the progression in geometric dimensionality and make 3D the next dominant UI style.)
The Newest Paradigm
Paradigm 3: Intent-Based Outcome Specification
I doubt that the current set of generative AI tools, like ChatGPT, Bard, etc., are representative of the UIs we’ll be using in a few years because they have deep-rooted usability problems. Their problems led to the development of a new role — the “prompt engineer.” Prompt engineers exist to tickle ChatGPT in the right spot so it coughs up the right results.
This new role reminds me of how we used to need specially trained query specialists to search through extensive databases of medical research or legal cases. Then Google came along, and anybody could search. The same level of usability leapfrogging is needed with these new tools: better usability of AI should be a significant competitive advantage. (And if you’re considering becoming a prompt engineer, don’t count on a long-lasting career.)
The current chat-based interaction style also suffers from requiring users to write out their problems as prose text. Based on recent literacy research, I deem it likely that half the population in rich countries is not articulate enough to get good results from one of the current AI bots.
That said, the AI user interface represents a different paradigm of the interaction between humans and computers — a paradigm that holds much promise.
As mentioned, in command-based interaction, the user issues commands to the computer one at a time, gradually producing the desired result. At least if the design has sufficient usability to allow people to understand what commands to issue at each step. Inherent to this paradigm, the computer is fully obedient and does exactly what it’s told. The downside is that low-usability systems often cause users to issue commands that do something different than what the user really wants.
With the new AI systems, the user no longer tells the computer what to do. Rather, he or she tells the computer what outcome they want. A classic example is “Make me a drawing suitable for the cover of a pulp SF magazine, showing a cowboy in a space suit on an airless planet with two red moons in the heavens.” Try ordering Photoshop circa 2021 to do that! Back then, you would have issued hundreds of commands to bring forth the illustration for your magazine client gradually. Today, Bing Image Creator made me four suggested images in a few seconds.
Image produced by Bing Image Creator from the above prompt.
The third UI paradigm, represented by current generative AI, is intent-based outcome specification. The user tells the computer the desired result but does not specify how this outcome should be accomplished, such as the steps to be executed. Compared to traditional command-based interaction design, this completely reverses the locus of control. I doubt we should even describe this user experience as an “interaction” because there is no turn-taking or gradual progress.
That said, in my science-fiction illustration example, I’m not happy with the space suits. This might be fixed by another round with the AI. Such rounds of gradual refinement are a form of interaction that is currently poorly supported, providing rich opportunities for usability improvements for those AI vendors who bother doing user research to discover better ways for average humans to control their systems.
“Do what I mean, not what I say” is a seductive UI paradigm — as mentioned, users often order the computer to do the wrong thing. On the other hand, assigning the locus of control entirely to the computer does have downsides, especially with current AI, which is prone to including erroneous information in its results. When users don’t know how something was done, it can be harder for them to correct the problem.
The intent-based paradigm doesn’t rise to the level of noncommand systems which I introduced in 1993. A true noncommand system doesn’t require the user to specify intent, because the computer acts as a side effect of the user’s normal actions. As an example, consider unlocking a car by pulling on the door handle: this is a noncommand unlock because the user would perform the same action whether the car is locked or unlocked. (In contrast, a car operated by voice-recognition could unlock the door as a result of the user stating “I want the car to be unlocked” which would be an intent-based outcome specification. And an old-fashioned car could be operated by the explicit command to unlock the door by inserting and twisting the key.)
Whether AI systems can achieve high usability within the intent-based outcome specification paradigm is unclear. I doubt it because I am a GUI bigot. Clicking things on a screen is an intuitive and essential aspect of user interaction that should not be overlooked. Thus, the 2nd UI paradigm will survive, albeit in a less dominant role. Future AI systems will likely have a hybrid user interface that combines elements of both intent-based and command-based interfaces while still retaining many GUI elements.
Infographic Summarizing this Article
Feel free to copy or use this infographic as much as you want, provided that you credit this URL as the source.
Quiz: Check Your Understanding of This Article
Take the Test, image by Midjourney.
Check your comprehension. Here are 7 questions about ideas and details in this article. The correct answers are given after the author’s biography.
Question 1: What is the main difference between batch processing and command-based interaction?
A. Batch processing involves back-and-forth interaction with the computer
B. Command-based interaction requires a complete workflow of instructions
C. Command-based interaction allows users to reassess the situation and modify future commands
D. Batch processing allows users to adjust their approach to the problem at hand
Question 2: What is the main advantage of graphical user interfaces over command-based systems?
A. They allow for batch processing
B. They show the status after each command
C. They require a complete workflow of instructions
D. They are based on punched cards
Question 3: What is the main characteristic of the third UI paradigm represented by AI?
A. Users specify a complete workflow of everything they want the computer to do
B. Users issue commands to the computer one at a time
C. Users tell the computer the desired result but do not specify how this outcome should be accomplished
D. Users interact with the computer using a graphical interface
Question 4: What is the main challenge of the intent-based outcome specification paradigm?
A. It requires users to issue commands one at a time
B. It is based on punched cards
C. It requires users to specify a complete workflow of instructions
D. It can be harder for users to correct problems when they don't know how something was done
Question 5: What is a noncommand system?
A. A system that requires the user to specify intent
B. A system that acts as a side effect of the user’s normal actions
C. A system that requires the user to issue commands one at a time
D. A system that is based on punched cards
Question 6: What is the author's prediction about the future of AI systems?
A. They will completely replace the second UI paradigm
B. They will have a hybrid user interface that combines elements of both intent-based and graphical user interfaces
C. They will be based on punched cards
D. They will require users to specify a complete workflow of instructions
Question 7: What is the main usability problem with early command-line systems?
A. They provided too much feedback from the computer
B. They did not provide any feedback unless there was an error
C. They required a complete workflow of instructions
D. They were based on punched cards
More on AI UX
This article is part of a more extensive series I’m writing about the user experience of modern AI tools. Suggested reading order:
AI Vastly Improves Productivity for Business Users and Reduces Skill Gaps
Ideation Is Free: AI Exhibits Strong Creativity, But AI-Human Co-Creation Is Better
AI Helps Elite Consultants: Higher Productivity & Work Quality, Narrower Skills Gap
The Articulation Barrier: Prompt-Driven AI UX Hurts Usability
UX Portfolio Reviews and Hiring Exercises in the Age of Generative AI
Analyzing Qualitative User Data at Enterprise Scale With AI: The GE Case Study
Navigating the Web with Text vs. GUI Browsers: AI UX Is 1992 All Over Again
UX Experts Misjudge Cost-Benefit from Broad AI Deployment Across the Economy
ChatGPT Does Almost as Well as Human UX Researchers in a Case Study of Thematic Analysis
“Prompt Engineering” Showcases Poor Usability of Current Generative AI
About the Author
Jakob Nielsen, Ph.D., is a usability pioneer with 40 years experience in UX and co-founded Nielsen Norman Group. He founded the discount usability movement for fast and cheap iterative design, including heuristic evaluation and the 10 usability heuristics. He formulated the eponymous Jakob’s Law of the Internet User Experience. Named “the king of usability” by Internet Magazine, “the guru of Web page usability" by The New York Times, and “the next best thing to a true time machine” by USA Today. Before starting NN/g, Dr. Nielsen was a Sun Microsystems Distinguished Engineer and a Member of Research Staff at Bell Communications Research, the branch of Bell Labs owned by the Regional Bell Operating Companies. He is the author of 8 books, including the best-selling Designing Web Usability: The Practice of Simplicity (published in 22 languages), Usability Engineering (26,240 citations in Google Scholar), and the pioneering Hypertext and Hypermedia. Dr. Nielsen holds 79 United States patents, mainly on making the Internet easier to use. He received the Lifetime Achievement Award for Human–Computer Interaction Practice from ACM SIGCHI.
· Subscribe to Jakob’s newsletter to get the full text of new articles emailed to you as soon as they are published.
Quiz Answers
Question 1: What is the main difference between batch processing and command-based interaction?
Correct Answer: C. Command-based interaction allows users to reassess the situation and modify future commands
Question 2: What is the main advantage of graphical user interfaces over command-based systems?
Correct Answer: B. They show the status after each command
Question 3: What is the main characteristic of the third UI paradigm represented by AI?
Correct Answer: C. Users tell the computer the desired result but do not specify how this outcome should be accomplished
Question 4: What is the main challenge of the intent-based outcome specification paradigm?
Correct Answer: D. It can be harder for users to correct problems when they don't know how something was done
Question 5: What is a noncommand system?
Correct Answer: B. A system that acts as a side effect of the user’s normal actions
Question 6: What is the author's prediction about the future of AI systems?
Correct Answer: B. They will have a hybrid user interface that combines elements of both intent-based and graphical user interfaces
Question 7: What is the main usability problem with early command-line systems?
Correct Answer: B. They did not provide any feedback unless there was an error
Fascinating article. In times of frustration when operating my computer I've often dreamed of what would be a potential 4th era - where a thought occurs in ones mind, and a computer machine renders it as dreamed. However this would possibly require some hardware device implanted into the body - the thought of which fills me with dread...