The shortest path from thought to action

The shortest path from thought to action

A reappraisal of Fitts’ law in the era of multimodal interfaces

By Julian Scaff

Press enter or click to view the full size image

A collage showing the evolution of user interfaces, including early monochrome desktop systems, Windows and Mac interfaces, smartphone touch gestures, and modern augmented reality widgets floating in space.
From early graphical desktops to touch, spatial and gesture-based systems, human-computer interaction continues to expand beyond screens into multimodal, immersive and embodied experiences.

When psychologist Paul Fitts published his paper on human motor control in 1954, he probably had no idea that his insights would one day guide the design of everything from smartphones to virtual worlds. Fitts conducted his experiments using simple physical devices, such as levers, styluses and illuminated targets, to measure how quickly participants could move and point to targets of different sizes and distances. These experiments were the precursors to the pointing and selection tasks that would later define human-computer interaction.

What emerged was called “Fitts’ law,” which describes the relationship between the distance to a target and the size of that target, predicting how long it takes a person to move and select it. The law states that the time to acquire a target increases with increasing distance and decreases with increasing size. Basically, the closer and larger a button is, the faster and easier it is to click.

Press enter or click to view the full size image

Diagram showing the experimental setup of Fitts' law, where a person uses a stylus to tap between targets of different width and distance, next to a diagram showing the target width (W) and distance (D) that define the movement amplitude in Fitts' model.
Illustration of Paul Fitts’ original experiments and model, showing how target width and distance determine movement time, a fundamental principle in human-computer interaction and interface design. (Fitts, 1954.)

This deceptively simple relationship became one of the cornerstones of human-computer interaction (HCI). Early graphical user interfaces, developed at Xerox PARC and refined by Apple, were designed around Fitts’ insight. Menus anchored to the screen edges, large icons for frequently used commands, and cursor acceleration algorithms are all there to minimize “moving costs.” Even today, every pixel and millisecond in a mouse-based interface carries the legacy of Fitts’ original experiments with pointing and tapping.

Press Enter or click to view the image at full size

Side-by-side screenshots showing the 1979 Xerox Star interface and the first Apple Macintosh operating system from 1984, with overlapping windows, icons, folders, and the desktop workspace that defined the early GUI era.
Early graphical user interfaces: Xerox Star (1979) on the left and Apple Macintosh System Software 1.0 (1984) on the right, both pioneers of the desktop metaphor that shaped modern interaction design.

From mouse to touch

With the rise of touch interfaces, the engine model changed, but the principle remained. The pointer was no longer a mouse cursor, it was the human hand itself. Target acquisition became a function of thumb reach, finger size, and hand posture. Fitts’ law still applied, but designers now had to deal with occlusion (the finger covers what it selects) and inaccuracy (thick fingers versus fine cursors). This gave rise to mobile design heuristics such as minimum touch target sizes, thumb zones, and one-handed reach maps. The designer’s task shifted from minimizing pointer movements to reducing physical strain and maximizing ergonomic comfort.

Press enter or click to view the full size image

Diagram showing thumb reach zones for the left, right and combined hands on smartphone screens, color-coded for natural, stretched and hard-to-reach areas, with additional illustrations showing that 49% of users operate the phone with one hand, 36% in the cradle and 15% with two hands.
Thumb zone maps that highlight the natural, stretched and hard-to-reach parts of smartphone screens, along with data on common hand postures, show how Fitts’ law influences the ergonomic design of mobile interfaces.

Gestures and spatial computing

In gesture-based and spatial computing systems, such as those used in XR headsets, mixed-reality displays, and motion-controlled environments, Fitts’ law is moving toward three dimensions. Targets are no longer flat areas, but volumetric zones in space. The ‘distance’ may involve moving your hand through the air or shifting your gaze, while the ‘size’ becomes a matter of both spatial volume and perceptual salience.

Designers must consider factors such as muscle fatigue, tracking accuracy, and depth perception. The principle of minimizing effort still applies, but now effort is distributed across space, muscle groups and sensory modalities. In spatial systems, Fitts’ law becomes a law of embodied range, not a measure of pixels, but of proprioception and fatigue.

Press enter or click to view the full size image

Four images of Apple VisionOS: a user wearing the Vision Pro headset viewing large virtual screens, floating app icons in a living room, an AR calendar widget interface, and a diagram illustrating hand gestures such as tap, pinch, drag, zoom, and rotate used to interact with spatial content.
Examples of Apple VisionOS’ spatial interface and gesture design, featuring immersive multitasking environments, floating 3D app windows, and core hand gesture patterns that define mixed reality interaction. Please note that these are marketing images from Apple and while they give a general impression of the interface patterns, they do not exactly represent the actual user experience.

Voice User Interfaces (VUI)

With voice interfaces, the idea of ​​physical distance disappears altogether, but the underlying cognitive pattern remains. When a user says, “Turn off the lights,” there is no target to touch or point to, but there is still some form of interaction distance, the mental and temporal gap between intention and response. Misrecognition, latency, or unclear feedback widen this gap, creating friction analogous to a small or distant button.

Fitts’ law becomes metaphorical: designers should minimize the cognitive and linguistic effort required to achieve a goal. The best VUIs reduce “speech traffic” by supporting natural phrasing, context awareness, and confirmation cues.

Press enter or click to view the full size image

Voice user interfaces (VUIs) make interaction invisible, shifting the design focus from visual goals to timing, tone, and conversational flow between humans and machines. (Image from Pexels.)

Towards neural and agentive interfaces

Looking further ahead, brain-computer interfaces (BCI) and agentive AI systems challenge Fitts’ model even more deeply. When a thought alone can cause an action, or when an intelligent system predicts intent before a command is issued, the concept of “goal acquisition” becomes almost instantaneous.

Yet Fitts’ Law still whispers beneath the surface: Every layer of mediation, from neural decoding errors to AI misinterpretations, adds new forms of interaction friction. The task for designers will be to minimize these invisible distances, not spatially or manually, but semantically and affectively, so that the path from intention to effect feels seamless, reliable and human.

Designing semantic and affective interfaces means paying attention not only to the mechanisms of interaction, but also to the meaning and emotion embedded in them. A semantic interface understands the why behind a user’s action and interprets intent through context, language and behavior rather than waiting for explicit commands. It bridges gaps in understanding by aligning system logic with human mental models, anticipating needs, and communicating in ways that feel natural and readable.

An affective interface, meanwhile, responds to emotional tone and state, recognizes frustration, joy, or hesitation and modulates its feedback, pace, or empathy accordingly. Together, these layers represent a new frontier for interaction design: systems that read nuances, convey intent, and maintain emotional resonance. In this new paradigm, minimizing friction means designing not only for efficiency, but also for coherence, ensuring that what the system does, means and feels aligns with the user’s goals and inner experience.

Affective or emotional computing remains largely underutilized in contemporary UX design practice. Today’s AI agents still lack the reliability and sensitivity needed to create consistent emotional coherence in interactions. Achieving this will require advances not only in software engineering, but also in the fundamentals of interaction design.

Press enter or click to view the full size image

A participant wearing an Emotiv EEG headset sits at a computer desk next to a small white collaborative robot (cobot). The cobot is connected to software that runs on the screen, while a 3D diagram of emotional data points based on the PAD (Pleasure-Arousal-Dominance) model appears on the left. The system monitors the user's affective state to dynamically adjust the robot's behavior.
Experimental setup from “Exploring cognition and affect during human-cobot interaction” (Canete, Gonzalez-Sanchez and Guerra-Silva, 2024). The image shows a participant wearing an Emotiv Insight 5-channel EEG headset while communicating with an Elephant Robotics myCobot 280 robotic arm. Brainwave data is analyzed through the Pleasure-Arousal-Dominance (PAD) model (visualized on the left), which allows the cobot to adjust its motor speed and lighting in real time based on the operator’s emotional and cognitive states, such as stress and focus.

The sustainable principle

Across modalities, Fitts’ Law continues to embody a deeper truth: design is about reducing the resistance between a user’s intent and the system’s response. Whether that resistance is physical, perceptual, cognitive or emotional, the role of the designer is to smooth the path of interaction, to make technology feel like an effortless extension of the body and mind. In this sense, Fitts’ law has transcended its original context in psychophysics and become a universal principle of interaction design: the shorter the distance between thought and action, the better the experience.

#shortest #path #thought #action

Similar Posts

Leave a Reply

Your email address will not be published. Required fields are marked *