Design

Hands Best Practices

Updated: Feb 19, 2026
The following best practices guide the use of hands as inputs and interactions in VR app design and development. They aim to enhance user experience and ensure optimal performance.

Involve designers early and experiment

Hands are still a major emerging input modality in immersive and VR applications. Designers should be involved early to craft and test interaction metaphors, gesture vocabularies, and user flows. Input design is something that needs a lot of prototyping and experimentation, and hands input is one of the most challenging areas. Give time and resources to test a variety of solutions early. Expect a lot of assumptions to be proven wrong and certain design challenges to take a variety of approaches to find the best solution.
It’s important to remember:
  • Designing inputs is a process.
  • Progress is more important than perfection.
  • Stay open, explore, and embrace learning.

Hands advantages for immersive experiences

Hands as an input method offer unique advantages that enhance user engagement and interaction, but also creates some challenges you should keep in mind, such as:
Advantages

  • Natural and intuitive interactions: Interacting with objects and touchscreens in VR with your virtual hands is just like interacting with objects and touchscreens in the real world. Hands are highly approachable and low-friction input due to this familiarity, attracting the widest range of users. Users less familiar with gaming controllers often find using hands more natural than hardware input. Natural hand movements reduce the need for complex controls and memory for button mapping.
  • Immersive and social presence: Seeing their virtual hands move just like real hands helps users feel present. Using hands naturally in virtual environments helps users feel more connected to others and to the experience itself. This method is also highly intuitive for non-gamers, as it leverages familiar, real-world gestures and movements, reducing the learning curve for new users.
  • Expressiveness: hands convey emotions and social cues effectively. Solo and social experiences are emotionally richer when users can use your real hands to express themselves.
  • Availability and flexible: No additional hardware is needed for hand input. Unlike hardware devices, hands are present as soon as a headset is donned. Hands are free when used as an input, they are free to interact with physical objects, such as adjusting hardware buttons on a headset or stoking a pet.
  • Real-world multitasking: hand tracking enables users to interact with virtual content while still being able to perform tasks in their physical environment. This is useful for apps that use passthrough.
  • Direct input for up-close interactions: hands excel at direct manipulation and touch-based input, making them ideal for tasks that require precision or are performed at close range.

Challenges

  • Users Expectations: Virtual hands do not work identically to real hands nor any specific fantasy universe, which can contradict user expectations rooted in both reality and sci-fi stories.
  • Tracking: Hand tracking relies on the headset’s cameras. The reliance introduces technological challenges such as tracking volume, occlusion, and environmental lighting conditions.
  • Tactility: Virtual objects don’t provide tactile feedback that happens when interacting with real-life objects. Feedback such as weight, touch, and material is all missing in virtual reality, which reduces the cues users expect when interacting with hands.
  • Limited input complexity: hands offer fewer distinct input types than controllers, which can limit the number of simultaneous actions or complex controls available to users. This can also create a slower pace of input due to sequential movements.
  • Distance interactions: Hands are less effective for interactions at a distance compared to controllers or ray-based input methods.
  • Fatigue: Extended use of hands, especially with arms held above the heart, can lead to user fatigue. Designing interactions that allow for arms to rest at the sides can help mitigate this challenge.
  • Gesture recall and variability: Gestures must be taught and memorized, and users perform gestures differently. This can lead to challenges in gesture recognition and recall, requiring forgiving design and in-app gesture reminders.

Designing with hands best practices

Start simple and with hands-first

If you are thinking about adding hands inputs it’s best to start designing your experience from a hands-first perspective and avoid adapting a controller-based design. Many interactions require special design considerations to create a high-quality hands experience. Controller interactions like pressing X to pick up an item are easy to learn. However, using hands for these actions can be harder to master and remember. Problems with user orientation, switching between movement and engagement, and targeting items often cause fatigue and frustration.

Start simple

When prioritizing hands as the primary input method: Simplicity is key.
For example, hand inputs should prioritize direct manipulation of physical objects over 2D menu interactions, because natural interactions like grasping, pointing, and gesturing suit hands better. While actions like scrolling will suit controller inputs better.

Sequential inputs with hands vs. Simultaneous inputs with controllers

Users are often more accustomed to performing multiple inputs simultaneously when using controllers. For example, you can often move with the left stick while turning the camera with the right stick at the same time.
In contrast, hand input typically requires sequential actions with one hand, such as using a thumb tap microgesture to teleport and then a swipe microgesture to snap turn. This sequential process is much slower than using controllers, due to both the nature of single-hand input and the time required for gesture detection. While it is possible to have simultaneous hand inputs from both the left and right hands, this approach is still slower than controllers, which allow multiple inputs from each device at once. These differences should be carefully considered when designing the pacing of your experience or gameplay.

Hands as the primary input during development

Use hands as the primary input while working on your experience as this will help surface issues and design opportunities early. For example, consider disabling controller input for a time or gating it behind a debug setting. Hands are different for everyone
But keep in mind: Just because it works for you doesn’t mean it works for everyone. Everyone’s hands are different, length, strength, orientation, and movement styles vary.
  • Test with a diverse group and expect to get things wrong before you get them right and that’s okay.
  • Prototyping with real users from the start helps develop a natural language for hand implementation.
  • When moving from controller-first contexts, explore interaction models like poke, grab, and ray casting to find the best fit while considering the player’s cognitive load.
This collaboration helps produce a more polished, intuitive, and enjoyable experience that meets user expectations and technical limits.

Use natural hand interactions

Users expect hand interactions to feel natural and intuitive, not just a direct mapping of controller buttons to hand gestures. Exact replication of controller inputs with hands often creates awkward, unintuitive experiences and forces users to unlearn natural behaviors. Users learn and remember more easily when their experience closely matches and mimics real-world hand use.
For example: Don’t push rope. Rope is good for pulling, not pushing. Or grab and turn objects instead of using microgestures.
This approach enhances immersion by letting users:
  • Interact as they would in the real world, rather than forcing controller-like behaviors.
  • Push buttons, if they are close, and only resort to raycast, if it is farther away.
  • Have more natural and responsive interactions.

Teach one core interaction, use it everywhere

Hands-first gameplay increases cognitive load, so users can have less bandwidth to learn complex controls. Instead, aim for one core interaction so users only have to remember one. For example, you could focus on grabbing as the primary interaction, with only a few additional gestures.
This helps because:
  • Users can master one motion and feel successful quickly.
  • You can polish this one core interaction for a great experience.
  • Automate wherever else possible to help reduce cognitive load.
    • For example, instead of one hand gesture to chop down a tree and another to harvest the wood, have the wood automatically be harvested.
  • Removes extraneous interaction and focuses on what’s vital to your experience.

Audio or visual feedback is critical

The absence of physical buttons and haptic feedback in controllers makes clear visual and audio feedback essential for hand interactions.
Feedback helps users:
  • Know when their gestures are recognized.
  • What actions occur, reducing confusion.
Visual cues, such as highlighting points or objects of interest and animations; both where the interactions happen, combined with subtle sounds, compensate for the lack of physical touch. These cues make interactions feel more responsive and engaging.

Different contexts require different visuals

Though keep in mind that there are some experiences or action where you should tailor hand visuals to improve immersion and reduce confusion, for example:
  • Hide virtual hands during activities like yoga poses to avoid distraction.
  • Keep hands visible when interacting with the UI to enhance presence and control.
  • Choose visualizations that optimize usability and comfort.
    • For example, minimize visuals flashing on and off. As this can be jarring, unsafe, and cause distraction.

Minimize physical strain

When creating experiences with hand input, it’s important to recognize that these interactions can be more physically demanding than other modalities, even with careful design.
To ensure comfort and accessibility, consider the following:
  • Incorporate rest periods.
  • Optimize interaction zones to reduce fatigue.
  • Avoid time-based or reflex-heavy events, unless the experience is intentionally fitness-focused.
  • Let users progress at their own speed and avoid unnecessary physical demands.
    • For example, do not penalize users for taking their time or needing breaks.
  • Avoid requiring hands to be held above heart level or excessive reaching.
  • Prioritize comfort.

Tailor user challenges carefully

When designing hand interactions, consider the overall challenge level of your experience. Below is a breakdown of different scenarios based on how challenging they are when using hands input:
Less challenging scenarios:
  • Few simultaneous input types
  • Stationary player position
  • Slower pace of gameplay
  • Few or no timing-critical inputs
More challenging scenarios:
  • Many simultaneous input types
    • For example: moving while shooting and reloading
  • Fast hand/arm motion
  • Smooth locomotion
  • Faster pace of gameplay
  • Timing-critical inputs
  • Real-time multiplayer
Though you should always keep in mind that hands are different for everyone and these recommendations might range more widely for your experiences target user demographic.

Gesture design best practices

Effective gesture design in hand-tracked experiences begins with three core principles:
  • Reliability: Gestures should consistently succeed and fail as expected.
  • Comfort: Support frequent and extended use without causing fatigue.
  • Intuition: Gestures feel natural to users, closely matching the intended actions.
Evaluating gestures across these axes helps ensure a balanced and accessible experience. Below are some additional factors when designing for gestures.

Hit detection

Currently, pinch gestures, especially those involving the fingertips and thumb tip, are more accurate for hit detection, with the index finger providing the highest precision. However, accuracy diminishes when using the middle, ring, or pinkie fingers. Keep in mind, aiming with hand or gesture-based input is inherently challenging due to the imprecise directional control these methods offer.

Latency

You should account for the latency between a player’s intention and the execution of motion gestures. For example, snapping actions can enhance the user experience, but may introduce input lag.
Hand tracking technology itself brings additional challenges, such as delays, reduced accuracy during rapid movements, and the lack of haptic feedback. To mitigate these issues, your experience should be forgiving, emphasizing the general direction rather than its exact position. It’s important to recognize that simply porting existing experiences to hand tracking is rarely effective. Instead, create new input variations that leverage the strengths of hand tracking hardware, balancing a smaller movement area for improved tracking with enough movement to keep players engaged.

Gesture variability

When implementing gestures, you should consider factors such as camera angle and transition speed. Rapid transitions can result in unintended intermediate states, and while adding latency may help, it can also reduce responsiveness. Since gesture performance varies from person to person, recognition systems should be lenient and limit the recognition window to accommodate individual differences. Finally, including too many gestures in a single game can be confusing or overwhelming, increasing the likelihood of misrecognition between gestures.

Limitations and mitigations

This section examines the inherent limitations and challenges that impact performance and usability and offers strategies for mitigation through design or code improvements.

Tracking volume

The headset’s sensors track within a specific area known as the tracking volume. Hands outside this area, such as resting by the thighs or placed too close to the device like on the cheek, will not be detected. The hand-tracking volume exceeds the display field of view, allowing for tracking even when hands are not visible to the user. The tracking volume varies depending on the hardware used.
Diagram showing tracking volume
Mitigation - Code: To compensate for hands moving outside the camera’s field of view, body tracking techniques are employed to infer the positions of the hands.
Mitigation - Design: To ensure that all relevant user motions are accurately tracked, it is recommended to design interactions that keep users’ movements within the tracking volume, avoiding the need for them to reach beyond it. For more guidance, refer to the comfort page.

Occlusion

Sensors play a vital role in tracking hand movements in virtual environments. They perform best when they have a clear view of the user’s hands. However, when parts of the hand are not visible to the sensors, the system attempts to estimate their positions. If significant portions of the hand are obscured, the virtual representation of the hand will disappear. The hand pose prediction model provides two key outputs: the estimated hand pose and the level of confidence in that prediction. If the confidence level falls below a certain threshold, the system will not display the hand. Obstacles that block the sensors’ view can include elements from the surrounding environment or the user’s own body, known as Self-Occlusions
Occlusions happen when hands are behind obstacles or interact with physical objects. Tracking accuracy can be affected by common obstructions such as long sleeves, hands in pockets, large rings, or black bandaids.
Self-Occlusions include scenarios where fingers are covered by other fingers, fingers point outward obscuring them from the headset-mounted display (HMD) cameras, or when hands overlap. These factors collectively influence the effectiveness of hand tracking technology.
Illustration of woman looking at hands showing occlusion
Mitigation - Design: Aim to design interactions with minimal hand overlap; for example, having the index finger interact with a wrist button or palm menu on the other hand is a minimal overlap. The more one hand covers the other, the lower the accuracy. Additionally, interactions that can be performed with just one hand are always a good choice, as they not only enhance accessibility but also require less physical effort. This approach allows the hand tracking engine to effectively track both hands independently.

Lighting

The quality and intensity of light can impact the system’s ability to accurately detect hand movements. Hand tracking has different lighting requirements than inside-out (head) tracking. In some situations, this could result in functional differences between head tracking and hand tracking, where one may work while the other has stopped functioning.
Image showing lighting levels
Mitigation - Design: Immersive experience headsets adjust to different light conditions. For optimal functionality, advise users to be aware of their environment’s lighting, perhaps by including a note when launching an application.

Movement speed

Rapid or erratic hand movements can challenge the system’s tracking accuracy. For example, when throwing a virtual object, the hand moves quickly, and the camera may not immediately capture the exact point of release. Nevertheless, the Move Fast app successfully demonstrates that playing fast-paced rhythm games using hand gestures is indeed possible.
Mitigation - Design: To improve tracking accuracy with fast-paced hand movements, consider these strategies:
  • Centralize targets and encourage keeping hands at a moderate distance from the headset.
  • Maintain a predictable target speed.
  • Note that hand movements directed away from the player (such as punches) are tracked more effectively than movements parallel to the view (such as chops).
  • Use visual cues to guide users in maintaining appropriate hand speeds and positions. For example, visually slowing down hands or targets can encourage users to adapt their speed.
By integrating these design elements, it is possible to significantly enhance the tracking performance of hand movements, particularly in fast-paced applications.

Characteristics

Tracking accuracy may be influenced by individual user traits. Factors such as long fingernails, specific types of nail polish, tattoos on the hands, accessories like rings, long sleeves, and diverse hand sizes and shapes can all impact performance.
Mitigation - Design: Enhance user interaction by offering a variety of input modalities and interaction choices. For instance, while long nails may hinder poke interactions, pinch actions using ray casting could be more effective. Alternatively, using a controller can serve as a reliable fallback option. For more information, please refer to the section on accessibility.

Hand interaction do and don’t

Below is a list of our recommended practices when using hands in an immersive experience:
DO Provide a multimodal experience so that users can use different input methods seamlessly.
DON'T use your headset outdoors. Sunlight can affect the tracking accuracy and damage the lenses and display of your headset.
DO encourage users to make gestures that are clearly visible to the HMD.
DON'T promote gestures that are too close (< 10cm) to the HMD.
DO encourage users to make gestures that are clearly visible to the HMD.

Next steps

More design resources on hands

Designing experiences

Explore more design guidelines and learn how to design great experiences for your app:

Developing experiences

For technical information, start from these development guidelines:

Meta Spatial SDK

Unity

Unreal

Did you find this page helpful?