Hands are typically rendered using
OVRSkeleton or
OVRCustomSkeleton. However, since those components visualize hand data from
OVRHand directly, they won’t match hand data that has been modified.
To solve this issue, Interaction SDK uses the
HandVisual component to render hand data. The
HandVisual component takes an
IHand as its data source. This means that you can include several
HandVisual components in your scene and attach them to various stages of the hand data modifier chain. For instance, you can visualize both raw hand tracking data and synthetic hand data at once by creating two
HandVisual components and setting them to the associated
Hand Modifiers.
Note: If you are just getting started with this Meta XR feature, we recommend that you use
Building Blocks, a Unity extension for Meta XR SDKs, to quickly add features to your project.