Saad Faruque via Compfight
Positional tracking is very important towards achieving immersion and presence in virtual reality. Whether it is the head, arms, fingers or objects (such as a weapon), positional tracking can deliver multiple benefits:
- Change the viewpoint of the user to reflect actions such as jumping, ducking or leaning forward.
- Show hands and other objects in the displayed image. A common complaint of users in virtual reality is that they can’t see their hands.
- Connect the physical and virtual world. For instance, by detecting hand position, a software program can implement the option to move virtual objects by touching them.
- Detect gestures. By analyzing position over time, a gesture can be detected. For instance, a user might draw the number “8” in air and have the software detect it.
There are several methods of tracking position and I felt it is worthwhile to describe some of them. This post focuses on tracking for virtual reality applications, so we will not look at vehicle tracking, tracking of firemen in buildings and so forth. In no particular order, here are some of the popular tracking methods include magnetic, inertial, optical and acoustic tracking as well as hybrid tracking that combines multiple methods.