We’ve seen some incredible things emerging from Leap Motion since the company hired visionary filmmaker Keiichi Matsuda last year as a vice president leading a new London research studio. Most recently, a video emerged from the company showing how a desk can become interactive when paired with the company’s finger and hand tracking technology, as well as an AR headset design called North Star the company is open-sourcing.
The video shows the nuanced way digital objects can slip into the real world, and how you might interact with these objects using natural gestures. In case you are unfamiliar, Matsuda created the impactful 2016 short film Hyper-Reality which imagined a future overtaken by AR.
While the videos Leap Motion produced in recent weeks show off some interesting concepts for AR software design, I wonder how far this interaction paradigm can go without offering haptic feedback to users. Valve, for instance, is developing controllers which strap to your hand for realistic grasp and release sensations, and these “Knuckles” controllers can track your finger movements with a fair degree of accuracy too. If you’ve ever held a bow in VR, you know how satisfying it can feel to pull back a bow string with even simple vibration feedback.
Adding more realistic feelings of grasping the bow and pinching the arrow may make it feel even more immersive. Here’s Matsuda’s answer to that question: For some experiences, total immersion is a clear goal; simulations of physical experiences, gaming, etc.
These all improve when we engage more senses. I believe that those cases will be niche, and that AR and VR will have to make do with sound and vision for the foreseeable future. Read more from uploadvr.com…
thumbnail courtesy of uploadvr.com