Early this week I took a few hours to write an EigenD agent for the Leap. Currently it only supports sending out the x, y and z axis of two palms relative to the Leap device itself. The API is very intuitive and it shouldn’t take long to gradually add support for hand directions and fingers, I wanted to first play around with it for a while though.
In this experiment I control two effects in Native Instruments’ Guitar Rig 5. It’s processing one of my songs that’s playing from iTunes and you can hear the recorded result. I did have to manually align the audio to the video, so that might not be perfect.
http://www.youtube.com/watch?v=Ys6Imj_hmEc
This developer version of the Leap device supports a little bit less range and a restricted field of view, which is why it’s sometimes missing some detection at natural boundaries. The production versions will not suffer from this.