[sf-lug] how can we use leapmotion with linux?
Kai Chang
kai.salmon.chang at gmail.com
Wed Mar 27 11:40:19 PDT 2013
I borrowed an OS X computer to develop.
Unfortunately I can't share details of the developer program, but
there are Linux drivers that will probably be released before May.
I'll email the list when this happens.
I can share the data I collected. These are the JSON files that I
recorded off the device. They include all the 3-D and orientation
information as well as fingertip positions.
https://github.com/syntagmatic/leap-play/tree/master/data/gestures
All of my open-source JavaScript examples are in that directory as well.
Supported languages are C++, C#, Objective-C, Python, Java,
JavaScript, and Unity.
The device uses two cameras and three infrared emitters, with some
onboard processing. You can find YouTube videos of early dev kits that
don't have the shiny casing.
http://youtu.be/LY3Ya__6BHw?t=2m42s
I'd also be happy to bring the device to user group meetup when the
Linux drivers are released.
On Wed, Mar 27, 2013 at 11:23 AM, Michael Shiloh
<michaelshiloh1010 at gmail.com> wrote:
> But very interesting nonetheless. I was hoping to use the Leap but was
> disgusted with the process of becoming a developer. I still haven't received
> confirmation that I was accepted.
>
> I (and I suspect the larger Linux community) would be very interested in
> learning more.
>
> Did you develop under Linux?
>
> In what language(s) can you program?
>
> Is there a well documented API? Is it a library or a client/server
> arrangement? Can you share the API with us?
>
> What else can you tell us?
>
> Michael
>
>
>
> On 03/27/2013 10:49 AM, Kai Chang wrote:
>>
>> This is a similar concept to the Leap. It's not open-source.
>>
>> https://www.leapmotion.com/
>>
>> I used a developer version of the device to record some hand gestures.
>> These are the X/Y positions of the fingertips (in the plane of the
>> screen).
>>
>> http://fleetinbeing.net/leap-play/recorder-gallery.html
>>
>> Here's a video of me using the device to rotate/scale a map
>> projection. This one uses the normal vector of the palm, and Y
>> position of the hand.
>>
>> http://www.youtube.com/watch?v=9PZs7VHhypk
>>
>> Anyways, probably not too helpful to your original problem of creating
>> a device from scratch.
>>
>> Cheers,
>> Kai
>>
>> On Wed, Mar 27, 2013 at 10:17 AM, Michael Shiloh
>> <michaelshiloh1010 at gmail.com> wrote:
>>>
>>> i'm having trouble figuring out the best search words to find what i'm
>>> looking for.
>>>
>>> basically, i'm looking for crude kinect-like behavior. i don't expect
>>> perfect recognition or tracking, but surely something along these lines
>>> is
>>> available.
>>>
>>> my plan is to mount 2 or 3 webcams in a small box. the user inserts
>>> his/her
>>> hand in the box, and i'd like to track the tips of their fingers. i would
>>> mount the webcams at right angles so as to track location in both the x
>>> and
>>> y (and z) planes. the inside of the box can be painted any color to
>>> differentiate from the hand, which i can require be free of gloves or
>>> clothing. i can light the box, although later it might be nicer to have
>>> the
>>> box dark and use infrared lighting.
>>>
>>> there are no other objects in the box, and only one hand at a time.
>>>
>>> thoughts?
>>>
>>> _______________________________________________
>>> sf-lug mailing list
>>> sf-lug at linuxmafia.com
>>> http://linuxmafia.com/mailman/listinfo/sf-lug
>>> Information about SF-LUG is at http://www.sf-lug.org/
More information about the sf-lug
mailing list