Simple 3d Motion Capture with Kinect & Synapse
First time poster- I’ve been able to find a lot of topics on this subject but nobody has been able to showcase whether this is possible or not? I have a .dae rigged human figure that I want to mimic the actions of a real world person. I can pull the XYZ locations (in a 3d world) of specific joints via synapse however I’m having a lot of trouble translating that to my rigged model (in which I can only track the positions relations to the nodes?). Any thoughts?
You might wanna have a look at the dp.kinect external instead of Synapse. It also gives you rotations of joints. http://hidale.com/dp-kinect/
Just to warn that the dp.kinect eternal only works with Windows 7 & 8, I can’t see any Mac options.
Then there is the jit.openni external by the same author. Not sure if it also has joint orientation. I guess it does.
maybe a better question to ask would be how to magnetize joint/nodes to a XYZ point in a 3d space?
What are you using to draw your body? jit.gl.mesh, jit.gl.multiple, jit.gl.sketch,…?
I’m no expert at this but I think that’s the wrong approach. AFAIK jit.gl.model lets you import a static model (or one with predefined animation). Modulating it might be possible but it’s gonna be hard. I think you need to look at the jit.anim.* of jit.phys.* objects. I’m sure I’ve seen an example or tutorial animating a puppet figure. That should get you going in the right direction.
There are also some discussions in the forum like: http://cycling74.com/forums/topic/trouble-with-animating-a-3d-puppet-from-kinect/
There, I’ve found it: http://cycling74.com/2012/03/20/jitter-recipes-book-four/2/