Since February 2010 I’ve worked on a series of applications which allow motion capture for performance as part of the Visuosonics research project at Southampton Solent University. The research:
uses real-time sound-image interactivity to create immersive environments in which the art gallery, the concert hall, the theatre and the cinema converge into a ‘total’ performance space.
My contribution involved creating bridging applications in Max/MSP which would allow hardware to be used with a variety of software front-ends. This included:
- Vision motion capture: USB or Firewire camera → capture software → OSC receiving software.
- Gesture motion capture: Acceleglove → capture software → OSC receiving software.
For details on how to use an Acceleglove in Max/MSP, please see the Acceleglove capture in Max/MSP tutorial.
To motion track through a camera and (via mxj) to connect to an Acceleglove data glove.