Motion sensing Max/MSP Jitter experiences...

    Jun 04 2006 | 3:35 pm
    Hello forum users,
    I'm currently working on a patch that brings motion and sound together. I ultimately want to create a synth, VST plug-in, or both, with movement of the body acting as an external controller, linked to various parameters of a synth/plug-in.
    I'm curious if anyone has any past eperiences of of such work within Cycling 74 products if you could point me towards, old/new project that bring motion and sound together.
    Mad Max.

    • Jun 04 2006 | 4:32 pm
      WOW< taths awosme, never heard of anything like that. do and report back
    • Jun 04 2006 | 6:51 pm
      On Jun 4, 2006, at 11:32 AM, yosh wrote:
      > WOW< taths awosme, never heard of anything like that.
      You might want to have a look at the Hipno plug-ins from Cycling '74. There are several plug-ins in the collection that use video camera input for a variety of effects.
    • Jun 04 2006 | 7:04 pm
      Hello guys,
      Thanks for the posts! I wouldn't say this was a 'WOW' project, i'm sure it has been done many times before, which is what i'm looking for. Examples, not patch examples, but project examples. I've done the usual google/yahoo searches but can't seem to find much out there.
      I'm asking here on the Jitter forums as the project ultimately uses video for sound and I would have thought someone has used Jitter before for a similar task?
      Gregory, i've looked at the Hipno plug-ins, a little out of my price range and too late to be trying out demos now i'm afraid! The theremin plug-in looks cool though!
      All suggestions welcome, thanks for your replies so far.
    • Jun 04 2006 | 7:27 pm
      It's generally the case that you'll have much better luck asking for examples of specific tasks [rather than asking someone to hand over an entire project they've spent a gazillion hours on], and even better luck than that if it's clear that you're working on those tasks yourself [and nothing says, "I'm working on this stuff and don't mind sharing" like posting a working subunit and asking for advice or feedback]. This is generally true on these forums, and the larger the thing you ask for, the more true it tends to be. :-)
      Barring that, a verbal sketch may serve you well - you may find that any number of posters are quite willing to share their own experiences with a given technique as prose - e.g, "I'm thinking of using jit.3m to derive some basic min/mean/max information about a slice of an incoming matrix and using that as control information. Will I get a broader or more varied and useful range of data by doing analysis on an individual plane, working in black and white, or RGB?" It's quite likely you'll get some sort of response of the nature "I tried that, but found that it's not so much which plane or planes I analyzed that made the difference as it was a case of smoothing the data...." If that seems like a rather specific example, it is. A real one of my own.
    • Jun 04 2006 | 7:36 pm
      all you really want is to localize 12 or so 0 -128s from the data field with bodies then route them to music generators/effects, cv.jit canhelp with its motion tracking static, I;ve wanted to further my own experiements that compare the tracked environment to a templete of conciveable human postures so that main esecence of the motion sensing are routed to controlling music head to hi hit, feet to bass line arms to delay shadows kinda thing, also you might enjoy putting color spray painted foam balls on your arms and legs to see what the optimum conversions of such information is with tap.motionor such,before you give it up to the computer to decide where the nodes are gonna lazy chaos track
      check out my performance patching works to see what im talking about
      if I could merge with there would be crazy vector version of body forms that are outline evolution musics -derek