Motion sensing Max/MSP Jitter experiences…
Hello forum users,
I’m currently working on a patch that brings motion and sound together. I ultimately want to create a synth, VST plug-in, or both, with movement of the body acting as an external controller, linked to various parameters of a synth/plug-in.
I’m curious if anyone has any past eperiences of of such work within Cycling 74 products if you could point me towards, old/new project that bring motion and sound together.
WOW< taths awosme, never heard of anything like that. do and report back
On Jun 4, 2006, at 11:32 AM, yosh wrote:
> WOW< taths awosme, never heard of anything like that.
You might want to have a look at the Hipno plug-ins
from Cycling ’74. There are several plug-ins in the
collection that use video camera input for a variety
Thanks for the posts! I wouldn’t say this was a ‘WOW’ project, i’m sure it has been done many times before, which is what i’m looking for. Examples, not patch examples, but project examples. I’ve done the usual google/yahoo searches but can’t seem to find much out there.
I’m asking here on the Jitter forums as the project ultimately uses video for sound and I would have thought someone has used Jitter before for a similar task?
Gregory, i’ve looked at the Hipno plug-ins, a little out of my price range and too late to be trying out demos now i’m afraid! The theremin plug-in looks cool though!
All suggestions welcome, thanks for your replies so far.
It’s generally the case that you’ll have much better
luck asking for examples of specific tasks [rather
than asking someone to hand over an entire project
they've spent a gazillion hours on], and even
better luck than that if it’s clear that you’re working on
those tasks yourself [and nothing says, "I'm working
on this stuff and don't mind sharing" like posting a
working subunit and asking for advice or feedback].
This is generally true on these forums, and the larger
the thing you ask for, the more true it tends to be. :-)
Barring that, a verbal sketch may serve you well – you
may find that any number of posters are quite willing
to share their own experiences with a given technique
as prose – e.g, "I’m thinking of using jit.3m to derive some
basic min/mean/max information about a slice of an
incoming matrix and using that as control information.
Will I get a broader or more varied and useful range
of data by doing analysis on an individual plane, working
in black and white, or RGB?" It’s quite likely you’ll get some
sort of response of the nature "I tried that, but found that
it’s not so much which plane or planes I analyzed that
made the difference as it was a case of smoothing the
data…." If that seems like a rather specific example, it
is. A real one of my own.
all you really want is to localize 12 or so 0 -128s from the data field with
bodies then route them to music generators/effects, cv.jit canhelp with
its motion tracking static, I;ve wanted to further my own experiements that
compare the tracked environment to a templete of conciveable human postures
so that main esecence of the motion sensing are routed to controlling music
head to hi hit, feet to bass line arms to delay shadows kinda thing, also
you might enjoy putting color spray painted foam balls on your arms and legs
to see what the optimum conversions of such information is with
tap.motionor such,before you give it up to the computer to decide
where the nodes are
gonna lazy chaos track
check out my performance patching works to see what im talking about
if I could merge http://www.vcnet.com/~franz/max/fastfill.htm
there would be crazy vector version of body forms that are outline evolution