I’m attempting to create a touch interface out of cloth. I’ll have a movie being rear projected onto the cloth and I’d like users to be able to come up and interface with it much like they would a screen.
Does this sound doable with Max/MSP?
My thought is to somehow track the bumps that will happen as people touch the screen and translate that to the mouse X and Y coordinates.
Am I crazy with this? Anyone have any ideas?
Thanks so much!
You might want to have a look at this approach…
look at front DI
Check out the Khronos Projector project-
And the video that explains how the project works-
The project was made using the Processing language (http://www.processing.org), but I think that something similar could be very doable using Jitter and Max.
I think you could totally do this. If it were me, I would go with the NUI route of using open source C based vision tracking, then max to easily make interface patches.
You also might want to check out mathieu chamgane’s MMF (http://www.mathieuchamagne.com) which will let you quickly use max patches with multi-touch. I use this method (open source, computer vision tracking, with MMF) all the time and find it very reliable and flexible using my vision based multi-touch screens.
C74 RSS Feed | © Copyright Cycling '74