I’m looking to trigger different sound effects stored in a ~buffer. I would like be able to assign sound effects to different pixels area in a video. Imagine rectangles overlaid on a video feed separating areas. When the pixels change in those rectangles (someone moves their hand in front of it) I want the buffers to trigger.
Any ideas/tutorials/3rd party add-ons that I should look at?
an easy way without any actual "tracking" per say would be to get the frame difference (jit.op @op absdiff) then use jit.3m to get the mean. From there it is easy to set a trigger point. You could then divide the screen into smaller segments via jit.submatrix or jit.scissors to several areas…
just one quick, simple possibility.
hope it helps.