Forums > Jitter

feedback delay in OpenGL…

November 22, 2010 | 8:33 pm


I thought I had this figured out, but the temporary solution I thought I had found is unsatisfactory. I need a simple example of how to set up feedback in OpenGL, preferably using Andrew Benson’s method in Jitter Recipe no.40 SceneWarp. I understand the concept but must be missing a step. Any help clarifying the process would be much appreciated. I’ve posted an example of what I’m trying to do below. Thank you!


– Pasted Max Patch, click to expand. –

November 23, 2010 | 2:50 am

Wow, no wonder nobody replied, this is so simple.

– Pasted Max Patch, click to expand. –

November 23, 2010 | 4:47 am

Yeah that’s pretty much the easiest way to do it..the scenewarp stuff gets a little more complicated with gl.sketch and other drawing contexts

November 23, 2010 | 6:13 pm

Looking at your patch, I would recommend against that jit.matrix object in there. You can use a object in its place. Based on what you seem to be working on, you’d be well-served to look at the Video Processing System articles:

The SceneWarp is really best for situations where you are trying to do video feedback involving a 3D scene, and introduces the idea of using to manage complex rendering setups.

November 26, 2010 | 10:48 pm

Thanks Andrew, I looked over those when you published them, but definitely didn’t take it all in. Thanks!

Viewing 5 posts - 1 through 5 (of 5 total)