why does jit.sobel look so much better than cf.sobel.jxs?

gpvillamil's icon

I've recently been porting a project from all Jitter to OpenGL, and discovered that the rendering quality is much poorer when using OpenGL shaders, specifically for edge detection (jit.sobel vs cf.sobel.jxs)

The OpenGL looks very noisy and artifact-y, and with lower vertical resolution.

Max Patch
Copy patch and select New From Clipboard in Max.

Here's a patch and a screenshot

312.jitter_vs_opengl.jpg
jpg
Rob Ramirez's icon

try adjusting the with parameter to the sobel shader in the slab.
( "param width $1" in a message box).

values less than the default of 1 seemed to get it closer to the CPU version.

gpvillamil's icon

That seems to help, though the results are never quite identical.

Is there anyway of defining the default shader parameters in the jit.gl.slab object (eg. @param width 0.4)? It looks like param is a message, not an attribute, so it doesn't seem to work.

Andrew Benson's icon

try typing exactly what you wrote into jit.gl.slab.

;)