matrix GPU processing to texture : error initializing capture

    Aug 15 2013 | 2:11 pm
    I am trying to move some matrix processing onto the GPU to be used as an intensity texture but cannot get it to work. What I have so far (based on an example by Wesley Smith) works when doing the processing on the CPU but when trying to do it on the GPU I get the following error message: unsupported framebuffer format! error initializing capture
    I am rather new to OpenGL (although quite a long time Max user) so I may miss something obvious. Does anyone have some hint ?
    Thanks in advance.
    Best, Julien

    • Aug 15 2013 | 7:05 pm outputs 4 plane textures, so i don't think this will work with @colormode intensity.
      if you really don't want to use jit.op for this, you could try
    • Aug 15 2013 | 7:07 pm
      rather, you could try combining the rgb values from one texture with the noise values from another texture. will also only output 4 plane textures.