Closed for Company Meetings: Between September 18 (5pm PDT) and September 22 (12pm PDT), Support and Sales requests will be delayed. Only time-sensitive issues will be addressed during that time. Thanks for your patience.
Indeed it was... have to say having forced myself to sit and play with jit.gl.pix that it is a truly awesome addition to jitter. Being able to fiddle with shaders in graphical land or use codeboxes as required rather than code in textedit and then reload your slab is well wicked, you can just reroute the output to whatever bit of the patch you're trying to understand, so quick to muck about with.
How do you know when you've reached the limit your card can handle, right now I've got a shader that does scale bias, brightness, contrast and saturate. When I try to add more (I want a colour splitter and a ripple repos) I start by adding a [sample] but it won't let me connect to it.
Also I start getting messages at the max window that lua isn't happy.
In the gen world, the only thing you connect to the input of [sample] is an [in] operator. It won't let you connect anything else by design. The [in] operators represent two different concepts simultaneously, which is probably what's causing some confusion. They are:
1) The current pixel/cell being processed
2) a proxy for the entire input matrix/texture
When you connect [in] to any operator except [sample], you're connecting the current pixel being processed to that operator (1). When you connect [in] to [sample], you're connecting the entire input matrix to it (2). This is because [sample] can pull an arbitrary pixel from a given input. All other ops can only work with the current pixel flowing through the gen patcher.
In your genjit patches, I noticed you had [in 1] -> [sample] with a [norm] connected to the second [sample] inlet. You actually don't need this. The following are equivalent:
If you run the two gen patchers above, you'll see that using [sample] in this way and simply passing through [in] are the same.
If you want to chain effects like a brcosa and then a repos type displacement, you'll have to chain separate gl.pix objects together. You can't sample from anything except the input, so you can't sample from the end of a processing chain like you were trying to do. The entire processed image simply isn't available at that time.
Hope this all makes sense. If anything is confusing, please ask.