Thanks!
I've managed to avoid copying the pixel arrays several time with that. However the time I'm saving is tiny compared to the cost of moving from GPU to CPU.
Thing is in the deep learning process, those pixels are moved back to GPU for faster computations. Guess I have to figure out a way to make it stay on the GPU, but I have no idea how for now. Not very familiar with GPU computations, plus I'd have to pass GPU data from C++ SFML to Python.
Well, if someone here ever did something somewhat similar, please let me know!