Introduction to Pixray
A simple explanation for what happens under the scene.
Last updated
A simple explanation for what happens under the scene.
Last updated
The main function of Pixray is the use of CLIP to guide image generation from text.
Pixray uses CLIP (Contrastive Language-Image Pre-Training) by OpenAI as its perception engine. Check out CLIP's and ) to learn more about CLIP.
The vqgan drawer uses Vector Quantized Generative Adversarial Networks, or simply VQGAN, by Esser et al. from the paper . The original code for VQGAN is in.
The clipdraw drawer uses by s, from the paper .
The diffusion drawer uses , and models are trained by . The code we use for diffusion is in .
The fft drawer is from the Aphantasia library by . The original code for fft drawer is in .
The pixel drawer is created by , the author of pixray.
The algorithm for generation is encoding prompts through CLIP, generating embeddings, which are compared against the output image for a loss that can backpropagate through the generation engine to learn a latent space that can minimize the loss. This minimal loss result is the output that most resembles the text which it was prompted.
The difference between the drawers lies between how the image itself is generated, and what the latent space of the model represents. It is possible to generate an image without an underlying model, but simply optimizing a tensor.