Warning: Some posts on this platform may contain adult material intended for mature audiences only. Viewer discretion is advised. By clicking ‘Continue’, you confirm that you are 18 years or older and consent to viewing explicit content.
It is just a simple prompt in Flux on Comfy UI. It is just an open source model running on my hardware. It is slow because it is such a large model (Flux Dev gguf Q4). You can find examples in the ComfyUI documentation and the model manager add-on to the base Comfy setup has all the models in the downloads menu.
At present, it only works on GPU and 16 GB is like 2+ minutes per image. It would be awesome in they split the chunks with the CPU to generate faster, but that is not implemented yet. It means you basically need 16+ GB to run it on your own hardware. There is a smaller model version, but that is not compatible in results quality. There is a larger model that is online only. Flux is actually FluX as in X-AI as in Musk. The weights for Flux-dev are open source, and that is what I care about for now.
How was this done? It’s pretty good!
It is just a simple prompt in Flux on Comfy UI. It is just an open source model running on my hardware. It is slow because it is such a large model (Flux Dev gguf Q4). You can find examples in the ComfyUI documentation and the model manager add-on to the base Comfy setup has all the models in the downloads menu.
At present, it only works on GPU and 16 GB is like 2+ minutes per image. It would be awesome in they split the chunks with the CPU to generate faster, but that is not implemented yet. It means you basically need 16+ GB to run it on your own hardware. There is a smaller model version, but that is not compatible in results quality. There is a larger model that is online only. Flux is actually FluX as in X-AI as in Musk. The weights for Flux-dev are open source, and that is what I care about for now.