Any plans to integrate in diffusers just like v1 https://github.com/huggingface/diffusers/pull/10148 With diffusers it becomes available for low VRAM users using pipe.enable_sequential_cpu_offload() and quantization like nf4. I have v1 working on as low as 6-8 GB VRAM, slow but it works. Thank you