Add NewbiePipeline and NextDiT_3B_GQA_patch2_Adaln_Refiner_WHIT_CLIP transformer #12789
+1,887
−1
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.


This PR introduces a new text-to-image pipeline named NewbiePipeline, as well as a new
NextDiT-based transformer architecture,
NextDiT_3B_GQA_patch2_Adaln_Refiner_WHIT_CLIP, fully implemented following
Diffusers' pipeline and model design principles.
🚀 Main additions
• New pipeline
Adds
NewbiePipelineunderdiffusers.pipelines.newbie/.The pipeline follows the standard Diffusers structure (DiffusionPipeline subclass) and
supports loading via
from_pretrained.• New transformer architecture
Adds
transformer_newbie.py, implementing:The transformer inherits from
ModelMixin, enabling standard save/load, weightserialization and integration with Diffusers utilities.
• RMSNorm implementation
Adds
RMSNormtodiffusers.models.components, using a PyTorch fallback and supportingApex fused RMSNorm if available.
• Scheduler compatibility
The pipeline is compatible with
FlowMatchEulerDiscreteSchedulerwithout requiringadditional custom scheduler code.
🧩 Motivation
This PR provides an implementation of a modern NextDiT-style text-to-image architecture
with high-resolution capability and strong conditioning support.
The goal is to enable researchers and users to load, run, and fine-tune this model
directly through Diffusers with minimal friction.
📁 Files added
src/diffusers/models/components.py
src/diffusers/models/transformers/transformer_newbie.py
src/diffusers/pipelines/newbie/pipeline_newbie.py
src/diffusers/pipelines/newbie/init.py
shell
Copy code
📁 Files modified
src/diffusers/init.py
src/diffusers/models/init.py
src/diffusers/models/transformers/init.py
src/diffusers/pipelines/init.py
yaml
Copy code
✔ Notes
from_pretrainedand produces expected outputsFixes # (no issue linked)
Before submitting
Who can review?
Tagging pipeline & transformer reviewers:
@asomoza @yiyixuxu @sayakpaul