{ "patch_size": 1, "in_channels": 64, "num_layers": 4, "num_single_layers": 4, "attention_head_dim": 128, "num_attention_heads": 16, "joint_attention_dim": 4096, "pooled_projection_dim": null, "guidance_embeds": false, "axes_dims_rope": [ 16, 56, 56 ], "rope_theta": 10000, "time_theta": 10000, "_use_default_values": [ "rope_theta", "time_theta", "pooled_projection_dim" ], "_class_name": "BriaTransformer2DModel", "_diffusers_version": "0.30.2", "num_mode": 6 }