Categories: Image

Z Image Base Knows Things and Can Deliver

Just a few samples from a lora trained using Z image base. First 4 pictures are generated using Z image turbo and the last 3 are using Z image base + 8 step distilled lora

Lora is trained using almost 15000 images using ai toolkit (here is the config: https://www.reddit.com/r/StableDiffusion/comments/1qshy5a/comment/o2xs8vt/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button ). And to my surprise when I use base model using distill lora, i can use sage attention like i normally would using turbo (so cool)

I set the distill lora weight to 0.9 (maybe that’s what is causing that “pixelated” effect when you zoom in on the last 3 pictures – need to test more to find the right weight and the steps – 8 is enough but barely)

If you are wondering about those punchy colors, its just the look i was going for and not something the base model or turbo would give you if you didn’t ask for it

Since we have distill lora now, I can use my workflow from here – https://www.reddit.com/r/StableDiffusion/comments/1paegb2/my_4_stage_upscale_workflow_to_squeeze_every_drop/ – small initial resolution with a massive latent upscale

My take away is that if you use base model trained loras on turbo, the backgrounds are a bit messy (maybe the culprit is my lora but its just what i noticed after many tests). Now that we have distill lora for base, we have best of both worlds. I also noticed that the character loras i trained using base works so well on turbo but performs so poorly when used with base (lora weight is always 1 on both models – reducing it looses likeness)

The best part about base is that when i train loras using base, they do not loose skin texture even when i use them on turbo and the lighting, omg base knows things man i’m telling you.

Anyways, there is still lots of testing to find good lora training parameters and generation workflows, just wanted to share it now because i see so many posts saying how zimage base training is broken etc (i think they talk about finetuning and not loras but in comments some people are getting confused) – it works very well imo. give it a try

4th pic right feet – yeah i know. i just liked the lighting so much i just decided to post it hehe

submitted by /u/Major_Specific_23
[link] [comments]

AI Generated Robotic Content

Share
Published by
AI Generated Robotic Content
Tags: ai images

Recent Posts

Update: Distilled v1.1 is live

We've pushed an LTX-2.3 update today. The Distilled model has been retrained (now v1.1) with…

12 hours ago

How to Implement Tool Calling with Gemma 4 and Python

The open-weights model ecosystem shifted recently with the release of the

12 hours ago

Structured Outputs vs. Function Calling: Which Should Your Agent Use?

Language models (LMs), at their core, are text-in and text-out systems.

12 hours ago

Cram Less to Fit More: Training Data Pruning Improves Memorization of Facts

This paper was accepted at the Workshop on Navigating and Addressing Data Problems for Foundation…

12 hours ago

How to build effective reward functions with AWS Lambda for Amazon Nova model customization

Building effective reward functions can help you customize Amazon Nova models to your specific needs,…

12 hours ago

How to find the sweet spot between cost and performance

At Google Cloud, we often see customers asking themselves: "How can we manage our generative…

12 hours ago