20

They say you shouldn't train on synthetic data, still worth a shot.

you are viewing a single comment's thread
view the rest of the comments
[-] pennomi@lemmy.world 8 points 1 year ago

The whole “don’t train on synthetic data” thing only holds true if you want to represent real world data, and even then it can be extremely useful.

But if the synthetic data is what you’re trying to replicate anyway, train away!

[-] zorlan@lemmy.world 6 points 1 year ago

I feel like it's similar to image compression, you lose a bit every iteration. Consider that the original model was weighted towards common aspects across the training set. Even with some creative prompting for your source images you could unintentionally introduce bias and reduce variations across images generated by your new model. You also get any mistakes or inconsistencies baked in.

[-] Even_Adder@lemmy.dbzer0.com 3 points 1 year ago

As long as the distortions aren't noticeable, no one can complain.

[-] Even_Adder@lemmy.dbzer0.com 5 points 1 year ago

I didn't know that. People make it seem like your model will explode if you do that.

[-] pennomi@lemmy.world 5 points 1 year ago

Yep. What’s the worst that can happen? You bias your model towards what you’re actually trying to achieve?

It’s really only a problem for targeting realism, because generated images simply aren’t real.

this post was submitted on 07 Oct 2023
20 points (79.4% liked)

Stable Diffusion

4305 readers
12 users here now

Discuss matters related to our favourite AI Art generation technology

Also see

Other communities

founded 1 year ago
MODERATORS