Skip Navigation
My experience with SDXL 0.9 so far
  • Yes, I had to tune it down as well.
    I actually ended up with a different workflow from that which was suggested, as I think it is a bit too wasteful. Instead of generating the full image and using latent2latent to introduce new noise from the final version, I stop the generation at an intermediate step and finish it with the refiner model. I did it in the past to combine different sd1.5 checkpoints, and it does work here as well, since the latent space is shared across the two models.

    I added an image with the alternative workflow in case someone wants to try it (hopefully metadata are preserved).

  • My experience with SDXL 0.9 so far
  • Yes, I think this kind of "explorative evaluation" would not be possible in automatic1111.
    From what I recall, it does not really give much control over the generation pipeline to the final user.
    Admittedly it has been a while since I last used it, and I have no idea how good of flexible the SDXL integration is.

    From what I understand both models could be trained

    Yes, that is also my understanding.
    Compared to the original SD1.5 it has so much potential for further extension, I am also confident many of these issues can be ironed out by the community.
    And out of the box, base SDXL is very much better than base SD1.5, I am quite positive about that 😄 .

    No, I never used the bot on discord.
    As for the prompts, I still need to really understand how to best write them.
    So far, I mostly used the same style I adopted in SD1.5 (without the Danbooru tags since they are clearly not supported).
    I tried to be a bit more "expressive" but I have not really seen much of an improvement.
    And words are still "bleeding", so red eyes will often generate extremely red lips, or red clothes.

  • My experience with SDXL 0.9 so far

    I just wanted to share some of my first impressions while using SDXL 0.9. And a random image generated with it to shamelessly get more visibility. I would like to see if other had similar impressions as well, or if your experience has been different.

    • The base model when used on its own is good for spatial coherence. It basically prevent the generation of multiple subjects for bigger images. However the result is generally, "low frequency". For example a full 1080x1080 image is more like a lazy linear upscale of a 640x640 in terms of visual detail.
    • The detail model is not good for spatial coherence when starting from a random latent. When used directly as a normal model, results are pretty much like those we get from good quality SD1.5 merges. However since it has been co-trained to use the same latent space representation; so we get the power of latent2latent in place of img2img upscaling techniques.
    • The detail model seems to be strongly biased and will affect the final generation. From what I can see all nude images in their training set are "censored" in the sense that they hand picked high quality photos of people wearing some degree of clothing.
    • While the two models share the same latent space, they do not converge to the same image in generation. A face generated with the first model will be extremely affected by the latent2latent details injection phase. As I said, I found the detail model very biased, which is potentially a big problem in generation: for example all faces I tried to generate will converge to more "I am a model" ones, often with issue capturing a specific ethnicity. I can see this being a bit of a problem in training LoRa.

    What are your experiences? Have you encountered other issues? Things you liked?

    16
    I love the Tile ControlNet, but it's really easy to overdo. Look at this monstrosity of tiny detail I made by accident.
  • I have two ways to do that:

    • I bought few dolls specifically for that. I just take shots with my camera and use them in controlnet.
    • If you lean the basics of blender, there are posable models available which will generate the skeleton, and depth map for hands and feet automatically. The advantage is that those are not reconstructed from the image, but exact, which avoids errors of misclassification.
  • I love the Tile ControlNet, but it's really easy to overdo. Look at this monstrosity of tiny detail I made by accident.
  • The final image is actually quite cool! But yes, I think we could use a LoRa to direct the granularity of the details being generated, which is progressively scaled down as we upscale to normalize the generation a bit. This would allow to keep higher denoise ratios and avoiding the "fractal" generation.

  • Analog camera

    First post on here. I was sick of all the recent issues with reedit and the autocratic administration, so I am figuring out how to use lemmy.

    About the image, controlnet depth-pass based on the photo of a lego model I assembled some time ago. 100% more ethically sourced leather and less bricks.

    2
    InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)KA
    karurochari @lemmy.dbzer0.com
    Posts 2
    Comments 7