• ubermeisters@lemmy.world
    link
    fedilink
    English
    arrow-up
    71
    arrow-down
    1
    ·
    edit-2
    1 year ago

    Seems like a bit of a stretch to call 4 seconds per frame, on a 3060, “realtime” / “as fast as you can type”.

    • Domi@lemmy.secnd.me
      link
      fedilink
      English
      arrow-up
      26
      ·
      1 year ago

      I tried it on a 6900 XT recently and generation time was well under half a second.

      Results are not as good as with SDXL but for the time it needs it’s very impressive.

      • ubermeisters@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        ·
        edit-2
        1 year ago

        A rapid dark-tan mammalian with a bushy tail, propels itself upward off the ground, to an elevation above (or greater) than that of the canine resting below, whom has a disposition contrary to productivity.

    • barsoap@lemm.ee
      link
      fedilink
      English
      arrow-up
      14
      ·
      edit-2
      1 year ago

      I’d guess that the ‘realtime’ is a quote from StabilityAI and of course they’re running that stuff on an A100. A couple of seconds is still interactive rate as generally speaking you want to think about the changes you’re making to your conditioning.

      Haven’t tried yet but if individual steps of XL Turbo take ballpark as much time as LCM steps then… well, it’s four to eight times faster. As quality generally isn’t production-ready we’re generally speaking about rough prompt prototyping, testing out an animation pipeline, such stuff, but that has the caveat that increasing step size often leads to markedly different results (complete change of composition, not just details) so the information you gain from those preview-quality images is limited.

      Oh, “production ready quality”: image quality being roughly en par with 4-step LCM means that it’s nowhere near production grade. For the final render you still want to give the model more steps. OTOH I’ve found that some LCM-based merges do in 30 steps what other models need 80 steps for so improvements are always welcome. But I’m also worried about these distilled models being less flexible, pruning only slightly trodden paths that you actually might want the model to take.

      EDIT: Addendum: I’m not seeing anything about using this stuff as a Lora. The nice thing about LCM is that you can take any model you have on your disk and turn it pretty much instantly into a model that can generate fast previews. Also, VAE decoding already can be slower than generation with LCM, so, yeah. I guess having something in between the full VAE and TAESD would be nice, TAESD is fast but is quite limited both when it comes to details, so much that you might not even be able to see what kind of texture SD generated. Oh and it also tends to get colours wrong, at least in my experience it tends to be oversaturated.

    • simple@lemm.ee
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      2
      ·
      1 year ago

      Well, it is technically as fast as you can type if you’re running a better GPU. The 3060 is pretty mid-tier at this point.

      • TheGrandNagus@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        1
        ·
        edit-2
        1 year ago

        Low end card.

        I’ll get crucified for saying that because people will interpret that as an attack on their PC or something daft like that. It’s not.

        It’s Ampere, a GPU architecture from 3.5 years ago. And even then, here’s what the desktop stack was like:

        1. 3090 Ti (GA102)

        2. 3090 (GA102)

        3. 3080 Ti (GA102)

        4. 3080 12GB (GA102)

        5. 3080 (GA102)

        6. 3070 Ti (GA102/GA104)

        7. 3070 (GA104)

        8. 3060 Ti (GA104/GA103)

        9. 3060 (GA106/GA104)

        10. 3050 (GA106/GA107)

        It was almost at the bottom of Nvidia’s stack 3 years ago. It was a low end card then (because, you know, it was at the bottom end of what they were offering). It’s an even more low end card now.

        People are always fooled by Nvidia’s marketing and thinking they’re getting a mid range card when in reality Nvidia’s giving people the scraps and pretending they’re giving you a great deal. People need to demand more from these companies.

        Nvidia takes a low end card, slaps a $400 price tag on it, calls it mid range, and people lap it up every time.

        • Blackmist@feddit.uk
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          2
          ·
          1 year ago

          The pricing makes it a mid range card, because the budget end is just gone these days.

          • TheGrandNagus@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            1 year ago

            Nvidia conning people into paying what used to be mid range/high end pricing for a low end card does not make it a low end card.

            The 3060 was always a low end card. Because it was on the low end of the product stack, both for Nvidia and against AMD.

        • simple@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          I know it’s low-end when compared to the newer generations but if we call a 3060 low-end then what do we call people with older GPUs like a 1070?

          • TheGrandNagus@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            ·
            1 year ago

            Should we not compare the 3060 against its own generation/the current one? To me that makes more sense than including the 1000 series or 900 series or something. How far would we go back? Are all cards sold now high end because they’re faster than a GTX 960? Earlier?

            Personally my cut off was cards still on sale either right now or very recently, say within the past year.

  • L3ft_F13ld!@links.hackliberty.org
    link
    fedilink
    English
    arrow-up
    56
    arrow-down
    11
    ·
    edit-2
    1 year ago

    XL Turbotastic Mega Ginormous, etc. Hate naming schemes like this. Why not just make it v2.0 or the Pro version instead? Why use multiple words that make it sound bigger and better? Marketing BS that just sounds dumb.

    • simple@lemm.ee
      link
      fedilink
      English
      arrow-up
      50
      arrow-down
      2
      ·
      edit-2
      1 year ago

      Not sure why you have a problem with it, the naming here makes a lot of sense if you know the context.

      Stable Diffusion --> The original SD with versions like 1.5, 2.0, 2.1 etc

      Stable Diffusion XL --> A version of SD with much bigger training data and support for much larger resolutions (hence, XL)

      Stable Diffusion XL Turbo --> A version of SDXL that is much faster (hence, Turbo)

      They have different names because they’re actually different things, it’s not exactly a v1.0 --> v2.0 scenario

      • at_an_angle@lemmy.one
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        3
        ·
        1 year ago

        Naming schemes that aren’t clear are absolute garbage.

        What if you’re new to it, and there are 6 different recent versions of something all named with a description instead of version number? Is Jumbo newer than Mega?

        Fuck it, I’m ranting about this because it still upsets me.

        I wanted to buy a 3DS to play Shovel Knight and Binding of Issac. Reading up on them, BoI would only play on a New 3DS XL. Cool.

        Went to the store and bought a new 3DS XL only to find out I got the wrong one. What I wanted was a NEW 3DS XL, and what I got was a 3DS XL that was new. There is a difference, and it took me 4 days to notice, and I was working out of town for the next month. So I can’t return it. FUN!

        So screw naming new versions of things with names instead of numbers. But somehow, Microsoft screwed that one up.

        KISS: Keep it simple, stupid.

        • simple@lemm.ee
          link
          fedilink
          English
          arrow-up
          5
          ·
          edit-2
          1 year ago

          Sure, 3DS names are dumb, but this is definitely not the case here. Using version numbers instead of different names for different things causes insane confusion and having to over-explain what it is.

          See: DLSS

          DLSS 2 is just DLSS 1 but better. DLSS 3 is frame generation that isn’t compatible with most hardware. DLSS 3.5 is similar to DLSS 2 but includes enhanced raytracing denoising.

          It’s a nightmare. Making a version 2, 3, 4 etc of something also makes it sound like there’s no reason to use the old version, whereas a lot of people are still using the regular stable diffusion over stable diffusion XL.

          Imagine if the discussion was “Hey don’t use Stable Diffusion 3 since you need a lot of VRAM, you should be using Stable Diffusion 1.5 or Stable Diffusion 2.1, but also it’s worth getting a new GPU for Stable Diffusion 4 cuz it’s very fast but has lower quality than version 3”

      • Phoenixz@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Yeah but the next version has yet a bigger training set, so what then? XXL? and what about the next ? Turbo was already used, so now we call it Nitro? This is not the “new kids” movies, you know…

    • grue@lemmy.world
      link
      fedilink
      English
      arrow-up
      35
      arrow-down
      2
      ·
      1 year ago

      Why not just make it v2.0 or the Pro version instead?

      “Pro version” is equally cringe.

      • L3ft_F13ld!@links.hackliberty.org
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        Yeah I get that. Would just have made more sense given that it’s widely used. Though I’ve been told why the name is so weird and it makes some sense now

      • TherouxSonfeir@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 year ago

        Here are my suggestions:

        Stable Diffusion Free

        Stable Diffusion Paid with Limitations

        Stable Diffusion Paid Unlimited

    • DABDA@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      ·
      1 year ago

      I agree with you in general, but for Stable Diffusion, “2.0/2.1” was not an incremental direct improvement on “1.5” but was trained and behaves differently. XL is not a simple upgrade from 2.0, and since they say this Turbo model doesn’t produce as detailed images it would be more confusing to have SDXL 2.0 that is worse but faster than base SDXL, and then presumably when there’s a more direct improvement to SDXL have that be called SDXL 3.0 (but really it’s version 2) etc.

      It’s less like Windows 95->Windows 98 and more like DOS->Windows NT.

      That’s not to say it all couldn’t have been better named. Personally, instead of ‘XL’ I’d rather they start including the base resolution and something to reference whether it uses a refiner model etc.

      (Note: I use Stable Diffusion but am not involved with the AI/ML community and don’t fully understand the tech – I’m not trying to claim expert knowledge this is just my interpretation)

      • barsoap@lemm.ee
        link
        fedilink
        English
        arrow-up
        3
        ·
        1 year ago

        AFAIU SDXL is actually an erm genetic descendant of SD1.5, with its architecture expanded, weights transferred from 1.5, and then trained on bigger inputs (512x512 in the end is awfully small). SD2.0 is a completely new model, trained from scratch and as far as I’m aware noone’s actually using it. Also noone is using the SDXL refiner if you go to civitai it’s all models with detailer capabilities baked in, what you do see is workflows that generate an image, add some noise at the very end and repeat the last couple of steps. Using the base sdxl refiner on the output of other sdxl models is sometimes right-out comical because it sometimes has no idea what it’s looking at and then produced exquisitely surface texture details of the wrong material. Say a silk keyboard because it doesn’t realise that it’s supposed to be ABS and, well, black silk exists.

    • foggy@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 year ago

      Im just glad we’re moving away from purposely misspelled product SEO hacks.

  • LifeInOregon@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    4
    ·
    1 year ago

    And the resulting faces still all have lazy eyes, asymmetric features, and significantly uncanny issues.

      • LifeInOregon@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        1
        ·
        1 year ago

        These features are abnormally asymmetric to the point of being off-putting. General symmetry of features is a significant part of what attracts people one to another, and why facial droops from things like Bells Palsy or strokes can often be psychologically difficult for the patient who experiences them.

        General symmetry, not exact symmetry.

      • Apothecary@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        1 year ago

        Anecdote: I think Denzel Washington is supposed to have one of the most symmetrical faces.

      • atocci@kbin.social
        link
        fedilink
        arrow-up
        5
        arrow-down
        1
        ·
        1 year ago

        What’s out there that actually works offline? Stable Diffusion is the only one I’ve heard about, everyone else is more interested in exclusively selling AI as a service.

        • barsoap@lemm.ee
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          edit-2
          1 year ago

          Llama etc, reduced ChatGPT models. Never tried them but they’re out there. There’s also plenty of support stuff that may or may not be interesting, e.g. turning images into depth maps in case you don’t have enough angles for actual photogrammetry. controlnet-aux for comfyui has a good selection of that analysis stuff.

      • TimeSquirrel@kbin.social
        link
        fedilink
        arrow-up
        22
        arrow-down
        1
        ·
        1 year ago

        You might be waiting a long time. We aren’t going back and this is one of those things that are not going back into the box. So now we must prepare for it and learn to live with it as the best course of action and make sure it’s not used to oppress us.

      • SCB@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        1
        ·
        1 year ago

        Lmao I’m old enough to remember “the internet is just a fad”

  • Stalinwolf@lemmy.ca
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    2
    ·
    1 year ago

    I’ve tried to install this multiple times but always manage to fuck it up somehow. I think the guides I’m following are outdated or pointing me to one or more incompatible files.

    • barsoap@lemm.ee
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      1 year ago

      Tough luck running any code published by people who put out models, it’s research-grade software in every sense of the word. “Works on my machine” and “the source is the configuration file” kind of thing.

      Get yourself comfyui, they’re always very fast when it comes to supporting new stuff and the thing is generally faster and easier on VRAM than A1111. Prerequisite is a torch (the python package) enabled with CUDA (nvidia) or rocm (AMD) or whatever Intel uses. Fair warning: Getting rocm to run on not officially supported cards is an adventure in itself, I’m still on torch-1.13.1+rocm5.2 newer builds just won’t work as the GPU I’m telling rocm I have so that it runs in the first place supports instructions that my actual GPU doesn’t, and they started using them.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    1 year ago

    This is the best summary I could come up with:


    Stability detailed the model’s inner workings in a research paper released Tuesday that focuses on the ADD technique.

    One of the claimed advantages of SDXL Turbo is its similarity to Generative Adversarial Networks (GANs), especially in producing single-step image outputs.

    Stability AI says that on an Nvidia A100 (a powerful AI-tuned GPU), the model can generate a 512×512 image in 207 ms, including encoding, a single de-noising step, and decoding.

    This move has already been met with some criticism in the Stable Diffusion community, but Stability AI has expressed openness to commercial applications and invites interested parties to get in touch for more information.

    Meanwhile, Stability AI itself has faced internal management issues, with an investor recently urging CEO Emad Mostaque to resign.

    Stability AI offers a beta demonstration of SDXL Turbo’s capabilities on its image-editing platform, Clipdrop.


    The original article contains 553 words, the summary contains 138 words. Saved 75%. I’m a bot and I’m open source!

  • You999@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    2
    ·
    1 year ago

    This is great news for people who make animations with deforum as the speed increase should make Rakile’s deforumation GUI much more usable for live composition and framing.

  • Gabu@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    2
    ·
    edit-2
    1 year ago

    Does it actually run any faster though? For instance, if I manually spun a model with the diffusers library and ran it locally on dml, would there be any difference?

    Edit: Assuming we’re normalizing the output to something reasonable, e.g. a recognizable picture of a dog.