• simple@lemm.ee
    link
    fedilink
    English
    arrow-up
    49
    ·
    6 months ago

    Stability AI crashed and burned so fast it’s not even funny. Their talent is abandoning ship they’ve even been caught scraping images from Midjourney, which means they probably don’t have a proper dataset.

    • Even_Adder@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      27
      arrow-down
      1
      ·
      6 months ago

      The model should be capable of much better than this, but they spent a long time censoring the model before release and this is what we got. It straight up forgot most human anatomy.

  • postmateDumbass@lemmy.world
    link
    fedilink
    English
    arrow-up
    27
    arrow-down
    2
    ·
    6 months ago

    Almost like the issues with repressing sex and nudity are harming the development of intelligence. Just like real life.

    • egeres@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      6 months ago

      I was going to say this, their new architecture seems to be better than previous ones, they have more compute and I’m guessing, more data. The only explanation for this downgrade is that they tried to ban porn. I haven’t read online info about this at the time anyways, I’m just learning this recently

  • leekleak@lemmy.world
    link
    fedilink
    English
    arrow-up
    24
    arrow-down
    7
    ·
    edit-2
    6 months ago

    Honestly I think that it’s models like these that output things that could be called art.

    Whenever a model is actually good, it just creates pretty pictures that would have otherwise been painted by a human, whereas this actually creates something unique and novel. Just like real art almost always ilicits some kind of emotion, so too do the products of models like these and I think that that’s much more interesting that having another generic AI postcard.

    Not that I’m happy to see how much SD has fallen though.

  • j4k3@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    6 months ago

    ? They are all bad at first for the average person that uses surface level tools, but SD3 won’t have the community to tune it because it is proprietary junk and irrelevant now.

        • fruitycoder@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 months ago

          I’m not seeing about the lead researcher leaving because of that, just they are leaving. With the expenses far exceeding revenue right now being a suspected reason.

    • TheRealKuni@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      6 months ago

      SD3 won’t have the community to tune it because it is proprietary junk and irrelevant now.

      What changed between SDXL and SD3? I’m out of the loop on this one.

      • randon31415@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        6 months ago

        They realized that no matter how much they charged as a one time fee, the people the got the one time fee enterprise license would eventually cost them more in computational costs them the fee. So they switched it to 6000 image generations, which wasn’t enough for most of the community that made fixes and trained loras, so none of the “cool” community stuff will work with SD3.

        • interdimensionalmeme@lemmy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 months ago

          Have they considered a community sponsored “group buy” of compute, to just train the model as far as the community will bear ? SDXL was so great, surely 100k people could put 5$ a month toward making monthly improvement open source checkpoints happen ? I don’t see any other financing model work out if the output is open source. It simply can’t be financed after publication. And it won’t get the community support if it’s behind a paywall.

  • kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    ·
    6 months ago

    Basically, any time a user prompt homes in on a concept that isn’t represented well in the AI model’s training dataset, the image-synthesis model will confabulate its best interpretation of what the user is asking for.

    I’m so happy that the correct terminology is finally starting to take off in replacing ‘hallucinate.’

  • db2@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    6 months ago

    Also from reddit, with zero irony:

    Kudos to Stablility AI for releasing ANOTHER excellent model for FREE.

    💀

  • BetaDoggo_@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    6 months ago

    The model does have a lot of advantages over sdxl with the right prompting, but it seems to fall apart in prompts with more complex anatomy. Hopefully the community can fix it up once we have working trainers.