• kitnaht@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    56
    ·
    edit-2
    2 months ago

    Honestly, he’s wrong though.

    I know tons of full stack developers who use AI to GREATLY speed up their workflow. I’ve used AI image generators to put something I wanted into the concept stage before I paid an artist to do the work with the revisions I wanted that I couldn’t get AI to produce properly.

    And first and foremost, they’re a great use in surfacing information that is discussed and available, but might be buried with no SEO behind it to surface it. They are terrible at deducing things themselves, because they can’t ‘think’, or coming up with solutions that others haven’t already - but so long as people are aware of those limitations, then they’re a pretty good tool to have.

    It’s a reactionary opinion when people jump to the ‘but they’re stealing art!’ – isn’t your brain also stealing art when it’s inspired by others art? Artists don’t just POOF, and have the capability to be artists. They learn slowly over time, using others as inspiration or as training to improve. That’s all stable diffusors do - just a lot faster.

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      ·
      edit-2
      2 months ago

      Speaking as someone who worked on AI, and is a fervent (local) AI enthusiast… it’s 90% marketing and hype, at least.

      These things are tools, they spit out tons of garbage, they basically can’t be used for anything where the output could likely be confidently wrong, and the way they’re trained is still morally dubious at best. And the corporate API business model of “stifle innovation so we can hold our monopoly then squeeze users” is hellish.

      As you pointed out, generative AI is a fantastic tool, but it is a TOOL, that needs some massive changes and improvements, wrapped up in hype that gives it a bad name… I drank some of the kool-aid too when llama 1 came out, but you have to look at the market and see how much fud and nonsense is flying around.

    • antonim@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 months ago

      they’re a great use in surfacing information that is discussed and available, but might be buried with no SEO behind it to surface it

      This is what I’ve seen many people claim. But it is a weak compliment for AI, and more of a criticism of the current web search engines. Why is that information unavailable to search engines, but is available to LLMs? If someone has put in the work to find and feed the quality content to LLMs, why couldn’t that same effort have been invested in Google Search?

      • kitnaht@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        2 months ago

        If someone has put in the work to find and feed the quality content to LLMs, why couldn’t that same effort have been invested in Google Search?

        I’d rather a world where 10 companies can compete with google search with AIs, than where they dump money into a monopoly.

        • antonim@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 months ago

          If you don’t feel like discussing this and won’t do anything more than deliberately miss the point, you don’t have to reply to me at all.

          • kitnaht@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            1 month ago

            The content is not unavailable to search engines. AI LLMs simply are better at surfacing it. I don’t know what point you were trying to make that I missed, it wasn’t on purpose, I assure you.

            • antonim@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 month ago

              AI LLMs simply are better at surfacing it

              Ok, but how exactly? Is there some magical emergent property of LLMs that guides them to filter out the garbage from the quality content?

              • kitnaht@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 month ago

                Yeah. Money. Google has an incentive to make search results less accurate to get you to click around and interact with more ads. As it currently stands, AI models aren’t inserting advertisements; though I suspect that’s only a matter of time.

                • antonim@lemmy.dbzer0.com
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  1 month ago

                  And that’s more or less what I was aiming for, so we’re back at square one. What you wrote is in line with my first comment:

                  it is a weak compliment for AI, and more of a criticism of the current web search engines

                  The point is that there isn’t something that makes AI inherently superior to ordinary search engines. (Personally I haven’t found AI to be superior at all, but that’s a different topic.) The difference in quality is mainly a consequence of some corporate fuckery to wring out more money from the investors and/or advertisers and/or users at the given moment. AI is good (according to you) just because search engines suck.

                  • kitnaht@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    edit-2
                    1 month ago

                    That’s kind of how things work you know.

                    AI is good (according to you) just because search engines suck.

                    Yeah, would you say the original iPhone is any good today? No. Because everything got better. That’s how things work. AI of today, in 20 years is probably going to be considered to suck.

                    That’s how that works. When things are better than other things, we consider them good.

    • DacoTaco@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      2 months ago

      He isnt wrong. This comes from somebody who technically uses ai daily to help develop ( github copilot in visual studio to assist in code prediction based on the code base of the solution ), but AI is marketed even worse than blockchain back in 2017. Its everywhere, in every product, even if it doesnt have ai or has nothing to do with it. Monitor ai shit? Mouse with ai? Hell, ive seen a sketch of a fucking toaster with ‘ai’.
      There is shit like microsoft recall, apple intelligence, bing co pilot, office co pilot, …
      All of those are just… Nothing special or useful. There are also chatbots which bring nothing new to the table either.
      Everyone and everything wants to market there stuff with ai and its disgusting.
      Does that mean that current ai tech cant bring anything to the table? No, it totally can, but 90% of ai stuff out there is, just like linus says, marketing bullshit.

    • Valmond@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      2
      ·
      2 months ago

      If you are just blatantly copying art, well yeah you’re stealing it.

    • AreaKode@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      9
      ·
      2 months ago

      AI can give me a blueprint for my logic. Then I, as a developer, make the code run. Cuts my scripting time in half.

      • Wrench@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 months ago

        Rofl. As a developer of nearly 20 years, lol.

        I used copilot until finally getting fed up last week and turning it off. It was a net negative to my productivity.

        Sure, when you’re doing repetitive operations that are mostly copy paste and changing names, it’s pretty decent. It can save dozens of seconds, maybe even a minute or two. That’s great and a welcome assist, even if I have to correct minor things around 50% of the time.

        But when an error slips through and I end up spending 20 minutes tracking down the problem later, all that saved time vanishes.

        And then the other times where my IDE is frozen because the plugin is stuck in some loop and eating every last resource and I spend the next 20 minutes cursing and killing processes, manually looking for recent updates that hadn’t yet triggered update notifications, etc… well, now we’re in the red, AND I’m pissed off.

        So no, AI is not some huge boon to developer productivity. Maybe it’s more useful to junior developers in the short term, but I have definitely dealt with more than a few problems that seem to derive from juniors taking AI answers and not understanding the details enough to catch the problems it introduced. And if juniors frequently rely on AI without gaining deep understanding, we’re going to have worse and worse engineers as a result.