• Optional@lemmy.world
    link
    fedilink
    English
    arrow-up
    38
    ·
    10 months ago

    . . . NVIDIA in-house counsel Nikki Pope: In a panel on “governing” AI risk, she cited internal research that showed consumers trusted brands less when they used AI.

    This gels with research published last December that found only around 25 percent of customers trust decisions made by AI over those made by people. One might think an executive with access to this data might not want to admit to using a product that would make people trust them less.

    Indeed.

    • Bigmouse@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      10 months ago

      25% is an abnormally large number considering the current techonological inability to the same thing as a human could. In my experience current “AI” is mostly useful for very specific tasks with very narrow guidelines.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      What’s interesting is the research where when humans don’t know that the output is generated by an AI, they prefer and trust it more than output from actual humans.

  • Optional@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    ·
    10 months ago

    Among the early adopter set, people couldn’t wait for the chance to hand over more of their agency to a glorified chatbot.

    Spot on.

  • BetaDoggo_@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    ·
    10 months ago

    Anywhere speculative investment is involved there are cult like patterns. If your investors don’t believe that your product is going to revolutionize its field you’re not going to get the kind of funding these startups want.

  • Lifecoach5000@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    arrow-down
    1
    ·
    10 months ago

    …an AI-assisted fleshlight for premature ejaculators.

    Sounds straight out of idiocracy lol

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      10 months ago

      I was a professional tech futurist and while I normally made more like ~5yr forecasts, around fifteen years ago I wrote a story for fun that was a further out prediction structured around a narrative taking place in the early 2030s.

      In it, in addition to there being AR computing interfaces and self driving cars, the key tech advance was AI having been developed around a decade earlier - outside of these three things most of the world was the same.

      By this time the AI was shoved into everything from toasters to musical instruments, and the story followed a new class of job that was solely focused on getting AI to do what people wanted by using natural language (what we’d now call a “prompt engineer”).

      The main antagonists were a modern resurgence of the Luddite movement which had grown in popularity as AI had grown.

      The story even had an AI powered dildo.

      It’s been a pretty fucking surreal past few years watching what’s been taking place.

  • TheDarkKnight@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    1
    ·
    10 months ago

    I love the tech but have much the same feelings. AI maybe improve the world eventually, but I predict a painful future in the intervening time. I hope investors turn sooner than later to slow this train but we’ll see. Lot of big players betting the farm on AI, to the point where they’ll do everything to see it through.

    • hansl@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      9
      ·
      10 months ago

      Every advance in technology (see all the Luddites in history) have been accompanied with a wake of pain.

      • Telodzrum@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        1
        ·
        10 months ago

        Not every new piece of technology is actually an advancement. You have an extreme case of selection bias in your assessment.

        • Adalast@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          4
          ·
          10 months ago

          Name 5 that did not have sweeping adverse consequences, with accompanying sources. I will even accept Wikipedia pages if they have attributions. Make sure they are major ones that really shaped the course of human existence moving forward from their introduction.

        • TempermentalAnomaly@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          10 months ago

          This is probably more of a failing of infrastructure and planning than technology. But I think if we only handle advances in technology as a thing on paper and not a thing in society used by people, then we miss an important, but simple point. Technologies are used by people and they is the only way they can change society.

          Any case, toilets ruined London for a couple of decades:

          As the population of Britain increased during the 19th century, the number of toilets did not match this expansion. In overcrowded cities, such as London and Manchester, up to 100 people might share a single toilet. Sewage, therefore, spilled into the streets and the rivers.

          This found its way back into the drinking water supply (which was brown when it came out of the pipes) and was further polluted by chemicals, horse manure and dead animals; as a result, tens of thousands died of water-borne disease, especially during the cholera outbreaks of the 1830s and 1850s.

          In 1848, the government decreed that every new house should have a water-closet (WC) or ash-pit privy. “Night soil men” were engaged to empty the ash pits. However, after a particularly hot summer in 1858, when rotting sewage resulted in “the great stink (pictured right in a cartoon of the day)”, the government commissioned the building of a system of sewers in London; construction was completed in 1865. At last, deaths from cholera, typhoid and other waterborne diseases dropped spectacularly.

          The Great Stink

          The Great Stink only arises because of the development of a sewer system that piped all the sewage to the Thames. And it didn’t stop with the stink:

          Despite Bazalgette’s ingenuity, the system still dumped tons of raw sewage into the Thames - sometimes with unfortunate results. The death toll from the sinking of the pleasure boat Princess Alice in 1878 would certainly have been smaller if it had sunk elsewhere on the Thames. As it was, it went down close to one of the main sewage outfalls. Approximately 640 passengers died, many poisoned rather than drowned. Horror at the deaths was instrumental in the building of a series of riverside sewage treatment plants. [Science Museum]

          So that’s just one example of toilet technology causing a mess. I bet there are others such as the need for an ‘S’ pipe. But ultimately, technological improvements require a little foresight, insight, feedback and a lot of social power.

          • TheDarkKnight@lemmy.world
            link
            fedilink
            English
            arrow-up
            3
            arrow-down
            1
            ·
            10 months ago

            Haha wow I had no idea, just was being a smart ass and got a fun fact out of it. Thanks for sharing!

  • PutangInaMo@lemmy.world
    link
    fedilink
    English
    arrow-up
    11
    ·
    10 months ago

    Behind the bastards is doing a 2 part podcast on this too. Those AI cult mofos are fucking crazy…

  • Crafter72@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    4
    ·
    10 months ago

    Have read the article.

    Maybe call me ignorant but as someone from Eastern part of the world, sometimes I wonder why would these people worry when all of these AI stuff are still prompted from human input, in a sense that We are the one who creates them and dictates its actions. All in all they’re just closed loop automata that happens to have better feedback input compared to your ordinary Closed loop system machines.

    Maybe these people worried because these (regular people) don’t know how these things works or simply they don’t have or lack of self control in first place which what makes them feels like having no control about what happening.

    I understand the danger of AI too, but those who prompted them also human too, in which it is just human nature by itself.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      arrow-down
      1
      ·
      edit-2
      10 months ago

      While this is true in aggregate, consider Elon’s Grok which then turned around and recognized trans women as women, black crime stats as nuanced, and the “woke mind virus” as valuable social progress.

      This was supposed to be his no holds barred free speech AI and rather than censor itself it told his paying users that they were fucking morons.

      Or Gab’s Adolf Hitler AI which, when asked by a user if Jews were vermin, said they were disgusting for having suggested such a thing.

      So yes, AI is a reflection of human nature, but it isn’t necessarily an easily controlled or shaped reflection of that.

      Though personally I’m not nearly as concerned about that being the continuing case as most people it seems. I’m not afraid of a world in which there’s greater intelligence and wisdom (human or otherwise) but one in which there is less.

      • Crafter72@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        10 months ago

        AI is a reflection of human nature, but it isn’t necessarily an easily controlled or shaped reflection of that.

        This partly true especially models that deployed on public and uses large samples gathered from large amount of peoples. Now the parts that we can’t control is if the model is trained with skewed dataset that benefits certain outcomes.

        • kromem@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 months ago

          It depends on which stage of training. As the recent Anthropic research showed, fine tuning out behavior isn’t so easy.

          And at the pretrained layer you really can’t get any halfway decent results with limited data sets, so you’d only be able to try to bias it at the fine tuned layer with biased sourcing, but then per the Anthropic findings (and the real world cases I mentioned above) you are only biasing a thin veneer over the pretrained layer.

    • kingthrillgore@lemmy.ml
      link
      fedilink
      English
      arrow-up
      4
      ·
      10 months ago

      This is not some “computers will outpace us” Terminator shit. Algorithms are human dictated. We are the sole architect of our demise, not something else.

  • nifty@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    3
    ·
    edit-2
    10 months ago

    Hype cycles are nothing new! Way back in the 1800s they used to have world tech fairs as well which were full of inventions that you’d think were full of shit or utterly dystopian. But adoption mostly depends on the masses, and if they’re not going to jive with something, then it doesn’t matter how many nerds are into it. XR/AR is a good example for now, but maybe that’ll change as the form factor of the tech improves.

    I think the main thing that comes out of the AI hype might be digital assistants which know you well enough to assist you like a real assistant, or can do easy but timesink tasks. ChatGPT-based assistants, Cortana, Siri, Alexa are not flexible enough to replace to an actual executive assistant, for example. Any current digital assistant requires a lot of hand holding.

    For people who can’t shell out an executive assistant salary, but need one (almost everyone who works), this will be awesome. For people who can afford an executive assistant, their life is complex enough that they’ll assign the EA to something else.