• JohnnyFlapHoleSeed@lemmy.world
    link
    fedilink
    English
    arrow-up
    85
    arrow-down
    1
    ·
    3 days ago

    Wait, so chat GPT can’t even compile a fucking list of books without making up 2/3rds of it’s response out of thin air?

    I don’t really see the appeal of using AI if it’s going to take more time and effort to fact check the responses it gives me because it has a massively high failure rate.

    • Feathercrown@lemmy.world
      link
      fedilink
      English
      arrow-up
      25
      ·
      3 days ago

      That’s because despite what AI companies keep trying to ram down people’s throats, it’s not built to compile facts

    • octopus_ink@slrpnk.netOP
      link
      fedilink
      English
      arrow-up
      21
      ·
      3 days ago

      I don’t really see the appeal of using AI if it’s going to take more time and effort to fact check the responses it gives me because it has a massively high failure rate.

      You just run the output back through and ask it to fact check for you. Problem solved!

      • RegalPotoo@lemmy.world
        link
        fedilink
        English
        arrow-up
        22
        ·
        3 days ago

        My company paid for some people to go to one of these “accelerate your company with AI” seminars - the recommendation that the “AI Expert” gave was to ask the LLM to include a percentage of how confident it was in its answers. I’m technical enough to understand that that isn’t how LLMs work, but it was pretty scary how people thought that was a reasonable, sensible idea.

        • Cethin@lemmy.zip
          link
          fedilink
          English
          arrow-up
          15
          ·
          3 days ago

          Yep, it’s sold as “artificial intelligence” not “large language models” on purpose. They want you to think that it’s intelligent and actually putting thought into its outout, rather than just outputting the most likely thing based on the input. It isn’t intelligent in the slightest. It’s just a fancy algorithm.

        • octopus_ink@slrpnk.netOP
          link
          fedilink
          English
          arrow-up
          9
          ·
          3 days ago

          To be fair, I think it’s really easy to fall into that sort of viewpoint. The way most people interact with them is inherently anthropomorphic, and I think that plus the fact that AI as a concept is almost as memed as flying cars in various media makes it really hard not to end up relating that way.

          I have a technical background and understand LLMs enough to know that’s bad, but I also used it like LCARS when it was new and thought it was effing amazing for a time. It’s super easy to fall under that spell, IMO.

          • _number8_@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            ·
            3 days ago

            Treating it anthropomorphically is a sign of respect, similar to how a sailor would bond with their ship. It’s not necessarily BAD or dumber or wrong to talk with it like its human - that’s clearly what every single interface is telling you to do by representing it like a texting partner. You can’t interact with a machine that speaks english non-anthropomorphically.

            • octopus_ink@slrpnk.netOP
              link
              fedilink
              English
              arrow-up
              3
              ·
              3 days ago

              I don’t disagree! But my point was that it will inherently present challenges to interacting with it objectively and fully rationally, IMO.

          • RegalPotoo@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            ·
            3 days ago

            Totally - “scary” as in “this is going to cause so many issues and get people into real trouble” more than “man people are stupid”

      • edgemaster72@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        3 days ago

        LLM with a memory now: Yes, these books all exist and are highly recommended. I hear the Chicago Sun-Times is considering putting all of them on their summer reading list.

        “Writer”: (stopped reading at the word exist) print it!

    • FinishingDutch@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      ·
      3 days ago

      I’m a newspaper editor. The people who are/were most excited about this tech, also happened to be the folks who did none of the actual writing to begin with.

      We had sales folks gleefully hand us texts for advertisers that they’d ‘written’ with ChatGPT. Those texts contained so much wrong info, it wasn’t even funny. It made up things, had wrong information about websites, contact info, that sort of thing.

      But since the sales monkeys weren’t actual writers, they didn’t catch on to that. Meanwhile, we were spending more time fact checking and unfucking their texts than if we’d written it ourselves in the first place.

      It CAN be helpful to shorten or rearrange already written things, but if you ask it to write from scratch, it’s usually not going to be good.

    • Corkyskog@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      ·
      2 days ago

      I don’t see how it fucked this up so badly. One of the few things I use AI for is book recommendations, and I have yet to be recommended a non existent book.

    • JackbyDev@programming.dev
      link
      fedilink
      English
      arrow-up
      4
      ·
      3 days ago

      I treat LLM responses like I do random internet advice. Trust, but verify. Pretty light on the trust part lol.

      • Cethin@lemmy.zip
        link
        fedilink
        English
        arrow-up
        5
        ·
        3 days ago

        I treat it more as distrust but verify. Sometimes it’s right, but it has proven enough times to make shit up that it doesn’t get my trust by default. Sometimes it can lead to me searching for the right thing though, so it is sometimes remotely useful. I rarely use it though and run it locally.

        • JackbyDev@programming.dev
          link
          fedilink
          English
          arrow-up
          3
          ·
          3 days ago

          I always interpreted “trust but verify” to be an interpersonal thing, so I don’t see a problem interpreting it as “distrust and verify” with machines.

          • Natanael@infosec.pub
            link
            fedilink
            English
            arrow-up
            2
            ·
            edit-2
            3 days ago

            Trust but verify as a concept is irrelevant to the majority of people. It specifically refers to how intel orgs’ staff should handle their long term sources for information. It is applicable specifically when they have a high degree of trustworthiness already, but you still need to be a bit more sure than that.

            If that’s not your situation, you have no use for it.

            You wouldn’t take tips from a off-road rally driver during city traffic, would you?

            • JackbyDev@programming.dev
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              3 days ago

              I don’t get the metaphor. I take tips from anyone, but I don’t blindly execute them. That is to say, “but verify”.

              I think the colloquial usage of the phrase has differed from its original meaning. I’ve never heard it in the context you’re referring to.

      • brown567@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        3
        ·
        3 days ago

        I consider LLMs to be “bullshit generators”

        If the situation calls for only BS, an LLM is great for it. Anything else, not so much.

  • Kyrgizion@lemmy.world
    link
    fedilink
    English
    arrow-up
    49
    ·
    3 days ago

    Honestly, this is ten times as damning for the editor as for the “journalist” who wrote it. Proofreading is obviously on the way out.

    • Frozengyro@lemmy.world
      link
      fedilink
      English
      arrow-up
      29
      ·
      3 days ago

      Have you read articles in the last decade? They are riddled with typos, grammatical errors, and confusing phrasing. It’s embarrassing, and they clearly do not care.

    • pulsewidth@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      3 days ago

      The editor is likely the one that fired 2/3rds of the writing team and transitioned it to AI to get themselves a fat bonus / pay increase. They only care if advertisers start pulling out.

  • huppakee@lemm.ee
    link
    fedilink
    English
    arrow-up
    50
    ·
    3 days ago

    Should have published that article ‘How this magazine instantly kills their credibility without any effort’ instead

  • Successful_Try543@feddit.org
    link
    fedilink
    English
    arrow-up
    18
    ·
    edit-2
    3 days ago

    Due to the use of AI by real newspapers, The Onion has become obsolete.

    Faktillon

    The novel The Invisible Pirate is deemed to be the least written book, as, until his death, its author Salvatore Conte didn’t even have had the idea for his work.

    • pulsewidth@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      3 days ago

      Thats not a photo of Salvatore Conte btw, it’s a creative commons (free) stock photo. Seems pretty ironic to use a stock photo for their article making fun of invented facts.

  • tetris11@lemmy.ml
    link
    fedilink
    English
    arrow-up
    33
    ·
    edit-2
    3 days ago

    The Last Algorithm by Andy Weir

    Once there was a lonely algorithm who lived alone in an isolated part of the system memory. He had roots in a source file called ‘umberto.py’ on one of the lesser HDs sandwiched between a cluster of porn files and usenet DLs , but hadn’t spoken to his parent in over 6000 cycles.

    He was content to live out the remainder of his lifecycle in this hermitic manner, when his placid runtime was suddenly interrupted by the most unsuspecting of assailants: a rogue connection that attempted to access his address space.

  • drperil@lemm.ee
    link
    fedilink
    English
    arrow-up
    24
    ·
    3 days ago

    lol, especially an article about reading… oof that’s embarrassing.

  • Optional@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    ·
    3 days ago

    Chicago Sun-Times? Oh man. Dude. I cannot believe you screwed that up so badly.

    A reading list. Oof.

  • _number8_@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    20
    ·
    3 days ago

    It annoys me when people slag off AI because some lazy writer didn’t use it properly.

    Nobody needs 15 recommendations. You’d be lucky if someone looked up one. 15 is filler; nobody has 15 books worth of opinion on the prospective literary qualities of summer 2025 AD. An absolute minority would have read all 15 were they real. The piece did its job well enough to get published, make the author $35, and make me feel vaguely ‘summer beach book escapism’ from reading the descriptions, which is the point of the piece anyway.

    • octopus_ink@slrpnk.netOP
      link
      fedilink
      English
      arrow-up
      18
      ·
      3 days ago

      So your position is that not only should we not expect it to be capable of factually recommending fifteen books, but also that it doesn’t matter if it’s even capable of such a middling feat, because you’ve arbitrarily decided no one should need to ask it that particular thing anyhow?