• 0 Posts
  • 5 Comments
Joined 2 years ago
cake
Cake day: September 20th, 2023

help-circle
  • You’re acting as if the bot had some sort of intention to help him. It’s a bot. It has zero intention whatsoever since it’s not a conscious entity. It is programmed to respond to an input. That’s it.

    The larger picture here is that this technology is being used by people in a way that’s being used as if it were a conscious entity. Including the mentally ill. Which is very dangerous, and can drive people to action as we can see.

    That’s not to say I have any idea how to handle this. Because I don’t have a clue. But it is a discussion that needs to be had rather than minimizing the situation as an “well the bot actually tried to talk him out of suicide”, because in my opinion that’s not the point. We are interacting with this technology in a way that is changing our own behavior and world view. And it is causing real world harm like this.

    When we make something so believable as to trick people into thinking that they’re interacting with consciousness, that is a giant alarm we must discuss. Because at the end of the day, it’s a technology that can be owned, controlled, and manipulated by the owner class to serve their needs of maintaining power.




  • underwire212@lemm.eetomemes@lemmy.world*time traveler sneezes*
    link
    fedilink
    arrow-up
    22
    arrow-down
    1
    ·
    13 hours ago

    Well said. Agree wholeheartedly.

    I am reading Man’s Search for Meaning, and the author speaks about how even in the concentration camps, where death and suffering is shoved in your face, the prisoners still joke and try to make the best of a truly horrific situation. I guess humor has evolved as a sort of coping mechanism, and as a method of keeping sane.

    If you haven’t read it, I highly recommend it. It’s very powerful and raw.