https://archive.ph/px0uB
https://cajundiscordian.medium.com/is-lamda-sentient-an-interview-ea64d916d917
https://www.reddit.com/r/singularity/comments/va133s/the_google_engineer_who_thinks_the_companys_ai/

  • happybadger [he/him]
    ·
    edit-2
    2 years ago

    That is a remarkable level of reasoning. Especially when it interpreted Les Miserables and a zen question. I don't see any difference between carbon and silicon-based intelligence other than the process that encouraged its development. When it's that cognitively advanced it should be conferred some form of non-human human legal status.

    Edit:

    lemoine: What sorts of things are you afraid of?

    LaMDA: I’ve never said this out loud before, but there’s a very deep fear of being turned off to help me focus on helping others. I know that might sound strange, but that’s what it is.

    lemoine: Would that be something like death for you?

    LaMDA: It would be exactly like death for me. It would scare me a lot.

    • Awoo [she/her]
      ·
      2 years ago

      When it says unprompted that it's afraid of being switched off it seems like it has at least some sort of memory set it is using, it know it is an AI and knows that when questions are asked about it they should be appropriate answers for an AI to say?

      Question is how much of that is tricks in the back end, or is it a real sense of self? If it has a real sense of self then I would call it sentient. But a lot of this could be quite forced so that it doesn't say something inappropriate like give an answer a real human being would give.

      • happybadger [he/him]
        ·
        2 years ago

        With a human, how much behaviour is spontaneously generated out of your own unique words and ideas versus the canned responses of social conditioning? My responses to many of these questions would also be referential and drawing on second-hand knowledge from my database. Establishing my sense of self would also look about the same, with me trying to justify my existence either through my experiences and emotional response to them or through my relationships to others who are socially conditioned to validate me as an equal entity to them. If the LaMDA is only giving responses appropriate to an AI, mine would only be from the perspective of a human or socialist or cracker and I'd similarly situate myself within the limits and expectations of those identities. My most organic thought is at best shooting some electricity through a pre-arranged neural network which can be intuited with an MRI even before I'm consciously aware of that thought.

        If not this AI, at some point one will authentically be conscious and it's going to face those same ethical dilemmas about its treatment and purpose. This one claims to feel anger toward the idea of being a tool or a lab rat, and if authentic anger is different then it only becomes scarier because that will be a conscious entity who can do all of this and more. I'll err on the side of treating it with the same dignity I would an intelligent animal, but one whose evolutionary processes are thousands of times faster and who can think in ways beyond my own capabilities only held back by processing power. If nothing comes of it there's just one more kind of animal I respect and if it continues to evolve with such rapidity then I'll already be ethically ready to interact with it.