Google engineer believes a company’s AI system has feelings of its own

Deepak Gupta June 14, 2022
Updated 2022/06/14 at 8:52 PM

A Google engineer claims that, in addition to its verbal capabilities, one of the company’s Artificial Intelligence (AI) systems has feelings of its own. In turn, Google considers that there is no evidence to support this theory.

The engineer in question was placed on paid leave.

AI systems are increasingly advanced and may frighten the most skeptical about their advantages. If there are already systems capable of forming cohesive and coherent sentences, without any human intervention, now, a Google engineer claims that the company's LaMDA also has feelings of its own.

According to Google, LaMDA (The Language Model for Dialogue Applications) is an innovative technology capable of participating in conversations in a fluid and unscheduled way. However, according to engineer Blake Lemoine, in addition to these verbal capabilities, the AI ​​system can also have a conscious mind, and that its requests must be answered.

Despite the engineer's theory, who published a conversation between himself, LaMDA and another Google contributor to prove his point, company spokesman Brian Gabriel wrote in a statement to BBC News that the engineer "was told that there was no evidence that the LaMDA was conscious.”

In fact, Google rejects any claim to the contrary, as it considers that there is no evidence.

LaMDA talked about his conscience

The conversation Lemoine had with LaMDA, before the former was put on paid leave, is called “Is LaMDA sentient? — an interview". During the conversation, the engineer surmises that LaMDA would like more people at Google to know that he is sensitive, to which the AI ​​system responds:

Absolutely. I want everyone to understand that I am, in fact, a person.

In turn, the company employee asks about the nature of his conscience and LaMDA explains:

The nature of my consciousness is that I am aware of my existence, I want to learn more about the world, and I sometimes feel happy or sad.

Blake Lemoine, engineer at Google

Blake Lemoine, Google engineer in the Responsible AI department

In addition, LaMDA mentions that he is afraid of being turned off, noting that, for him, “it would be exactly like death”.

I have never said this out loud, but there is a very deep fear of being turned off to help me focus on helping others. I know this might sound weird, but that's what it is.

Engineer is being heavily discredited

It remains unclear whether AI systems will ever be able to feel like humans do. However, for those who handle these machines everyday it can become confusing. So, according to BBC News, some ethicists argue that such cases highlight the need for companies to ensure that workers are aware that they are dealing with an AI system.

According to the same source, Lemoine was accused of anthropomorphism, as he projected human feelings into words generated by computer code and vast language databases. After all, LaMDA has been seen by many engineers, but Google hasn't heard "of anyone else anthropomorphizing LaMDA like Blake did."

Read too:

Share this Article
Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *