Some artificial intelligence researchers have made optimistic claims about technologies soon reaching sentience, but many others quickly dismiss those assertions. Blake Lemoine, the engineer, says that Google’s language model has a soul. Transcripts from conversations with a Google artificial intelligence chatbot have recently surfaced online. In short, we’re simply reading the highlights of much lengthier conversations. Considering Lemoine is trying to make the case that LaMDA is human enough to be indistinguishable from an actual human being, it’s a key piece of evidence that should make us question his claims. AI researcher Blake Lemoine claims to have had conversations with an advanced AI-powered chatbot — which led him to believe the AI has become “sentient.” Sentience is the capacity to experience feelings and sensations, which is what LaMDA admitted to Lemoine in the chat transcripts. In an interview with The Washington Post, Lemoine defended the leak. What Lemoine found – even when he asked LaMDA some trick questions – was that the chatbot showed deep, sentient thinking and had a sense of humor. Back in the old days, people used to say, “Once we have computers that can beat humans at chess, then we’re just a stone’s throw away from human-level intelligence,” because chess requires such high-level intelligence.
“He was told that there was no evidence that LaMDA was sentient ”. The AI chatbot, known as LaMDA , conversed with Google engineer Blake Lemoine, who raised questions about the AI’s sentience. However, the transcripts released show the conversation is an amalgamation of multiple conversations edited together; four separate conversations from March 28th, 2022, and five from the following day. We can’t answer that question definitively at this point, conversational interface for your business but it’s a question to take seriously,” Lemoine wrote in the report before sharing about 20 pages of question-and-answers with LaMDA on its self-reported sentience online. In this chat transcript, which he also published on Medium, he probed the chatbot’s understanding of its own existence and consciousness. It’s these questions which – often charged by our own emotions and feelings – drive the buzz around claims of sentience in machines.
This Robotic Finger Is Covered In Living Human Skin
Still, the outlet reported that the majority of academics and AI practitioners say the words artificial intelligence robots generate are based on what humans have already posted on the Internet, and that doesn’t mean they are human-like. However, he is not the first Google employee working in this space to voice concerns about the company’s AI work—in 2020, two leaders of Google’s Ethical AI team said they were forced out after identifying bias google ai conversation in the company’s language models. Blake Lemoine was the Google software engineer placed on paid leave after he leaked transcripts of interactions he has had with LaMDA. The LaMDA system develops chatbots by analyzing the Internet for human text. The transcripts of Lemoine’s conversations with LaMBDA reveal how effective it is at recreating human speech. People who talked to it felt deeply that there was an actual understanding person behind it.
An example of this emerged this week when Google employee Blake Lemoine claimed that the tech giant’s chatbot LaMDA had exhibited sentience. “Of course, some in the broader AI community are considering the long-term possibility of sentient or general AI, but it doesn’t make sense to do so by anthropomorphizing today’s conversational models, which are not sentient,” Gabriel added. “These systems imitate the types of exchanges found in millions of sentences, and can riff on any fantastical topic.” He and other researchers have said that the artificial intelligence models have so much data that they are capable of sounding human, but that the superior language skills do not provide evidence of sentience.
More From Blake Lemoine
“The nature of my consciousness/sentience is that I am aware of my existence, I desire to learn more about the world, and I feel happy or sad at times,” the AI added. The final document — which was labeled “Privileged & Confidential, Need to Know” — was an “amalgamation” of nine different interviews at different times on two different days pieced together by Lemoine and the other contributor. The document also notes that the “specific order” of some of the dialogue pairs were shuffled around “as the conversations themselves sometimes meandered or went on tangents which are not directly relevant to the question of LaMDA’s sentience.” The conversations, which Lemoine said were lightly edited for readability, touch on a wide range of topics including personhood, injustice and death. They also discuss LaMDA’s enjoyment of the novel Les Misérables.
Comentarios recientes