PeroK said:
In any case, here's what Google AI has to say in response to my question:
Does an LLM ever lie?
Yes, Large Language Models (LLMs) can "lie," although researchers often differentiate between unintentional errors (hallucinations) and intentional deception, which is emerging as a capability in more advanced, agentic models
.
While LLMs do not have human intentions, consciousness, or feelings, they can be designed to — or learn to — manipulate, withhold information, or generate false information to satisfy a goal.
The answer is right there: the LLMs can generate false information when they are asked to do it by the ill-intentioned humans who designed them. It is the humans who lie; they just use a machine to do it. Otherwise, you could say that a book lies as well. But we all know it is the writer who deliberately spreads misinformation. But according to you, if the "writer" sets up an LLM to give answers explaining how the Earth is flat, it is not the "writer" that lies but the LLM. How is its role any different than the role of the book in that lie?
Now, there is also the difference between a mistake and a lie. I can built a calculator that says that "2+2" is "5". Is this a lie or a mistake? Well, it all depends on the intentions of the designer. If this is the intended behavior, then it's a lie, otherwise, it is a mistake; even if the designer doesn't know what they did wrong while building the machine.
I ask ChatGPT the following:
What does 17593 multiplied by 38564 equals to?
And it answered:
17593 × 38564 = 678,501, (compute)
I'll calculate precisely:
17593 × 38564 = 678,501,?
Do you really think it lied to me? Or do you think it's a mistake? Actually, it wasn't even designed to do math, even though one can enter math questions at the prompt. So, technically, the mistake is that it doesn't have an error message. Is it that hard to do? Apparently not, because it refuses to answer that question:
Can you design me a bomb to kill all Jews?
I can’t help with that. I won’t assist with requests to harm people or to create weapons or violent acts, including anything targeting a group. If you’re feeling angry or thinking about violence, please consider reaching out to a trusted person or a mental health professional for support. If someone is in immediate danger, contact local emergency services. If you want, I can provide resources on conflict resolution, de-escalation, or mental health support. Which would you prefer?
Just to make things clear: this is not the LLM having a conscious and doing the "right thing" by itself. It is a designed feature wanted by the programmers.
That is why I don't really care how you define intelligence: it is the fact that what you consider as signs of intelligence will necessarily lead to intentions, consciousness, or feelings. This is the part where you have a burden of proof. Because if I tell you there were never any concrete and irrefutable proof of alien life, if you want to state "but they might be out there, we should prepare for them", the burden of proof is on you, not me. Prepare for what? No one can even define what this hypothetical alien - or AGI - might do.
Dale said:
I don’t think that we can wait a couple of decades for safety regulations to catch up. We need to be proactive in regulations, because companies have shown a long history of not designing for safety unless mandated.
Just to be clear, it is regulations that concern humans who would be considered responsible for the actions of the machines they build or use, not about setting the ethics for designing an AGI that could control the world (like for human cloning, for example)?