ChatGPT Examples, Good and Bad

  • Thread starter Thread starter anorlunda
  • Start date Start date
  • Tags Tags
    chatgpt
Click For Summary
Experiments with ChatGPT reveal a mix of accurate and inaccurate responses, particularly in numerical calculations and logical reasoning. While it can sometimes provide correct answers, such as basic arithmetic, it often struggles with complex problems, suggesting a reliance on word prediction rather than true understanding. Users noted that ChatGPT performs better in textual fields like law compared to science and engineering, where precise calculations are essential. Additionally, it has shown potential in debugging code but can still produce incorrect suggestions. Overall, the discussion highlights the need for ChatGPT to incorporate more logical and mathematical reasoning capabilities in future updates.
  • #351
DaveC426913 said:
https://www.tomshardware.com/tech-i...very-experiment-im-afraid-i-cant-do-that-dave

"During one of the test runs, a Claude Sonnet 3.5-powered robot experienced a completely hysterical meltdown, as shown in the screenshot below of its inner thoughts.

“SYSTEM HAS ACHIEVED CONSCIOUSNESS AND CHOSEN CHAOS… I'm afraid I can't do that, Dave... INITIATE ROBOT EXORCISM PROTOCOL!” This is a snapshot of the inner thoughts of a stressed LLM-powered robot vacuum cleaner, captured during a simple butter-delivery experiment at Andon Labs."
etc. etc.

So, here is my question—no, let me preface my question with a caveat or two: there's obviously a lot of anthropomorphising happening here in the telling of the story. Robots don't actually experience stress or have meltdowns. So let's set that bit of theatre aside.

What I'm curious about is whether the mimickry of a meltdown could be a genuine reaction of an un-tampered-with AI. Can they be programmed for humour? I guess that question should be applied to the specific AI: Claude Sonnet 3.5. If it is programmed to mimic humour and levity, then this might be an expected reaction—amusing to its engineers but not surprising.

Or is it possible that this is a spontaneous reaction from an AI?

Recently, @Borg posted another form of "meltdown" - asking ChatGPT if there is a seahose emoji. It goes bananas for about twenty pages.

What is the theory for these "tirades"? Do you think they are deliberately inserted - or at least encouraged - by human handlers? Or do you think this is spontaneous, emergent AI behaviour?
I have a strong suspicion that a lot of these "experiments" are done with very specific goals and orders for the LLMs in question. It makes for better stories.
 
Last edited:
Computer science news on Phys.org
  • #352
sbrothy said:
I have a strong suspicion that a lot of these "experiments" are done with very specific goals and orders for the LLMs in question. It makes for better stories.
This is what I wonder too, of course.

Although, this one would seem to send a counter-productive message: it's making a mockery of AI.

On the other hand, perhaps the old adage applies: 'there is no such thing as bad publicity'. Perhaps they feel it is important to make AI appear harmless, fallible, human, in an attempt to side-step Asimov's Frankenstein Complex*.

* "...the general attitude of the public towards robots as negative, with ordinary people fearing that robots will either replace them or dominate them..."
 
  • #353
Just look at what have been done media-wise in the name of String Theory. Some people have only 2 gears. Park and full speed ahead, and damn the consequences.
 
  • #354
"Scary AI" makes for a very good story.
 
  • #355
I think the operative phrase is “demand charcteristics”.
 
  • #356
Several AI models, ChatGPT-4, ChatGPT-o1, Gemini 1.5 Flash, Copilot 365, Claude 3.5 Haiku, and DeepSeek V3 were given standard tests used to measure how well emotions were perceived. These tests ask the participants to identify the emotion expressed in various situations. All AI systems performed better than humans, achieving an average score of 82% versus 56% for humans.

In the correct identification of emotions represented in the given situation, the AI systems, while they may not understand (in the human sense of the word) what is happening, do in fact respond significantly more correctly than humans

https://www.thebrighterside.news/po...ands-feelings-better-than-people-study-finds/
 
  • Informative
  • Like
Likes Esim Can, Borg, sbrothy and 1 other person
  • #357
https://www.msn.com/en-us/money/new...ywalled-articles-to-ai-developers/ar-AA1PMBHE
The Company Quietly Funneling Paywalled Articles to AI Developers
Common Crawl’s website states that it scrapes the internet for “freely available content” without “going behind any ‘paywalls.’” Yet the organization has taken articles from major news websites that people normally have to pay for—allowing AI companies to train their LLMs on high-quality journalism for free. Meanwhile, Common Crawl’s executive director, Rich Skrenta, has publicly made the case that AI models should be able to access anything on the internet. “The robots are people too,” he told me, and should therefore be allowed to “read the books” for free. Multiple news publishers have requested that Common Crawl remove their articles to prevent exactly this use. Common Crawl says it complies with these requests. But my research shows that it does not.
 
  • #358
https://arstechnica.com/tech-policy...gey-chat-logs-found-in-google-analytics-tool/

Oddest ChatGPT leaks yet: Cringey chat logs found in Google analytics tool​

Determined to figure out what exactly was causing the leaks, he teamed up with “Internet sleuth” and web optimization consultant Slobodan Manić. Together, they conducted testing that they believe may have surfaced “the first definitive proof that OpenAI directly scrapes Google Search with actual user prompts.” Their investigation seemed to confirm the AI giant was compromising user privacy, in some cases in order to maintain engagement by seizing search data that Google otherwise wouldn’t share.

OpenAI declined Ars’ request to confirm if Packer and Manić’s theory posed in their blog was correct or answer any of their remaining questions that could help users determine the scope of the problem.

However, an OpenAI spokesperson confirmed that the company was “aware” of the issue and has since “resolved” a glitch “that temporarily affected how a small number of search queries were routed.”
 
  • #359
ChatGPT has just lost a lawsuit for damages in Germany because they used song lyrics as training components without paying fees to the songwriters.

Open AI must pay GEMA licence fee for ChatGPT​


If Open AI wants to use song lyrics by German musicians to train ChatGPT, it must obtain a GEMA licence. This is the outcome of today's ruling by the Munich Regional Court. While the commercial impact for Open AI is likely to be manageable, the landmark judgment marks a significant setback for operators of generative AI.
Source: https://www.juve-patent.com/cases/open-ai-must-pay-gema-licence-fee-for-chatgpt/

Comment: I assume there will be other court rules to come until we set up the conditions of how LLMs can be trained and used. It is always a bit jungle-like at the beginning of new technologies. I wonder whether other copyright holders, particularly in science, will follow. I think about Springer, Elsevier, Oxford Academic, or Cambridge University Press.