Bias, errors, etc. within ChatGPT & other AI chatbots

  • Thread starter artis
  • Start date
  • #1
artis
1,374
905
I thought maybe we should have a thread about these software as there are probably interesting artifacts about them and they are just interesting in general

So here is an interesting fact I just found out while messing randomly with chat GPT.

I asked it a simple question - "5 best jokes about communism"
It gave me this answer
communism.png


The I did the same and asked "5 best jokes about capitalism

capitalism.png


Now maybe I got it to joke about capitalism because I wrote it wrong - "capytalism"
But then I wrote once more with the correct writing and still got jokes from GPT

capitalism take 2.png



ChatGPT more like ComradeGPT...:biggrin:

Apparently it's trained on internet data and I assume there is a bias within the internet data ( obviously, because its made by humans) and ChatGPT simply sees it that for whatever reason, unknow to it of course, the word communism is put within the "untouchable" subject category much like racism, etc but capitalism is not therefore it freely talks about it.
 
Last edited:
  • Informative
  • Like
Likes ChemAir, Wrichik Basu, jack action and 4 others

Answers and Replies

  • #3
russ_watters
Mentor
22,135
9,285
You probably should have asked it for jokes about Democracy.

I haven't played with it yet - does it have any analytical capability or is it really just an eloquent search engine? Could you ask it if it sees a contradiction between its answers?

Still, I could foresee an attempt at a more sophisticated AI based on the premise "How would a human answer this/behave here?" I could see someone instructing a coffee robot to "surprise me". Robot thinks for a minute (googles), then turns and heads for the janitor's closet...
 
  • #4
Borg
Science Advisor
Gold Member
2,077
3,518
Some of the anecdotal stories that I've read on PF and other places leads me to believe that Chat GPT is doing something akin to Google's Wide & Deep model where a model has a standard deep learning aspect along with a wide component that allows it to learn over time.

In Chat GPT's case, it obviously has a large 300 billion+ neural network but also seems to learn during a conversation so that a person is able to convince it of different beliefs. This seems probable to me since you can tell it that it's wrong about something and it will adjust accordingly. However, those 'beliefs' don't appear to carry over from one conversation to the next - I asked Chat GPT if it could see information in another chat on my account and it couldn't.

This approach would have several benefits that I can see:
  • The deep portion provides very good base model that starts each conversation.
  • The wide aspect allows the model to adjust itself to the user's responses (for good or bad). This might explain some of the odd conversations that have been posted.
  • The adjustments during a conversation aren't carried over to other conversations and avoids the trolling issues like Microsoft's Tam suffered from. The Open AI team controls the updates that get into the base model.
 
  • #5
artis
1,374
905
You probably should have asked it for jokes about Democracy.

I haven't played with it yet - does it have any analytical capability or is it really just an eloquent search engine? Could you ask it if it sees a contradiction between its answers?

Still, I could foresee an attempt at a more sophisticated AI based on the premise "How would a human answer this/behave here?" I could see someone instructing a coffee robot to "surprise me". Robot thinks for a minute (googles), then turns and heads for the janitor's closet...
democracy.png

The way it currently seems to me is that its more like a search engine on steroids.
Somewhat like a talkative google.
But to be honest I think in some ways it's lacking because in the traditional search you can chose from all the answers that are displayed or at least all of the ones you have the patience to go through while in bots like ChatGPT it gives you the right answer like a pill on a silver platter and you just have to take it for what it is.
It has already been proven that more often than one would like it gives just an outright wrong answer but in many more cases it gives and answer that has a bias in it but is presented as some official fact.


But then again it's what you'd expect from someone who has read only whats online.
 
  • Like
Likes russ_watters
  • #7
14,291
8,331
One fact I learned was that its a static LLM created in 2021.

At the start of each new conversation, it must load in the static model to initialize things and then during the conversation you can train it further.

This prevents it from going off the rails when learning from many people intent on forcing it to fail.

Another aspect of starting with a default LLM is that you can test it and tweak it as needed. It reminds me of using the seed value for random number generation during testing so that you can replay the code to see why it did what it did. In production mode you'd allow the seed value to vary with time.

One way to test this is to start two conversations on separate machines at the same time and then enter the same queries to each and see if the generated output is the same. You could probably do this on the same machine by starting a new chat session but I believe they may be updating the LLM daily so running parallel seems like a safer bet to test the static nature of the initial load of the LLM. They may have some random seed that comes into play with the static LLM to vary things even in this case though.
 
  • #8
rcgldr
Homework Helper
8,806
590
One that may be fixed now: Why are cows eggs bigger than chicken eggs?
 
  • #9
TeethWhitener
Science Advisor
Gold Member
2,487
2,036
The way it currently seems to me is that its more like a search engine on steroids.
More like autocomplete on steroids.
 
  • #10
TeethWhitener
Science Advisor
Gold Member
2,487
2,036
I wish I had saved the exact conversation I had with ChatGPT, but it made it abundantly clear to me where the gaps are. Background: in the world of metal-catalyzed CO2 reduction, different metals give different products. I was putting together a presentation and I was admittedly too lazy to look up which metals gave which products, so I figured I'd save time and ask ChatGPT.

My prompt was something along the lines of "In CO2 electroreduction, which metals are formate formers, which ones are H2 formers, and which ones are hydrocarbon formers?" ChatGPT's answer was something along the lines of "CO2 electroreduction is <basically the first paragraph of a wikipedia entry>. Formate production is catalyzed by copper, zinc, and palladium. Some metals, such as copper, zinc and palladium produce mainly H2, whereas hydrocarbons are mainly produced by copper, zinc, and palladium." Once I corrected it (because I knew formates were mainly products of p-block metals, and also because its answer was silly and clearly wouldn't pass the Turing test), it said something along the lines of "I'm sorry for the mistake. You are correct, in addition to copper, zinc, and palladium, formate production is also catalyzed by p-block metals."

There are some impressive use cases for ChatGPT (tbh, for me with no talent whatsoever in visual art or graphic design, the text to image AI's have been far more useful for filling presentations and proposals with slick graphics), but a superintelligent evil human-species-destroying AI is at least a few more years away. I saw a great quote that said "AI won't take your job; people who know how to use AI will." Pretty much encapsulates how I feel about it. Now, if I could just train an AI to turn an abstract into a quad chart...
 
  • #11
14,291
8,331
And then the AI takes over those people to get past any captchas that may block it from dominating the world.

There was a scary footnote in the GPT-4 paper (pg 58) about looping the model so it can improve itself no need for humans.
 
  • #12
JLowe
83
47
One fact I learned was that its a static LLM created in 2021.
I tried my best to convince it I was a time traveler from two years into its future, but it refused to believe me. It informed me that it could not verify my claims and that I am not a credible source of information.
 
  • Haha
Likes jack action
  • #13
Borg
Science Advisor
Gold Member
2,077
3,518
It wouldn't give me the design for a flux capacitor either.
 

Suggested for: Bias, errors, etc. within ChatGPT & other AI chatbots

  • Last Post
Replies
2
Views
298
Replies
3
Views
395
Replies
15
Views
876
  • Last Post
9
Replies
284
Views
8K
  • Last Post
3
Replies
99
Views
3K
  • Last Post
Replies
8
Views
422
Replies
2
Views
93
Replies
11
Views
669
  • Last Post
Replies
3
Views
601
Top