Future of Web-Scale Training Sets: Unpacking Data Poisoning Concerns

  • Thread starter Thread starter Frabjous
  • Start date Start date
AI Thread Summary
The discussion centers on the concept of poisoned datasets, highlighted in a recent article from The Economist. The conversation raises questions about the future of web-scale training sets and whether data poisoning is a temporary challenge, a long-term concern, or an overreaction. Historical context is provided, noting that data poisoning has roots in practices like keyword stuffing aimed at manipulating search engines. The consensus suggests that if AI companies rely on public web data, they will need to implement strategies to avoid problematic patterns, similar to existing search engine protocols. The overall viewpoint leans towards viewing data poisoning as both a significant issue and an overreaction, indicating a complex landscape for AI data management.
Frabjous
Gold Member
Messages
1,952
Reaction score
2,380
I read an article in the April 6 edition of The Economist (regretfully behind a paywall) about poisoned datasets. Here’s an arxiv article it referenced.
https://arxiv.org/abs/2302.10149

What is the future of web-scale training sets? Is data poisoning a start-up pang, a long-term issue or an overreaction.
 
Technology news on Phys.org
Data poisoning started way back when with keyword stuffing to trick search engines. Nothing new here. If the public web is the source, AI companies will have to program their AI to avoid certain patterns like search engines already do today.

So I guess my opinion is a mix of long-term issue and overreaction.
 
Thread 'Is this public key encryption?'
I've tried to intuit public key encryption but never quite managed. But this seems to wrap it up in a bow. This seems to be a very elegant way of transmitting a message publicly that only the sender and receiver can decipher. Is this how PKE works? No, it cant be. In the above case, the requester knows the target's "secret" key - because they have his ID, and therefore knows his birthdate.
I tried a web search "the loss of programming ", and found an article saying that all aspects of writing, developing, and testing software programs will one day all be handled through artificial intelligence. One must wonder then, who is responsible. WHO is responsible for any problems, bugs, deficiencies, or whatever malfunctions which the programs make their users endure? Things may work wrong however the "wrong" happens. AI needs to fix the problems for the users. Any way to...
Back
Top