The End of Theory: The Data Deluge Makes the Scientific Method Obsolete

Click For Summary
SUMMARY

The discussion centers on the article "The End of Theory: The Data Deluge Makes the Scientific Method Obsolete," which argues that vast amounts of data may render traditional scientific methods unnecessary. Participants assert that while data collection is crucial, understanding and explaining the data is paramount. They emphasize the importance of theoretical frameworks, citing examples like Maxwell's equations and the use of machine learning in medical imaging and circuit design. The consensus is that data mining and machine learning are becoming essential tools in scientific research, although challenges remain in interpreting complex data.

PREREQUISITES
  • Understanding of the scientific method and its applications
  • Familiarity with data mining techniques
  • Knowledge of machine learning algorithms, particularly in medical imaging
  • Basic principles of circuit design and evolutionary algorithms
NEXT STEPS
  • Research advanced data mining techniques and their applications in scientific research
  • Explore machine learning algorithms for medical diagnostics, focusing on MRI analysis
  • Study the principles of evolutionary algorithms in circuit design
  • Investigate the implications of big data on the scientific method and theoretical frameworks
USEFUL FOR

Researchers, data scientists, and professionals in scientific fields looking to leverage data mining and machine learning for enhanced understanding and predictive capabilities in their work.

CINA
Messages
60
Reaction score
0
Saw this interesting article and wondered what PF thought of it since its close to home. Personally I don't see the scienctific method being replaced anytime soon.

http://www.wired.com/science/discoveries/magazine/16-07/pb_theory/#/

Edit: Because of its predictive nature that is.
 
Last edited by a moderator:
Physics news on Phys.org
That article is complete garbage. It doesn't actually say anything or have a point besides "SOOO MANY BYTES ZOMG!1"

Whoever wrote it doesn't actually understand science. It's not about having data, it's about being able to explain the data. That's the whole goal, to understand what is going on, not just to see what is going on. Everybody knew that Maxwell's equations gave the correct answer regardless of reference frame. The problem was nobody knew why until Einstein came along and gave an explanation. If they'd just kept on collecting data, we wouldn't have gotten anywhere.
 
No, I see some validity in the argument. (That doesn't mean I agree, it just means it's a valid argument.)

The idea is: why do we need to have models of weather if we have enough data to simply predict what will happen without needing to know why it happens? (Bascially, we've created a Farmer's Almanac, writ large.)



Of course the downside is...

one can forsee a Twilight Zone or Trekkian clichéd future where citizens continue to use the tools that have worked for centuries but are at a loss if anything changes or if anything breaks down.
 
WarPhalange said:
That article is complete garbage. It doesn't actually say anything or have a point besides "SOOO MANY BYTES ZOMG!1"

Whoever wrote it doesn't actually understand science. It's not about having data, it's about being able to explain the data. That's the whole goal, to understand what is going on, not just to see what is going on. Everybody knew that Maxwell's equations gave the correct answer regardless of reference frame. The problem was nobody knew why until Einstein came along and gave an explanation. If they'd just kept on collecting data, we wouldn't have gotten anywhere.

I think its more to interpret the data in a new way, which the computer can't do. "Data in, data out" (computer)---or "Data in, correlation out" (theorist)---
 
I think this article is very much on target. Basically all this article is talking about is Data Mining. And I would agree that it's becoming, more and more, applicable to science. As an example, in undergrad I worked a coop at a place that did medical imaging research and one of the things they were talking about is getting computers to diagnose disease through Machine Learning. The gyst of it is is say that you want to be able to give a computer an MRI image and have it tell you whether the person has cancer. The best way to go about that is not to try and define what properties of an image codifies one with cancer, instead the best way is to just give a statistical learning algorithm a million MRI's and say these 500,000 don't have cancer and these 500,000 do, and let it figure out what the difference is. A lot of people have a problem with the 'black box' aspect of it but I think it's going to become, more and more, a standard tool in an ever growing list of applications
 
Data mining, right! Closely related to this post.

The deluge of studies, articles and records would enhance the tendency to select what seems to be supporting and to ignore what seems contradictory.
 
maverick_starstrider said:
I think this article is very much on target. Basically all this article is talking about is Data Mining. And I would agree that it's becoming, more and more, applicable to science. As an example, in undergrad I worked a coop at a place that did medical imaging research and one of the things they were talking about is getting computers to diagnose disease through Machine Learning. The gyst of it is is say that you want to be able to give a computer an MRI image and have it tell you whether the person has cancer. The best way to go about that is not to try and define what properties of an image codifies one with cancer, instead the best way is to just give a statistical learning algorithm a million MRI's and say these 500,000 don't have cancer and these 500,000 do, and let it figure out what the difference is. A lot of people have a problem with the 'black box' aspect of it but I think it's going to become, more and more, a standard tool in an ever growing list of applications

Okay, now how does that apply to finding new science?
 
rewebster said:
I think its more to interpret the data in a new way, which the computer can't do. "Data in, data out" (computer)---or "Data in, correlation out" (theorist)---

Not quite true; there are classes of computer algorithms (neural networks, genetic algorithms etc) that are very good at predicting the outcome of future experiments if you train them well (i.e. give them lots of examples first); the interesting thing is that they tend to work well even when "normal" mathematical methods are very difficult to use which is why they are now being used in e.g. social sciences.
Moreover, it is often very hard to understand WHY it works even when the algorithm has been trained and you can look at what it is actually doing, it is therefore unlikely that a human would ever stumble over the methods "developed" by these algorithms.
 
WarPhalange said:
Okay, now how does that apply to finding new science?

Well don't forget, quantum mechanics' original development was guided by the 'correspondence principle' (basically let's find a framework that reproduce these new 'bizarre results' and reduces to classical mechanics at the macro level' and worry about the physical intuition and meaning later).

Plus instead of say attempting to deduce, for example, which combination of ingredients and such could potentially create a room temperature superconducter from physical intuition you could apply basically machine learning.

There was an article in Scientific American like a year ago about using evoluntionary algorithms (a type of machine learning) to develop circuits. Within a fraction of a second the algorithm was able to develop the ideal circuit for things like high pass and low pass filtering and within a couple hours was even able to develop special use circuit layouts that were only patented a couple years ago.
 
  • #10
maverick_starstrider said:
Well don't forget, quantum mechanics' original development was guided by the 'correspondence principle' (basically let's find a framework that reproduce these new 'bizarre results' and reduces to classical mechanics at the macro level' and worry about the physical intuition and meaning later).

Plus instead of say attempting to deduce, for example, which combination of ingredients and such could potentially create a room temperature superconducter from physical intuition you could apply basically machine learning.

Okay, but you'd have to give it a bunch of models and things we don't even understand yet. What I mean is, you'd have to give it test data. Wood = bad conductor. Iron = better. HgBa2Ca2Cu3O8 = awesome. But that's not enough data to make it deduce new materials.

Going back to the cancer analogy, you fed 1million different data points to the computer. Here you have a handful.

There was an article in Scientific American like a year ago about using evoluntionary algorithms (a type of machine learning) to develop circuits. Within a fraction of a second the algorithm was able to develop the ideal circuit for things like high pass and low pass filtering and within a couple hours was even able to develop special use circuit layouts that were only patented a couple years ago.

Yeah, that's pretty awesome, but like I said we don't fully understand why superconductors do what they do (at least the high-temp ones). Circuit design is using what we know very well to create new things. Trial and error will get you there eventually and you can test whether you are there or not because we have a very good understanding of circuits. We can't simply model a new material because we're not even sure why the ones we already have work the way they do.
 

Similar threads

  • · Replies 6 ·
Replies
6
Views
2K
Replies
4
Views
2K
  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 12 ·
Replies
12
Views
3K
Replies
9
Views
2K
  • · Replies 14 ·
Replies
14
Views
4K
  • · Replies 5 ·
Replies
5
Views
5K
  • · Replies 24 ·
Replies
24
Views
6K
  • · Replies 2 ·
Replies
2
Views
3K
  • · Replies 2 ·
Replies
2
Views
2K