Einstein simultaneity: just a convention?

Click For Summary
The discussion centers on the nature of Einstein's simultaneity and its implications for the Lorentz transformation, debating whether it is an arbitrary convention or a fundamental aspect of time and motion. Participants express differing views on whether the isotropic speed of light is a law of nature or merely a convenient coordinate choice stemming from Einstein's framework. There is a concern about the pedagogical approach to teaching special relativity, particularly regarding the historical context of the ether and the implications of simultaneity. Some argue for a more refined teaching method that emphasizes the mathematical and physical significance of Einstein's synchronization convention. Ultimately, the conversation highlights the complexities in understanding and teaching the foundational principles of relativity.
  • #121
Ken G said:
Which does it need to do, predict within the well tested regime, or beyond it? If I have a theory that predicted in the tested regime, then I can build bridges with it.

A theory cannot set arbitrary limits on itself. That can only be done when it is understood as an approximation to another more general theory. Consider my example of the Fibonacci series. We notice a pattern that seems to hold for all the terms we have looked at. So the formula a_n = a_{n-1} + a_{n-2} will be our theory which unifies all the observations we have made. We might have looked at the terms from 200 to 300. Then we would expect that it will hold for the term 301, because that's what our information is suggesting to us. By the same logic we will expect it to hold for the terms 5000, 5001 etc too. We have no information that tells us that it is less likely to hold in the ~5000 regime than in the ~400 regime.

If we've tested the theory in the 10s and the 30s, then we believe it will hold in the 20s as strongly as we believe that it will hold in the 500s. That's because we have no information that distinguishes them. There are various things that will make the 20s more likely than the 500s. For example, we may believe from other experience that theories only become inaccurate gradually. We may believe in continuity in some general sense, but it is important to realize that that is extra information, which, if used, is part of your theory.

Ken G said:
Exactly why do I need to be able to extend my predictions beyond that regime? I would call that nothing but the formation of a hypothesis, and we don't need a theory to do that, we can try anything we like.

Let's say that what we expect has consequences. Maybe we want to send some machine to another galaxy. Then we will construct it according to what our theories tell us, although we're not sure if everything will work the way we expect. Just because we don't know doesn't mean that trying some arbitrary hypothesis is as good as going with the theories we currently have.


Ken G said:
But that's the definition of the Fibonacci series. It's only a theory if you don't know where that series comes from (this is the problem we have in physics).

Thats what I meant. I was considering a situation where the terms of the Fibonacci series are outcomes of an experiment.

Ken G said:
If the series is coming from observations of some kind where you are not just getting out what you put in, then you have no reason to expect the form will continue indefinitely. If it worked for a million terms, it seems probable it will work for many thousands more, because why should they be special, but what about a billion more?

Whats different about a thousand more and a billion more? There's no difference. Both are outside the tested regime and you have no information about them. If your expectations had consequences then the best thing is to stick to the pattern you noticed, i.e. the theory. Again, like I said before you may have some general ideas about how things become inaccurate only gradually, but that idea would also be part of your theory.

Ken G said:
It is to avoid fooling oneself. Feynman has a great quote that science is about learning how to not fool yourself, given that you are the easiest person you can fool. I see science falling into that same trap, it is not taking its own principles far enough if it keeps causing us to fool ourselves into false expectations, and then having "revolutions" later on.
I would say that is exactly the way science should be done, yes. Granted, it is no simple matter to define what is meant by an "untested regime", but for now we'll simply allow there is such a concept even if we can't be terribly precise about what it is.

I think what you're trying to say is that we should be aware that the concepts and theories we use are not absolute truths, and also be aware of where they have been tested and where they have not been etc. Is that right?

Ken G said:
I claim your analysis is using the symmetry of the coin, and that's why it seems "arbitrary" to do anything else. But if you know the coin has a symmetry, you are indeed using knowledge of the coin. Write that same argument but for a conical hat.

No, my analysis did not use the symmetry of the coin because the relevance of the symmetry is not known to me. For that I would have needed to know the details of Newtonian mechanics, gravity, center of mass etc. That is all information that I did not have. All I knew was that there were two possible outcomes, heads and tails. With no other information, I would have to expect both equally.

Ken G said:
True, but that would not lead me to expect a 50-50 chance, it would lead me to simply say I have no meaningful way to assess the probability. Probability requires a great deal of knowledge about what variables are outside your control-- if you don't even know that, it is a meaningless concept.

I was talking about expectation, not probability. I don't know if you think probability is something absolute about the system or not, but we don't need to go into that. There is always a meaningful way to form expectation. The way is this - "take all the information you have into account, and nothing else".

Ken G said:
I don't agree, all experiments can have two possible outcomes-- a particular one, and anything else. Shall we start with the assumption, then, that any outcome you can name has a 50-50 chance of happening, on the grounds that we have no other information about the probabilities of "all other outcomes"? We always have to group outcomes, there's no absolute sense of "the possible outcomes of an experiment".

The fact that you can group all possible outcomes into "this" and "everything else" has nothing to do with what you should logically expect. If that's all the information you had, then yes, you should assign an equal expectation to both. But, if you knew that there were six possible outcomes, then you cannot ignore that information. If I told you that there were some number from 1 to 10 elephants in the cage, then you would assign equal expectation to each number from 1 to 10 because you don't have any other information. You cannot group it into 1 and {2,..,10}, and then assign equal expectation to those sets because that would mean that you're ignoring some of the information that you were given. The best possible expectation given a certain amount of information is the expectation that takes all information into account, and nothing else.

Ken G said:
Even if you are flipping a coin, there is the location of every other particle involved in that experiment. You can say you don't care about them, so you are grouping outcomes.

It's not that I don't care about them, its that I don't know about them. If I did know about them, then my expectation will take them into account. I think your familiarity with coins and the immense experience you have that could be relevant is preventing you from thinking clearly. Remember, all I know is that there are two possibilities. That information may not be true. But given the information, the logical expectation is to assign equal expectation to both.

Ken G said:
If the "other planet" is a neutron star, we get a breakdown, and if it's like the planets we built the theory for, we don't.

Which has nothing to do with what you should expect. When you say you expect something you are not saying that it is true. If you didn't know the relevant difference between a neutron star and a normal planet, then you should expect the same for both. If you had to send some kind of machine to do experiments to a neutron star and a planet, and if you didn't know what the difference was, you would build it according to your current expectations (which are based on your current knowledge). Any random hypothesis would not be just as good, because it would be ignoring information that is available to you.
 
Physics news on Phys.org
  • #122
Hurkyl said:
Actually, it's not that simple; the problem of a priori probabilities is a significant philosophical issue -- this is a common and often useful convention, but it's far from clear that it would be the "most logical expectation"

In this case it is that simple. This is all the information you have :

1. The experiment has two possible outcomes.

Let's leave aside the question of whether you should expect anything at all. If your life depended on it, then the best thing is to expect them equally. Any other expectation would not be justified logically. There is no way to logically assign unequal expectations to the two outcomes without using some information that is not contained in (1).

Hurkyl said:
And it's not strictly necessary anyways -- one can view the scientific process as merely determining which theories have the stronger Bayes factors, rather than trying to determine which theories are most probable. i.e. science seeks to accumulate evidence, not to uncover truth

Science seeks to unify experience, and to answer meaningful questions about things one has not experienced yet. The second depends on the first.

Say a theory A unifies a certain domain of your experience. Then that theory will have something to say about questions about that domain which can be meaningfully asked within the structure of that theory. Given that you believe in the theory, you should logically "expect" what it tells you. You may not believe that the theory applies in that case. And if you have any reasons for that belief, then it is also part of your theory. Your expectations are ideally products of the sum of your knowledge.

I don't know what you mean by "which theories are most probable". If you mean which theories have the most evidence, then yes that is exactly what we are trying to find. We are interested in the ideas that have the most support.
 
  • #123
dx said:
If we've tested the theory in the 10s and the 30s, then we believe it will hold in the 20s as strongly as we believe that it will hold in the 500s.
Why would we do that? It simply isn't logical. If you have some information about the series that tells you the 500s are no different from the 10s, then you have something more than a physical theory actually gets to have. That's why a mathematical series is not actually a good example of what we are talking about-- in math, you only get out what you put in, though the trick is to figure out everything you put in without realizing it.
Maybe we want to send some machine to another galaxy. Then we will construct it according to what our theories tell us, although we're not sure if everything will work the way we expect. Just because we don't know doesn't mean that trying some arbitrary hypothesis is as good as going with the theories we currently have.
I agree that an "arbitrary" hypothesis would not be a good guide-- our theories do give us a guide for making hypotheses that are likely to be of use. That's quite a bit different from the expectation that our theories will work, however. In your example, our "expectation" would be that our theories will provide no more than a basis to conduct an experiment. What we expect is to be surprised, and to require modifications. But the key point is, we will try to apply theories that are actually within their regime of applicability, or we are just kidding ourselves.

Take the Wright brothers. Yes they did rely on theory, and yes they did have to do a lot of trial-and-error modifications of that theory, but the theory they were using was a theory based on fluid mechanics quite similar to air passing over a wing. They were not using equations about high-viscosity fluids and applying them to air flows, for example, simply for lack of anything better. Had they been doing the latter, their "expectations" wouldn't be worth a plugged nickel. They had to know something experimental about air to make the expectations useful.

Thats what I meant. I was considering a situation where the terms of the Fibonacci series are outcomes of an experiment.
In any such situation, knowledge of the first 30 numbers affords very little information about the 500th. We could do the same thing in the Earth's atmosphere, say. We could do an experiment where we travel upwards by 100 feet, measure the density, write it down, then go up another 100 feet. If we know nothing about the atmosphere from other experiences with it, then we'd have nothing to go on expect a very obvious if approximate pattern that would govern about the first 50 numbers in that sequence. By your logic, it is then natural to expect it to apply also to the next 500 as well. Of course it will actually break down completely in this example, but my point is that this is what we should actually expect-- the example is typical.

Similarly, you could write down the energies of the first 30 levels in hydrogen, and see the pattern. Will that apply to the 500th? Probably not, you'd need very low densities and weak fields. Again, breakdown of the pattern is the wiser expectation in general.

Whats different about a thousand more and a billion more? There's no difference.
See above.

I think what you're trying to say is that we should be aware that the concepts and theories we use are not absolute truths, and also be aware of where they have been tested and where they have not been etc. Is that right?
Yes, the main idea would be to equip every theory with a sense of what regime it has been tested in, by experiment. That way, no theory would ever be considered independent from the experimental data it unifies. Just think of how much wild and useless speculation over the years, and even today, we could avoid with that approach. No "determinism", no "landscape", no "quantum suicide"-- just the experiments and the theories that explain them.
No, my analysis did not use the symmetry of the coin because the relevance of the symmetry is not known to me.
Then your analysis can be applied to a conical hat as well as to a coin. Is it any good for a conical hat?

For that I would have needed to know the details of Newtonian mechanics, gravity, center of mass etc. That is all information that I did not have.
No, you'd just need a fairly straightforward understanding of symmetry.
The fact that you can group all possible outcomes into "this" and "everything else" has nothing to do with what you should logically expect.
No, we'll always have to make such groupings. You may think "heads" plus the location of every person in the room is an obviously different grouping than tails and such, but that's just your grouping. I can just as easily group heads and me in any position but one with all the tails, and leave the other possibility as heads and me in one position. Why is that a "wrong" way to divide the outcomes? Your logic applies to that grouping as well as the other-- why would it not?
 
  • #124
Ken G said:
Why would we do that? It simply isn't logical. If you have some information about the series that tells you the 500s are no different from the 10s, then you have something more than a physical theory actually gets to have.

No, we don't have any information that tells us that the 500s are different from the 10s. That's the point. The only information you have is what you have observed. And that information is not telling us that the 500s are different from the 10s. We're considering an idealized experiment where the observations in that experiment is all we know. The question is, what can we expect (and to what degree) about further observations (based only on that information)?

To make it concrete,

1. You start out with no knowledge at all (about anything whatsoever).
2. You make a series of observations (physical observations, not mathematical. the Fibonacci series is ubiquitous in nature, and many conceivable observations could result in it.)

1 -> 1
2 -> 1
3 -> 2
4 -> 3
etc. say up to a million.

Then there's a meaningful question that we may ask, i.e. million + 1 -> ?

The pattern we noticed could be extended to make a prediction of what it would be. Now, the question is, to what degree do we believe that prediction? "I don't care" is not an option. We must choose on a scale from "I don't believe in it at all" to "I believe it completely". What do you think? (the only information we have is the set of observations up to a million. Physical observations, not mathematical!)


Ken G said:
I agree that an "arbitrary" hypothesis would not be a good guide-- our theories do give us a guide for making hypotheses that are likely to be of use. That's quite a bit different from the expectation that our theories will work, however.

You seem to still be confusing "expecting our theory will work" with "I know the theory will work". That is not the sense in which I'm using the word expect. Expectation can have degree. You can expect strongly, weakly, completely, not at all, or anything in between.

Ken G said:
In your example, our "expectation" would be that our theories will provide no more than a basis to conduct an experiment. What we expect is to be surprised, and to require modifications. But the key point is, we will try to apply theories that are actually within their regime of applicability, or we are just kidding ourselves.

You should expect to be surprised in all cases where you don't have perfect knowledge, i.e. all the time. But that does not mean you can't expect certain things to certain degrees. They are complementary. If your expectation that the plane will fly on a scale of 0-1 is 0.9, then your expectation that you will be surprised is 0.1.

Ken G said:
They were not using equations about high-viscosity fluids and applying them to air flows, for example, simply for lack of anything better. Had they been doing the latter, their "expectations" wouldn't be worth a plugged nickel. They had to know something experimental about air to make the expectations useful.

But to know that your expectations are not useful, you have to try them out first. When you're creating a machine to send into some unknown regime, you don't have any information about that regime. This is the situation where what you expect becomes crucial. You may not have the chance to do it again, so what should you do based on what you already know?


Ken G said:
By your logic, it is then natural to expect it to apply also to the next 500 as well. Of course it will actually break down completely in this example, but my point is that this is what we should actually expect-- the example is typical.

Yes, without further information that's what you should expect.


Ken G said:
Then your analysis can be applied to a conical hat as well as to a coin. Is it any good for a conical hat?

Again, you're confusing reality with information. Whether the analysis is any good in the sense of whether it is an accurate picture or not is not the question. You have no way of knowing that other than obtaining more information. The most logical expectation does not necessarily match reality. The point is to make use of all information you have. There are even cases where a logical analysis of the information you have will lead you to expect something that is completely false. But it is still the best you can do with that information.

Ken G said:
No, you'd just need a fairly straightforward understanding of symmetry.

Which is not information I possess. I've said this many times. You can use only that information that you have. In my example, the only information you have is that there are two possible outcomes, and also the general shape of the coin. You don't understand symmetry, i.e. you don't know the relevance of the shape for the experiment you are performing. You don't even know what "tossing" means. That's why the symmetry cannot tell you anything at all about the outcomes.

Ken G said:
I can just as easily group heads and me in any position but one with all the tails, and leave the other possibility as heads and me in one position. Why is that a "wrong" way to divide the outcomes? Your logic applies to that grouping as well as the other-- why would it not?

You can divide the outcomes in that way as long as you don't forget what each group consists of, since that would be throwing away information. Please read the following carefully.

1. You have a box with some balls in it.
2. There are 10 balls, each with a number from 1-10 on it.

This is the information you have. Now if you pick one without looking (i.e. you're not collecting any more information while you're picking), then you should assign the same expectation to each of the balls. You cannot group them into {1} and {2,...,10} and then assign equal expectation to these sets, because this grouping does not represent the same information you had before. You would have to forget some information, i.e. the fact that the second group has more than one ball in it.
 
  • #125
dx said:
Then there's a meaningful question that we may ask, i.e. million + 1 -> ?
But the issue we've been talking about is not the million+1 case, that's clearly a case where you have no reason to think you crossed a regime boundary. I'm saying that you should have no expectation at all about the billionth term in the series, unless you also have some reason to expect it will continue to follow the trend (as in the case of a mathematical example, which is why they are not appropriate to physics analogies). Consider instead the levels of a real hydrogen atom, for example.
We must choose on a scale from "I don't believe in it at all" to "I believe it completely". What do you think? (the only information we have is the set of observations up to a million. Physical observations, not mathematical!)
I already gave two examples of that, and why my expectation would be "I don't believe in it at all", but here we're talking about the billionth entry, not the million+1.

You can expect strongly, weakly, completely, not at all, or anything in between.
I'm fine with the degrees, I'm saying the appropriate degree is determined not by any element of the theory itself, but only by the degree to which the regime you are considering connects with regimes you have already experimented on. Gravity is a good example of this. We know pretty well how it works on scales of boulders and planets, but we already got one surprise extending it to the density of a neutron star, and another surprise extending it to the scale between galaxy clusters. We have no idea how it works on scales smaller than boulders, say for atoms, but my money says we should have "no confidence at all" that our current theory can treat the gravity in an atom-- should we ever be able to observe that.
Yes, without further information that's what you should expect.
My examples already show why that expectation would be baseless, as they are quite generic examples. You can add the third example of gravity, as above. Granted, gravity worked well going from a solar system to a galaxy cluster, and that's a significant increase over orders of magnitude. But the theory was built to unify observations where M/R never went above what you find in the solar system. If you exceed that by about 8 orders of magnitude, you come to the M/R for a black hole, or for the whole observable universe, both of which appear to give significant deviations from Newton's theories. We should have expected that, since M/R exceeds our experimental experience by 8 orders of magnitude.
Whether the analysis is any good in the sense of whether it is an accurate picture or not is not the question.
It certainly is the question, it's precisely the question.
The most logical expectation does not necessarily match reality.
Then on what basis do you claim it is the "most logical expectation"? What do you think that phrase means?

The point is to make use of all information you have. There are even cases where a logical analysis of the information you have will lead you to expect something that is completely false. But it is still the best you can do with that information.
Not if the best you can do with the information is recognize that it is insufficient to draw any conclusions whatsoever. That is the actual logical thing to do.
You can use only that information that you have.
Obviously, the point is, one way you can use that information is to say you cannot say anything useful.
In my example, the only information you have is that there are two possible outcomes, and also the general shape of the coin. You don't understand symmetry, i.e. you don't know the relevance of the shape for the experiment you are performing.
Then you cannot use shape information at all, and your argument has to apply just as well for a conical hat as for a coin. Does it?
You don't even know what "tossing" means. That's why the symmetry cannot tell you anything at all about the outcomes.
If that is true, you should have no expectation at all.
You can divide the outcomes in that way as long as you don't forget what each group consists of, since that would be throwing away information. Please read the following carefully.

1. You have a box with some balls in it.
2. There are 10 balls, each with a number from 1-10 on it.

This is the information you have. Now if you pick one without looking (i.e. you're not collecting any more information while you're picking), then you should assign the same expectation to each of the balls. You cannot group them into {1} and {2,...,10} and then assign equal expectation to these sets, because this grouping does not represent the same information you had before.
Sure it does, the grouping loses no information, it merely groups it. You have already done such a grouping, when you assumed I can distinguish a "1" from a "2" without knowing the orientation of the ball. You group all "1" results together, regardless of orientation. So how do you know there "really are" 10 possibilities here? You are the one who has imposed that on the experiment, the actual experiment will generate a virtual infinity of distinguishable outcomes. I'm saying that the way you do the grouping does not appear anywhere in your argument, so I am free to group all 1's with the 2's except for 1's in a single precise orientation when I remove it from the box, for example. What about your logic precludes that?

You would have to forget some information, i.e. the fact that the second group has more than one ball in it.
I don't need to forget that information, I am well aware of it. So what, what about your logic requires there be an equal number of balls in each group? To say that, you must be assuming that each ball is equally likely, then using that assumption to reason that we should expect each ball to be equally likely. That is precisely the argument you are giving.
 
  • #126
Ken G said:
But the issue we've been talking about is not the million+1 case, that's clearly a case where you have no reason to think you crossed a regime boundary. I'm saying that you should have no expectation at all about the billionth term.

Why the billionth? Why not the trillionth? Give me precisely the piece of information from your set of observations that tells you where exactly you start "not believing at all". How do you determine that boundary?

Ken G said:
It certainly is the question, it's precisely the question.

No, it's not. You will not know whether your expectation matches reality satisfactorily until you make an observation to test it. But question was, "what should you expect before you make the observation?". Once you make the observation, you will have more information, which you use to update your expectations. But before you make the observation, there is no way to tell if your expectation will match reality. That in no way prevents you from using the information you have to make a guess. Even when you're predicting within the regime, you're still guessing. It may be a well supported guess, but it's still a guess.
Ken G said:
Then on what basis do you claim it is the "most logical expectation"? What do you think that phrase means?

The phrase means the expectation that follows logically from the information you have.

Ken G said:
Not if the best you can do with the information is recognize that it is insufficient to draw any conclusions whatsoever. That is the actual logical thing to do.

We are not drawing conclusions! We are assigning degrees of belief to possibilities. If the information was sufficient to make a conclusive deduction, then you will "know", not "expect". We are considering what we should do when we don't have enough information to solve the problem deductively. We can realize that the information is not sufficient to tell for sure, but that doesn't mean we cannot use the information we have to make a guess. And the best guess would be the one that takes into account all the information we have.

Ken G said:
Obviously, the point is, one way you can use that information is to say you cannot say anything useful.Then you cannot use shape information at all, and your argument has to apply just as well for a conical hat as for a coin. Does it?

You don't know whether you can say anything useful until you do the experiment. The idea is to make the best of what you have. It turns out that in the case of the coin it's accurate, and in the case of a conical hat its not. But you cannot tell before you do the experiment.

Ken G said:
If that is true, you should have no expectation at all.

Why not? How do you know whether the information you have is enough for an accurate picture or not until you do the experiment? You cannot decide before the experiment that "I don't have enough information to say anything useful, so I won't say anything". You cannot know how accurate your picture of reality is before the experiment.

Ken G said:
Sure it does, the grouping loses no information, it merely groups it.

The grouping by itself doesn't lose information. But when you assign equal probability to the two groups you are ignoring the fact that you have some information about the difference between the two groups. Assume that I know that the coin will land heads or tails. Then I will expect them equally because I don't have information that tells me that I should expect one or the other more. But If you group the balls as {1} and {2,..,10}, then you do have information that allows you to distinguish between them, i.e. the fact that the second group has more than one possibility. If you assign equal probability to the groups now, you would be ignoring the fact that you have relevant information that distinguishes between the two groups.

Ken G said:
You have already done such a grouping, when you assumed I can distinguish a "1" from a "2" without knowing the orientation of the ball. You group all "1" results together, regardless of orientation. So how do you know there "really are" 10 possibilities here?

No, the orientation of the ball etc. are not information I have, nor am I collecting it. I am only concerned with the number of the ball. And I know that there are 10 balls. That is the information I have. I'm not imposing that on the experiment, I just know it. Maybe someone told me. I am concerned with using that information to form beliefs. Not belief as in "I know". But belief as in "I believe this to a certain degree, given this information".
Ken G said:
I don't need to forget that information, I am well aware of it. So what, what about your logic requires there be an equal number of balls in each group? To say that, you must be assuming that each ball is equally likely, then using that assumption to reason that we should expect each ball to be equally likely. That is precisely the argument you are giving.

No it's not. I don't need to group anything at all. I have been told that there are 10 possible types of balls, and nothing else. That is information that I may know from a previous experiment, or whatever. And I also know that when I perform a particular experiment, the possible outcomes are {1,2...,10}. The experiment does not measure the orientation of the balls. I only look at the number on the ball. So {1,2..,10} is the set of outcomes. There can be no other possible outcome that is not in this set. I am only looking at the number. What I look at is precisely the outcome of the experiment.
 
Last edited:
  • #127
dx said:
Why the billionth? Why not the trillionth? Give me precisely the piece of information from your set of observations that tells you where exactly you start "not believing at all". How do you determine that boundary?
It is a difficult boundary to determine, most likely we need a theory about that too. I'd draw on a combination of the dynamical range of previous successes and the precision of those successes to build expectation about the appropriate regime for sustaining precision going forward. I never said it was easy, but I did say that we have to do precisely this all the time. If you drive a car, for example, you know you could kill yourself, and you know the chances get worse at higher speed and in poorer conditions. Every time we drive, the conditions are a little different than they have ever been before. If it gets extremely foggy, those are very different conditions indeed-- so at what point do you judge the conditions are unsafe, on the grounds that your past safety record is not applicable in the new conditions? We are forced to make such determinations all the time, I am merely applying the same principle to physical theories.

But before you make the observation, there is no way to tell if your expectation will match reality. That in no way prevents you from using the information you have to make a guess.
Certainly, but what you are failing to do is to assess the validity of your guess. Some things really are pure guess, and have no more value than a guess. We need to recognize when that is the case, as it avoids the "revolution" problem. Every physics "revolution" was actually just a case of someone guessing wrong who had no business expecting to be right, pure and simple-- no revolution there when you see it in that light.

Even when you're predicting within the regime, you're still guessing. It may be a well supported guess, but it's still a guess.
Labelling everything a guess is of no value, calling something a guess means you have a low opinion of its likelihood of being correct. That's how most people use the term, and how I am using it too.
We can realize that the information is not sufficient to tell for sure, but that doesn't mean we cannot use the information we have to make a guess. And the best guess would be the one that takes into account all the information we have.
Yes, but I'm saying, what if the "best guess" really still ends up being just a guess? What good is it to identify it as the "best guess", if the best guess is still completely worthless? That's what I'm saying, I'm not disputing what the best guess is, only what the meaning of a best guess is. We always need to supplement the best guess with some concept of confidence in that guess, or it is a truly useless concept. No one is forcing us to guess at all-- the sensible approach is to test, not to guess. Ergo, we should set up tests that best narrow down the possibilities, completely irrespective of what any meaningless version of an "expectation" would say.

Let's return to the coin and/or the conical hat problem. I think I now see what you and granpa were saying. If there are two outcomes that you have chosen to distinguish, then your "best guess" is that if you pick one, you have a 50% chance of being right, if you have no other information. That's true, but that's very different from saying that the object is expected to demonstrate a 50/50 distribution of outcomes over many trials. The latter would not be the logical expectation, barring some other information (such as the symmetry of a coin versus a hat).

Furthermore, the 50% chance of being right only works if you make your choice at random. But if you follow some theory to arrive at your expectation, it is no longer a random choice, and therefore you can no longer say that you have a 50% chance of guessing correctly in the absence of any other information beyond the theory you used. I'm not sure why the whole Bayesian argument was brought up as relevant in the first place, but these are two important limitations to bear in mind whatever that purpose was.

Let me give another example to underscore this. Let's say there are three coins, two pennies and a quarter, in a jar, and someone is going to shake that jar until one coin comes out. You know nothing about the coins except that two are worth 1 cent and one is worth 25 cents. You are to receive the coin that shakes out-- how much money do you expect to receive, say over the course of 100 repeated trials? Now, if you pick randomly between "I get a penny" and "I get a quarter", you will be right half the time, and it makes no difference at all anything else about the experiment. If you can distinguish the pennies into penny #1 and penny #2, then you may instead decide to randomly choose between receiving each of the three coins, and now you'll be right 1/3 of the time, again independent of any other information you may have about the experiment. However, nothing that has been said so far may be used to form a meaningful expectation value for the amount you'll receive in 100 trials. There is simply no expectation that is not a pure guess, and logically you should not expect any such guess to converge on something correct over any number of trials. It simply doesn't mean a thing.
You don't know whether you can say anything useful until you do the experiment. The idea is to make the best of what you have.
But why is forming an expectation making the best of what you have? What possible benefit are you deriving from that? If the expectation is of no usefulness, then it being "best" among all the useless ways to form an expectation isn't saying much.
It turns out that in the case of the coin it's accurate, and in the case of a conical hat its not. But you cannot tell before you do the experiment.
Right, so it is an expectation that means nothing at all, there's no point in even forming it prior to the experiment.
Why not? How do you know whether the information you have is enough for an accurate picture or not until you do the experiment? You cannot decide before the experiment that "I don't have enough information to say anything useful, so I won't say anything". You cannot know how accurate your picture of reality is before the experiment.
Yet, as I pointed out in the driving in a fog example, we are called upon to do precisely that, all the time. We always need to have a concept of the confidence we can place in our expectations. What I'm saying is, we've been doing a particularly bad job of that in science when we treat all predictions on an equal footing, and that has led to all the "revolutions"-- and all the misunderstandings about the fallibility of science.

Assume that I know that the coin will land heads or tails. Then I will expect them equally because I don't have information that tells me that I should expect one or the other more.
That is not the logical expectation, you are asserting that we should expect a flat distribution after many trials. That will generally only be true in situations that exhibit a symmetry of some kind, and if you have no knowledge of any such symmetry, you should reject that expectation on the grounds that it is not generic. Instead, you should expect an unequal distribution after many trials, but you don't know which outcome will be the favored one, or by how much-- you simply have no expectation there at all. You still have a 50% chance of picking the right outcome if you do so randomly, but that tells you something about your guessing strategy, and nothing about the experiment.
But If you group the balls as {1} and {2,..,10}, then you do have information that allows you to distinguish between them, i.e. the fact that the second group has more than one possibility. If you assign equal probability to the groups now, you would be ignoring the fact that you have relevant information that distinguishes between the two groups.
Yet I can still be right 50% of the time by choosing randomly between those groups. That much is no different than if I randomly choose between 1-5 and 6-10, my chance of being right is 50% either way, completely independently from anything about the experiment other than I can associate any outcome with either one or the other of those groups. And I still can say nothing about what I expect the distributions after many trials to look like over the {1, 2-10} grouping versus the {1-5,6-10} grouping, other than the latter will have at least as high a left/right ratio, and likely higher, than the former. That's all that can be said, it is simply wrong to expect 1-5 to come up 5 times more often than 1, and it will generally not hold in the absence of a symmetry that leads us to expect it.

So {1,2..,10} is the set of outcomes.
If you choose it to be so. I can just as easily say there are two outcomes, {1} and {2-10}. That is the "information" I can go on, and apply all the same logic as your coin situation. If you doubt that, imagine a 20-sided die with two indistinguishable 1's and eighteen indistinguishable 2's on it, but I don't know the die has 20 sides. How many outcomes shall I count in that situation? That's all I've done with my grouping, and I will apply all your logic to that situation. Why would you expect that logic to work any better in your situation than in mine? Is yours more generic somehow?
 
Last edited:
  • #128
Ken G said:
Every physics "revolution" was actually just a case of someone guessing wrong who had no business expecting to be right
Prove it.
 
  • #129
Hurkyl said:
Prove it.
Fair enough, I'll be happy to. Of course it depends on what any individual would label a "revolution", so I'll just restrict to some of the more uncontroversial ones and we can always extend the list. My charge is to cite why someone was guessing wrong about something they had no reason to expect to be right about:

1) The geocentric universe: This was based primarily on concepts of gravity that would make a stationary point at the center special in some way, along with the absence of stellar parallax, indicating a stationary Earth if the stars are not too far away. So the guess was made that gravity really did pick out a special point at the center, and the stars really were not that far away. Neither of those guesses had the slightest shred of supporting evidence, they had no business expecting them to be right, they merely served a purpose of unifying the existing data at the time. As I've said, one must try a little harder when the goal is not just to unify the data, but also do it in a way that does not introduce unnecessary and unwarranted guesses.

2) Determinism: this is an element of Newton's formulation of physics, and worked extremely well in a wide array of situations ranging from gas beakers to the cosmos. However, it was never more than a useful model, and of course there is no way to test if determinism is "real", because no observations have suitable precision to be able to make that claim. It was just a guess that the universe "really is" deterministic-- and some would say not a very good guess at all. All we can really say is that application of deterministic models works well in situations in which they have been shown to work well, depending on the goal of the model and the details of the application. Extending it to a philosophical truth about reality, as by DesCartes and so forth, was a pure guess and it is no surprise it has produced nothing of value in our understanding of our place in things. We had no business making that leap, and still don't, as we cannot support it.

3) Special Relativity: the subject of this thread. I maintain that we had no idea that Newton's laws would be extendable to arbitrary relative velocity, so it should come as no surprise that they cannot. It was pure assumption, with zero observational backing, that reality did not embed a characteristic velocity scale that would be reflected in its dynamics. One does not assume that the absence of evidence is an evidence of absence, and we had no business making the guess that we could.

4) General Relativity: it was pure guess that "action at a distance" was really a physically plausible thing. All we could really say is that whatever was mediating that action, it was happening very fast, and with some ability to accommodate motions in the future that were similar to motions in the past (both constant velocity and constant acceleration are in effect "accommodated" by gravity to mimic instantaneous response). It was pure guess that such an accomodation would extend to all types of dynamics, as required by action-at-a-distance. (As I recall, even Newton himself was bothered by that assumption--he himself did not even expect it to hold true!) We had no business thinking action-at-a-distance was a fundamentally real property of the universe, we only knew that it worked in the situations tested, much like low-speed Newtonian mechanics worked without requiring a characteristic speed be embedded into reality.

5) Quantum mechanics: it was pure guess that the dynamics that ruled the cosmos could also rule an atom in a similar way. We had no idea what the scales of the forces would be, or if new forces would appear, or even completely new physics (like wave mechanics). It was a totally new regime on a vastly different scale and nothing similar was used to constrain any of Newton's mechanics, so there's no reason it should obey Newton's mechanics.

6) Wave/particle duality and the quantization of light: it was pure guess that we had no business making that just because we observed a clear difference in the behavior of macroscopic particles and waves, that this clear distinction would survive at all scales. Indeed, it is quite common for physicists today to take the opposite default assumption-- that everything that appears different on one scale can be unified at a deeper or higher energy scale. Neither of these assumptions have any basis, they are just guesses, and it was pure guess that waves and particles were fundamentally unconnected, just as it is pure guess today that the strong force and gravity are fundamentally connected. (The search for such a unification is good science-- the expectation that it is there is not.)

7) Spin: classical mechanics does not allow for particles to have internal degrees of freedom that store angular momentum, so it may be viewed as a "revolution" when it was discovered that they do. Again, classical mechanics never said anything about the internal degrees of freedom of fundamental particles, it was totally uncharted territory, and we had no business expecting the absence of strange new properties like spin because it was only our minds that separated particles from the rest of the universe. We now see the connections between the properties of particles and the symmetries of the universe, it was pure guess that there would be no such connection.

8) Dark matter: the easiest of all. If we use light to track matter, it is pure guesswork that we won't miss anything important. We never had any reason to expect the universe would not contain dark matter, the only real surprise is why the amount is not completely different from the amount of baryonic matter.

9) Dark energy: it was a complete guess that the gravity that works in galaxies would also work over the vast scales of the whole universe. We had no business guessing that gravity could only come from matter, simply because the only gravity we had seen came from matter. Again this is the difference between including what you have seen in your theory, and expecting that you have not left something out. What a sillly thing for science to do.

10) Evolution: let's get a non-physics topic in here. It was pure guess that species had to be created by a supreme power, we had zero evidence for that scenario. So the discovery that natural processes could lead to speciation was just that-- a discovery. Seeing it as a "revolution" that threw off the old power structure is simply a recognition that the Emperor had no clothes, a fact I am asserting we should simply build into our understanding of how science works, until there is no need to see everything as a "revolution" instead of what it really is-- another piece of the puzzle.
 
Last edited:
  • #130
Ken G said:
Certainly, but what you are failing to do is to assess the validity of your guess. Some things really are pure guess, and have no more value than a guess.

Again, you have no way of telling whether your guess has any value or not before experiment. The validity of the guess can only be assessed with the information you receive from the experiment. If you use all the information you have to form the best possible opinion that the information allows(in cases when must form an opinion), then you are doing the most logical thing. You may find after your experiment that your guess was not valid to an acceptable degree, and the new information will be used to update your guess. Of course, in all this the guesses also have degrees of belief that are determined by your current information.

Ken G said:
Labelling everything a guess is of no value, calling something a guess means you have a low opinion of its likelihood of being correct. That's how most people use the term, and how I am using it too.

I've used the words "expectation", "probability", "guess" and "opinion". If you don't like any of those words, then suggest a new one. The idea is "the opinion/guess/expectation/conclusion that is most supported by the information you have."

Ken G said:
Yes, but I'm saying, what if the "best guess" really still ends up being just a guess? What good is it to identify it as the "best guess", if the best guess is still completely worthless?

How do you know that the best guess is worthless? If guesses are worthless why do you make them? Guesses are not random fantasies. They are, ideally, logically drawn opinions that use all the information that is available to you. When you guess something, do you know whether it is a useless guess? No. You only know that after experiment, i.e. testing the guess. You can guess that a girl likes you, based on conversations/gestures/expressions etc. You are using the information available to make that guess. It could be wrong, but you don't know that until you ask her. You cannot decide that the guess is worthless. You can definitely believe to a certain degree that a particular guess has a low likelihood based on your information, but you cannot decide for certain that it is "wrong" or "useless" or "worthless".

Ken G said:
That's what I'm saying, I'm not disputing what the best guess is, only what the meaning of a best guess is. We always need to supplement the best guess with some concept of confidence in that guess, or it is a truly useless concept.

That's exactly what I've been trying to say.

Ken G said:
No one is forcing us to guess at all-- the sensible approach is to test, not to guess.

There are countless cases where a guess is crucial. In fact, every time you apply physics to predict something, you are guessing. Even in the tested regime. The guesses in the tested regime are very well supported, so your confidence in them is high, but they're still guesses. And by guess I don't mean "they have a low likelihood of being correct" as you seem to think. There are many guesses you make everyday that you believe almost as fact, and have good reason to do so.

Ken G said:
Let me give another example to underscore this. Let's say there are three coins, two pennies and a quarter, in a jar, and someone is going to shake that jar until one coin comes out. You know nothing about the coins except that two are worth 1 cent and one is worth 25 cents. You are to receive the coin that shakes out-- how much money do you expect to receive, say over the course of 100 repeated trials? Now, if you pick randomly between "I get a penny" and "I get a quarter", you will be right half the time, and it makes no difference at all anything else about the experiment.

The problem with examples like this is that our familiarity with them misleads us. Let us analyze this one closely. Here's the information you have.

1. The jar has 3 coins.
2. two are worth 1 cent and one is worth 25 cents.

The experiment is to choose one coin, or shake out one coin. You have no information about this process of choosing, other than the fact that it must result in anyone of 3 coins. So you must expect each equally. So even though you cannot distinguish between the two pennies, you must still expect to get a penny 2/3 of the time.
Ken G said:
If you can distinguish the pennies into penny #1 and penny #2, then you may instead decide to randomly choose between receiving each of the three coins, and now you'll be right 1/3 of the time, again independent of any other information you may have about the experiment.

See above.

Ken G said:
However, nothing that has been said so far may be used to form a meaningful expectation value for the amount you'll receive in 100 trials. There is simply no expectation that is not a pure guess, and logically you should not expect any such guess to converge on something correct over any number of trials. It simply doesn't mean a thing.

Yet a meaningful expectation can be formed. Before you do the 100 trials you have a certain amount of information. That information can be used to form a meaningful guess (not an arbitrary guess, a guess that uses the information you have). You are likely to obtain information during the 100 trials that could considerably change your guess, but before you do the trials you have to go by the information you have. The detailed analysis of problems like this, to determine the guess that uses all the relevant information you have and nothing more and also how strongly you must believe in this guess, is difficult generally, but it can be done.

Ken G said:
But why is forming an expectation making the best of what you have? What possible benefit are you deriving from that?

You are deriving whatever benefits that the information you have can give you. It may not be much, or it may be a lot. You don't know until you try.

Ken G said:
If the expectation is of no usefulness, then it being "best" among all the useless ways to form an expectation isn't saying much.Right, so it is an expectation that means nothing at all, there's no point in even forming it prior to the experiment.

Again, you don't know if the expectation is "of no usefulness". You have made it as useful as it can be by using the information you have. What more can you ask? If you don't want to form any expectation at all before the experiment, then you are saying you don't care about the information that you currently have. You want to have certain answers to everything, and you will only believe something once it has been verified exactly, which of course can never be done. So, in effect, you are saying physics is useless, since physics itself is just a system of belief that is based on experimental information; various beliefs that you believe to varying degrees due to varying degrees of informational support.

Ken G said:
We always need to have a concept of the confidence we can place in our expectations. What I'm saying is, we've been doing a particularly bad job of that in science when we treat all predictions on an equal footing, and that has led to all the "revolutions"-- and all the misunderstandings about the fallibility of science.

You may be right. I don't know what scientists in the 19th century thought about the validity of their theories.
Ken G said:
If you choose it to be so. I can just as easily say there are two outcomes, {1} and {2-10}. That is the "information" I can go on

You cannot choose what information you have. The information is given to you, or has been collected by you before. If you know that there are 10 numbered balls, you cannot just say, "I choose to go by the information that there are only two outcomes".

Ken G said:
Imagine a 20-sided die with two indistinguishable 1's and eighteen indistinguishable 2's on it, but I don't know the die has 20 sides. How many outcomes shall I count in that situation? That's all I've done with my grouping, and I will apply all your logic to that situation.

If I don't know that there are 20 sides, then I cannot use that information to form my beliefs. If all I knew was that if I pick up your die and look at it, I will see either 1 or 2, then I would assign equal likelihood to both. You must realize that until after you've done this experiment, you don't know to what degree these beliefs are accurate.
 
Last edited:
  • #131
Not if the best you can do with the information is recognize that it is insufficient to draw any conclusions whatsoever. That is the actual logical thing to do.

you can't draw any conclusion about the probability but you can draw a conclusion about the expectation.

the expectation is not a guess. bayesian probability has been precisely mathematically defined.
 
  • #132
Ken G said:
Fair enough, I'll be happy to.
You're missing something very important -- justifications for your assertions. What criteria are you applying to judge, for example, that we had no business making an inferences based on all of the empirical support for the wave nature of light.

I notice that you're also making presumptions even in your critiques -- for example
it was pure guess that "action at a distance" was really a physically plausible thing. All we could really say is that whatever was mediating that action,​
you seem to have presumed that all actions are 'mediated'. Why is that justified?
 
  • #133
granpa said:
you can't draw any conclusion about the probability but you can draw a conclusion about the expectation.
The mean is something 'about the probability'.

bayesian probability has been precisely mathematically defined.
Yes -- the a posteriori values are defined in terms of the a priori values. Without assigning a priori probabilities, you cannot have a posteriori values.

Of course, the Bayes factor generally wouldn't require such a choice. Is that to what you're referring?
 
  • #134
The mean is something 'about the probability'.

the expectation is not the same as the mean. that's where you are going wrong.

if someone offers me a bag with an unknown number of black and white marbles in it and asks me to predict what color marble i would draw from it at random then my expectation is that i can predict it 50% of the time by choosing black or white at random. the probability is unknown but if enough random people make me this offer with enough random bags of marbles then i expect the probability to average 50%. its the probable probability. that is totally different from the probability.

but as i keep drawing marbles from anyone bag i gain experience. suppose i keep drawing black marbles. then with each draw the bayesian probability (the expectation) that i will draw another black marble goes up. with an infinite number of trials the bayesian probability=the probability.
 
Last edited:
  • #135
granpa said:
The mean is something 'about the probability'.

the expectation is not the same as the mean.
I'm having a mental blank then -- the mean is the only technical meaning for the word 'expectation' that I can imagine at the moment. Could you spell out for me what you mean?
 
  • #136
the mean is the only technical meaning for the word 'expectation' that I can imagine at the moment.

https://www.physicsforums.com/showthread.php?p=1738079&highlight=expectation#post1738079

"But on what basis do you say that it is equally likely?

on the basis that there are 2 sides. if you knew nothing else about it you would still expect to be able to predict the outcome 50% of the time just by choosing at random. but notice that 50% is not the probability. the probability is unknown. the probability could be 100% or 0% or anywhere in between. if its not the probability what is it? the logical thing to call it would be the expectation but that word is already taken. so its called the bayesian probability."
 
  • #137
dx said:
Again, you have no way of telling whether your guess has any value or not before experiment.
Of course we do, and we do it all the time. Every single moment of our lives is a unique experiment, it has never happened before and never will again. Yet we are not forced to live our lives with "no way of telling" the value of the predictions we use to function.

The validity of the guess can only be assessed with the information you receive from the experiment.
I'm talking about justifiable confidence in a "guess" (your usage), not "validity".

I've used the words "expectation", "probability", "guess" and "opinion".
As I said, "guess" carries the connotation that the prediction is not at all reliable. That is the common usage, if you mean something else by it, it is you who need the new word. Further, I have noted that the concept of "expectation" is completely useless without an associated concept of "justifiable confidence in the expectation". That seems clear enough in how we function daily.

How do you know that the best guess is worthless?
Again, we are called upon to make such assessments constantly all the time. Why object now? If I ask you, what is your best guess for the team that will win the World Series in 2020, what is your answer? What is the value of that "expectation", and what odds would you take? How can you decide that, if you claim there's no way to assess the usefulness of the expectation?

If guesses are worthless why do you make them?
That's the point. We should not-- unless we are also willing to assess the degree of reliability of the guess. Otherwise, in betting we would lose our money, and in science, we would engender misconceptions about what science can tell us.

When you guess something, do you know whether it is a useless guess? No.
I can certainly evaluate the usefulness of the guess, and am called on to do so all the time.
There are countless cases where a guess is crucial. In fact, every time you apply physics to predict something, you are guessing.
Again, you are using the term "guess" as if it had no different meaning for predicting the World Series winner in 2020 versus predicting that a ball wiill fall if I release it. I am not using "guess" that way, I am using the standard meaning of the term.

And by guess I don't mean "they have a low likelihood of being correct" as you seem to think.
I am well aware you are not using the standard meaning of the word. The real question is, why do you deny that the reliability of a prediction can be assessed? Why, when the reliability can be assessed to be low, do you still think it is important to form an "expectation"? That makes no sense, at some point the reliability is so low that there is simply no use for the prediction in the first place, except as a kind of "benchmark" to know when a theory has broken down. That, for example, is how one debunks astrological predictions.

There are many guesses you make everyday that you believe almost as fact, and have good reason to do so.
And none of them count as "guesses" in the standard usage.
You have no information about this process of choosing, other than the fact that it must result in anyone of 3 coins. So you must expect each equally.
No, there is no such requirement on your expectations. Indeed, it is far more reasonable to expect different frequencies of occurence, but different in an unknown way. For example, if I say I will do 3 million trials, with Poisson noise of something like 1,700 outcomes, and I allow you to win the quarter if it either appears within 5,000 outcomes of 1 million occurrences, or if it does not, on what basis can you say it is logical to choose the former? There is no basis for that expectation, if you know nothing about how the coin is chosen in each trial.
So even though you cannot distinguish between the two pennies, you must still expect to get a penny 2/3 of the time.
But if all I tell you is that the outcome is either a penny or a quarter, you would apply your logic to expect to get a penny 1/2 the time. I, on the other hand, claim that is a meaningless expectation.

Before you do the 100 trials you have a certain amount of information. That information can be used to form a meaningful guess (not an arbitrary guess, a guess that uses the information you have).
The problem is, calling a guess "meaningful" any time it is not "arbitrary" is an extremely low standard for meaning, and in practice will be a worthless standard. It's not so much a problem of how little information you possess, or even how much information you don't possess, it's more an issue of how little information you possesses about the information you don't possess. A probability is meaningful when you have clear information about what you don't know, but when you don't even have that, there is no meaning to a probability estimate. When there is no meaning to a probability estimate, there is also no usefulness to making predictions-- other than as benchmarks to tell when a prediction failed (again as an assessment of a predictive scheme, like astrology).

Again, you don't know if the expectation is "of no usefulness".
I don't see why you hold to that position, in contradiction to our almost daily applicaton of its inverse.
You have made it as useful as it can be by using the information you have.
Of course, and that usefulness might well be squat, as in the case of predictions overturned by scientific "revolutions".

What more can you ask?
I can ask to restrict to expectations that actually have some merit behind them, like the ones made by a person building a bridge, or by the Wright brothers as they tried to make an airplane-- and unlike the expectation that reality is fundamentally deterministic based on the success of deterministic models in various situations, or the expectation that reality actually fragments into "many worlds" based on the success of models involving unitary time evolution of closed systems in between couplings to an experimental apparatus. These are classic examples of expectations that are useless by virtue of their unreliability.

If you don't want to form any expectation at all before the experiment, then you are saying you don't care about the information that you currently have.
It means that I do not think the information I have is useful enough to form a useful expectation. If that's your meaning of "don't care", then yes.

So, in effect, you are saying physics is useless, since physics itself is just a system of belief that is based on experimental information; various beliefs that you believe to varying degrees due to varying degrees of informational support.
I'm certainly not saying physics is useless, on the contrary I'm pointing out a key requirement for it to be useful-- the ability to gauge the reliability of an expectation. It is your version that would be useless, wherein all we can ever do is form the best guess we can from the information we have, and know nothing about the reliability of that expectation until we do the experiment.

You may be right. I don't know what scientists in the 19th century thought about the validity of their theories.
A science historian might see it differently, but it is my general impression that scientists throughout history, from the Greeks right up to today, have tended to perceive the "body of scientific knowledge" with a considerable degree of certainty. That was valid then, as now, only insofar as we are keeping track of the experimental regimes we actually have direct knowledge of. In other words, we must always recognize the difference between the statements "to be wrong, that theory would require some experiment that hasn't been done to come out different from X", versus saying, "to be wrong, that theory would require some experiment that has been done to come out differently from X, where X is the result we got."
 
  • #138
Hurkyl said:
You're missing something very important -- justifications for your assertions. What criteria are you applying to judge, for example, that we had no business making an inferences based on all of the empirical support for the wave nature of light.
My criterion for judging that is the absence of a justification for concluding the inverse. In other words, who needs such a criterion more, the person who imagines that successfully modeling some dynamics with wave mechanics implies that light is fundamentally a wave and not a particle(which is what I presume you mean by "wave nature", because it was obvious that light exhibits wave properties), or the person who notes that such success demonstrates no such thing? It is not my position that requires such a criterion-- I am pointing to the absence of a justification for the competing idea. Ergo, it is actually your claim that I have insufficient justification that is the claim with insufficient justification here-- my stance is simply the skeptical one.

I notice that you're also making presumptions even in your critiques -- for example
it was pure guess that "action at a distance" was really a physically plausible thing. All we could really say is that whatever was mediating that action,​
you seem to have presumed that all actions are 'mediated'. Why is that justified?
Pick any word you like-- all I mean by "mediated" is that the action "acts" in some way, which has to be instantaneous in the case of action at a distance. It hardly seems a "presumption" that an action must act.
 
Last edited:
  • #139
granpa said:
on the basis that there are 2 sides. if you knew nothing else about it you would still expect to be able to predict the outcome 50% of the time just by choosing at random. but notice that 50% is not the probability. the probability is unknown.
Right, the 50% tells you everything about your guessing strategy (split all the possible results into two distinct classes and guess one of the two at random), and nothing at all about either the experiment or even the possible outcomes of the experiment, other than that it is possible to distinguish them into two exclusive all-encompassing groups and choose the correct group 50% of the time with that strategy.
if its not the probability what is it?
It's just the result of a choosing strategy, call it a game theory. It doesn't connect to the experiment under study in any useful way.
 
  • #140
i never said otherwise.
 
  • #141
Ken G said:
I can certainly evaluate the usefulness of the guess, and am called on to do so all the time.

You cannot evaluate the usefulness. You can only evaluate the degree to which you believe in it. Whether the guess is close enough to reality to be useful or not cannot be evaluated before the experiment. You may have a lot of evidence to support your claim that "It is unlikely to be useful", but that is also a guess based on information. You are only assigning a degree to your belief.
 
  • #142
granpa said:
i never said otherwise.
OK, just making sure.
 
  • #143
dx said:
You may have a lot of evidence to support your claim that "It is unlikely to be useful", but that is also a guess based on information. You are only assigning a degree to your belief.
Yes, except I would replace your "guess" with "useful prediction", and remove your "only".
 
  • #144
Ken G said:
In other words, who needs such a criterion more,
Huh? "Needing a criterion more"? What the heck are you talking about?

(which is what I presume you mean by "wave nature", because it was obvious that light exhibits wave properties)
"Obvious"? What does 'obvious' mean? Is it anything other than an expression of the biases in how we interpret what we see?

that light is fundamentally a wave and not a particle
(Classical) waves, of course, cannot be (classical) particles.

My criterion for judging that is the absence of a justification for concluding the inverse.
So how do you justify dismissing all of the evidence for light being a wave? Keep in mind that your argument is heavily dependant upon you being able to devise a justification that is not also applicable to things like the obviousness of light exhibiting wave properties, or whether or not the sun will rise tomorrow.


Pick any word you like
Fine; I pick the phrase "caused by invisible pink unicorns". So are you seriously criticizing classical mechanics because it doesn't permit gravity to be caused by invisible pink unicorns?
 
Last edited:
  • #145
Hurkyl said:
Huh? "Needing a criterion more"? What the heck are you talking about?
Simple enough. You are claiming I need a criterion to claim that physicists should not have, for example, used the success of Newton's laws as evidence that the universe was fundamentally deterministic. I am saying, no, I do not need such a criterion-- not nearly as much as physicist need a criterion to make that undemonstrated leap of faith. Ergo, it is your position that "needs a criterion more", so it is you who need to explain said criterion.
"Obvious"? What does 'obvious' mean? Is it anything other than an expression of the biases in how we interpret what we see?
Obvious means that if you use wave mechanics successfully to understand some aspects of light behavior, then wave mechanics can be used successfully to understand some of light's behavior. That's obvious, yes, and that is just what I said. The point is, if that's all you mean by "wave nature of light", then it was obvious. So I presume you instead mean, "light is fundamentally a wave, in contrast to, say, a particle". If you do not explain what you mean by phrases like "wave nature", it forces me to fill in the blanks in your argument, and I am merely explaining how I'm doing that.

(Classical) waves, of course, cannot be (classical) particles.
This is a definition of the word "classical", I hardly see what this tells us about the concepts of waves and particles.

So how do you justify dismissing all of the evidence for light being a wave?
Your own words belie your argument. "Light being a wave"? What does that mean? Light is light, last I checked.

Keep in mind that your argument is heavily dependant upon you being able to devise a justification that is not also applicable to things like the obviousness of light exhibiting wave properties, or whether or not the sun will rise tomorrow.
I see no challenge there at all. Yes, we can analyze much of the behavior of light using wave mechanics. That's what we do in physics, we make mathematical idealizations and use them to create models that describe certain aspects of what we observe. The entire leap that I am arguing against is that we even use language like "ilght being a wave". Science has no idea how to define the meaning of that sentence, yet we persist in using it, like it said something more than "we successfully apply wave models to unify the results of the following list of experiments, which we assume extends to analogous situations but we have no idea how hard we can push it into new regimes." Now, that didn't hurt, did it?
Fine; I pick the phrase "caused by invisible pink unicorns". So are you seriously criticizing classical mechanics because it doesn't permit gravity to be caused by invisible pink unicorns?
If you see a parallel between the word "mediate" and the words "caused by invisible pink unicorns", then I suppose you can go ahead and imagine that if you like. As for my (and Newton's own) criticism of classical mechanics, it had nothing to do with the agent we ascribe to action at a distance, it had merely to do with the necessity that the agent act instantaneously with no description of how the influence got from point A to point B. It is the issue of causality that is at stake here, a deeply held principle given its vast unifying potential, even before the advent of relativity.
 
Last edited:
  • #146
Ken G said:
Simple enough. You are claiming I need a criterion to claim that physicists should not have, for example, used the success of Newton's laws as evidence that the universe was fundamentally deterministic.
Well, yeah. That's how rational discussion works -- rather than expect us to accept all of your claims as a matter of faith, you attempt to provide rational justification for them. And since you are apparently denying empiricism, one of the cornerstones of the scientific process, your justification needs to be pretty darn good.



I see no challenge there at all. ... "we successfully apply wave models to unify the results of the following list of experiments, which we assume extends to analogous situations but we have no idea how hard we can push it into new regimes." Now, that didn't hurt, did it?
You lose; your form of argument also denies the claim "the sun will rise tomorrow".


If you see a parallel between the word "mediate" and the words "caused by invisible pink unicorns", then I suppose you can go ahead and imagine that if you like.
Please clarify -- do you criticize classical mechanics on the grounds that it does not allow for the possibility that all observed phenomena are simply the actions of invisible pink unicorns?

If you do make this criticism, then do you really expect me to take you seriously?
If you do not make this criticism, then why not?


As for my (and Newton's own) criticism of classical mechanics, it had nothing to do with the agent we ascribe to action at a distance, it had merely to do with the necessity that the agent act instantaneously with no description of how the influence got from point A to point B.
You assume, a priori, that influences have to "get" from one point to another. You also assume, a priori, that the process is not 'fundamental', but instead can be described in terms of some other processes that you find more palatable.

Why should I make such a prior assumptions? More importantly, why should a scientist?


It is the issue of causality that is at stake here, a deeply held principle given its vast unifying potential, even before the advent of relativity.
:confused: Causality has nothing to do with whether or not 'action at a distance' is possible. (That is a consequence of the principles of special relativity, but not something inherent in the notion of causality itself)
 
Last edited:
  • #147
Ken G said:
I'm curious about how people here view Einstein's prescription for determining simultaneity in an inertial frame, and...
I don't really understand what it is that you're concerned about, or what sort of answer you're looking for, but I'm going to explain how I think about these things:

SR is just the claim that space and time can be represented mathematically by Minkowski space. That's the one and only "axiom" of the theory. The Minkowski metric has a non-trivial group of isometries called the Poincaré group. We call the members of the Poincaré group Lorentz transformations, or inertial frames. (They are diffeomorphisms that map the manifold onto itself, and since the manifold is R4, they are also global coordinate systems on the manifold).

Pick any two events that are space-like separated. There's always an inertial frame x that assigns the same time coordinate to those two events. The two events are said to be simultaneous in any such frame.

You asked if the invariance of the speed of light is a law of nature. The word "law" usually refers to some small part of a theory that can be expressed in one sentence or one equation. In this case we're definitely talking about a part of a theory that can be expressed succinctly, so I think the world "law" is appropriate.
 
  • #148
Hurkyl said:
Well, yeah. That's how rational discussion works -- rather than expect us to accept all of your claims as a matter of faith, you attempt to provide rational justification for them.
I fear you are rather missing the definition of "skepticism". Let me try another approach. When Newton's laws were discovered, what would you describe as the basic for the subsequent prevailing view by scientists that the universe was fundamentally deterministic? On what basis was that view held? Finally, on what basis can it be challenged? You see, it is you who need to answer these questions, for mine are quite simple-- there was no sufficient basis, and the basis can be challenged from any angle you choose-- it must meet all challenges. Did it?

And since you are apparently denying empiricism, one of the cornerstones of the scientific process, your justification needs to be pretty darn good.
I have no idea what you mean by the word "deny" in that sentence, but I certainly did not deny the usefulness or importance of empiricism in science. Indeed, all I did is require that we keep track of what empiricism is really about-- and what it is not about (to wit, it is not about forming philosophical expectations about the fundamental nature of the universe outside of what we have ever actually tested, that would have to be counted as a tremendous miscarriage of empiricism-- indeed that is pretty much my whole point).

You lose; your form of argument also denies the claim "the sun will rise tomorrow".
Only if you are following it badly. Perhaps my statement above has clarified that empirical statements about what we have found to be true already are valid ways of predicting what will continue to be true, insofar as past testing is relevant to future similar tests (a core assumption of all science). I thought I had been quite clear on that when I distinguished "usefully reliable predictions" made "inside the box" of past experience from philosophical expectations "outside the box"-- the latter being an example of mistaking a simple hypothesis for a reliable prediction.
Please clarify -- do you criticize classical mechanics on the grounds that it does not allow for the possibility that all observed phenomena are simply the actions of invisible pink unicorns?
Classical mechanics certainly does "allow" that possibility, why would you claim it doesn't? All it claims is that whatever is the agent, your unicorns or my "mediators", they act at a distance and instantaneously. That's all it requires, but of course this is a limited model that is not going to stand up as a fundamental truth of reality, even Newton suspected that strongly. There was never any reason to think it would stand up, that's my point, and the fact that we still use action at a distance even after we know it doesn't stand up proves that this aspect never even needed to stand up as a fundamental truth. Ergo, it was never a necessary part of the theory, this is the point.

Indeed, I have not criticized classical physics at all, it's a lovely model for what it's good at. I criticized the foolish tendency to think any theory is something more than what it is, so that we get "surprised" when we find out that the philosophical scaffolding we imagined was supporting that theory turns out to not be a required element of the theory's usefulness. That's been the theme of the whole thread, in fact.
If you do make this criticism, then do you really expect me to take you seriously?
If you do not make this criticism, then why not?
A good theory does not include any extraneous and unnecessary elements, like pink unicorns, until they are needed. Nothing I've said in the least contradicts that. What I've said is actually quite the opposite-- that once you have a useful theory that unifies some set of observations, you should never mistake any optional conveniences associated with the application of the theory for a demonstrated part of the theory itself. Instead, you should endeavor to find the minimum theory, that makes the fewest possible assumptions, yet achieves the same unification of the data. Obviously including pink unicorns in classical physics doesn't do that-- and neither does assuming that the universe is fundamentally deterministic. That's what I'm saying, your objections are to things I have not said nor would.

You assume, a priori, that influences have to "get" from one point to another.
On the contrary, I make no such assumption, indeed I've no doubt that picture will suffer from limitations as well. All I have ever said is that simply because classical mechanics met with success by modeling forces as actions at a distance, it gives us no reason to conclude that the influences don't have to get from one point to another. I would also take the same attitude toward any particular model for how the influences get from one point to another-- this is the purpose of skepticism. We need to keep track of what is really necessary in a theory-- and none of Newton's theories required anything beyond a very rapid, possibly instantaneous, propagation of influences (using propagation loosely, as I said specific ideas about what that means will also just be models). When one has a wide range of possibilities like that, it is of course natural to expect that more careful scrutiny will yield a more generic answer than an extreme limiting case. So it is not a surprise that this is just what we find.
You also assume, a priori, that the process is not 'fundamental', but instead can be described in terms of some other processes that you find more palatable.
You are mistaken to think that this is an issue of "palatability". That would be subjective. What I am talking about is simply being true to the ways we define what constitutes the action of science, and the goals of that action-- nothing more. I find it odd that you think I'm breaking from empiricism-- my goal is to restore it from how badly we tend to stray when we become too enamored of our own idealizations. The lure of the "warm fuzzy feeling" of understanding is indeed great, but science is about achieving it without fooling ourselves in the process, not tripping over ourselves in our rush to achieve it by fooling ourselves. As I said before, there are other ways to achieve the latter which require far less educational investment than science.

Why should I make such a prior assumptions? More importantly, why should a scientist?
You, and they, shouldn't, if the data don't require it-- do you not see that this is the whole point?
:confused: Causality has nothing to do with whether or not 'action at a distance' is possible.
I can't agree, but that's another discussion entirely.
 
Last edited:
  • #149
Fredrik said:
SR is just the claim that space and time can be represented mathematically by Minkowski space. That's the one and only "axiom" of the theory.
I tend to agree with you, but then the question of the thread becomes, why is that not what you will find in virtually any physics textbook? Instead, you will find two postulates, the first being that the laws of physics are the same in any inertial frame, and the second that the speed of light is constant. Also, you find the Einstein simultaneity convention tacked on as the means for deciding what speed means. So it's really three axioms, or two axioms and a coordinate definition, depending on how you count it. That doesn't sound like your one axiom to me.
The Minkowski metric has a non-trivial group of isometries called the Poincaré group. We call the members of the Poincaré group Lorentz transformations, or inertial frames. (They are diffeomorphisms that map the manifold onto itself, and since the manifold is R4, they are also global coordinate systems on the manifold).
Yes, this is also the crucial elements of the picture emerging from my discussion with Hurkyl. Mathematics is often a good way to really see what are the core elements of a theory, and what are unknown philosophical statements. Which brings us to...
You asked if the invariance of the speed of light is a law of nature. The word "law" usually refers to some small part of a theory that can be expressed in one sentence or one equation. In this case we're definitely talking about a part of a theory that can be expressed succinctly, so I think the world "law" is appropriate.
So what part of your "one axiom" requires this "law"? You see, the question is not whether or not this succinct statement is normally used in SR, we know that it is. The question is, is it a necessary part of SR, such that the observations that support SR support this "law"? Or is it purely a statement of philosophical preference, engendering a non-generic aspect to the theory simply because we don't know we can't, not because we need to (examples of that in the past include geocentrism and determinism, to name just a few)? I'm arguing that attaching such non-generic philosophical preferences to our theories is the best way to get "surprised" down the road, leading to yet another unnecessary "revolution".
 
Last edited:
  • #150
Ken G said:
I tend to agree with you, but then the question of the thread becomes, why is that not what you will find in virtually any physics textbook?
I suspect that some authors just don't know any better, and that the others are choosing the traditional path because it's easier (for them).

Ken G said:
Instead, you will find two postulates, the first being that the laws of physics are the same in any inertial frame, and the second that the speed of light is constant.
I actually find it quite ridiculous that these postulates are always presented as if they are mathematical axioms from which you can derive everything else, when they are in fact ill-defined. The biggest problem is that the concept of an "inertial frame" hasn't been defined in advance. I gave this some thought a few months ago, and I came to the conclusion that any reasonable definition must actually include these "postulates" in some way. This is what makes them so ill-defined. They are a part of a definition of a concept they depend on!

So it really doesn't make any sense to think of them as axioms. They should be thought of as items on a check list. Once we have learned about the two "postulates", our next task is to find a theory of space and time that implies some version of them. What the "postulates" are really saying, is that we're only going to consider theories that do.

There is of course nothing wrong with this way of finding a theory. Once the theory has been found, it can be tested in experiments, and if the experiments fail to disprove the theory, it doesn't matter how we found it. It's still a good scientific theory.

I do however have a problem with the traditional presentation, because it gives the student the impression that Einstein's postulates are sufficient to define the theory, that they are the axioms of a theory, and that all those calculations that the book and their teacher goes through is part of an actual derivation of time dilation, the Lorentz transformation, and so on, when in fact those calculations are just there to help us guess what the real axioms of the theory are (and to improve our general understanding of relativistic effects).

I would have thought that an author who really understands this would be inclined to actually say these things, but they never do, so I sometimes wonder if any of them really understand it. Maybe the smart ones do, and just assume that this is obvious to everyone.

Ken G said:
So what part of your "one axiom" requires this "law"? You see, the question is not whether or not this succinct statement is normally used in SR, we know that it is. The question is, is it a necessary part of SR, such that the observations that support SR support this "law"?
Define the velocity associated with a curve C and a point p on the curve, in an inertial frame, as the 3-vector we get by taking the spatial components of the tangent vector of C at p and dividing them by the magnitude of the temporal component. Define the speed as the magnitude of the velocity. These definitions and the properties of Minkowski space imply that the speed associated with any null line, at any point on the line, and in any inertial frame, is =1.
 
Last edited:

Similar threads

  • · Replies 16 ·
Replies
16
Views
2K
  • · Replies 32 ·
2
Replies
32
Views
4K
  • · Replies 51 ·
2
Replies
51
Views
5K
  • · Replies 127 ·
5
Replies
127
Views
9K
  • · Replies 34 ·
2
Replies
34
Views
2K
  • · Replies 6 ·
Replies
6
Views
927
  • · Replies 9 ·
Replies
9
Views
1K
  • · Replies 12 ·
Replies
12
Views
2K
  • · Replies 4 ·
Replies
4
Views
933
  • · Replies 57 ·
2
Replies
57
Views
7K