Understanding Refractive Phase Differences in Electromagnetic Waves

Click For Summary
SUMMARY

This discussion centers on the refractive phase differences in electromagnetic waves, specifically addressing the behavior of light and radio waves as they pass through different media. It is established that while the wavelength and velocity of light change in a refractive medium, the frequency remains constant. The concept of optical path length is introduced as a method to analyze these changes. Additionally, the interaction of electromagnetic waves with electrons in a medium is discussed, emphasizing the classical explanation of how these interactions lead to a delay in wave propagation.

PREREQUISITES
  • Understanding of electromagnetic wave theory
  • Familiarity with the concept of refractive index (n)
  • Knowledge of optical path length calculations
  • Basic principles of wave-particle duality in physics
NEXT STEPS
  • Research the optical path length technique in detail
  • Study the interaction of electromagnetic waves with matter at the microscopic level
  • Explore the differences in refractive indices for various materials
  • Investigate the implications of wave-particle duality on electromagnetic wave behavior
USEFUL FOR

Physicists, optical engineers, and students studying wave optics or electromagnetic theory will benefit from this discussion, particularly those interested in the behavior of light and radio waves in different media.

ajdecker1022
Messages
10
Reaction score
0
I'm wondering about small changes of phase due to a refractive medium. For example, suppose there is an emitter of radio or light waves, and two detectors equidistant from the emitter. One of the detectors is behind a medium, while the other is a straight shot through a vacuum.

How would the difference in the detectors evolve over time? Would it just be a constant phase shift, or would the angular frequency of the refracted wave change?
How can I think about this properly?

Thanks in advance.
 
Science news on Phys.org
Draw a diagram ... if one wave is subject to a refractive medium it will have a change of wavelength in that medium because v=c/n, but the frequency will not change.

The one through vacuum has n=1, so v=c and there is no refraction.
 
Hmm, perhaps my question was not explained well. The detector behind the medium is not in the medium. I know that the wavelength and the velocity will change, but I'm not sure whether the detected phase change will by proportional to time or constant. My confusion is coming from two perspectives:

1. As an analogy, the speed of the convoy is the slowest ship. Since the speed of light is slowed through the medium, all of the light going through the medium should travel at a slower rate. Let's say that that in the vacuum, n = 2. Then c / 2 is proportional to the number of photons (regardless of frequency in the spectrum?) hitting the detector. It seems like the detector behind the medium will detect photons half as as fast, which in turn seems to imply that somehow the frequency has changed?

Number of photons = c * k1 * t
Number of photons = c / 2 * k1 * t
where k1 is some proportion relating speed of light and number of photons. As time goes on, the difference in photons detected at each receiver will increase.

2. Since the frequency is constant and determined by the emitter, the frequency should not change. Thus, there should be a constant difference of photons detected.

Number of photons = c * k1 * t
Number of photons = c * k1 * (t - delta)
where delta is the constant shift.

Any advice is appreciated.

EDIT: Additionally, I'm not sure how to think about it because light can be thought of as a particle or a wave. Does the particle model apply to radio waves as well?
 
The frequency never changes while passing through a medium - it is always the wavelength.

Radio and light are both electromagnetic waves, so the theory is the same, but have many practical differences.
One is that the index of refraction barely exists for radio waves because almost everything is through air ... hence n=1.

But using light you can use the "optical path length" technique: http://en.wikipedia.org/wiki/Optical_path_length

This should provide your answer.
 
UltrafastPED said:
The frequency never changes while passing through a medium - it is always the wavelength.

Radio and light are both electromagnetic waves, so the theory is the same, but have many practical differences.
One is that the index of refraction barely exists for radio waves because almost everything is through air ... hence n=1.

But using light you can use the "optical path length" technique: http://en.wikipedia.org/wiki/Optical_path_length

This should provide your answer.

What is the description of the mechanism of how the wavelength changes in the medium ?
The geometry was first shown by Huygen : λ2 = λ v/c
I understand that the frequency has to be the same because of the boundary conditions.
So the apparent slowing of light transmission in refraction is because of a shorter wavelength .Im not looking for a geometric optics explanation. Rather the exact physical mechanism at the micro scale in terms of the interaction of the time varying electric field of the EM wave with the electrons in the medium.
 
Last edited:
Would as 'classical' explanation satisfy you? You could think in terms of a distribution of charges with mass, loading the wave in its progress and delaying the perturbations as the wave progresses through the medium as they are forced to oscillate by the impressed varying EM fields and then re radiate a bit later.
 

Similar threads

  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 10 ·
Replies
10
Views
3K
  • · Replies 41 ·
2
Replies
41
Views
5K
  • · Replies 1 ·
Replies
1
Views
5K
  • · Replies 2 ·
Replies
2
Views
1K
  • · Replies 4 ·
Replies
4
Views
4K
  • · Replies 16 ·
Replies
16
Views
4K
  • · Replies 4 ·
Replies
4
Views
3K