Understanding Fourier Optics: From Spatial to Temporal Transformations

AI Thread Summary
Fourier optics involves spatial transformations where the image formed by a lens system is a scaled Fourier transform of the object, directly observable on a screen. Unlike temporal Fourier transforms, which require instruments to observe frequency spectra, spatial transformations maintain physical presence in space. The Fourier transform is complex-valued, with intensity recorded on film, while phase information is crucial for understanding wavefront behavior, though not directly measurable with standard equipment. Fourier optics is primarily applicable in far-field scenarios, and its assumptions may not hold in near-field conditions or with complex optical systems. Practical challenges such as aliasing and loss of shift invariance arise with CCD detectors, but some analysis remains feasible.
KFC
Messages
477
Reaction score
4
I am reading some introduction on Fourier optics. In the text, the simplest example would be a single len system (focal length f) with an object f(x,y) sitting on the front focal plane of the len while the image is the corresponding spatial Fourier transformation F(\frac{x}{\lambda f}, \frac{y}{\lambda f}) where \lambda is the wavelength.

This spatial transformation is quit confusing while being compared temporal case. In temporal case, Fourier transform of f(t) is given as F(\omega) which is defined in frequency domain. Hence, we cannot directly observe that in physical space, we need to put F(\omega) into some instrument to observe the frequency spectrum. But
for the spatial case, it seems that the Fourier transform is nothing but just a rearrangement or mapping and scaling from the source to the image, but anyway the image are still in the physical space because the image can be directly observed on the
rear focal plane with screen or film. I wonder if my understand is correct?

Another question on the complex function. In most cases, the Fourier transform will be complex-valued. And for len system, F(\frac{x}{\lambda f}, \frac{y}{\lambda f}) could also be complex, but I think physically the film will only record the intensity instead. But what about the phase? What's the role of the phase in this case? Or I ask in this way. If I keep the intensity for every point on the image unchanged but I shift the phase by a number or randomly change the phase, how difference the image will be? in what way?

Thanks.
 
Science news on Phys.org
KFC said:
But
for the spatial case, it seems that the Fourier transform is nothing but just a rearrangement or mapping and scaling from the source to the image, but anyway the image are still in the physical space because the image can be directly observed on the
rear focal plane with screen or film. I wonder if my understand is correct?

Yes... what's happening is that what is *physically* on the screen is a scaled Fourier transform, with the scaling depending on the distance to the screen and wavelength. (The frequency in the Fourier transform here isn't a temporal frequency... it's a spatial frequency.)

KFC said:
Another question on the complex function. In most cases, the Fourier transform will be complex-valued. And for len system, F(\frac{x}{\lambda f}, \frac{y}{\lambda f}) could also be complex, but I think physically the film will only record the intensity instead. But what about the phase?

The phase is exactly what you would expect... if you had a measurement device at the screen that was capable of precisely measuring the electric field in the light, you would find that the fields at two different points on the screen would be out of phase by the amount predicted by the imaginary part of the Fourier transform.

Films and screens can't measure to this precision though, so we end up with the intensity instead, as you noted.
 
KFC said:
I am reading some introduction on Fourier optics.
<snip>

This spatial transformation is quit confusing while being compared temporal case.
<snip> it seems that the Fourier transform is nothing but just a rearrangement or mapping and scaling from the source to the image, but anyway the image are still in the physical space because the image can be directly observed on the
rear focal plane with screen or film. I wonder if my understand is correct?

Another question on the complex function. In most cases, the Fourier transform will be complex-valued. And for len system, F(\frac{x}{\lambda f}, \frac{y}{\lambda f}) could also be complex, but I think physically the film will only record the intensity instead. But what about the phase? What's the role of the phase in this case? Or I ask in this way. If I keep the intensity for every point on the image unchanged but I shift the phase by a number or randomly change the phase, how difference the image will be? in what way?

Thanks.

Good questions. First, the spatial transform: if you like, the fourier-conjugate variables are *angles*: absent a screen, the far-field diffraction pattern (the Fourier transform of the aperture) is in terms of angle. The scaling factors relate the image size (in distance) to the angles when a screen is present.

Second, you are correct- the Fourier transform/point spread function/optical transfer function is complex; (nonabsorptive) aberrations are modeled as changes to the phase of the wavefront. This gets into the difference between coherent and incoherent imaging, the reason why the cutoff frequency for imaging is either given by the pupil function directly or the autocorrelation of the pupil function, differences between optical transfer function and modulation transfer function, etc.

An instructive example is to work the Fourier transform for two pure sine apertures: 1 in amplitude (transmission) only and 1 in phase only. The far-field diffraction pattern is quite different.
 
Thanks both of you. That explains what I doubt. And it makes sense that "far-field diffraction pattern (the Fourier transform of the aperture) is in terms of angle", thanks Andy.

I still have one question. In the text, it either use single len system as example or consider a field propagating in free space. In both case, far field sitiutation is considered. So does it mean Fourier Optics in only valid in case of far field? I am curious how to apply Fourier transformation if I place a very complicated "transparency" or media in-between object and image!?
 
Fourier optics analysis is based on the assumption of a linear shift-invariant system. So you can model a very complex optical system under some conditions- Fourier optics will fail for the near field, for example, and must be extended to a vector formalism when the numerical aperture becomes large.

But perhaps the most practical complication is the use of CCD detector arrays- then shift invariance is lost. Aliasing, spurious frequency responses, etc. will occur. However, some analysis is still possible.
 
Thanks a lot. I do learn something :)

Andy Resnick said:
Fourier optics analysis is based on the assumption of a linear shift-invariant system. So you can model a very complex optical system under some conditions- Fourier optics will fail for the near field, for example, and must be extended to a vector formalism when the numerical aperture becomes large.

But perhaps the most practical complication is the use of CCD detector arrays- then shift invariance is lost. Aliasing, spurious frequency responses, etc. will occur. However, some analysis is still possible.
 
Thread 'Simple math model for a Particle Image Velocimetry system'
Hello togehter, I am new to this forum and hope this post followed all the guidelines here (I tried to summarized my issue as clean as possible, two pictures are attached). I would appreciate every help: I am doing research on a Particle Image Velocimetry (PIV) system. For this I want to set a simple math model for the system. I hope you can help me out. Regarding this I have 2 main Questions. 1. I am trying to find a math model which is describing what is happening in a simple Particle...
I would like to use a pentaprism with some amount of magnification. The pentaprism will be used to reflect a real image at 90 degrees angle but I also want the reflected image to appear larger. The distance between the prism and the real image is about 70cm. The pentaprism has two reflecting sides (surfaces) with mirrored coating and two refracting sides. I understand that one of the four sides needs to be curved (spherical curvature) to achieve the magnification effect. But which of the...
Back
Top