Undergrad Factors that influence depth of field

  • Thread starter Thread starter Oldhouse
  • Start date Start date
  • Tags Tags
    Dof Photography
Click For Summary
The discussion centers on the factors influencing depth of field (DOF) in photography and videography, particularly the claims made in a YouTube video regarding sensor size and image projection. The video asserts that sensor format does not affect DOF, which some participants challenge by emphasizing the importance of the circle of confusion and how image size impacts perceived focus when displayed. Key points include that the size of the entrance pupil, distance to the subject, and viewing conditions all play significant roles in determining DOF. Participants also highlight that while the video may provide a basic understanding, it lacks depth in explaining the physics behind DOF. Overall, the conversation seeks clarity on the accurate factors that influence DOF in various contexts.
  • #31
sophiecentaur said:
OK. It's good to hear from someone with actual experience. Perhaps you could give me an example of how that definition (or any definition) is actually implemented (objectively) in practice, say when setting up shots of a scene with two different focus settings, moving from one to another pat of the scene with the camera static. What numbers are used? How is 'focus' checked and also the amount of de-focus in another part of the scene. Does the cinematographer actually measure it or does he/she look at the ranges involved? There will be studio monitors which the director can look at and make a decision at the time, using the shots. But how are the choices of apertures made, other than from past experience? What 'calculations' are done?
At a more general level, I do know that TV vision suites use a monitor which shows the resulting output from the studio. This has to involve the sum of artefacts from scene to broadcast signal. I know that, in the days when (analogue) Colour studios were required to produce a Compatible monochrome output, there was also a mono monitor. In sports programmes, the resolution in the compatible mono picture was what determined the field of the colour cameras. When they finally stopped using 405 lines, there were more players in shot. And so on to 4K.

"What DOF is, is clearly defined/established". Could you be more detailed here and tell me an example with definition and numbers - say when assessing when two objects are both sufficiently well focussed AND when the differential focus between two objects is sufficient. As a creative, I would imagine you would have been through that process. Are you old enough to remember using tape measures for film? There must have been many rules of thumb.

Having looked into astrophotography, i am aware that ther Airey Disc is used when deciding on resolution between to point objects - but the situation is limited to the objective lens and the sensor array and we don't go for de-focus very often. Everything is at infinity, too.
Generally speaking, the following applies to narrative filmmaking: You usually light a scene for a specific f-stop (whole scene will be shot at that same f-stop). The director of photography might decide that he wants to shoot the scene at f/4 and instructs his crew about the light levels for individual areas of the scene (example, key light on actor at f/4, fill light 2 stops under, BG 3 stops under etc.). The f-stop chosen is usually a creative choice based on experience from previous shoots. The director is often also involved in this choice because it can have a major influence on the overall look of the film. Calculations on DOF aren't really done at this stage. This is just all from experience and looking at reference images with the director.
While filming, focus is adjusted by the 1st AC. During rehearsals, he will use a tape measure or laser range finder to measure the distance from the film plane to the actor/important objects. Usually, he will put marks on the floor where actors are standing etc. He will also mark the follow focus or lens scale directly so he can quickly change focus from one mark to another. As the scene plays out, he will manually focus the lens, using the distance marks on the floor and on the follow focus/lens to keep the object/actor of interest in focus. This takes a lot of practice to do well. It becomes increasingly difficult the more movement you have in the scene (actor might move, as well as the camera). It is actually pretty rare that you calculate your DOF.
A typical instance would be if you are doing a car scene with somebody in the front seat and somebody in the back seat and the director wants both actors in focus at the same time. In the early days, you would consult DOF tables to figure out the DOF you have available. These days you use an app like PCam. For the car scene example, you might decide to put the focus point somewhere in between the two actors (can be risky but might be the only option depending on the available DOF). BTW, those DOF tables are written with a specific CoC in mind, and if you have an app like PCam you can even adjust the CoC used for the calculation (for each format, there is a recommended size for the CoC that has been calculated so it gives you a somewhat accurate idea of how much DOF you will have if the film will be screened on an "average sized cinema screen". For example, the established permissible coc for shooting on s35 is 21.08µ. Some are currently debating if this has to be changed because we have sharper lenses available and higher resolution, better projections standards etc.). If you are shooting for iMax or some other "out of the ordinary screen" you might have to adjust the permissible CoC for more accurate results. Another instance where you might check how much DOF you have available, is in low light conditions, especially if there is a lot of movement which makes the job hard for the 1st AC. If you know that you only have a few inches of DOF, you might put some extra focus marks down or in some extreme cases (when shooting extreme close-ups), the actor might have to be informed that he has to keep his movements to a minimum (this happens quite often on low budget shots that don't have large lighting packages and run out of daylight). BTW, the 1st AC usually doesn't check the focus on a screen (the screen on the camera is often way too small to do that and in some cases there isn't even a screen available). When shooting film (actual film), the only person that can actually judge if the "focus was on" is the operator that looks through the camera. The video tap (beamsplitter with a low resolution videocamera in the optical path of the viewfinder) only gives you a reference picture for framing. With digital cameras, you could pull focus of a production monitor but it usually isn't done that way. In order to predict the actors movement better, it is easier to just stand next to the camera and pull focus based on the distances you marked during rehearsals (if you pull focus of a monitor, chances are that you are always one step behind... if you see the image going out of focus it is already too late). Hope that all makes sense.

As mentioned, the cases where you actually calculate your available DOF are relatively rare. Most movies are shot between f/2.8 and f/5.6 (s35 format). In most regular shooting conditions, you have plenty of DOF to keep your object of interest in focus. When it comes to some more specialty fields like filming miniatures or filming stuff for visual effects, DOF calculations become very important (Can't really talk much about those fields as my knowledge in them is mainly theoretical).

Sorry for the super long post...
 
  • Like
Likes sophiecentaur
Science news on Phys.org
  • #32
Oldhouse said:
We are going in circles here.... My original question is precisely that: Are the statements true if we use the established definition of DOF?
Based on my limited knowledge and what we've posted here in-thread, no, the statements saying that DOF only depends on the properties of the optical system and not the size of the sensor are not correct in general, as it ignores how the image will end up being displayed.
 
  • #33
Drakkith said:
Based on my limited knowledge and what we've posted here in-thread, no, the statements saying that DOF only depends on the properties of the optical system and not the size of the sensor are not correct in general, as it ignores how the image will end up being displayed.
Thanks. I know it might sound as if I just wanted somebody to confirm that I'm right for the shake of being right. This is not the case; after having a long discussion with the guy in the video, we basically agreed on all the technical aspects, but he was still insisting that everything that is said in the video is correct. This genuinely confused/irritated me as I couldn't figure out how he could come to that conclusion after we agreed on all the technicalities. I then started to doubt myself... maybe I'm just not properly understanding/interpreting what is actually being said in the video (English is not my native tongue). It just bothered me... BTW, I also posted the same question on https://photo.stackexchange.com/ and it looks like somebody was able to pinpoint the issue at hand. His explanation makes a lot of sense and adds some additional context.

Here is the answer from Steven Kersting:

Your understanding is correct; the author is confusing depth of focus with depth of field... it's a fairly common confusion as the two are interrelated, but only one is common terminology.

The depth of focus is the relative sharpness of details at the image plane, and depth of focus is a fixed characteristic of an image. It is the non-variable component of depth of field; and it is dictated only by the physical size of the aperture opening/entrance pupil (not the f/#).

Depth of field is not a fixed aspect of an image; so when people talk about depth of field as a fixed aspect, they are really talking about the depth of focus.

Depth of field is dictated only by magnification... it is how apparent the depth of focus is made to the viewer. Magnification includes all of the other variables... focal length, subject distance, sensor area/cropping/enlargement, viewing distance, and even the viewer's visual acuity.

If the final/total magnification causes the image to have the same composition and relative size to the viewer, then the depth of field will remain the same and be dependent only on the depth of focus recorded. That's why the standard for image sharpness (and the calculators), assumes a standard viewing distance approximately equal to the image diagonal; and to where the image occupies the human's ~ 45˚ primary field of view.

source: https://photo.stackexchange.com/que...-dof-and-the-factors-that-influence-it#134817
 
  • Like
Likes Drakkith
  • #34
Oldhouse said:
Are the statements true if we use the established definition of DOF?
The definition of DoF relies on the definition of the CoC which is rather fuzzy and application dependent:
https://en.wikipedia.org/wiki/Circle_of_confusion
wikipedia said:
The CoC limit can be specified on a final image (e.g. a print) or on the original image (on film or image sensor).
If you define the CoC based on the original image (on film/sensor), then it is independent of the display parameters (print/screen size and viewing distance). So I would suggest to clarify how the author of the video defines the CoC.
 
  • Like
Likes Drakkith and Oldhouse
  • #35
A.T. said:
If you define the CoC based on the original image (on film/sensor), then it is independent of the display parameters (print/screen size and viewing distance).
Ah, I missed this when I read that wikipedia article. Defining the CoC in reference to a spot size on the sensor itself is more in line with what we deal with in astrophotography.
 
  • #36
Oldhouse said:
Your understanding is correct; the author is confusing depth of focus with depth of field... it's a fairly common confusion as the two are interrelated, but only one is common terminology.

The depth of focus is the relative sharpness of details at the image plane, and depth of focus is a fixed characteristic of an image. It is the non-variable component of depth of field; and it is dictated only by the physical size of the aperture opening/entrance pupil (not the f/#).

Depth of field is not a fixed aspect of an image; so when people talk about depth of field as a fixed aspect, they are really talking about the depth of focus.

Depth of field is dictated only by magnification... it is how apparent the depth of focus is made to the viewer. Magnification includes all of the other variables... focal length, subject distance, sensor area/cropping/enlargement, viewing distance, and even the viewer's visual acuity.
Be careful... that's only partially correct.

"Depth of focus" refers to the allowed mechanical tolerance of placing the sensor at the plane of best focus. Therefore, it is independent of the sensor.

"Depth of field" refers to the range of object distances imaged 'in focus' at the image plane, and depends on all the stuff we have been discussing here.

Unfortunately, many people use the terms interchangeably.
 
  • Informative
Likes berkeman
  • #37
A.T. said:
The definition of DoF relies on the definition of the CoC which is rather fuzzy and application dependent:
https://en.wikipedia.org/wiki/Circle_of_confusion

If you define the CoC based on the original image (on film/sensor), then it is independent of the display parameters (print/screen size and viewing distance). So I would suggest to clarify how the author of the video defines the CoC.
Yes, of course it doesn't matter if you define the CoC at the sensor or on the final printed image. However, when defining the CoC, you have to take the final viewing conditions into account (no matter if you define the permissible CoC on the sensor or in the final print), else you end up with a completely useless measure. The CoC (no matter if defined on the sensor or the printed image) is usually chosen so when the picture is viewed, a "singular point" is still perceived as a "singular point" and not as a blurry point.
 
  • #38
Oldhouse said:
However, when defining the CoC, you have to take the final viewing conditions into account (...), else you end up with a completely useless measure.
I can see some uses for it:

If the final viewing conditions are unknown, one still might want to compare different optics in terms of DoF, for example based on a CoC defined as a fraction of the original image size.

If the footage is used for automated digital image processing that reads the sensor directly, one still might need to know the DoF, for example based the a CoC in terms of sensor pixels.

In any case, the definition of the CoC is just too fuzzy, to argue about right and wrong.
 
  • #39
A.T. said:
I can see some uses for it:

If the final viewing conditions are unknown, one still might want to compare different optics in terms of DoF, for example based on a CoC defined as a fraction of the original image size.

If the footage is used for automated digital image processing that reads the sensor directly, one still might need to know the DoF, for example based the a CoC in terms of sensor pixels.

In any case, the definition of the CoC is just too fuzzy, to argue about right and wrong.
I probably wouldn't call this comparing optics in terms of DoF, but just comparing optics in terms of the the CoC as a fraction of the sensor size (which is how the "Zeiss Formula" defines the CoC). This now becomes all semantics... I agree, not really a point in arguing over it.
 
  • #40
Oldhouse said:
I probably wouldn't call this comparing optics in terms of DoF, but just comparing optics in terms of the the CoC as a fraction of the sensor size (which is how the "Zeiss Formula" defines the CoC).
The Zeiss Formula gives you the acceptable CoC based on sensor size, which you then can plug into the DoF formula to compute and compare the DoF for different optics. None of that requires knowledge of the actual viewing conditions.
 
  • #41
A.T. said:
The Zeiss Formula gives you the acceptable CoC based on sensor size, which you then can plug into the DoF formula to compute and compare the DoF for different optics. None of that requires knowledge of the actual viewing conditions.
No, you forget that the Zeiss Formula was derived from the DOF markings of an Zeiss Triotar lens, and the DOF markings were put on the lens with a CoC calcululated with a viewing distance that is equal to the picture diagonal. I'm done talking about this... We are going in circles.
 
  • Like
Likes sophiecentaur
  • #42
Oldhouse said:
No, you forget that the Zeiss Formula was derived from the DOF markings of an Zeiss Triotar lens, and the DOF markings were put on the lens with a CoC calcululated with a viewing distance that is equal to the picture diagonal.
Of course the formula was defined based on some practical consideration and not just a random ratio. But when it is applied, you don't need any data on the viewing conditions, just the sensor size. Also, for relative comparison of two optical systems it doesn't matter if you use the Zeiss-CoC or say twice the value, because it's about the ratio of the resulting DoF.

And for automated digital image processing there is no human viewing intended at all. Some algorithm reads the pixel values directly, but is limited by the CoC in terms of pixels, which then gives you the limits in terms of DoF.
 
  • #43
Oldhouse said:
TL;DR Summary: What are the actual factors that influence depth of field in photography?

The DOF we get in a picture is different if we record it on a small piece of film or on a large piece of film because we need to enlarge the small piece of film more than we enlarge the large piece of film to display it on our TV. Therefore the statement "it doesn't matter how big the wall is, it hasn't changed the depth of field" because "the image has already been formed, as it goes through the lens, before it hits the sensor" seams to be nonsensical.

then you misunderstand depth of field

The depth of focus is the relative sharpness of details at the image plane, and depth of focus is a fixed characteristic of an image. It is the non-variable component of depth of field; and it is dictated only by the physical size of the aperture opening/entrance pupil (not the f/#).

Depth of field is not a fixed aspect of an image; so when people talk about depth of field as a fixed aspect, they are really talking about the depth of focus.

I agree with Andy, having done photography for more years than I can to remember.

Depth of field and depth of focus are essentially the same thing, and the guy's comments in the video are correct when stating that the sensor size is irrelevant.

Dave
 
  • Like
Likes Gleb1964
  • #44
davenn said:
then you misunderstand depth of field



I agree with Andy, having done photography for more years than I can to remember.

Depth of field and depth of focus are essentially the same thing, and the guy's comments in the video are correct when stating that the sensor size is irrelevant.

Dave
I would say you misunderstand depth of field. In post #15, another member even posted the formula which makes it obvious that sensor size has an influence on DOF. Furthermore, you can simply use any depth of field calculator (for example: https://dofsimulator.net/en/) and test it yourself. Leave all parameters untouched except for the sensor size and you will see yourself that DOF changes.
 
  • #45
At the risk of beating a dead horse, I thought of a different perspective that might clarify how the size of the image (a non-subjective measure) can impact the depth of field (a subjective measure*).

* it is possible in some circumstances to quantify a maximum permissible defocus error; in fact this calculation is required for machine vision applications.

Specifically, consider display technology. In this analogy, the minimum size of a printed 'dot' or a display pixel approximately corresponds to the size of a circle of confusion. In this analogy, complexities of optical systems and image recording are not relevant; additionally, a quantitative metric of interest (dots per inch or pixel pitch) is always normalized to the subjective nature of human vision and the "optimal" viewing distance.

I looked up the specifications for the following and as best I could, converted everything to dots per inch (dpi):

Las Vegas Sphere (outside): 0.8 dpi
The Humungotron (located in my hometown): 100 dpi
Typical roadside billboard displays: 20 dpi
desktop/laptop monitor(**): 100-200 dpi
phone screen: 300-600 dpi
high-end printing services: 600-1200 dpi

(**) :) if you are looking at this on a specialized 8k or 12k display, you should already understand CoC so why are you wasting your time reading this? :)

The important observation is that for each of these, the image is considered "sufficiently sharp". For example, the ratio of laptop and phone dpi is nearly equal to the ratio of viewing distances for the two, showing how viewing distance (image size) impacts depth of field.

I think this analogy helps understand how depth of field (which depends on the size of the circle of confusion) scales with image size, which is a concept that can seem nonsensical or at least counter-intuitive.
 
  • #46
Drakkith said:
That assumption is the crux of the issue. If both images were displayed on the same monitor but at their native resolutions then the DOF of both would be the same. But since there's an assumption that the second image will be magnified to be displayed at the same size as the first then we must admit that this will reduce the DOF since it involves magnifying the image and all the blurring inherent in that image.
To this thread I chime in with my own skepticism/misapprehensions.

In my ancient world, DoF is about the maximum physical distance between objects in focus.

Stick a stake in the ground every foot from camera to horizon. Near stakes will be out of focus, far ones too. DoF defines the one in focus.

That is determined in-camera.

I fail to see how print resizing or DPI of printout will change which stakes are in focus and which stakes are out of focus. OK, unless a distant stake is only 2 printer dots wide...

Likewise, I can see how sensor size might cause some antialiasing artifacts that will overlap with DoF, since a distant stake that's only 2px wide will seem out of focus


... but I wouldn't have confabulated DPI or PPI with DoF. That seems sloppy to me.
 
  • #47
DaveC426913 said:
That is determined in-camera.

I fail to see how print resizing or DPI of printout will change which stakes are in focus and which stakes are out of focus. OK, unless a distant stake is only 2 printer dots wide...

It is not. Go to post #8 and read the posted link for a pretty easy to understand explanation.
 
  • #48
Oldhouse said:
It is not. Go to post #8 and read the posted link for a pretty easy to understand explanation.
OK. so I was correct. Today's DoF is a mixture of what I will call 'optical DoF' versus 'perceptual DoF'.

"Despite the debate, there is a standard in the photography industry : when looking at a print of 17cm*25cm at an optimum viewing distance of 25cm, a blur circle of less than 0.2mm diameter is seen as a dot and not a circle anymore. This is the diameter of the Circle of Confusion, the largest circle still perceived as a sharp point by the human eye. By default, the DoF is defined relatively to this degree of sharpness."

IOW, if it looks blurry (for whatever reason); it is declared to be blurry.

I still say this seems sloppy. Or at least application-specific. (By that I mean, it is less important what the actual causes of DoF are than what the effects/consequences of it are in the final format.) Pjut another way, it;s become a practical, engineer-y factor, rather than a theoretical science-y factor. A loss of data there.

The implication, as I see it, is that, by this definition, DoF can be affected by almost innumerable subjective, spurious artefactal factors:
  • looking at a photo in dim light will affect its DoF (because: pupils)
  • looking at a photo that has been printed with poor colour-calibraton on an inkjet printer will affect its DoF (by smearing out the dots. All the dots.)
  • looking at a photo through a translucent sheet of paper will change the DoF, because the smallest discernible dot-as-opposed-to-circle is now huge.
  • There is literally no limit to number of ways and conditons that can potentially alter the size of the smallest discernible dot in a picture when viewed after-the-fact.

These factors above affect the entire image equally - the entire pic equally - regardless of pic subject. (Every dot is enlarged by the same factor, it's just that some dots cross the CoC threshold and some do not.)

But we already have a term for how a given image's overall sharpness is defined. It's called sharpness.

Why redefine a term that already had a specific useful meaning? DoF used to be about a differential blurriness between disparate parts of the same photo's semantic subject (eg. foreground vs background), as oppsoed to the photo as an item. Does that still have a name?

Before you pound on me: This is a rant. I acknowledge the information inherent in this definition of DoF is useful to the modern photographer; I just don't know why they need to muddle perfectly cromulant technical terms.

I'm an old school photographer, and for me, DoF is an in-camera factor (though it can be rendered in post - by differentially changing the dot sizes). I am perfectly happy to talk about dot-size and sensor size and print rez and all those things. I'm just happy to use their specific technical labels so that they do not get lost the the melting pot with other similar, yet distinct, technical artefacts.


Question: did such a definition for DoF precede or follow the advent of digital photography?
 
Last edited:
  • #49
DaveC426913 said:
I fail to see how print resizing or DPI of printout will change which stakes are in focus and which stakes are out of focus. OK, unless a distant stake is only 2 printer dots wide...
Blow up the photograph to a poster that's 20 feet wide and 40 feet tall and viewed from 1 foot away. Now none of the field appears to be in focus.
 
  • #50
Drakkith said:
Blow up the photograph to a poster that's 20 feet wide and 40 feet tall and viewed from 1 foot away. Now none of the field appears to be in focus.
No it doesn't. Out-of-focus is distinctive*. The giant poster appears in-focus but no longer sharp.

* For example, bokeh is a depth-of-focus-effect and depends on actal distance in image, not on image-wide dot-size.

We already have terms that fit the phenomena.

Sharpness affects every element in the image, whether 2 inches or 200 miles. in the aboe example they're all one inch diameter dots. Even the Moon, 200,000 miles away. Why call it depth of field? It makes no sense.

In this definition, your 20x40 photo has a DoF of exactly zero. Nothing is "inside the field of focus". That's a nonsensical outcome.

To-wit:

Man at 20 feet: "I love the photographer's use of depth of field. The bees are in focus while the field of flowers is not."
Man at one foot: "False. The depth of field has collapsed and vanished. This is the worst exhibit ever."

🤔
 
Last edited:
  • #51
DaveC426913 said:
False. It appears in-focus but no longer sharp.
Those are essentially the same thing. Besides, at least half or more of the descriptions or definitions I can find of depth of field says something similar to: the size of the area in your image where objects appear acceptably sharp.
 
  • #52
DaveC426913 said:
In this definition, your 20x40 photo has a DoF of exactly zero. Nothing is "inside the field of focus". That's a nonsensical outcome.
Of course it's a sensible outcome. A badly misaligned camera lens can produce photos that never look sharp unless you shrink down the image to where it is very small. If you viewed a picture from this camera at normal size on your monitor you would say that the entire image is out of focus. If the entire image is out of focus, or unsharp, then you could certainly say it has zero depth of field.
 
  • #53
Drakkith said:
Those are essentially the same thing.
They're not. One is an subject property. It affects different parts of the (conceptual) image.
The other is a material property. It affects the entire (actual) image unilaterally.

In the subject image with small DoF, two areas arbirarily close together, can have widely varying degrees of blurriness (say, a hair, seen against a field).

Halving the DPI of the print doesn't prefentially affect the bg focus and ignore the foreground.

Drakkith said:
Besides, at least half or more of the descriptions or definitions I can find of depth of field says something similar to: the size of the area in your image where objects appear acceptably sharp.
"area"?? It's a distance, strictly radially, in the subject, from the lens.
 
  • #54
Drakkith said:
Of course it's a sensible outcome. A badly misaligned camera lens can produce photos that never look sharp unless you shrink down the image to where it is very small. If you viewed a picture from this camera at normal size on your monitor you would say that the entire image is out of focus. If the entire image is out of focus, or unsharp, then you could certainly say it has zero depth of field.
Sure. If the entire pic was shot out of focus, then it follows that the depth of field is zero.
But if I view a sharp pic through a piece of wax paper I can declare, with a straight face, "this photograph has a zero DoF"?


Again:

Man at 20 feet: "I love the photographer's use of depth of field. The bees are in focus while the field of flowers is not."
Man at one foot: "False. The depth of field has collapsed and vanished. This is the worst exhibit ever."


Likewise:

1714703442228.png

Left pic: high depth-of-field
Right pic: zero depth of field
?

1714703529393.png


Even worse:
Left pic: high depth-of-field
Right pic: zero depth of field
?
(I guess we'll just have to invent a new term to describe pics where the small range of distances in the image are in-focus while other ranges are not? A pity. The technique - which has a front-row role in creative photography - is now nameless.) 🤔
 
Last edited:
  • #55
@DaveC426913 You are getting this completely wrong. Please define "sharpness". If possible, what would the formula be? What permissible size circle of confusion are you choosing, and what is the process of settling on a specific size.
 
  • #56
DaveC426913 said:
Likewise:

View attachment 344410
Left pic: high depth-of-field
Right pic: zero depth of field
?

View attachment 344412

Even worse:
Left pic: high depth-of-field
Right pic: zero depth of field
Yes, it is indeed true that the right picture has less DOF than the left... absolutely!
 
  • #57
DaveC426913 said:
I still say this seems sloppy. Or at least application-specific. (By that I mean, it is less important what the actual causes of DoF are than what the effects/consequences of it are in the final format.) Pjut another way, it;s become a practical, engineer-y factor, rather than a theoretical science-y factor. A loss of data there.

It is not sloppy at all. It is like that by definition:
Drakkith said:
After looking into this more, the issue boils down to how DOF is defined:

##DOF \approx \frac{2u^2Nc}{f^2}##

Here ##u## is distance to subject, ##N## is the f-number of the system, ##c## is the acceptable circle of confusion, and ##f## is the focal length of the system. The value ##c##, the maximum acceptable circle of confusion, is what is important to understand here. It turns out that the circle of confusion DOES change as sensor size changes. Per wikipedia:

Image sensor size affects DOF in counterintuitive ways. Because the circle of confusion is directly tied to the sensor size, decreasing the size of the sensor while holding focal length and aperture constant will decrease the depth of field (by the crop factor). The resulting image however will have a different field of view. If the focal length is altered to maintain the field of view, the change in focal length will counter the decrease of DOF from the smaller sensor and increase the depth of field (also by the crop factor).

This also makes total sense when it comes to practical applications: If I film two people standing at different distance to the camera, and if I want them both to be in focus, I need to know if the film will be shown in an iMax theater or just on a home TV. Else it can happen that the two actors are in focus (sharp) on the home TV but when shown in iMax, one of the actors is out of focus.
 
  • Like
Likes hutchphd
  • #58
Oldhouse said:
Yes, it is indeed true that the right picture has less DOF than the left... absolutely!
That is insane*.

No. That's happy.
I have decided that 'happy' has been redefined to mean something new. The concept-formerly-known-as-insane now has no word to describe it until a new term is invented. , no matter how useful the CFKaI remains.

This doesn't have to be logically defensible. As long as I get enough people to agree with me (argumentum ad populum) I can get any word repurposed. And thus, any concept ... *ahem* double plus ** unexpressable.


*no personal insult intended

**pointed literary reference
 
  • #59
Oldhouse said:
This also makes total sense when it comes to practical applications: If I film two people standing at different distance to the camera, and if I want them both to be in focus, I need to know if the film will be shown in an iMax theater or just on a home TV. Else it can happen that the two actors are in focus (sharp) on the home TV but when shown in iMax, one of the actors is out of focus.
Categorically and demonstrably false.

1714704870734.png

1714705196785.png


The foreground and background of this (or any similar) image will have the variances in depth of field between fg and bg preserved from a 4-inch phone screen to a 400-foot silver screen.
 
  • #60
DaveC426913 said:
That is insane*.

No. That's happy.
I have decided that 'happy' has been redefined to mean something new. The concept-formerly-known-as-insane now has no word to describe it until a new term is invented. , no matter how useful the CFKaI remains.

This doesn't have to be logically defensible. As long as I get enough people to agree with me (argumentum ad populum) I can get any word repurposed. And thus, any concept ... *ahem* double plus ** unexpressable.

Ok, let's break this down, and you tell me how that is illogical:
See attached image for the formula of DOF.

"c" stands for the circle of confusion.

Now lets look at how the permissible CoC is defined:

"In photography, the circle of confusion diameter limit (CoC limit or CoC criterion) is often defined as the largest blur spot that will still be perceived by the human eye as a point, when viewed on a final image from a standard viewing distance."
https://en.wikipedia.org/wiki/Circle_of_confusion

The important part here is "largest blur spot that will still be perceived by the human eye as a point".

Now please tell me how you can add grain, or noise without changing the "blur spot"?
If you scale the image, change the viewing distance, alter the resolution etc. you obviously change the quality/size of the "blur spots". Therefore you change DOF.

If you still can't wrap your head around it... just go to https://dofsimulator.net/en/ and change the sensor size while leaving everything else the same.... DOF changes as you see. So why is that in your opinion?
 

Attachments

  • Screen Shot 2024-05-03 at 4.55.20 AM.png
    Screen Shot 2024-05-03 at 4.55.20 AM.png
    3.3 KB · Views: 56

Similar threads

  • · Replies 14 ·
Replies
14
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 25 ·
Replies
25
Views
3K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 226 ·
8
Replies
226
Views
16K
Replies
35
Views
10K
Replies
5
Views
8K
  • · Replies 3 ·
Replies
3
Views
5K
  • · Replies 1 ·
Replies
1
Views
3K
  • · Replies 47 ·
2
Replies
47
Views
4K