Hey, I was just doing some questions on thin lenses, and thought of a question I really want answered just can't find anything on in the text or anywhere.
So the question involves a lens with a focal length of f = 18.18 cm
now the question asks if the lens would be able to produce an image on a screen that is 35 meters away. Now I can solve this by
1/f - 1/q = 1/p
stating that if you place the object 18.27 cm from the lens you can get an object at 35 meters that will be magnified and inverted, correct ?
now my question is.. say we make the screen 300 meters away
this equation shows that you will still get an image at 300 meters away by placeing the object 18.19 cm away from the lens? however the object will be hugely magnified, is this correct? I am haveing trouble grasping this.. will the image still be in focus at that length ? it just doesn't seem practical.. even moving it to 500 meters i could still get an image.. when does it end? does light interference eventually wash out the image?