Reconstructing an image out of a white surface light information.

AI Thread Summary
The discussion explores the theoretical possibility of reconstructing images of surrounding objects using light information reflected off a white surface, such as a rectangular piece of paper. It references advancements in computational photography, which can separate direct and indirect light, and lensless image sensors that gather light from all directions for later image computation. Participants express the need for clearer questions and understanding, as the original inquiry lacks specificity. The consensus suggests that without controlled illumination, accurately reconstructing images from a white surface is challenging due to mixed light information. Overall, the feasibility of this concept remains uncertain, highlighting the complexities of light analysis in image reconstruction.
Ovden151
Messages
5
Reaction score
0
After watching these videos about Computational photography:

and lensless image sensors:


was wondering if it is at least theoretically possible to calculate the light of the surroundings of a white surface just from the light information on that surface.

For instance taking a picture of a rectangular white paper and then compute how all the things in front of it looked like just by extracting the light information on the white paper.

Are there any algorithms or papers about this?
 
Science news on Phys.org
Sorry for the bump... If this subject needs a better explanation just let me know.
 
Your question is a little unclear. What does "calculate the light of the surroundings" and "compute how all the things in front of it looked" mean? Are you trying to use the white paper as a type of "mirror"?
 
Those videos are too long to expect us to watch before answering your question. So you need to ask the question in a way that we will understand without watching them.
 
Thanks for the heads up guys. Here is a graphic describing the idea, and a short video of what's described in the second video above.

wwi4np.jpg




The first video of the main post is about computational photography and shows some examples of techniques from that field.
For instance it shows how computationally it can be separated the direct light and the global illumination affecting different objects. They came up with algorithms that can separate bounced indirect light, sub-surface scattered light, from all the direct light. They also show how they could remove the motion blur of all moving objects no matter the speed difference and direction of each object.

And the second shows a lensless image sensor which gathers all the light around, it doesn't generate the final image, but later the image is computed analyzing the light gathered by the sensor. It seems to gather light from all directions, and with the light information saved, it can even allow to chose the focusing of the image *after* capturing the light information.
 
In this video they show a way to infer the image from a point of view different than the one of the camera.

 
Ovden151 said:
For instance taking a picture of a rectangular white paper and then compute how all the things in front of it looked like just by extracting the light information on the white paper.

If I understand the last video correctly, no. It seems to me that they illuminated one section of the playing card at a time using the projector. Without the ability to control the illumination you can't get a clear image from the white surface since light from the entire card is mixed together.
 
Back
Top