1. The problem statement, all variables and given/known data Here's the question: A monochromatic point source of light radiates 25 W at a wavelength of 5000 angstroms. A plate of metal is placed 100 cm from the source. Atoms in the metal have a radius of 1 angstrom. Assume that the atom can continually absorb light. The work function of the metal is 4 eV. How long is it before an electron is emitted from the metal? 2. Relevant equations My attempt at an answer. I used the 100 cm placement away from the source for the radius and put that in to 4*pi*r^2 to get the total surface area. Divided the energy of the source by the total surface area. Multiplied that by the square of the atom spacing to get how much energy per sec was falling on that atom. I then used the work function of 4eV divided by the energy falling on that area per sec to find out how long it takes to reach the work function energy level. 3. The attempt at a solution My only issue is, is that the time they are looking for or is there some other concept I am missing. I know using the normal photo electric effect theory it is hf - work function= energy of emitted electron. If I stop timing at the point I reach the work function value then in theory would I not be emitting an electron with 0 energy. (tech impossible). If its absorbing continuously would I add the time to allow for one more wave cycle above that of the energy and time needed to get to the work function.