1. The problem statement, all variables and given/known data If a black hole and a "normal" star orbit each other, gases from the normal star falling into the black hole can have their temperature increased by millions of degrees due to frictional heating. When the gases are heated that much, they begin to radiate light in the X-ray region of the electromagnetic spectrum (high-energy light photons). Cygnus X-1, the second strongest known X-ray source in the sky, is thought to be one such binary system; it radiates at an estimated power of 4.00x10^31 W. If we assume that 0.84 percent of the in-falling mass escapes as X ray energy, at what rate is the black hole gaining mass? 2. Relevant equations E=m*c^2 3. The attempt at a solution Wondering if someone could verify this because this question confuses me. I'm assuming that the black hole absorbs 100-.84=99.16% of the power. I take 4.00x10^31 * .9916 = 3.9664x10^31 J Using E=m*c^2 => m = 3.9664x10^31/((3x10^8)^2) = 4.407x10^4 kg/s Does this seem a logical conclusion, am I interpreting the question correctly? Thanks in advance.