The book we are using (Physics 101) has shown quit well that there will be a greater quantity of heat loss (and faster), of a substance to it’s environment, if the difference of temperatures between the two is greater. Also, there will be initially a faster rate of heat loss / time due to this greater difference in temperature. Can one compare this effect to that of the half-life of radioactive decay? I believe I understand correctly, that a “Hot” radioactive substance will lose (decay) more energy during the same time frame (half-life cycle) than a same sample that is “Cold” as there is more energy to be cut in half during the same amount of time. If this is correct, then it shows that as a substance looses heat, the slower it looses heat, as it gets cooler. The reason I say this is we had a discussion question that asked when to add cream to hot coffee. If there was going to be a wait prior to drinking it, do you add the the cream as soon as possible, or just before you drink it? The idea is the best way to have the hottest coffee that has sat for length of time. Does anyone have any thoughts about this?