Dismiss Notice
Join Physics Forums Today!
The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

Mass loss in relation to radius of a star

  1. Nov 15, 2013 #1
    One of the most popular mass loss equations of a star, developed by D. Reimers, is given by:
    dM/dt = -(4x10^-13) * η(L/(gR)) solar masses per year

    Where η is a free parameter close to unity and L, g, and R are the luminosity of the star, surface gravity of the star, and the radius of the star, respectively.

    What I am curious about is that when R increases, the amount of mass lost decreases. This seems counterintuitive, as when the radius increases, the density will decrease and the pull of gravitational energy will decrease so you would expect there to be greater mass loss.

    Is there a qualitative reason why the star loses more mass as the radius decreases?
     
    Last edited: Nov 15, 2013
  2. jcsd
  3. Nov 15, 2013 #2

    D H

    User Avatar
    Staff Emeritus
    Science Advisor

    In a sense it doesn't. Look at Reimers' law more closely:
    $$\frac{dM}{dt} = -4\cdot10^{-13} \, \eta \frac {L_{\ast}} {gR_{\ast}}$$
    That g in the denominator is surface gravity relative to that of the Sun: ##g=M_{\ast}/R_{\ast}^2##. Thus another way to write Reimers' law is
    $$\frac{dM}{dt} = -4\cdot10^{-13} \, \eta \frac {L_{\ast}R_{\ast}} {M_{\ast}}$$
     
Know someone interested in this topic? Share this thread via Reddit, Google+, Twitter, or Facebook