Mass loss in relation to radius of a star

  • #1
One of the most popular mass loss equations of a star, developed by D. Reimers, is given by:
dM/dt = -(4x10^-13) * η(L/(gR)) solar masses per year

Where η is a free parameter close to unity and L, g, and R are the luminosity of the star, surface gravity of the star, and the radius of the star, respectively.

What I am curious about is that when R increases, the amount of mass lost decreases. This seems counterintuitive, as when the radius increases, the density will decrease and the pull of gravitational energy will decrease so you would expect there to be greater mass loss.

Is there a qualitative reason why the star loses more mass as the radius decreases?
 
Last edited:

Answers and Replies

  • #2
D H
Staff Emeritus
Science Advisor
Insights Author
15,393
685
Is there a qualitative reason why the star loses more mass as the radius decreases?
In a sense it doesn't. Look at Reimers' law more closely:
$$\frac{dM}{dt} = -4\cdot10^{-13} \, \eta \frac {L_{\ast}} {gR_{\ast}}$$
That g in the denominator is surface gravity relative to that of the Sun: ##g=M_{\ast}/R_{\ast}^2##. Thus another way to write Reimers' law is
$$\frac{dM}{dt} = -4\cdot10^{-13} \, \eta \frac {L_{\ast}R_{\ast}} {M_{\ast}}$$
 

Related Threads on Mass loss in relation to radius of a star

Replies
3
Views
22K
Replies
5
Views
3K
Replies
3
Views
2K
  • Last Post
Replies
2
Views
2K
Replies
5
Views
1K
  • Last Post
Replies
4
Views
3K
  • Last Post
Replies
1
Views
3K
Replies
4
Views
2K
Replies
15
Views
2K
Top