
#1
Nov411, 03:38 PM

P: 13

How do I prove that the mean of a random variable Z which is the sum of to other random variables X and Y is the sum of the mean of X with the mean of Y?




#2
Nov411, 03:48 PM

Sci Advisor
P: 5,942

Essentially the theorem is equivalent to the theorem that the integral of a sum is the sum of the integrals.




#3
Nov411, 03:54 PM

P: 13

well I obviously know that the integral of the sum is the sum of the integral but I don't know how I can relate that to the situation a mentioned, can you please be more specific?
I'm trying to prove it and I'm getting a convultion integral so far... thank you. 



#4
Nov411, 09:58 PM

P: 523

the mean of a sum of variables.Another approach that would work for the nonindependent case is to consider separately the joint distribution and marginal distributions of X and Y. 



#5
Nov511, 04:06 AM

P: 13

well but I'm not being able to prove it either for dependent or independent variables, can you please show me the proof or tell me where I can find it? thank you




#6
Nov511, 04:03 PM

Sci Advisor
P: 5,942

Two random variables.
E(X+Y)=∫∫(x+y)dF(x,y)=∫∫xdF(x,y) + ∫∫ydF(x,y). Integrate with respect to y in the first integral and integrate with respect to x in the second integral. You will be left with E(X) + E(Y). In the above F(x,y) is the joint distribution function. 


Register to reply 
Related Discussions  
Matlab Question: variables that depend on previous variables  Math & Science Software  5  
Complex conjugate variables as independent variables in polynomial equations  Linear & Abstract Algebra  0  
Angular Variables and Tangential Variables  Introductory Physics Homework  1  
how do you know if you should use 3 variables for 4 variables in IC design?  Electrical Engineering  2  
Solve for three variables with only two variables??  Introductory Physics Homework  8 