## Orthogonal complement of gradient field?

I am doing my research in probability. I have found some probability distribution of a random variable X on the n dimensional unit sphere. Let b be a smooth and lipschitz vector field mapping X to $R^n$. I have also found that for all continuous differentiable function f mapping X to $R$, the expectation of $\triangledown f\cdot b$ is zero. I have strong feeling that this implies b(X)=0 with probability 1, but I am not sure how I can prove it.
 PhysOrg.com science news on PhysOrg.com >> Ants and carnivorous plants conspire for mutualistic feeding>> Forecast for Titan: Wild weather could be ahead>> Researchers stitch defects into the world's thinnest semiconductor

 Quote by liguolong I am doing my research in probability. I have found some probability distribution of a random variable X on the n dimensional unit sphere. Let b be a smooth and lipschitz vector field mapping X to $R^n$. I have also found that for all continuous differentiable function f mapping X to $R$, the expectation of $\triangledown f\cdot b$ is zero. I have strong feeling that this implies b(X)=0 with probability 1, but I am not sure how I can prove it.
Hey liguolong and welcome to the forums.

The fact that X is a random variable defined over the n-dimensional unit sphere means that X is a pdf where the pdf is a function of n variables and the CDF is the integral over the n-dimensional space where you will have the limits represent the entangled nature (since you're defining over the surface of the sphere). This implies a continuous distribution which is important to note later on.

Now you then have the constraint about the gradient and it's expectation is zero.

Now if you want to prove that this transformation is only defined at b(X) = 0, then you will need to prove two things: E[b(X)] = 0 and VAR[b(X)] = 0. If you prove these two things then you will show that without a doubt the only value that will suffice will be at the point b(X) = 0.

Now your grad(f) . b = 0 implies that b is perpendicular to the gradient of f for all f. Now I haven't done vector calculus in a long time but I think that for this to happen then grad(b) . b = 0 since f includes b. This implies that f = 0. (I could be wrong so please correct me if I am).

So if f is the zero function then that means everywhere will be zero where b(X) maps to 0 for every value and hence E[b(X)] = 0 and VAR[b(X)] = 0 which means that P(b(X) = 0) should be one as you suggested.

I'm interested if any of my reasoning is wrong and again I haven't taken vector calculus in a long time.
 The fact that E[b(X)] = 0 is easy, just take f=x_i for each i it is ok. However, I am not sure how to prove Var[b(X)] = 0. I am not sure what is meant by grad(b).b=0 since b is a vector field, and grad is only defined for scalar functions.

## Orthogonal complement of gradient field?

 Quote by liguolong The fact that E[b(X)] = 0 is easy, just take f=x_i for each i it is ok. However, I am not sure how to prove Var[b(X)] = 0. I am not sure what is meant by grad(b).b=0 since b is a vector field, and grad is only defined for scalar functions.
This is what I mean by grad(f)