I'm taking a probability class where multivariate calculus was not a prerequisite, but some of it is coming up, I get the concept of, say integrating over a region, but get lost in some of the mechanics(adsbygoogle = window.adsbygoogle || []).push({});

Here is the problem (I don't want a full solution):

A point is uniformly distributed within the disk of radius 1. That is its density is

[itex]

f(x,y) = C \hspace{1cm} 0 \leq x^2 + y^2 \leq 1

[/itex]

Find the probability that its distance from the origin is less than x, 0 \leq x \leq 1

I'm pretty sure I have to set up an integral that integrates over a disc of radius x to get the probability

Something like this

[itex]

\int_A \int_B C \, dx \, dy

[/itex]

But I don't know what the intervals A and B are supposed to be.

Can someone point me in the right direction? I get confused because my attempts end up with x being involved in the limit of integration, but x is the dummy variable, which doesn't seem right.

**Physics Forums | Science Articles, Homework Help, Discussion**

Dismiss Notice

Join Physics Forums Today!

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

# Integrating over a disk

**Physics Forums | Science Articles, Homework Help, Discussion**