Determining the states in Markov chains

  • Thread starter Thread starter stripes
  • Start date Start date
  • Tags Tags
    States
Click For Summary
SUMMARY

This discussion focuses on determining the number of states in Markov chains based on previous events. The user correctly identifies that the number of states can be calculated using the formula mn, where n is the number of previous days and m is the number of possible events. For example, Joe's mood, which can be gloomy, so-so, or cheerful, results in 3 states for 1 day, while the weather, which can be rainy or dry over 2 days, results in 4 states (22). The user confirms their understanding through various examples, including a 3-day weather scenario yielding 8 states (23).

PREREQUISITES
  • Understanding of Markov chains and their properties
  • Basic probability concepts, including event outcomes
  • Familiarity with combinatorial calculations
  • Knowledge of state transition diagrams
NEXT STEPS
  • Study the concept of state transition matrices in Markov chains
  • Learn about the stationary distribution of Markov chains
  • Explore applications of Markov chains in real-world scenarios
  • Practice calculating states for various Markov chain examples
USEFUL FOR

Students in probability and statistics, data scientists modeling stochastic processes, and anyone interested in understanding Markov chains and their applications in real-world scenarios.

stripes
Messages
262
Reaction score
0

Homework Statement



This is not a homework question, just me trying to wrap my head around things. My probability class talked about Markov chains for less than 2 hours worth of lecture, and I've been super sick lately, so I'm still a little confused.

If we're considering real world examples where the probability of an event happening tomorrow depends on what happens in the previous n days, where each day, m different events could have happened, is the number of states just mn?

Two practice problems that were given to us were as follows:

Joe's mood depends on his mood from the previous day only (so we're going back the previous 1 day). He could be gloomy, so-so, or cheerful (3 possible events). This means we have a Markov chain with 31 states?

The probability that it rains tomorrow depends on whether it rained the past 2 days (going back the previous 2 days). If the weather on any day can either be rainy or dry, then the number of possible events is 2. So the number of states is 22 = 4?

One more example I came across was the weather one but with more days: suppose the probability that it rains tomorrow depends only on whether it rained the past 3 days (going back the previous 3 days). It can either be rainy or dry (2 possible events). So the number of states is 23 = 8?

I know that these answers are correct but is my way of thinking about right? I will be given such real world examples (involving previous days and what not) on my final exam, so when I'm trying to figure out the number of states in these real world examples, am I doing it correctly?

I guess I could just figure it out by going RRR (rain rain rain), RRD (rain rain dry), RDR, RDD, ... etc. I get this feeling that I'm totally missing the point (missing how to figure things out intuitively). I take it it's because I've missed some class due to illness.

If someone could let me know if I'm going about this correctly I would really appreciate it.

Homework Equations



--

The Attempt at a Solution



--
 
Physics news on Phys.org
stripes said:

Homework Statement



This is not a homework question, just me trying to wrap my head around things. My probability class talked about Markov chains for less than 2 hours worth of lecture, and I've been super sick lately, so I'm still a little confused.

If we're considering real world examples where the probability of an event happening tomorrow depends on what happens in the previous n days, where each day, m different events could have happened, is the number of states just mn?

Two practice problems that were given to us were as follows:

Joe's mood depends on his mood from the previous day only (so we're going back the previous 1 day). He could be gloomy, so-so, or cheerful (3 possible events). This means we have a Markov chain with 31 states?

The probability that it rains tomorrow depends on whether it rained the past 2 days (going back the previous 2 days). If the weather on any day can either be rainy or dry, then the number of possible events is 2. So the number of states is 22 = 4?

One more example I came across was the weather one but with more days: suppose the probability that it rains tomorrow depends only on whether it rained the past 3 days (going back the previous 3 days). It can either be rainy or dry (2 possible events). So the number of states is 23 = 8?

I know that these answers are correct but is my way of thinking about right? I will be given such real world examples (involving previous days and what not) on my final exam, so when I'm trying to figure out the number of states in these real world examples, am I doing it correctly?

I guess I could just figure it out by going RRR (rain rain rain), RRD (rain rain dry), RDR, RDD, ... etc. I get this feeling that I'm totally missing the point (missing how to figure things out intuitively). I take it it's because I've missed some class due to illness.

If someone could let me know if I'm going about this correctly I would really appreciate it.

Homework Equations



--

The Attempt at a Solution



--

Your 8 states RRR, RRD, ... are correct; that is the way these things are ususally done.
 

Similar threads

  • · Replies 8 ·
Replies
8
Views
8K
Replies
9
Views
3K
  • · Replies 1 ·
Replies
1
Views
2K
Replies
1
Views
12K
  • · Replies 12 ·
Replies
12
Views
3K
  • · Replies 3 ·
Replies
3
Views
1K
Replies
24
Views
4K
  • · Replies 3 ·
Replies
3
Views
2K
Replies
16
Views
2K
Replies
1
Views
2K