Gaussian Mixture Model Confusion

AI Thread Summary
The discussion centers on implementing the Gaussian Mixture Model (GMM) for background subtraction as outlined by Stauffer and Grimson. The main confusion arises regarding the update logic for mean and variance, particularly the recursive formula for exponential moving statistics. The participant notes that the parameter ρ tends to be very small, leading to slow convergence due to the initial assumption of large variances. It is highlighted that the updating process occurs under specific conditions, particularly when there are drastic changes in pixel values. Clarification on normalization or adjustments to improve convergence speed is sought.
mp6250
Messages
3
Reaction score
0
Hi All,

I'm trying to implement the Gaussian Mixture Model for background subtraction as described by Chris Stauffer and W.E.L Grimson in their paper "Adaptive background mixture models for real-time tracking."

I'm having a little trouble with the logic in the step that updates the mean and variance of the models. According to the paper, when new image data comes in, you follow a recursive formula to get exponential moving statistics for these parameters based on the following formulas:

μ_t = (1-ρ)μ_{t-1} + ρX_t
σ^2_t = (1-ρ)σ^2_{t-1} + ρ(X_t-μ_t)^T(X_t-μ_t)

where μ and σ are the mean and standard deviation of the model, X_t is the incoming data vector, and the subscript indicate the relative times between variables. ρ is defined as:

ρ=α \frac{1}{(2π)^{\frac{n}{2}}|Ʃ|^{\frac{1}{2}}} e^{-\frac{1}{2}(X_t-μ_t)^T Ʃ^{-1}(X_t-μ_t)}

where Ʃ is the covariance matrix (taken to be diagonal for simplicity) and α is a parameter that controls the learning rate.

My confusion is this, ρ will always be tiny. The algorithm assumes large variances to begin with and the tiny probabilities that come out of these functions will cause very slow convergence, regardless of the choice of alpha (usually taken to be around 0.05 or so). It's my understanding that you would never set α > 1.0, so where could this be corrected for? Is there a normalization I am missing somewhere?
 
Physics news on Phys.org
mp6250 said:
Hi All,

I'm trying to implement the Gaussian Mixture Model for background subtraction as described by Chris Stauffer and W.E.L Grimson in their paper "Adaptive background mixture models for real-time tracking."

http://www.google.com/url?sa=t&rct=...yJmHKPWgOWl9_zw&bvm=bv.44770516,d.aWM&cad=rja

My confusion is this, ρ will always be tiny. The algorithm assumes large variances to begin with and the tiny probabilities that come out of these functions will cause very slow convergence, regardless of the choice of alpha (usually taken to be around 0.05 or so).

Just glancing at that paper, the updating you described only happens on certain conditions. A drastic change in a pixel value is updated by a different procedure.
 
I was reading documentation about the soundness and completeness of logic formal systems. Consider the following $$\vdash_S \phi$$ where ##S## is the proof-system making part the formal system and ##\phi## is a wff (well formed formula) of the formal language. Note the blank on left of the turnstile symbol ##\vdash_S##, as far as I can tell it actually represents the empty set. So what does it mean ? I guess it actually means ##\phi## is a theorem of the formal system, i.e. there is a...
Back
Top