Recent content by Number Nine

  1. N

    Undergrad Fitting two models in sequence with parameters in common

    When I say it's probably not what you want, I mean it's probably a bad idea, and it doesn't really have much statistical justification. All of your problems stem from the fact that the PO model isn't a good model. You don't use it to estimate CL and V because it doesn't provide trustworthy...
  2. N

    Undergrad Fitting two models in sequence with parameters in common

    OP: You can do this pretty easily in R. Just write a function f(CL, V, F, ka) = \alpha E_\text{iv}(CL, V) + (1-\alpha) E_\text{po}(CL, V, F, ka) where E_\text{iv} is a function returning the sum of squared errors for the IV model given parameters CL, V, and E_\text{po} is the same thing for the...
  3. N

    Undergrad Fitting two models in sequence with parameters in common

    The MCMC approach is almost certainly (and the FME documentation confirms this) fitting a Bayesian model. The approach your colleague is suggesting is, I think, similar to what I alluded to in a previous post, where you would derive a posterior distribution for the parameters in the first model...
  4. N

    Undergrad Fitting two models in sequence with parameters in common

    I was confused about this as well. You can account for uncertainty in the parameters estimated in model one in a few ways (e.g. using their posterior distributions as priors in model two, or marginalizing over their uncertainty in model two), but none of them involve fixing them to exact values...
  5. N

    Undergrad Fitting two models in sequence with parameters in common

    I'm very bored, and I think I have a good way to do this, so give me an hour or two and I'll get back to you with some R code that I think will do what you want. EDIT: Can you explain your "condition A" in more detail? If the PO data provides information about the parameters CL and V, it would...
  6. N

    Undergrad Standard Deviation Versus Sample Size & T-Distribution

    A minor point: the "population standard deviation" (i.e. the square root of the sum of squared deviations from the mean, divided by n-1) is actually a biased estimate of the standard deviation. This follows from Jensen's inequality, since the square root is a concave function. It's fairly...
  7. N

    Undergrad The Standard Deviation of Sample vs. Population in Probability Calculations

    The "sample standard deviation" (the sum of squared deviations from the mean, divided by n) is a biased estimator of the population SD. It's actually very difficult to construct an unbiased estimator of the standard deviation, even for a normal distribution.
  8. N

    Graduate How to Optimize Predictive Models: Include Interactions or Not?

    This is actually a bit of a tricky question. The simplest approach would be to just compare all possible models and select the best one, which would be 2^7 = 128 models in your case, which might take a while if you have to code them manually. If you want to try something a little more...
  9. N

    Graduate How to Optimize Predictive Models: Include Interactions or Not?

    If you want to do prediction, why use AIC? Why not evaluate prediction error directly? Cross-validation is what you want. Just compare all the models you're interested in and choose the one with the best predictive performance. The AIC is actually asymptotically equivalent to leave-one-out cross...
  10. N

    Graduate Comparing Kullback-Leibler divergence values

    Keep in mind that the KL-divergence is non-commutative, and different "orders" correspond to different objective functions (and different research questions). The way you're fitting it (that is, KL(Q||P), where Q is being fit to P) is trying to match regions of high density, and it does seem to...
  11. N

    Undergrad Is the F-test Calculation Different for Longitudinal Data in Linear Regression?

    What are you trying to do, exactly? Generally you wouldn't use linear regression for longitudinal data without some modifications.
  12. N

    Undergrad Pie Charts & Error Bars: Which to Use?

    As an aside, it should be said that there is really no reason to ever use a pie chart, and most people involved in any kind of data visualization despise them. That's not just a personal preference; they're objectively terrible at communicating information. It's almost impossible to directly...
  13. N

    Graduate Should I Ignore Data Driven Models or Use Bayesian Methods for Model Selection?

    It should be noted that neither stepwise regression nor fit indices like the AIC really protect against overfitting. Stepwise regression in particular is kind of a joke among statisticians, though it's still (unfortunately) popular in the social/biological sciences. If you're concerned about...
  14. N

    Undergrad Central limit theorem, panel study

    The central limit theorem will not guarantee this; nor will anything, really. Given an appropriately chosen regression model, the errors may or may not be normally distributed. If they're not, then you'll have to either transform the data in some way, or fit a regression model with a different...
  15. N

    Studying Why do Walter Rudin's proofs in real analysis often seem so elusive and clever?

    "Little Rudin" is a classic, and is a good textbook for people who have a strong background in pure math, and are already familiar with the basics of real analysis. That said, I would absolutely not recommend it as an introduction. Rudin's proofs are very "nice", in the sense that they tend to...