1. Not finding help here? Sign up for a free 30min tutor trial with Chegg Tutors
    Dismiss Notice
Dismiss Notice
Join Physics Forums Today!
The friendliest, high quality science and math community on the planet! Everyone who loves science is here!

Prob/Stats Introduction to the Practice of Statistics by David Moore

  1. Strongly Recommend

    100.0%
  2. Lightly Recommend

    0 vote(s)
    0.0%
  3. Lightly don't Recommend

    0 vote(s)
    0.0%
  4. Strongly don't Recommend

    0 vote(s)
    0.0%
  1. Jan 20, 2013 #1
    The Basic Practice of Statistics by David Moore


    Table of Contents:
    Code (Text):

    [LIST]
    [*] To the Instructor: About This Book
    [*] To the Student: Statistical Thinking
    [*] Exploring Data
    [LIST]
    [*] Picturing Distributions with Graphs
    [LIST]
    [*] Individuals and variables
    [*] Categorical variables: pie charts and bar graphs
    [*] Quantitative variables: histograms
    [*] Interpreting histograms
    [*] Quantitative variables: stemplots
    [*] Time plots
    [/LIST]
    [*] Describing Distributions with Numbers
    [LIST]
    [*] Measuring center: the mean
    [*] Measuring center: the median
    [*] Comparing the mean and the median
    [*] Measuring spread: the quartiles
    [*] The five-number summary and boxplots
    [*] Spotting suspected outliers
    [*] Measuring spread: the standard deviation
    [*] Choosing measures of center and spread
    [*] Using technology
    [*] Organizing a statistical problem
    [/LIST]
    [*] The normal distributions
    [LIST]
    [*] Density curves
    [*] Describing density curves
    [*] Normal distributions
    [*] The 68-95-99.7 rule
    [*] The standard normal distribution
    [*] Finding normal proportions
    [*] Using the standard normal table
    [*] Finding a value given proportion
    [/LIST]
    [*] Scatterplots and Correlation
    [LIST]
    [*] Explanatory and response variables
    [*] Displaying relationships: scatterplots
    [*] Interpreting scatterplots
    [*] Adding categorical variables to scatterplots
    [*] Measuring linear association: correlation
    [*] Facts about correlation
    [/LIST]
    [*] Regression
    [LIST]
    [*] Regression lines
    [*] The least-squares regression line
    [*] Using technology
    [*] Facts about least-squares regression
    [*] Residuals
    [*] Influential observations
    [*] Cautions about correlation and regression
    [*] Association does not imply causation
    [/LIST]
    [*] Two-Way Tables
    [LIST]
    [*] Marginal distributions
    [*] Conditional distributions
    [*] Simpson's paradox
    [/LIST]
    [*] Exploring Data: Part I Review
    [LIST]
    [*] Part I summary
    [*] Review exercises
    [*] Supplementary exercises
    [*] EESEE case studies
    [/LIST]
    [/LIST]
    [*] From Exploration to Inference
    [LIST]
    [*] Producing Data: Sampling
    [LIST]
    [*] Observation versus experiment
    [*] Sampling
    [*] How to sample badly
    [*] Simple random samples
    [*] Other sampling designs
    [*] Cautions about sample surveys
    [*] Inference about the population
    [/LIST]
    [*] Producing Data: Experiments
    [LIST]
    [*] Experiments
    [*] How to experiment badly
    [*] Randomized comparative experiments
    [*] The logic of randomized comparative experiments
    [*] Cautions about experimentation
    [*] Matched pairs and other block designs
    [/LIST]
    [*] Commentary: Data Ethics
    [LIST]
    [*] Institutional review boards
    [*] Informed consent
    [*] Confidentiality
    [*] Clinical trials
    [*] Behavioral and social science experiments
    [/LIST]
    [*] Introducing Probability
    [LIST]
    [*] The idea of probability
    [*] Probability models
    [*] Probability rules
    [*] Discrete probability models
    [*] Continuous probability models
    [*] Random variables
    [*] Personal probability
    [/LIST]
    [*] Sampling Distributions
    [LIST]
    [*] Parameters and statistics
    [*] Statistical estimation and the law of large numbers
    [*] Sampling distributions
    [*] The sampling distribution of [itex]\overline{x}[/itex]
    [*] The central limit theorem
    [*] Statistical process control
    [*] [itex]\overline{x}[/itex] charts
    [*] Thinking about process control
    [/LIST]
    [*] General Rules of Probability
    [LIST]
    [*] Independence and the multiplication rule
    [*] The general addition rule
    [*] Conditional probability
    [*] The general multiplication rule
    [*] Independence
    [*] Tree diagrams
    [/LIST]
    [*] Binomial Distributions
    [LIST]
    [*] The binomial setting and binomial distributions
    [*] Binomial distributions in statistical sampling
    [*] Binomial probabilities
    [*] Using technology
    [*] Binomial mean and standard deviation
    [*] Then normal approximation to binomial distributions
    [/LIST]
    [*] Confidence Intervals: The Basics
    [LIST]
    [*] Estimating with confidence
    [*] Confidence intervals for the mean [itex]\mu[/itex]
    [*] How confidence intervals behave
    [*] Choosing the sample size
    [/LIST]
    [*] Tests of Significance: The Basics
    [LIST]
    [*] The reasoning of tests of significance
    [*] Stating hypotheses
    [*] Test statistics
    [*] P-values
    [*] Statistical significance
    [*] Tests for a population mean
    [*] Using tables of critical values
    [*] Tests from confidence intervals
    [/LIST]
    [*] Inference in Practice
    [LIST]
    [*] Where did the data come from?
    [*] Cautions about the [itex]z[/itex] procedures
    [*] Cautions about confidence intervals
    [*] Cautions about significance tests
    [*] The power of a test
    [*] Type I and Type II errors
    [/LIST]
    [*] From Exploration to Inference: Part II Review
    [LIST]
    [*] Part II summary
    [*] Review exercises
    [*] Supplementary exercises
    [*] Optional exercises
    [*] EESEE case studies
    [/LIST]
    [/LIST]
    [*] Inference about Variables
    [LIST]
    [*] Inference about a Population Mean
    [LIST]
    [*] Conditions for inference
    [*] The [itex]t[/itex] distributions
    [*] The one-sample [itex]t[/itex] confidence interval
    [*] The one-sample [itex]t[/itex] test
    [*] Using technology
    [*] Matched pairs [itex]t[/itex] procedures
    [*] Robustness of [itex]t[/itex] procedures
    [/LIST]
    [*] Two-Sample Problems
    [LIST]
    [*] Two-sample problems
    [*] Comparing two population means
    [*] Two-sample [itex]t[/itex] procedures
    [*] Examples of the two-sample [itex]t[/itex] procedures
    [*] Using technology
    [*] Robustness again
    [*] Details of the [itex]t[/itex] approximation
    [*] Avoid the pooled two-sample [itex]t[/itex] procedures
    [*] Avoid inference about standard deviations
    [*] The [itex]F[/itex] test for comparing two standard deviations
    [/LIST]
    [*] Inference about a Population Proportion
    [LIST]
    [*] The sample proportion [itex]\hat{p}[/itex]
    [*] The sampling distribution of [itex]\hat{p}[/itex]
    [*] Large-sample confidence intervals for a proportion
    [*] Accurate confidence intervals for a proportion
    [*] Choosing the sample size
    [*] Significance tests for a proportion
    [/LIST]
    [*] Comparing Two Proportions
    [LIST]
    [*] Two-sample problems: proportions
    [*] The sampling distribution of a difference between proportions
    [*] Large-sample confidence intervals for comparing proportions
    [*] Using technology
    [*] Accurate confidence intervals for comparing proportions
    [*] Significance tests for comparing proportions
    [/LIST]
    [*] Inference about Variables: Part III Review
    [LIST]
    [*] Part III summary
    [*] Review exercises
    [*] Supplementary exercises
    [*] EESEE case studies
    [/LIST]
    [/LIST]
    [*] Inference about Relationships
    [LIST]
    [*] Two Categorical Variables: The Chi-Square Test
    [LIST]
    [*] Two-way tables
    [*] The problem of multiple comparisons
    [*] Expected counts in two-way tables
    [*] The chi-square test
    [*] Using technology
    [*] Cell counts required for the chi-square test
    [*] Uses of the chi-square test
    [*] The chi-square distributions
    [*] The chi-square and the [itex]z[/itex] test
    [*] The chi-square test for goodness of fit
    [/LIST]
    [*] Inference for Regression
    [LIST]
    [*] Conditions for regression inference
    [*] Estimating the parameters
    [*] Using technology
    [*] Testing the hypothesis of no linear relationship
    [*] Testing lack of correlation
    [*] Confidence intervals for the regression slope
    [*] Inference about prediction
    [*] Checking the conditions for inference
    [/LIST]
    [*] One-Way Analysis of Variance: Comparing Several Means
    [LIST]
    [*] comparing several means
    [*] The analysis of variance [itex]F[/itex] test
    [*] Using technology
    [*] The idea of analysis of variance
    [*] Conditions for ANOVA
    [*] [itex]F[/itex] distributions and degrees of freedom
    [*] Some details of ANOVA: the two-sample case
    [*] Some details of ANOVA
    [/LIST]
    [*] Statistical Thinking Revisited
    [*] Notes and Data Sources
    [*] Tables
    [LIST]
    [*] Standard normal probabilities
    [*] Random digits
    [*] [itex]t[/itex] distribution critical values
    [*] [itex]F[/itex] distribution critical values
    [*] Chi-square distribution critical values
    [*] Critical values of the correlation [itex]r[/itex]
    [/LIST]
    [*] Answers to Selected Exercises
    [*] Index
    [/LIST]
    [*] Optional Companion Chapters (on the BPS CD and online)
    [LIST]
    [*] Nonparametric Tests
    [LIST]
    [*] Comparing two samples: the Wilcoxon rank sum test
    [*] The normal approximation for [itex]W[/itex]
    [*] Using technology
    [*] What hypotheses does Wilcoxon test
    [*] Dealing with ties in rank tests
    [*] Matched pairs: the Wilcoxon signed rank test
    [*] The normal approximation for [itex]W^+[/itex]
    [*] Dealing with ties in the signed rank test
    [*] Comparing several samples: the Kruskal-Wallis test
    [/LIST]
    [*] Statistical Process Control
    [LIST]
    [*] Processes
    [*] Describing processes
    [*] The idea of statistical process control
    [*] [itex]\overline{x}[/itex] charts for process monitoring
    [*] [itex]s[/itex] charts for process monitoring
    [*] Using control charts
    [*] Setting up control charts
    [*] Comments on statistical control
    [*] Don't confuse control with capability!
    [*] Control charts for sample proportions
    [*] Control limits for [itex]p[/itex] charts
    [/LIST]
    [*] Multiple Regression
    [LIST]
    [*] Parallel regression lines
    [*] Estimating parameters
    [*] Using technology
    [*] Inference for multiple regression
    [*] Interaction
    [*] The multiple linear regression model
    [*] The woes of regression coefficients
    [*] A case study for multiple regression
    [*] Inference for regression parameters
    [*] Checking the conditions for inference
    [/LIST]
    [*] Two-Way Analysis of Variance (available online only)
    [LIST]
    [*] Extending the one-way ANOVA model
    [*] Two-way ANOVA models
    [*] Using technology
    [*] Inference for two-way ANOVA
    [*] Inference for a randomized block design
    [*] Multiple comparisons
    [*] Contrasts
    [*] Conditions for two-way ANOVA
    [/LIST]
    [/LIST]
    [/LIST]
     
     
    Last edited: May 6, 2017
  2. jcsd
  3. Jan 22, 2013 #2

    I like Serena

    User Avatar
    Homework Helper

    I have tutored quite a few psychology students that were using this book by now.
    It taught me quite a bit about applied statistics while I was teaching them.
    It's known as Moore, McCabe, and Craig, or MMC for short.

    Btw, I believe high school math is sufficient as prerequisite.
    That is what my students have anyway.
     
    Last edited: Jan 22, 2013
Know someone interested in this topic? Share this thread via Reddit, Google+, Twitter, or Facebook




Similar Discussions: Introduction to the Practice of Statistics by David Moore
Loading...