Review:
1. DATA COLLECTION AND EXPLORING UNIVARIATE DISTRIBUTIONS Introduction. A model for problem solving and its application. Types of data and frequency distribution tables. Tools for describing data: Graphical methods. Graphing Categorical Data. Graphing Numerical Data. Visualizing distributions. Tool for Describing Data: Numerical measures. Measures of Center. Measures of Position. Measures of variation (or spread). Reading Computer Printouts. The effect of shifting and scaling of measurements on summary measures. Summary Measures and Decisions. The Empirical Rule. Standardized Values and z-scores. Boxplots. Detecting Outliers. Summary. Supplemental Exercises. 2. EXPLORING BIVARIATE DISTRIBUTIONS AND ESTIMATING RELATIONS Introduction. Two-way table for categorical data. Time series analysis. Scatterplots: Graphical analysis of association between measurements. Correlation: Estimating the strength of linear relation. Regression: Modeling linear relationships. The Coefficient of Determination. Residual Analysis: Assessing the adequacy of the model. Transformations. Reading Computer Printout. Summary. Supplemental Exercises. 3. OBTAINING DATA. Introduction. Overview of methods of data collection. Planning and Conducting Surveys. Planning and Conducting Experiments. Completely Randomized Design. Randomized Block Design. Planning and Conducting an Observational Study. Summary. Supplemental Exercises. 4. PROBABILITY. Introduction. Sample space and relationships among events. Definition of probability. Counting rules useful in probability. Conditional probability and independence. Rules of probability. Odds, odds ratios, and risk ratio. Summary. Supplemental Exercises. 5. DISCRETE PROBABILITY DISTRIBUTIONS. Introduction. Random variables and their probability distributions Expected values of random variables. The Bernoulli distribution. The Binomial distribution. The Geometric and Negative Binomial distributions. The Geometric distribution. The Negative Binomial distribution. The Poisson distribution. The hypergeometric distribution. The Moment-Generating Function. Simulating probability distributions. Summary. Supplementary Exercises. 6. CONTINUOUS PROBABILITY DISTRIBUTIONS. Introduction. Continuous random variables and their probability distributions. Expected values of continuous random variables. The Uniform distribution. The exponential distribution. The Gamma distribution. The Normal distribution. The Lognormal Distribution. The Beta distribution. The Weibull distribution. Reliability. The Moment-generating Functions for Continuous Random Variables. Simulating probability distributions. Summary. Supplementary Exercises. 7. MULTIVARIATE PROBABILITY DISTRIBUTIONS. Introduction. Bivariate and Marginal Probability Distributions. Conditional Probability Distributions. Independent Random Variables. Expected Values of Functions of Random Variables. The Multinomial Distribution. More on the Moment-Generating Function. Conditional Expectations. Compounding and Its Applications. Summary. Supplementary Exercises. 8. STATISTICS, SAMPLING DISTRIBUTIONS, AND CONTROL CHARTS. Introduction. The sampling distributions. The sampling distribution of X (General Distribution). The sampling distribution of X (Normal Distribution). The sampling distribution of sample proportion Y/n (Large sample). The sampling distribution of S^2 (Normal Distribution). Sampling Distributions: The multiple-sample case. The sampling distribution of (X1 - X2). The sampling distribution of XD. The sampling distribution of (^p1 - ^p2). The sampling distribution of S^21/S^22. Control Charts. The X-Chart: Known and s. The X and R-Charts: Unknown and s. The X and S-Charts: Unknown and s. The p-Chart. The c-chart. The u-chart. Process Capability. Summary. Supplementary Exercises. 9. ESTIMATION. Introduction. Point estimators and their properties. Confidence Intervals: The Single-Sample Case. Confidence Interval for : General Distribution. Confidence Interval for Mean: Normal Distribution. Confidence Interval for Proportion: Large sample case. Confidence interval for s^2. Confidence Intervals: The Multiple Samples Case. Confidence Interval for Linear Functions of Means: General Distributions. Confidence Interval for Linear Functions of Means: Normal Distributions. Large Samples Confidence Intervals for Linear Functions of Proportions. Confidence Interval for s^22/s^21: Normal distribution case. Prediction Intervals. Tolerance Intervals. The Method of Maximum Likelihood. Bayes Estimators. Summary. Supplementary Exercises. 10. HYPOTHESIS TESTING. Introduction. Terminology of Hypothesis Testing. Hypothesis Testing: The Single-Sample Case. Testing for Mean: General Distributions Case. Testing a Mean: Normal distribution Case. Testing for Proportion: Large Sample Case. Testing for Variance: Normal Distribution Case. Hypothesis Testing: The Multiple-Sample Case. Testing the Difference between Two means: General Distributions Case. Testing the Difference between Two means: Normal Distributions case. Testing the difference between the means for paired samples. Testing the ratio of variances: Normal distributions case. '^2 tests on Frequency data. Testing parameters of the multinomial distribution. Testing equality among Binomial parameters. Test of Independence. Goodness of Fit Tests. '^2 Test. Kolmogorov-Smirnov test. Using Computer Programs to Fit Distributions. Acceptance Sampling. Acceptance Sampling by Attributes. Acceptance Sampling by Variables. Summary. Supplementary Exercises. 11. ESTIMATION AND INFERENCE FOR REGRESSION PARAMETERS. Introduction. Regression models with one predictor variable. The probability distribution of random error component. Making inferences about slope. Estimating slope using a confidence interval. Testing a hypothesis about slope. Connection between inference for slope and correlation coefficient. Using the simple linear model for estimation and prediction. Multiple regression analysis. Fitting the model: The least-squares approach. Estimation of error variance. Inferences in multiple regression. A test of model adequacy. Estimating and testing hypothesis about individual parameters. Using the multiple regression model for estimation and prediction. Model building: A test for portion of a model. Other regression models. Response surface method. Modeling a time trend. Logistic regression. Checking conditions and some pitfalls. Checking conditions. Some pitfalls. Reading printouts. Summary. Supplemental Exercises. 12. ANALYSIS OF VARIANCE. Introduction. Review of Designed Experiments. Analysis of Variance (ANOVA) Technique. Analysis of Variance for Completely Randomized Design. Relationship of ANOVA for CRD with a t test and Regression. Equivalence between a t test and an F test for CRD with 2 treatments. ANOVA for CRD and Regression Analysis. Estimation for Completely randomized design. Analysis of Variance for the Randomized Block Design. ANOVA for RBD. Relation between a Paired t test and an F test for RBD. ANOVA for RBD and Regression Analysis. Bonferroni Method for Estimation for RBD. Factorial Experiments. Analysis of variance for the Factorial Experiment. Fitting Higher Order Models. Summary. Supplemental Exercises. APPENDIX. REFERENCES.
About the Author:
Richard L. Scheaffer, Professor Emeritus of Statistics, University of Florida, received his Ph.D. in statistics from Florida State University. Accompanying a career of teaching, research and administration, Dr. Scheaffer has led efforts on the improvement of statistics education throughout the school and college curriculum. Co-author of five textbooks, he was one of the developers of the Quantitative Literacy Project that formed the basis of the data analysis strand in the curriculum standards of the National Council of Teachers of Mathematics. He also led the task force that developed the AP Statistics Program, for which he served as Chief Faculty Consultant. Dr. Scheaffer is a Fellow and past president of the American Statistical Association, a past chair of the Conference Board of the Mathematical Sciences, and an advisor on numerous statistics education projects. Madhuri S. Mulekar, Professor of Statistics, University of South Alabama, received her Ph.D. from Oklahoma State University. She is involved in teaching, research, and consulting activities with the faculty from the college of medicine, University of South Alabama hospitals, and Mitchell Cancer Institute. She has authored an exam-preparation book for Advanced Placement (AP) Statistics and has been involved with the AP Statistics exam since it was first administered, serving on the test development committee. She is actively involved in the American Statistical Association's efforts for improving statistics education by serving on various national committees such as Advisory Committee on Continuing Education and ASA-MAA joint committee on undergraduate statistics. Prof. Mulekar has published many research as well as teaching related articles. Recipient of many grants, she directed undergraduate research program in statistics that was funded by NSF. Dr. Mulekar is a Fellow of American Statistical Association and a recipient of outstanding scholar award by Phi Kappa Phi, the national honor society.
"About this title" may belong to another edition of this title.