Continuing to emphasize numerical and graphical methods, An Introduction to Generalized Linear Models, Third Edition provides a cohesive framework for statistical modeling. This new edition of a bestseller has been updated with Stata, R, and WinBUGS code as well as three new chapters on Bayesian analysis. Like its predecessor, this edition presents the theoretical background of generalized linear models (GLMs) before focusing on methods for analyzing particular kinds of data. It covers normal, Poisson, and binomial distributions; linear regression models; classical estimation and model fitting methods; and frequentist methods of statistical inference. After forming this foundation, the authors explore multiple linear regression, analysis of variance (ANOVA), logistic regression, log-linear models, survival analysis, multilevel modeling, Bayesian models, and Markov chain Monte Carlo (MCMC) methods. Using popular statistical software programs, this concise and accessible text illustrates practical approaches to estimation, model fitting, and model comparisons. It includes examples and exercises with complete data sets for nearly all the models covered.
Generalized linear models provide a unified theoretical and conceptual framework for many of the most commonly used statistical methods. In the ten years since publication of the first edition of this bestselling text, great strides have been made in the development of new methods and in software for generalized linear models and other closely related models. Thoroughly revised and updated, An Introduction to Generalized Linear Models, Second Edition continues to initiate intermediate students of statistics, and the many other disciplines that use statistics, in the practical use of these models and methods. The new edition incorporates many of the important developments of the last decade, including survival analysis, nominal and ordinal logistic regression, generalized estimating equations, and multi-level models. It also includes modern methods for checking model adequacy and examples from an even wider range of application. Statistics can appear to the uninitiated as a collection of unrelated tools. An Introduction to Generalized Linear Models, Second Edition illustrates how these apparently disparate methods are examples or special cases of a conceptually simple structure based on the exponential family of distribution, maximum likelihood estimation, and the principles of statistical modelling.
Linear models are central to the practice of statistics and form the foundation of a vast range of statistical methodologies. Julian J. Faraway's critically acclaimed Linear Models with R examined regression and analysis of variance, demonstrated the different methods available, and showed in which situations each one applies. Following in those footsteps, Extending the Linear Model with R surveys the techniques that grow from the regression model, presenting three extensions to that framework: generalized linear models (GLMs), mixed effect models, and nonparametric regression models. The author's treatment is thoroughly modern and covers topics that include GLM diagnostics, generalized linear mixed models, trees, and even the use of neural networks in statistics. To demonstrate the interplay of theory and practice, throughout the book the author weaves the use of the R software environment to analyze the data of real examples, providing all of the R commands necessary to reproduce the analyses. All of the data described in the book is available at http://people.bath.ac.uk/jjf23/ELM/ Statisticians need to be familiar with a broad range of ideas and techniques. This book provides a well-stocked toolbox of methodologies, and with its unique presentation of these very modern statistical techniques, holds the potential to break new ground in the way graduate-level courses in this area are taught.
Author: Walter W. Stroup
Publisher: CRC Press
Release Date: 2016-04-19
Generalized Linear Mixed Models: Modern Concepts, Methods and Applications presents an introduction to linear modeling using the generalized linear mixed model (GLMM) as an overarching conceptual framework. For readers new to linear models, the book helps them see the big picture. It shows how linear models fit with the rest of the core statistics curriculum and points out the major issues that statistical modelers must consider. Along with describing common applications of GLMMs, the text introduces the essential theory and main methodology associated with linear models that accommodate random model effects and non-Gaussian data. Unlike traditional linear model textbooks that focus on normally distributed data, this one adopts a generalized mixed model approach throughout: data for linear modeling need not be normally distributed and effects may be fixed or random. With numerous examples using SAS® PROC GLIMMIX, this book is ideal for graduate students in statistics, statistics professionals seeking to update their knowledge, and researchers new to the generalized linear model thought process. It focuses on data-driven processes and provides context for extending traditional linear model thinking to generalized linear mixed modeling. See Professor Stroup discuss the book.
Now in its third edition, this classic book is widely considered the leading text on Bayesian methods, lauded for its accessible, practical approach to analyzing data and solving research problems. Bayesian Data Analysis, Third Edition continues to take an applied approach to analysis using up-to-date Bayesian methods. The authors—all leaders in the statistics community—introduce basic concepts from a data-analytic perspective before presenting advanced methods. Throughout the text, numerous worked examples drawn from real applications and research emphasize the use of Bayesian inference in practice. New to the Third Edition Four new chapters on nonparametric modeling Coverage of weakly informative priors and boundary-avoiding priors Updated discussion of cross-validation and predictive information criteria Improved convergence monitoring and effective sample size calculations for iterative simulation Presentations of Hamiltonian Monte Carlo, variational Bayes, and expectation propagation New and revised software code The book can be used in three different ways. For undergraduate students, it introduces Bayesian inference starting from first principles. For graduate students, the text presents effective current approaches to Bayesian modeling and computation in statistics and related fields. For researchers, it provides an assortment of Bayesian methods in applied statistics. Additional materials, including data sets used in the examples, solutions to selected exercises, and software instructions, are available on the book’s web page.
Author: James S. Hodges
Publisher: CRC Press
Release Date: 2016-04-19
A First Step toward a Unified Theory of Richly Parameterized Linear Models Using mixed linear models to analyze data often leads to results that are mysterious, inconvenient, or wrong. Further compounding the problem, statisticians lack a cohesive resource to acquire a systematic, theory-based understanding of models with random effects. Richly Parameterized Linear Models: Additive, Time Series, and Spatial Models Using Random Effects takes a first step in developing a full theory of richly parameterized models, which would allow statisticians to better understand their analysis results. The author examines what is known and unknown about mixed linear models and identifies research opportunities. The first two parts of the book cover an existing syntax for unifying models with random effects. The text explains how richly parameterized models can be expressed as mixed linear models and analyzed using conventional and Bayesian methods. In the last two parts, the author discusses oddities that can arise when analyzing data using these models. He presents ways to detect problems and, when possible, shows how to mitigate or avoid them. The book adapts ideas from linear model theory and then goes beyond that theory by examining the information in the data about the mixed linear model’s covariance matrices. Each chapter ends with two sets of exercises. Conventional problems encourage readers to practice with the algebraic methods and open questions motivate readers to research further. Supporting materials, including datasets for most of the examples analyzed, are available on the author’s website.
Broadening its scope to nonstatisticians, Bayesian Methods for Data Analysis, Third Edition provides an accessible introduction to the foundations and applications of Bayesian analysis. Along with a complete reorganization of the material, this edition concentrates more on hierarchical Bayesian modeling as implemented via Markov chain Monte Carlo (MCMC) methods and related data analytic techniques. New to the Third Edition New data examples, corresponding R and WinBUGS code, and homework problems Explicit descriptions and illustrations of hierarchical modeling—now commonplace in Bayesian data analysis A new chapter on Bayesian design that emphasizes Bayesian clinical trials A completely revised and expanded section on ranking and histogram estimation A new case study on infectious disease modeling and the 1918 flu epidemic A solutions manual for qualifying instructors that contains solutions, computer code, and associated output for every homework problem—available both electronically and in print Ideal for Anyone Performing Statistical Analyses Focusing on applications from biostatistics, epidemiology, and medicine, this text builds on the popularity of its predecessors by making it suitable for even more practitioners and students.
Bridging the gap between theory and practice for modern statistical model building, Introduction to General and Generalized Linear Models presents likelihood-based techniques for statistical modelling using various types of data. Implementations using R are provided throughout the text, although other software packages are also discussed. Numerous examples show how the problems are solved with R. After describing the necessary likelihood theory, the book covers both general and generalized linear models using the same likelihood-based methods. It presents the corresponding/parallel results for the general linear models first, since they are easier to understand and often more well known. The authors then explore random effects and mixed effects in a Gaussian context. They also introduce non-Gaussian hierarchical models that are members of the exponential family of distributions. Each chapter contains examples and guidelines for solving the problems via R. Providing a flexible framework for data analysis and model building, this text focuses on the statistical methods and models that can help predict the expected value of an outcome, dependent, or response variable. It offers a sound introduction to general and generalized linear models using the popular and powerful likelihood techniques. Ancillary materials are available at www.imm.dtu.dk/~hm/GLM
Author: Francisco J. Samaniego
Publisher: CRC Press
Release Date: 2014-01-14
Provides a Solid Foundation for Statistical Modeling and Inference and Demonstrates Its Breadth of Applicability Stochastic Modeling and Mathematical Statistics: A Text for Statisticians and Quantitative Scientists addresses core issues in post-calculus probability and statistics in a way that is useful for statistics and mathematics majors as well as students in the quantitative sciences. The book’s conversational tone, which provides the mathematical justification behind widely used statistical methods in a reader-friendly manner, and the book’s many examples, tutorials, exercises and problems for solution, together constitute an effective resource that students can read and learn from and instructors can count on as a worthy complement to their lectures. Using classroom-tested approaches that engage students in active learning, the text offers instructors the flexibility to control the mathematical level of their course. It contains the mathematical detail that is expected in a course for "majors" but is written in a way that emphasizes the intuitive content in statistical theory and the way theoretical results are used in practice. More than 1000 exercises and problems at varying levels of difficulty and with a broad range of topical focus give instructors many options in assigning homework and provide students with many problems on which to practice and from which to learn.
Designed for a graduate course in applied statistics, Nonparametric Methods in Statistics with SAS Applications teaches students how to apply nonparametric techniques to statistical data. It starts with the tests of hypotheses and moves on to regression modeling, time-to-event analysis, density estimation, and resampling methods. The text begins with classical nonparametric hypotheses testing, including the sign, Wilcoxon sign-rank and rank-sum, Ansari-Bradley, Kolmogorov-Smirnov, Friedman rank, Kruskal-Wallis H, Spearman rank correlation coefficient, and Fisher exact tests. It then discusses smoothing techniques (loess and thin-plate splines) for classical nonparametric regression as well as binary logistic and Poisson models. The author also describes time-to-event nonparametric estimation methods, such as the Kaplan-Meier survival curve and Cox proportional hazards model, and presents histogram and kernel density estimation methods. The book concludes with the basics of jackknife and bootstrap interval estimation. Drawing on data sets from the author’s many consulting projects, this classroom-tested book includes various examples from psychology, education, clinical trials, and other areas. It also presents a set of exercises at the end of each chapter. All examples and exercises require the use of SAS 9.3 software. Complete SAS codes for all examples are given in the text. Large data sets for the exercises are available on the author’s website.
Modern computer-intensive statistical methods play a key role in solving many problems across a wide range of scientific disciplines. This new edition of the bestselling Randomization, Bootstrap and Monte Carlo Methods in Biology illustrates the value of a number of these methods with an emphasis on biological applications. This textbook focuses on three related areas in computational statistics: randomization, bootstrapping, and Monte Carlo methods of inference. The author emphasizes the sampling approach within randomization testing and confidence intervals. Similar to randomization, the book shows how bootstrapping, or resampling, can be used for confidence intervals and tests of significance. It also explores how to use Monte Carlo methods to test hypotheses and construct confidence intervals. New to the Third Edition Updated information on regression and time series analysis, multivariate methods, survival and growth data as well as software for computational statistics References that reflect recent developments in methodology and computing techniques Additional references on new applications of computer-intensive methods in biology Providing comprehensive coverage of computer-intensive applications while also offering data sets online, Randomization, Bootstrap and Monte Carlo Methods in Biology, Third Edition supplies a solid foundation for the ever-expanding field of statistics and quantitative analysis in biology.
Since 1975, The Analysis of Time Series: An Introduction has introduced legions of statistics students and researchers to the theory and practice of time series analysis. With each successive edition, bestselling author Chris Chatfield has honed and refined his presentation, updated the material to reflect advances in the field, and presented interesting new data sets. The sixth edition is no exception. It provides an accessible, comprehensive introduction to the theory and practice of time series analysis. The treatment covers a wide range of topics, including ARIMA probability models, forecasting methods, spectral analysis, linear systems, state-space models, and the Kalman filter. It also addresses nonlinear, multivariate, and long-memory models. The author has carefully updated each chapter, added new discussions, incorporated new datasets, and made those datasets available for download from www.crcpress.com. A free online appendix on time series analysis using R can be accessed at http://people.bath.ac.uk/mascc/TSA.usingR.doc. Highlights of the Sixth Edition: A new section on handling real data New discussion on prediction intervals A completely revised and restructured chapter on more advanced topics, with new material on the aggregation of time series, analyzing time series in finance, and discrete-valued time series A new chapter of examples and practical advice Thorough updates and revisions throughout the text that reflect recent developments and dramatic changes in computing practices over the last few years The analysis of time series can be a difficult topic, but as this book has demonstrated for two-and-a-half decades, it does not have to be daunting. The accessibility, polished presentation, and broad coverage of The Analysis of Time Series make it simply the best introduction to the subject available.
Survival Analysis Using S: Analysis of Time-to-Event Data is designed as a text for a one-semester or one-quarter course in survival analysis for upper-level or graduate students in statistics, biostatistics, and epidemiology. Prerequisites are a standard pre-calculus first course in probability and statistics, and a course in applied linear regression models. No prior knowledge of S or R is assumed. A wide choice of exercises is included, some intended for more advanced students with a first course in mathematical statistics. The authors emphasize parametric log-linear models, while also detailing nonparametric procedures along with model building and data diagnostics. Medical and public health researchers will find the discussion of cut point analysis with bootstrap validation, competing risks and the cumulative incidence estimator, and the analysis of left-truncated and right-censored data invaluable. The bootstrap procedure checks robustness of cut point analysis and determines cut point(s). In a chapter written by Stephen Portnoy, censored regression quantiles - a new nonparametric regression methodology (2003) - is developed to identify important forms of population heterogeneity and to detect departures from traditional Cox models. By generalizing the Kaplan-Meier estimator to regression models for conditional quantiles, this methods provides a valuable complement to traditional Cox proportional hazards approaches.
Author: Thomas Leonard
Publisher: Chapman & Hall
Release Date: 2000-01
Developed from long experience in teaching categorical analysis to a multidisciplinary mix of undergraduate and graduate students, A Course in Categorical Data Analysis presents the easiest, most straightforward ways of extracting real-life conclusions from contingency tables. The author uses a Fisherian approach to categorical data analysis and incorporates numerous examples and real data sets. Although he offers S-PLUS routines through the Interest, readers do not need full knowledge of a statistical software package.
Im Labyrinth der Daten. Dieses Lehrbuch wendet sich an all jene, die in Studium und Beruf mit ökologischen sowie Landschafts- und Umweltdaten arbeiten. Es ist als Begleittext zu Lehrveranstaltungen und zum Selbststudium geeignet und bietet Hilfe bei der Datenanalyse in wissenschaftlichen Arbeiten. Der Lehrtext behandelt neben relevanten statistischen Grundlagen, Ordinationsverfahren, Klassifikationen und Permutationsverfahren zur Prüfung multivariater Beziehungen. Anschaulich erläutert er die Methoden anhand von Beispieldatensätzen und unter weitgehendem Verzicht auf Formeln. Einzigartiger Einstieg für Anwender zur Auswertung eigener Daten.