Methods and Applications of Statistics in Clinical Trials, Volume 2 -  - E-Book

Methods and Applications of Statistics in Clinical Trials, Volume 2 E-Book

0,0
177,99 €

-100%
Sammeln Sie Punkte in unserem Gutscheinprogramm und kaufen Sie E-Books und Hörbücher mit bis zu 100% Rabatt.

Mehr erfahren.
Beschreibung

Methods and Applications of Statistics in Clinical Trials, Volume 2: Planning, Analysis, and Inferential Methods includes updates of established literature from the Wiley Encyclopedia of Clinical Trials as well as original material based on the latest developments in clinical trials. Prepared by a leading expert, the second volume includes numerous contributions from current prominent experts in the field of medical research. In addition, the volume features: * Multiple new articles exploring emerging topics, such as evaluation methods with threshold, empirical likelihood methods, nonparametric ROC analysis, over- and under-dispersed models, and multi-armed bandit problems * Up-to-date research on the Cox proportional hazard model, frailty models, trial reports, intrarater reliability, conditional power, and the kappa index * Key qualitative issues including cost-effectiveness analysis, publication bias, and regulatory issues, which are crucial to the planning and data management of clinical trials

Sie lesen das E-Book in den Legimi-Apps auf:

Android
iOS
von Legimi
zertifizierten E-Readern

Seitenzahl: 2045

Veröffentlichungsjahr: 2014

Bewertungen
0,0
0
0
0
0
0
Mehr Informationen
Mehr Informationen
Legimi prüft nicht, ob Rezensionen von Nutzern stammen, die den betreffenden Titel tatsächlich gekauft oder gelesen/gehört haben. Wir entfernen aber gefälschte Rezensionen.



Contents

Cover

Half Title page

Title page

Copyright page

Contributors

Preface

Chapter 1: Analysis of Over- and Underdispersed Data

1.1 Introduction

1.2 Overdispersed Binomial and Count Models

1.3 Other Approaches to Account for Overdispersion

1.4 Underdispersion

1.5 Software Notes

References

Chapter 2: Analysis of Variance (ANOVA)

2.1 Introduction

2.2 Factors, Levels, Effects, and Cells

2.3 Cell Means Model

2.4 One-Way Classification

2.5 Parameter Estimation

2.6 The R(.) Notation—Partitioning Sum of Squares

2.7 ANOVA—Hypothesis of Equal Means

2.8 Multiple Comparisons

2.9 Two-Way Crossed Classification

2.10 Balanced and Unbalanced Data

2.11 Interaction Between Rows and Columns

2.12 Analysis of Variance Table

References

Chapter 3: Assessment of Health-Related Quality of Life

3.1 Introduction

3.2 Choice of HRQOL Instruments

3.3 Establishment of Clear Objectives in HRQOL Assessments

3.4 Methods for HRQOL Assessment

3.5 HRQOL as the Primary End Point

3.6 Interpretation of HRQOL Results

3.7 Examples

3.8 Conclusion

References

Further Reading

Chapter 4: Bandit Processes and Response-Adaptive Clinical Trials: The Art of Exploration Versus Exploitation

4.1 Introduction

4.2 Exploration Versus Exploitation with Complete Observations

4.3 Exploration Versus Exploitation with Censored Observations

4.4 Conclusion

References

Chapter 5: Bayesian Dose-Finding Designs in Healthy Volunteers

5.1 Introduction

5.2 A Bayesian Decision-Theoretic Design

5.3 An Example of Dose Escalation in Healthy Volunteer Studies

5.4 Discussion

References

Chapter 6: Bootstrap

6.1 Introduction

6.2 Plug-In Principle

6.3 Monte Carlo Sampling— The “Second Bootstrap Principle”

6.4 Bias and Standard Error

6.5 Examples

6.6 Model Stability

6.7 Accuracy of Bootstrap Distributions

6.8 Bootstrap Confidence Intervals

6.9 Hypothesis Testing

6.10 Planning Clinical Trials

6.11 How Many Bootstrap Samples Are Needed

6.12 Additional References

References

Chapter 7: Conditional Power in Clinical Trial Monitoring

7.1 Introduction

7.2 Conditional Power

7.3 Weight-Averaged Conditional Power or Bayesian Predictive Power

7.4 Conditional Power of a Different Kind: Discordance Probability

7.5 Analysis of a Randomized Trial

7.6 Conditional Power: Pros and Cons

References

Chapter 8: Cost-Effectiveness Analysis

8.1 Introduction

8.2 Definitions and Design Issues

8.3 Cost and Effectiveness Data

8.4 The Analysis of Costs and Outcomes

8.5 Robustness and Generalizability in Cost-Effectiveness Analysis

References

Further Reading

Chapter 9: Cox-Type Proportional Hazards Models

9.1 Introduction

9.2 Cox Model for Univariate Failure Time Data Analysis

9.3 Marginal Models for Multivariate Failure Time Data Analysis

9.4 Practical Issues in Using the Cox Model

9.5 Examples

9.6 Extensions

9.7 Softwares and Codes

References

Further Reading

Chapter 10: Empirical Likelihood Methods in Clinical Experiments

10.1 Introduction

10.2 Classical EL: Several Ingredients for Theoretical Evaluations

10.3 The Relationship Between Empirical Likelihood and Bootstrap Methodologies

10.4 Bayes Methods Based on Empirical Likelihoods

10.5 Mixtures of Likelihoods

10.6 An Example: ROC Curve Analyses Based on Empirical Likelihoods

10.7 Applications of Empirical Likelihood Methodology in Clinical Trials or Other Data Analyses

10.8 Concluding Remarks

Appendix

References

Chapter 11: Frailty Models

11.1 Introduction

11.2 Univariate Frailty Models

11.3 Multivariate Frailty Models

11.4 Software

References

Chapter 12: Futility Analysis

12.1 Introduction

12.2 Common Statistical Approaches to Futility Monitoring

12.3 Examples

12.4 Discussion

References

Further Reading

Chapter 13: Imaging Science in Medicine I: Overview

13.1 Introduction

13.2 Advances in Medical Imaging

13.3 Evolutionary Developments in Imaging

13.4 Conclusion

References

Chapter 14: Imaging Science in Medicine, II: Basics of X-Ray Imaging

14.1 Introduction to Medical Imaging: Different Ways of Creating Visible Contrast Among Tissues

14.2 What the Body Does to the X-Ray Beam: Subject Contrast From Differential Attenuation of the X-Ray Beam by Various Tissues

14.3 What the X-Ray Beam Does to the Body: Known Medical Benefits Versus Possible Radiogenic Risks

14.4 Capturing the Visual Image: Analog (20th Century) X-Ray Image Receptors

Chapter 15: Imaging Science in Medicine, III: Digital (21st Century) X-Ray Imaging

15.1 The Computer in Medical Imaging

15.2 The Digital Planar X-Ray Modalities: Computed Radiography and Digital Radiography and Fluoroscopy

15.3 Digital Fluoroscopy and Digital Subtraction Angiography

15.4 Digital Tomosynthesis: Planar Imaging in Three Dimensions

15.5 Computed Tomography: Superior Contrast in Three-Dimensional X-Ray Attenuation Maps

Chapter 16: Intention-to-Treat Analysis

16.1 Introduction

16.2 Missing Information

16.3 The Intention-to-Treat Design

16.4 Efficiency of the Intent-to-Treat Analysis

16.5 Compliance-Adjusted Analyses

16.6 Conclusion

References

Further Reading

Chapter 17: Interim Analyses

17.1 Introduction

17.2 Opportunities and Dangers of Interim Analyses

17.3 The Development of Techniques for Conducting Interim Analyses

17.4 Methodology for Interim Analyses

17.5 An Example: Statistics for Lamivudine

17.6 Interim Analyses in Practice

17.7 Conclusions

References

Chapter 18: Interrater Reliability

18.1 Definition

18.2 The Importance of Reliability in Clinical Trials

18.3 How Large a Reliability Coefficient Is Large Enough?

18.4 Design and Analysis of Reliability Studies

18.5 Estimate of the Reliability Coefficient—Parametric

18.6 Estimation of the Reliability Coefficient— Nonparametric

18.7 Estimation of the Reliability Coefficient—Binary

18.8 Estimation of the Reliability Coefficient—Categorical

18.9 Strategies to Increase Reliability (Spearman–Brown Projection)

18.10 Other Types of Reliabilities

References

Chapter 19: Intrarater Reliability

19.1 Introduction

19.2 Intrarater Reliability for Continuous Scores

19.3 Nominal Scale Score Data

19.4 Ordinal and Interval Score Data

19.5 Concluding Remarks

References

Further Reading

Chapter 20: Kaplan—Meier Plot

20.1 Introduction

20.2 Estimation of Survival Function

20.3 Additional Topics

References

Chapter 21: Logistic Regression

21.1 Introduction

21.2 Fitting the Logistic Regression Model

21.3 The Multiple Logistic Regression Model

21.4 Fitting the Multiple Logistic Regression Model

21.5 Example

21.6 Testing for the Significance of the Model

21.7 Interpretation of the Coefficients of the Logistic Regression Model

21.8 Dichotomous Independent Variable

21.9 Polytomous Independent Variable

21.10 Continuous Independent Variable

21.11 Multivariate Case

References

Chapter 22: Metadata

22.1 Introduction

22.2 History/Background

22.3 Data Set Metadata

22.4 Analysis Results Metadata

22.5 Regulatory Submission Metadata

References

Chapter 23: Microarray

23.1 Introduction

23.2 What is a Microarray?

23.3 Other Array Technologies

23.4 Define Objectives of the Study

23.5 Experimental Design for Microarray

23.6 Data Extraction

23.7 Microarray Informatics

23.8 Statistical Analysis

23.9 Annotation

23.10 Pathway, GO, and Class-Level Analysis Tools

23.11 Validation of Microarray Experiments

23.12 Conclusions

References

Chapter 24: Multi-Armed Bandits, Gittins Index, and Its Calculation

24.1 Introduction

24.2 Mathematical Formulation of Multi-Armed Bandits

24.3 Off-Line Algorithms for Computing Gittins Index

24.4 On-Line Algorithms for Computing Gittins Index

24.5 Computing Gittins Index for the Bernoulli Sampling Process

24.6 Conclusion

References

Chapter 25: Multiple Comparisons

25.1 Introduction

25.2 Strong and Weak Control of the FWE

25.3 Criteria for Deciding Whether Adjustment is Necessary

25.4 Implicit Multiplicity: Two-Tailed Testing

25.5 Specific Multiple Comparison Procedures

References

Chapter 26: Multiple Evaluators

26.1 Introduction

26.2 Agreement for Continuous Data

26.3 Agreement for Categorical Data

26.4 Summary and Discussion

References

Chapter 27: Noncompartmental Analysis

27.1 Introduction

27.2 Terminology

27.3 Objectives and Features of Noncompartmental Analysis

27.4 Comparison of Noncompartmental and Compartmental Models

27.5 Assumptions of NCA and Its Reported Descriptive Statistics

27.6 Calculation Formulas for NCA

27.7 Guidelines for Performance of NCA Based on Numerical Integration

27.8 Conclusions and Perspectives

References

Further Reading

Chapter 28: Nonparametric ROC Analysis for Diagnostic Trials

28.1 Introduction

28.2 Different Aspects of Study Design

28.3 Nonparametric Models and Hypotheses

28.4 Point Estimator

28.5 Asymptotic Distribution and Variance Estimator

28.6 Derivation of the Confidence Interval

28.7 Statistical Tests

28.8 Adaptations for Cluster Data

28.9 Results of a Diagnostic Study

28.10 Summary and Final Remarks

References

Chapter 29: Optimal Biological Dose for Molecularly Targeted Therapies

29.1 Introduction

29.2 Phase I Dose-Finding Designs for Cytotoxic Agents

29.3 Phase I Dose-Finding Designs for Molecularly Targeted Agents

29.4 Discussion

References

Further Reading

Chapter 30: Over- and Underdispersion Models

30.1 Introduction

30.2 Count Dispersion Models

30.3 Count Explanatory Models

30.4 Summary and Final Remarks

References

Chapter 31: Permutation Tests in Clinical Trials

31.1 Randomization Inference—Introduction

31.2 Permutation Tests—How They Work

31.3 Normal Approximation to Permutation Tests

31.4 Analyze as You Randomize

31.5 Interpretation of Permutation Analysis Results

31.6 Summary

References

Chapter 32: Pharmacoepidemiology, Overview

32.1 Introduction

32.2 The Case-Crossover Design

32.3 Confounding Bias

32.4 Risk Functions Over Time

32.5 Probabilistic Approach for Causality Assessment

32.6 Methods Based on Prescription Data

References

Chapter 33: Population Pharmacokinetic and Pharmacodynamic Methods

33.1 Introduction

33.2 Terminology

33.3 Fixed Effects Models

33.4 Random Effects Models

33.5 Model Building and Parameter Estimation

33.6 Software

33.7 Model Evaluation

33.8 Stochastic Simulation

33.9 Experimental Design

33.10 Applications

References

Further Reading

Chapter 34: Proportions: Inferences and Comparisons

34.1 Introduction

34.2 One-Sample Case

34.3 Two Independent Samples

34.4 Note on Software

References

Chapter 35: Publication Bias

35.1 Publication Bias and the Validity of Research Reviews

35.2 Research on Publication Bias

35.3 Data Suppression Mechanisms Related to Publication Bias

35.4 Prevention of Publication Bias

35.5 Assessment of Publication Bias

35.6 Impact of Publication Bias

References

Further Reading

Chapter 36: Quality of Life

36.1 Background

36.2 Measuring Health-Related Quality of Life

36.3 Development and Validation of HRQoL Measures

36.4 Use in Research Studies

36.5 Interpretation/Clinical Significance

36.6 Conclusions

References

Chapter 37: Relative Risk Modeling

37.1 Introduction

37.2 Why Model Relative Risks?

37.3 Data Structures and Likelihoods

37.4 Approaches to Model Specification

37.5 Mechanistic Models

References

Chapter 38: Sample Size Considerations for Morbidity/Mortality Trials

38.1 Introduction

38.2 General Framework for Sample Size Calculation

38.3 Choice of Test Statistics

38.4 Adjustment of Treatment Effect

38.5 Informative Noncompliance

References

Chapter 39: Sample Size for Comparing Means

39.1 Introduction

39.2 One-Sample Design

39.3 Two-Sample Parallel Design

39.4 Two-Sample Crossover Design

39.5 Multiple-Sample One-Way ANOVA

39.6 Multiple-Sample Williams Design

39.7 Discussion

References

Chapter 40: Sample Size for Comparing Proportions

40.1 Introduction

40.2 One-Sample Design

40.3 Two-Sample Parallel Design

40.4 Two-Sample Crossover Design

40.5 Relative Risk—Parallel Design

40.6 Relative Risk—Crossover Design

40.7 Discussion

References

Chapter 41: Sample Size for Comparing Time-to-Event Data

41.1 Introduction

41.2 Exponential Model

41.3 Cox’s Proportional Hazards Model

41.4 Log-Rank Test

41.5 Discussion

References

Chapter 42: Sample Size for Comparing Variabilities

42.1 Introduction

42.2 Comparing Intrasubject Variabilities

42.3 Comparing Intersubject Variabilities

42.4 Comparing Total Variabilities

42.5 Discussion

References

Chapter 43: Screening, Models of

43.1 Introduction

43.2 What is Screening?

43.3 Why Use Modeling?

43.4 Characteristics of Screening Models

43.5 A Simple Disease and Screening Model

43.6 Analytic Models for Cancer

43.7 Simulation Models for Cancer

43.8 Model Fitting and Validation

43.9 Models for Other Diseases

43.10 Current State and Future Directions

References

Chapter 44: Screening Trials

44.1 Introduction

44.2 Design Issues

44.3 Sample Size

44.5 Analysis

44.6 Trial Monitoring

References

Chapter 45: Secondary Efficacy End Points

45.1 Introduction

45.2 Literature Review

45.3 Review of Methodology for Multiplicity Adjustment and Gatekeeping Strategies for Secondary End Points

45.4 Summary

References

Further Reading

Chapter 46: Sensitivity, Specificity, and Receiver Operator Characteristic (ROC) Methods

46.1 Evaluating a Single Binary Test Against a Binary Criterion

46.2 Evaluation of a Single Binary Test: ROC Methods

46.3 Evaluation of a Test Response Measured on an Ordinal Scale: ROC Methods

46.4 Evaluation of Multiple Different Tests

46.5 The Optimal Sequence of Tests

46.6 Sampling and Measurement Issues

46.7 Summary

References

Chapter 47: Software for Genetics/Genomics

47.1 Introduction

47.2 Data Management

47.3 Genetic Analysis

47.4 Genomic Analysis

47.5 Other

References

Further Reading

Chapter 48: Stability Study Designs

48.1 Introduction

48.2 Stability Study Designs

48.3 Criteria for Design Comparison

48.4 Stability Protocol

48.5 Basic Design Considerations

48.6 Conclusions

References

Chapter 49: Subgroup Analysis

49.1 Introduction

49.2 The Dilemma of Subgroup Analysis

49.3 Planned Versus Unplanned Subgroup Analysis

49.4 Frequentist Methods

49.5 Testing Treatment by Subgroup Interactions

49.6 Subgroup Analyses in Positive Clinical Trials

49.7 Confidence Intervals for Treatment Effects within Subgroups

49.8 Bayesian Methods

References

Chapter 50: Survival Analysis, Overview

50.1 Introduction

50.2 History

50.3 Survival Analysis Concepts

50.4 Nonparametric Estimation and Testing

50.5 Parametric Inference

50.6 Comparison with Expected Survival

50.7 The Cox Regression Model

50.8 Other Regression Models for Survival Data

50.9 Multistate Models

50.10 Other Kinds of Incomplete Observation

50.11 Multivariate Survival Analysis

50.12 Concluding Remarks

References

Chapter 51: The FDA and Regulatory Issues

51.1 Caveat

51.2 Introduction

51.3 Chronology of Drug Regulation in the United States

51.4 FDA Basic Structure

51.5 IND Application Process

51.6 Drug Development and Approval Time Frame

51.7 NDA Process

51.8 U.S. Pharmacopeia and FDA

51.9 CDER Freedom of Information Electronic Reading Room

51.10 Conclusion

Chapter 52: The Kappa Index

52.1 Introduction

52.2 The Kappa Index

52.3 Inference for Kappa via Generalized Estimating Equations

52.4 The Dependence of Kappa on Marginal Rates

52.5 General Remarks

References

Chapter 53: Treatment Interruption

53.1 Introduction

53.2 Therapeutic TI Studies in HIV/AIDS

53.3 Management of Chronic Disease

53.4 Analytic Treatment Interruption in Therapeutic Vaccine Trials

53.5 Randomized Discontinuation Designs

53.6 Final Comments

References

Chapter 54: Trial Reports: Improving Reporting, Minimizing Bias, and Producing Better Evidence-Based Practice

54.1 Introduction

54.2 Reporting Issues in Clinical Trials

54.3 Moral Obligation to Improve the Reporting of Trials

54.4 Consequences of Poor Reporting of Trials

54.5 Distinguishing Between Methodological and Reporting Issues

54.6 One Solution to Poor Reporting: CONSORT 2010 and CONSORT Extensions

54.7 Impact of CONSORT

54.8 Guidance for Reporting Randomized Trial Protocols: SPIRIT

54.9 Trial Registration

54.10 Final Thoughts

References

Chapter 55: U.S. Department of Veterans Affairs Cooperative Studies Program

55.1 Introduction

55.2 History of the Cooperative Studies Program (CSP)

55.3 Organization and Functioning of the CSP

55.4 Roles of the Biostatistician and Pharmacist in the CSP

55.5 Ongoing and Completed Cooperative Studies (1972–2000)

55.6 Current Challenges and Opportunities

55.7 Concluding Remarks

References

Chapter 56: Women’s Health Initiative: Statistical Aspects and Selected Early Results

56.1 Introduction

56.2 WHI Clinical Trial and Observational Study

56.3 Study Organization

56.4 Principal Clinical Trial Comparisons, Power Calculations, and Safety and Data Monitoring

56.5 Biomarkers and Intermediate Outcomes

56.6 Data Management and Computing Infrastructure

56.7 Quality Assurance Program Overview

56.8 Early Results from the WHI Clinical Trial

56.9 Summary and Discussion

References

Chapter 57: World Health Organization (WHO): Global Health Situation

57.1 Introduction

57.2 Program Activities to the End of the Twentieth Century

57.3 Vision for the Use and Generation of Data in the First Quarter of the Twenty-First Century

Reference

Further Reading

Index

Methods and Applications of Statistics in Clinical Trials

WILEY SERIES IN METHODS AND APPLICATIONS OF STATISTICS

Advisory Editor

N. Balakrishnan

McMaster University, Canada

The Wiley Series in Methods and Applications of Statistics is a unique grouping of research that features classic contributions from Wiley’s Encyclopedia of Statistical Sciences, Second Edition (ESS, 2e) alongside newly written articles that explore various problems of interest and their intrinsic connection to statistics. The goal of this collection is to encompass an encyclopedic scope of coverage within individual books that unify the most important and interesting applications of statistics within a specific field of study. Each book in the series successfully upholds the goals of ESS, 2e by combining established literature and newly-developed contributions written by leading academics, researchers, and practitioners in a comprehensive and accessible format. The result is a succinct reference that unveils modern, cutting-edge approaches to acquiring, analyzing, and presenting data across diverse subject areas.

WILEY SERIES IN METHODS AND APPLICATIONS OF STATISTICS

Balakrishnan • Methods and Applications of Statistics in the Life and Health Sciences

Balakrishnan • Methods and Applications of Statistics in Business, Finance, and Management Science

Balakrishnan • Methods and Applications of Statistics in Engineering, Quality Control, and the Physical Sciences

Balakrishnan • Methods and Applications of Statistics in the Social and Behavioral Sciences

Balakrishnan • Methods and Applications of Statistics in the Atmospheric and Earth Sciences

Balakrishnan • Methods and Applications of Statistics in Clinical Trials, Volume 1: Concepts, Principles, Trials, and Designs

Balakrishnan • Methods and Applications of Statistics in Clinical Trials, Volume 2: Planning, Analysis, and Inferential Methods

Copyright © 2014 by John Wiley & Sons, Inc. All rights reserved.

Published by John Wiley & Sons, Inc., Hoboken, New Jersey. All rights reserved. Published simultaneously in Canada.

No part of this publication may be reproduced, stored in a retrieval system, or transmitted in any form or by any means, electronic, mechanical, photocopying, recording, scanning, or otherwise, except as permitted under Section 107 or 108 of the 1976 United States Copyright Act, without either the prior written permission of the Publisher, or authorization through payment of the appropriate per-copy fee to the Copyright Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, (978) 750-8400, fax (978) 750-4470, or on the web at www.copyright.com. Requests to the Publisher for permission should be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at http://www.wiley.com/go/permission.

Limit of Liability/Disclaimer of Warranty: While the publisher and author have used their best efforts in preparing this book, they make no representations or warranties with respect to the accuracy or completeness of the contents of this book and specifically disclaim any implied warranties of merchantability or fitness for a particular purpose. No warranty may be created or extended by sales representatives or written sales materials. The advice and strategies contained herein may not be suitable for your situation. You should consult with a professional where appropriate. Neither the publisher nor author shall be liable for any loss of profit or any other commercial damages, including but not limited to special, incidental, consequential, or other damages.

For general information on our other products and services or for technical support, please contact our Customer Care Department within the United States at (800) 762-2974, outside the United States at (317) 572-3993 or fax (317) 572-4002.

Wiley also publishes its books in a variety of electronic formats. Some content that appears in print may not be available in electronic format. For information about Wiley products, visit our web site at www.wiley.com.

Library of Congress Cataloging-in-Publication Data:

Methods and applications of statistics in clinical trials vol 2/ [edited by] N. Balakrishnan.     p.; cm. — (Methods and applications of statistics)  Includes bibliographical references and index. ISBN 978-1-118-30476-1 (cloth)  I. Balakrishnan, N., 1956– editor of compilation. II. Series: Wiley series in methods and applications of statistics.  [DNLM: 1. Clinical Trials as Topic. 2. Statistics as Topic. QV 771.4]  R853.C55 610.72’4—dc23

2013034342

Contributors

Per Kragh Andersen, University of Copenhagen, Copenhagen, Denmark, [email protected]

Garnet L. Anderson, Fred Hutchinson Cancer Research Center, Seattle, WA, [email protected]

Chul Ahn, University of Texas Southwestern Medical Center, Dallas, TX, [email protected]

Edgar Brunner, Professor Emeritus of Biostatistics, University Medical Center, Göttingen, Germany, [email protected]

Jürgen B. Bulitta, State University of New York at Buffalo, Buffalo, NY, [email protected]

Jianwen Cai, University of North Carolina, Chapel Hill, NC, [email protected]

Patrizio Capasso, University of Kentucky, Lexington, KY, [email protected]

Robert C. Capen, Merck Research Laboratories West Point, PA

Jhelum Chakravorty, McGill University, Montreal, QC, Canada, [email protected]

Chi Wan Chen, Pfizer Inc., New York, NY

David H. Christiansen, Christiansen Consulting, Boise, ID

Shein-Chung Chow, Duke University Durham, NC, [email protected]

Joseph F. Collins

Jason T. Connor, Berry Consultants, Orlando, FL, [email protected]

Richard J. Cook, University of Waterloo, Waterloo, ON, Canada, [email protected]

Xiangqin Cui, University of Alabama at Birmingham, Birmingham, AL. [email protected]

C. B. Dean, Western University, Western Science Centre, London, ON, Canada, [email protected]

Yu Deng, University of North Carolina, Chapel Hill, NC, [email protected]

Diane L. Fairclough, University of Colorado Health Sciences, Center Denver, CO, [email protected]

John R. Feussner, Medical University of South Carolina, Charleston, SC

Boris Freidlin, National Cancer Institute, Bethesda, MD, [email protected]

Patricia A. Granz, UCLA Jonsson Comprehensive Cancer Center, Los Angeles, CA, [email protected]

Courtney Gray-McGuire, Case Western Reserve University, Cleveland, OH, [email protected]

Birgit Grund, University of Minnesota, Minneapolis, MN, [email protected]

Kilem L. Gwet, Advanced Analytics, LLC, Gaithersburg, MD, [email protected]

H. R. Hapsara, World Health Organization, Geneva, Switzerland, [email protected]

William R. Hendee, Medical College of Wisconsin, Milwaukee, WI, [email protected]

William G. Henderson

Tim Hesterberg, Insightful Corporation, Seattle, WA

Nicholas H. G. Holford, University of Auckland, Auckland, New Zealand, [email protected]

Norbert Holländer, University Hospital of Freiburg, Freiburg, Germany, [email protected]

David W. Hosmer, University of Massachusetts, Amherst, MA, [email protected]

Alan D. Hutson, University at Buffalo, Buffalo, NY, [email protected]

Peter B. Imrey, Cleveland Clinic, Cleveland, OH, [email protected]

Elizabeth Juarez-Colunga, University of Colorado Denver, Aurora, CO, [email protected]

Seung-Ho Kang, Ewha Woman’s University, Seoul, South Korea

Jörg Kaufmann, AG Schering SBU Diagnostics & Radiopharmaceuticals, Berlin, Germany

Niels Keiding, University of Copenhagen, Copenhagen, Denmark, [email protected]

Célestin C. Kokonendji, University of Franche-Comté, Besançon, France, [email protected]

Helena Chmura Kraemer, Stanford University, Palo Alto, CA, [email protected]

John M. Lachin, George Washington University, Washington, DC, [email protected]

Philip W. Lavori, Stanford University School of Medicine, Standford, CA, [email protected]

Morven Leese, Institute of Psychiatry—Health Services and Population Research Department, London, UK

Stanley Lemeshow, Ohio State University, Columbus, OH, [email protected]

Jason J. Z. Liao, Merck Research Laboratories West Point, PA, [email protected]

Tsae-Yun Daphne Lin, Center for Drug Evaluation and Research, U.S. Food and Drug Administration, Rockville, MD, [email protected]

Qing Lu, Michigan State University, East Lansing, MI, [email protected]

Aditya Mahajan, McGill University, Montreal, QC, Canada, [email protected]

Michael A. McIsaac, University of Waterloo, Waterloo, ON, Canada, [email protected]

David Moher, Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, ON, Canada and Department of Epidemiology and Community Medicine, University of Ottawa, Ottawa, ON, Canada, [email protected]

Grier P. Page, RTI International, Research Triangle Park, NC, [email protected]

Peter Peduzzi, Yale School of Public Health, New Haven, CT, [email protected]

Ross L. Prentice, Fred Hutchinson Cancer Research Center, Seattle, WA, [email protected]

Philip C. Prorok, National Institutes of Health, Bethesda, MD, [email protected]

Michael A. Proschan, National Institute of Allergy and Infectious Diseases, Bethesda, MD, [email protected]

Frank Rockhold, GlaxoSmithKline R&D, King of Prussia, PA, [email protected]

Hannah R. Rothstein, City University of New York, NY, [email protected]

W. Janusz Rzeszotarski, U.S. Food and Drug Administration, Rockville, MD

Mike R. Sather, Department of Veterans Affairs, Albuquerque, NM, [email protected]

Tony Segreti, Research Triangle Institute, Research Triangle, NC

Larissa Shamseer, Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, ON, Canada and Department of Epidemiology and Community Medicine, University of Ottawa, Ottawa, ON, Canada, [email protected]

Joanna H. Shih, National Cancer Institute, Bethesda, MD

Richard M. Simon, National Cancer Institute, Bethesda, MD, [email protected]

Yeunjoo Song, Case Western Reserve University, Cleveland, OH

Chris Stevenson, Monash University, Victoria, Australia, [email protected]

Samy Suissa, McGill University, Montreal, QC, Canada, [email protected]

Ming T. Tan, Georgetown University, Washington, DC, [email protected]

Duncan C. Thomas University of Southern California, Los Angeles, CA, [email protected]

Susan Todd, University of Reading Reading, Berkshire, UK, [email protected]

Lucy Turner, Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, ON, Canada, [email protected]

Albert Vexler, University at Buffalo, Buffalo, NY, [email protected]

Hansheng Wang, Peking University Beijing, P. R. China, [email protected]

Xikui Wang, University of Manitoba, Winnipeg, MN, Canada, [email protected]

C. S. Wayne Weng, Chung Yuan Christian University, Chungli, Taiwan

Andreas Wienke, University Halle-Wittenberg, Halle, Germany, [email protected]

Anthony B. Wolbarst, University of Kentucky, Lexington, KY, [email protected]

Andrew R. Wyant, University of Kentucky, Lexington, KY, [email protected]

Yang Xie, University of Texas Southwestern Medical Center, Dallas, TX

Jihnhee Yu, University at Buffalo, Buffalo, NY, [email protected]

Antonia Zapf, University Medical Center, Göttingen, Germany, [email protected]

Donglin Zeng, University of North Carolina, Chapel Hill, NC, [email protected]

Yinghui Zhou, The University of Reading, Reading, Berkshire, UK

David M. Zucker, Hebrew University of Jerusalem Jerusalem, Israel, [email protected]

Preface

Planning, developing, and implementing clinical trials, have become an important and integral part of life. More and more efforts and care go into conducting various clinical trials as they have been responsible in making key advances in medicine and treatments to different illnesses. Today, clinical trials have become mandatory in the development and evaluation of modern drugs and in identifying the association of risk factors to diseases. Due to the complexity of various issues surrounding clinical trials, regulatory agencies oversee their approval and also ensure impartial review. The main purpose of this two-volume handbook is to provide a detailed exposition of historical developments and also to highlight modern advances on methods and analysis for clinical trials.

It is important to mention that the four-volume Wiley Encyclopedia of Clinical Trials served as a basis for this handbook. While many pertinent entries from this Encyclopedia have been included here, a number of them have been updated to reflect recent developments on their topics. Some new articles detailing modern advances in statistical methods in clinical trials and their applications have also been included.

A volume of this size and nature cannot be successfully completed without the cooperation and support of the contributing authors, and my sincere thanks and gratitude go to all of them. Thanks are also due to Mr. Steve Quigley and Ms. Sari Friedman (of John Wiley & Sons, Inc.) for their keen interest in this project from day one, as well as for their support and constant encouragement (and, of course, occasional nudges, too) throughout the course of this project. Careful and diligent work of Mrs. Debbie Iscoe in the typesetting of this volume and of Angioline Loredo at the production state, is gratefully acknowledged. Partial financial support of the Natural Sciences and Engineering Research Council of Canada also assisted in the preparation of this handbook, and this support is much appreciated.

This is the seventh in a series of handbooks on methods and applications of statistics. While the first handbook has focused on life and health sciences, the second handbook has focused on business, finance, and management sciences, the third has focused on engineering, quality control, and physical sciences, the fourth has focused on behavioral and social sciences, the fifth has focused on atmospheric and earth sciences, and the sixth handbook has concentrated on methods and applications of statistics to clinical trials. This is the second of two volumes describing in detail statistical developments concerning clinical trials, focusing specifically on planning, analysis, and inferential methods.

It is my sincere hope that this handbook and the others in the series will become basic reference resources for those involved in these fields of research!

PROF. N. BALAKRISHNAN McMASTER UNIVERSITY

Hamilton, Canada February 2014

Chapter 1

Analysis of Over- and Underdispersed Data

Elizabeth Juarez-Colunga and C. B. Dean

1.1 Introduction

In the analysis of discrete data, for example, count data analyzed under a Poisson model, or binary data analyzed under a binomial model quite often the empirical variance exceeds the theoretical variance under the presumed model. This phenomenon is called overdispersion. If overdispersion is ignored, standard errors of parameter estimates will be underestimated, and therefore p-values for tests and hypotheses will be too small, leading to incorrectly declaring a predictor as significant when in fact it may not be.

The Poisson and binomial distributions are simple models but have strict assumptions. In particular, they assume a special mean-variance relationship since each of these distributions is determined by a single parameter. On the other hand, the normal distribution is determined by two parameters, the mean μ and variance σ2, which characterize the location and the spread of the data around the mean. In both the Poisson and binomial distributions, the variance is fixed once the mean or the probability of success has been defined.

Hilbe [25] provides a very comprehensive discussion of what he calls apparent overdispersion, which refers to scenarios in which the data exhibit variation beyond what can be explained by the model and this lack of fit is due to several “fixable” reasons. These reasons may be omitting important predictors in the model, the presence of outliers, omitting important interactions as predictors, the need of a transformation for a predictor, and misspecifying the link function for relating the mean response to the predictors. Hilbe [25] also discusses how to recognize overdispersion, and how to adjust for it when it is present beyond apparent cases, and provides an excellent overall review of the topic.

It is important to note that if apparent overdispersion has been ruled out, in log-linear or logistic analyses, the point estimates of the covariate effects will be quite similar regardless of whether overdispersion is accounted for or not. Hence, treatment and other effects will not be aberrant or give a hint of the presence of overdispersion. As well, this suggests that adjusting for overdispersion can be handled through adjustments of variance estimates [35].

Evidence of apparent or real overdispersion exists when the Pearson or deviance residuals are too large [6]; the corresponding Pearson and deviance goodness-of-fit statistics indicate a poor fit. Several tests have been developed for overdispersion in the context of Poisson or binomial analyses [11, 12, 54], as well as in the context of zero-heavy data [30, 51, 53, 52],

1.2 Overdispersed Binomial and Count Models

1.2.1 Overdispersed Binomial Model

In the binomial context, overdispersion typically arises because the independence assumption is violated. This is commonly caused by clustering of responses; for instance, clinics or hospitals may induce a clustering effect due to differences in patient care strategies across institutions.

(1)

1.2.2 Overdispersed Poisson Model

A simple way to incorporate overdispersion is through the use of an individual-specific random effect vi. Given vi, and the covariate vector xi corresponding to the ith individual, the counting process Ni(t) may be modeled as a Poisson process with intensity function

(2)

(3)

where

(4)

and

(5)

(6)

is the likelihood for a mixed Poisson model based on the total counts observed for individual i. The likelihood L(θ) becomes the negative binomial if vi is gamma distributed (i.e., G(vi;.) is a gamma distribution). If there is a single panel, Lp(θ) [see Equation (3)] will reduce to the simple mixed Poisson kernel, L(θ), where the response is the total count of events in the entire follow-up time.

Overdispersed recurrent event counts are often encountered in trials where the main interest is to test whether certain treatments are effective in reducing the recurrences of events, as illustrated in the example Section 1.2.3. In this case, the βs are parametrized such that the treatment effects are measured relative to treatment 1, so that β1 reflects the overall mean and α describes the shape of the intensity function ρ(t, α); common forms of ρ(t, α) are exponential (exp(αt)) and Weibull (αtα−1).

1.2.3 Example

Consider a clinical trial, conducted by the Veterans Administration Co-operative Urological Research Group, that studied the effects of placebo pills, pyridoxine pills, and periodic instillation of thiotepa into the bladder on the frequency of recurrence of bladder cancer [8]. The data appear in Andrews and Herzberg [2]. All 116 patients had bladder cancer when they entered the study; the tumors were removed, and the patients were randomly assigned to one of the three treatments. Here we consider estimation of the treatment effect under both a design with continuous follow-up, as in the study, and an artificial design, for illustrative purposes, with 2 equally spaced scheduled follow-up visits over 64 months; for the panel design, we record information on event recurrences at the scheduled follow-up times and at termination times.

Table 1: Parameter estimates (Est) and their standard errors (SE), resulting from the Poisson and negative binomial (NB) likelihood fit to the bladder cancer data. The regression parameters β1, β2, β3 correspond to the three treatment groups, parametrized with respect to the placebo, and α parametrizes the baseline intensity function.

1.3 Other Approaches to Account for Overdispersion

1.3.1 Generalized Linear Mixed Model

A general class of models that encompasses the incorporation of several random effects, not necessarily independent, is generalized linear mixed models. This may include an individual-specific random effect, as discussed above and also more complex structures that can accommodate dependencies in outcome variables as well as in random effects. A generalized linear mixed model specifies that

(7)

where μ and xi are the mean of the response and the vector of covariates, corresponding to the ith individual, respectively; zi is a vector of covariates determining the random effects structure, and the vector of random effects γ is distributed with a mean of zero and finite variance matrix; g is the link function. Conditional on γ, the responses are assumed to have a distribution in the exponential family, for example, Poisson or binomial.

Maximum likelihood estimation involves q-dimensional integration, where q is the dimension of γ; often random effects are assumed to be Gaussian. Tuerlinckx et al. [47] provide a review of methods used for estimation of generalized linear mixed models, discussing methods used to approximate the integral when integrating over the random effects distribution and methods that approximate the integrand of the marginal likelihood. Within the first set of methods, quadrature, Monte Carlo-based numerical methods, and expectation-maximization algorithms are reviewed; within the second, which approximate the integrand, Laplace’s and quasi-likelihood methods are considered.

With overdispersion present, the use of the Poisson or binomial maximum likelihood equations for estimating the regression parameters in the mean is still valid. The usual likelihood equations obtained assuming a generalized linear model are unbiased, estimating equations regardless of any misspecification of the variance structure. Hence, an alternate approach to the use of generalized linear mixed models is to use the corresponding generalized linear model and adjust variance estimates. In this case, often as a final step, the variance is estimated by the sandwich estimator formula, which is an empirical estimator; this approach has become very popular in the last few decades [31, 46].

Nonparametric approaches for handling random effects have also been developed. Lindsay [32] provides a classic comprehensive source on the topic. More recently, Böhning and Seidel [3] provide a review of advances in estimation in mixture models, including nonparametric estimation, the EM algorithm, likelihood ratio tests for testing the number of components in the mixture, special mixtures such as zero-inflated Poisson models, multivariate mixtures, and testing and adjusting for heterogeneity. Groeneboom et al. [20] propose an algorithm, called the support reduction algorithm, to estimate M-estimators in mixture models through iterative unconstrained optimization. Wang [50] proposes three algorithms based on the constrained Newton method [49] to estimate semiparametric mixture models. In these, the mixture distribution G is left unspecified and a finite-dimensional parameter β is common to all mixture components. The three methods are based on (1) alternating estimation of parameters G and β, (2) profiling the likelihood, and (3) modifying the support set; they all use the constrained Newton method and an additional optimization algorithm for unconstrained problems.

There have been some efforts in combining models that account both for overdispersion and clustering effects, the latter perhaps arising from longitudinal measurements. Booth et al. [4] propose a negative binomial model to account for overdispersion, which incorporates random effects, in the linear predictor of the mean, to account for such clustering effects; numerical methods or the EM algorithm is proposed for estimation. Along the same lines, Molenberghs et al. [36] discuss a similar model with gamma and normal random effects to account for overdispersion and clustering effects and Molenberghs et al. [37] generalize the model to a family of generalized linear models for repeated measures with normal and conjugate random effects. Iddi and Molenberghs [27] discuss a marginalized model to account for overdispersion and longitudinal correlation.

Serial correlation may also be accommodated, in addition to overdispersion, through Gaussian time series [23]. Jowaheer and Sutradhar [28] use generalized estimating equations to account for autocorrelation structures as well as overdispersion in longitudinal counts. Parameters are estimated via a two-stage iterative procedure. Henderson and Shimakura [24] and Fiocco et al. [18] discuss a model that, conditional on a frailty, follows a Poisson distribution for counts of events and uses a gamma serially correlated process to model dependency between observations arising from the same individual. In this generalization of the individual frailty model, the random effects are first-order autocor-related. Henderson and Shimakura [24] estimate the parameters of the model using a composite likelihood method based on pairs of time points, while Fiocco et al. [18] discuss an alternative approach using a two-stage procedure. In the two-stage procedure all parameters except the frailty correlation are estimated at the first stage while, in the second stage, the correlation of the frailties is estimated, based on pairs of observations.

1.3.2 Zero-Inflated Models

Sometimes apparent overdispersion is induced by the presence of another mode in the data, often at 0. In these cases, the remedy is to fit a model that handles the extra zeros that cannot be accounted for through the Poisson distribution [7, 40]. However, it may also occur that there is overdispersion beyond zero-inflation, in which case models accounting for both extra zeros and overdispersion have been developed, for example, the zero-inflated negative binomial [19]. There has been great interest in the last decade in accounting as well for correlation structures such as longitudinal, cluster, or spatial components. Ainsworth [1] provides a review of zero-inflated models, pointing out several references, mainly in the field of environmental statistics, that address such challenges in zero-heavy models. Hall [22] considers the challenges of simultaneously modeling within—and between—subject heterogeneity, while Dobbie and Welsh [14] consider serial correlation; both of these are framed in the context of zero-heavy count data models. Along the same lines, Wan and Chan [48] discuss a modeling approach based on a geometric process that accounts for overdispersion in zero-heavy models and, additionally, can handle serial correlation.

1.4 Underdispersion

1.5 Software Notes

Software for incorporating overdispersion includes SAS [42], using, for instance, procedures LOGISTIC, GENMOD, GLIMMIX, and NLMIXED, and R [39] using, for example, packages glm, lmer, and lme4. Parametric mixture models can also be conducted in the MCMC framework using WinBUGS [34], OpenBUGS [33], JAGS [38], or the package mcmc in R.

References

[1] L. M. Ainsworth. Models and Methods for Spatial Data: Detecting Outliers and Handling Zero-Inflated Counts. PhD thesis, Simon Fraser University, 2007.

[2] D. F. Andrews and A. M. Herzberg. Data: A Collection of Problems from Many Fields for the Student and Research Worker. Springer-Verlag, New York, 2000.

[3] Dankmar Böhning and Wilfried Seidel. Editorial: recent developments in mixture models. Computational Statistics & Data Analysis, 41(3–4):349–357, January 2003.

[4] James G. Booth, George Casella, Herwig Friedl, and James P. Hobert. Negative binomial loglinear mixed models. Statistical Modelling, 3(3):179–191, October 2003.

[5] Ronald J. Bosch and Louise M. Ryan. Generalized poisson models arising from Markov processes. Statistics & Probability Letters, 39(3):205–212, August 1998.

[6] N. E. Breslow. Generalized linear models: checking assumptions and strengthening conclusions. Statistica Applicata, 8:23–41, 1996.

[7] Anne Buu, Runze Li, Xianming Tan, and Robert A. Zucker. Statistical models for longitudinal zero-inflated count data with applications to the substance abuse field. Statistics in Medicine, 31(29):4074–4086, July 2012.

[8] D. Byar, C. Blackard, and the Veterans Administration Co-operative Urological Research Group. Comparisons of placebo, pyridoxine, and topical thiotepa in preventing recurrence of stage I bladder cancer. Urology, 10:556–561, 1977.

[9] A. C. Cameron and P. Johansson. Count data regression using series expansions: with applications. Journal of Applied Econometrics, 12:203–223, 1997.

[10] R.J. Cook and J.F. Lawless. The Statistical Analysis of Recurrent Events. Springer, New York, 2007.

[11] C. Dean and J.F. Lawless. Tests for detecting overdispersion in Poisson regression models. Journal of the American Statistical Association, 84(406):467–472, June 1989.

[12] C. B. Dean. Testing for overdispersion in Poisson and binomial regression models. Journal of the American Statistical Association, 87(418):451–457, June 1992.

[13] Joan Del Castillo and Marta Pérez-Casany. Weighted Poisson distributions for overdispersion and underdispersion situations. Annals of the Institute of Statistical Mathematics, 50(3):567–585, 1998.

[14] Melissa J. Dobbie and A. H. Welsh. Modelling correlated zero-inflated count data. Australian & New Zealand Journal of Statistics, 43(4):431–444, December 2001.

[15] Bradley Efron. Double exponential families and their use in generalized linear regression. Journal of the American Statistical Association, 81(395):709–721, September 1986.

[16] M. J. Faddy. Extended Poisson process modelling and analysis of count data. Biometrical Journal, 39(4):431–440, 1997.

[17] M. J. Faddy and R. J. Bosch. Likelihood-based modeling and analysis of data underdispersed relative to the Poisson distribution. Biometrics, 57(2):620–624, June 2001.

[18] M. Fiocco, H. Putter, and J. C. Van Houwelingen. A new serially correlated gamma-frailty process for longitudinal count data. Biostatistics, 10(2):245–257, April 2009.

[19] Aldo M. Garay, Elizabeth M. Hashimoto, Edwin M. M. Ortega, and Victor H. Lachos. On estimation and influence diagnostics for zero-inflated negative binomial regression models. Computational Statistics & Data Analysis, 55(3):1304–1318, 2011.

[20] Piet Groeneboom, Geurt Jongbloed, and Jon A. Wellner. The support reduction algorithm for computing non-parametric function estimates in mixture models. Scandinavian Journal of Statistics, 35(3):385–399, September 2008.

[21] Gary K. Grunwald, Stephanie L. Bruce, Luohua Jiang, Matthew Strand, and Nathan Rabinovitch. A statistical model for under- or overdispersed clustered and longitudinal count data. Biometrical Journal, 53(4):578–594, June 2011.

[22] D. B. Hall. Zero-inflated Poisson and binomial regression with random effects: a case study. Biometrics, 56(4):1030–1039, December 2000.

[23] J. L. Hay and A. N. Pettitt. Bayesian analysis of a time series of counts with covariates: an application to the control of an infectious disease. Biostatistics, 2(4):433–444, December 2001.

[24] Robin Henderson and Silvia Shimakura. A serially correlated gamma frailty model for longitudinal count data. Biometrika, 90(2):355–366, June 2003.

[25] J. M. Hilbe. Negative Binomial Regression. Cambridge University Press, New York, 2nd edition, 2011.

[26] John Hinde and Clarice G. B. Demétrio. Overdispersion: Models and estimation. Computational Statistics & Data Analysis, 27(2):151–170, April 1998.

[27] S. Iddi and G. Molenberghs. A combined overdispersed and marginalized multilevel model. Computational Statistics & Data Analysis, 56(6):1944–1951, June 2012.

[28] Vandna Jowaheer and Brajendra C. Sutradhar. Analysing longitudinal count data with overdispersion. Biometrika, 89(2):389–399, June 2002.

[29] E. Juarez-Colunga. Recurrent Event Studies: Efficient Panel Designs and Joint Modeling of Events and Severities. PhD thesis, Simon Eraser University, 2011.

[30] Byoung Cheol Jung, Myoungshic Jhun, and Jae Won Lee. Bootstrap tests for overdispersion in a zero-inflated Poisson regression model. Biometrics, 61(2):626–628, June 2005.

[31] Kung-Yee Liang and Scott L Zeger. Longitudinal data analysis using generalized linear models. Biometrika, 73(1):13–22, 1986.

[32] Bruce G. Lindsay. Mixture models: theory, geometry and applications. In NSF-CBMS Regional Conference Series in Probability and Statistics, volume 5 of Institute of Mathematical Statistics, Hayward, 1995.

[33] David D. Lunn, David D. Spiegelhalter, Andrew A. Thomas, and Nicky N. Best. The BUGS project: Evolution, critique and future directions. Audio, Transactions of the IRE Professional Group on, 28(25):3049–3067, November 2009.

[34] David J. Lunn, Andrew Thomas, Nicky Best, and David Spiegelhalter. WinBUGS—A Bayesian modelling framework: Concepts, structure, and extensibility. Statistics and Computing, 10(4):325–337, 2000.

[35] Peter McCullagh and James A. Nelder. Generalized Linear Models. Chapman Hall, London, 2nd edition, 1989.

[36] Geert Molenberghs, Geert Verbeke, and Clarice G. B. Demétrio. An extended random-effects approach to modeling repeated, overdispersed count data. Lifetime Data Analysis, 13(4):513–531, December 2007.

[37] Geert Molenberghs, Geert Verbeke, Clarice G. B. Demétrio, and Afrnio M. C. Vieira. A family of generalized linear models for repeated measures with normal and conjugate random effects. Statistical Science, 25(3):325–347. August 2010.

[38] Martyn Plummer. JAGS: A program for analysis of Bayesian graphical models using Gibbs sampling, 2012.

[39] R Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, 2013.

[40] M. Ridout, C.G.B. Demetrio, and J. Hinde. Models for count data with many zeros. Proceedings of the XIXth International Biometric Conference. Cape Town, 1998.

[41] M. S. Ridout and P. Besbeas. An empirical model for underdispersed count data. Statistical Modelling, 4(1):77–89, April 2004.

[42] SAS Institute Inc. SAS/STAT Software, Version 9.3. Cary, NC, 2011.

[43] K. F. Sellers, S. Borle, and G. Shmueli. The COM-Poisson model for count data: a survey of methods and applications. Applied Stochastic Models in Business and Industry, 28:104–116, 2012.

[44] Kimberly F. Sellers and Galit Shmueli. A flexible regression model for count data. Annals of Applied Statistics, 4(2):943–961, November 2010.

[45] Galit Shmueli, Thomas P. Minka, Joseph B. Kadane, Sharad Borle, and Peter Boatwright. A useful distribution for fitting discrete data: Revival of the Conway-Maxwell-Poisson distribution. Journal of the Royal Statistical Society, Series C (Applied Statistics), 54(1):127–142, January 2005.

[46] Brajendra C. Sutradhar. Dynamic Mixed Models for Familial Longitudinal Data. Springer, New York, 2011.

[47] Francis Tuerlinckx, Frank Rijmen, Geert Verbeke, and Paul De Boeck. Statistical inference in generalized linear mixed models: a review. British Journal of Mathematical and Statistical Psychology, 59(Pt 2):225–255, November 2006.

[48] Wai Yin Wan and Jennifer S.K. Chan. A new approach for handling longitudinal count data with zero-inflation and overdispersion: Poisson geometric process model. Biometrical Journal. 51(4):556–570, August 2009.

[49] Yong Wang. On fast computation of the non-parametric maximum likelihood estimate of a mixing distribution. Journal of the Royal Statistical Society, Series B (Statistical Methodology), 69(2):185–198, January 2007.

[50] Yong Wang. Maximum likelihood computation for fitting semiparametric mixture models. Statistics and Computing, 20(1):75–86, March 2009.

[51] Liming Xiang, Andy H. Lee, Kelvin K. W. Yau, and Geoffrey J. McLachlan. A score test for overdispersion in zero-inflated Poisson mixed regression model. Statistics in Medicine, 26(7):1608–1622, 2007.

[52] F.-C. Xie, B.-C. Wei, and J.-G. Lin. Score tests for zero-inflated generalized Poisson mixed regression models. Computational Statistics & Data Analysis, 53(9):3478–3489, July 2009.

[53] Zhao Yang, James W Hardin, and Cheryl L. Addy. Testing overdispersion in the zero-inflated Poisson model. Journal of Statistical Planning and Inference, 139(9):3340–3353, September 2009.

[54] Zhao Z. Yang, James W. Hardin, Cheryl L. Addy, and Quang H. Vuong. Testing approaches for overdispersion in Poisson regression versus the generalized Poisson model. Biometrical Journal, 49(4):565–584, August 2007.

Chapter 2

Analysis of Variance (ANOVA)

Jörg Kaufman

2.1 Introduction

The development of analysis of variance (ANOVA) methodology has in turn had an influence on the types of experimental research being carried out in many fields. ANOVA is one of the most commonly used statistical techniques, with applications across the full spectrum of experiments in agriculture, biology, chemistry, toxicology, pharmaceutical research, clinical development, psychology, social science, and engineering. The procedure involves the separation of total observed variation in the data into individual components attributable to various factors as well as those caused by random or chance fluctuation. It allows performing hypothesis tests of significance to determine which factors influence the outcome of the experiment. However, although hypothesis testing is certainly a very useful feature of the ANOVA, it is by no means the only aspect. The methodology was originally developed by Sir Ronald A. Fisher [1], the pioneer and innovator of the use and applications of statistical methods in experimental design, who coined the name “Analysis of Variance—ANOVA.”

For most biological phenomena, inherent variability exists within the response processes of treated subjects as well as among the conditions under which treatment is received, which results in sampling variability, meaning that results for a subject included in a study will differ to some extent from those of other subjects in the affected population. Thus, the sources of variability must be investigated and must be suitably taken into account when data from comparative studies are evaluated correctly. Clinical studies are in particular a fruitful field for the application of this methodology.

The basis for generalizability of a successful clinical trial is strengthened when the coverage of a study is as broad as possible with respect to geographical area, patient demographics, and pretreatment characteristics as well as other factors that are potentially associated with the response variables. At the same time, heterogeneity among patients becomes more extensive and conflicts with the precision of statistical estimates, which is usually enhanced by homogeneity of subjects. The methodology of the ANOVA is a means to structure the data and their validation by accounting for the sources of variability such that homogeneity is regained in subsets of subjects and heterogeneity is attributed to the relevant factors. The ANOVA method is based on the use of sums of squares of the deviation of the observations from respective means (→ Linear Model).

The tradition of arraying sums of squares and resulting F-statistics in an ANOVA table is so firmly entrenched in the analysis of balanced data that extension of the analysis to unbalanced data is necessary. For unbalanced data, many different sums of squares can be defined and then be used in the numerators of F-statistics. providing tests for a wide variety of hypotheses. In order to provide a practically relevant and useful approach, the ANOVA through the cell means model is introduced below.

The concept of the cell means model was introduced by Searle [2,3]. Hocking and Speed [4], and Hocking [5] to resolve some of the confusion associated with ANOVA models with unbalanced data. The simplicity of such a model is readily apparent: No confusion exists on which functions are estimable, what their estimators are, and what hypotheses can be tested. The cell means model is conceptually easier, it is useful for understanding the ANOVA models, and it is, from the sampling point of view, the appropriate model to use.

In many applications, the statistical analysis is characterized by the fact that a number of detailed questions need to be answered. Even if an overall test is significant, further analyses are, in general, necessary to assess specific differences in the treatments. The cell means model provides within the ANOVA framework the appropriate model for a correct statistical inference and provides such honest statements on statistical significance in a clinical investigation.

2.2 Factors, Levels, Effects, and Cells

One of the principal uses of statistical models is to explain variation in measurements. This variation may be caused by the variety of factors of influence, and it manifests itself as variation from one experimental unit to another. In well-controlled clinical studies, the sponsor deliberately changes the levels of experimental factors (e.g., treatment) to induce variation in the measured quantities to lead to a better understanding of the relationship between those experimental factors and the response. Those factors are called independent and the measured quantities are called dependent variables. For example, consider a clinical trial in which three different diagnostic imaging modalities are used on both men and women in different centers. Table 1 shows schematically how the resulting data could be arrayed in a tabular fashion.

Table 1: Factors, Levels, and Cells (i j k)

2.3 Cell Means Model

In linear model form, it is written

(1)

where the errors eir are identically independent normal N(0, σ2) distributed (i.i.d. variables). Note that, a model consists of more than just a model equation: It is an equation such as Equation 1 plus statements that describe the terms of the equation. In the example above, μi is defined as the population mean of the ith population and is equal to the expectation of yir