73,50 €
Multivariate Analysis
Comprehensive Reference Work on Multivariate Analysis and its Applications
The first edition of this book, by Mardia, Kent and Bibby, has been used globally for over 40 years. This second edition brings many topics up to date, with a special emphasis on recent developments.
A wide range of material in multivariate analysis is covered, including the classical themes of multivariate normal theory, multivariate regression, inference, multidimensional scaling, factor analysis, cluster analysis and principal component analysis. The book also now covers modern developments such as graphical models, robust estimation, statistical learning, and high-dimensional methods. The book expertly blends theory and application, providing numerous worked examples and exercises at the end of each chapter. The reader is assumed to have a basic knowledge of mathematical statistics at an undergraduate level together with an elementary understanding of linear algebra. There are appendices which provide a background in matrix algebra, a summary of univariate statistics, a collection of statistical tables and a discussion of computational aspects. The work includes coverage of:
Although primarily designed as a textbook for final year undergraduates and postgraduate students in mathematics and statistics, the book will also be of interest to research workers and applied scientists.
Sie lesen das E-Book in den Legimi-Apps auf:
Seitenzahl: 943
Veröffentlichungsjahr: 2024
Cover
Table of Contents
Title Page
Copyright
Dedication
Epigraph
Preface to the Second Edition
Preface to the First Edition
Acknowledgments from First Edition
Notation, Abbreviations, and Key Ideas
Matrices and Vectors
Random Variables and Data
Parameters and Statistics
Distributions
Matrix Decompositions
Geometry
Main Abbreviations and Commonly Used Notation
1 Introduction
1.1 Objects and Variables
1.2 Some Multivariate Problems and Techniques
1.3 The Data Matrix
1.4 Summary Statistics
1.5 Linear Combinations
1.6 Geometrical Ideas
1.7 Graphical Representation
1.8 Measures of Multivariate Skewness and Kurtosis
Exercises and Complements
2 Basic Properties of Random Vectors
Introduction
2.1 Cumulative Distribution Functions and Probability Density Functions
2.2 Population Moments
2.3 Characteristic Functions
2.4 Transformations
2.5 The Multivariate Normal Distribution
2.6 Random Samples
2.7 Limit Theorems
Exercises and Complements
3 Nonnormal Distributions
3.1 Introduction
3.2 Some Multivariate Generalizations of Univariate Distributions
3.3 Families of Distributions
3.4 Insights into Skewness and Kurtosis
3.5 Copulas
Exercises and Complements
4 Normal Distribution Theory
4.1 Introduction and Characterization
4.2 Linear Forms
4.3 Transformations of Normal Data Matrices
4.4 The Wishart Distribution
4.5 The Hotelling Distribution
4.6 Mahalanobis Distance
4.7 Statistics Based on the Wishart Distribution
4.8 Other Distributions Related to the Multivariate Normal
Exercises and Complements
5 Estimation
Introduction
5.1 Likelihood and Sufficiency
5.2 Maximum‐likelihood Estimation
5.3 Robust Estimation of Location and Dispersion for Multivariate Distributions
5.4 Bayesian Inference
Exercises and Complements
6 Hypothesis Testing
6.1 Introduction
6.2 The Techniques Introduced
6.3 The Techniques Further Illustrated
6.4 Simultaneous Confidence Intervals
6.5 The Behrens–Fisher Problem
6.6 Multivariate Hypothesis Testing: Some General Points
6.7 Nonnormal Data
6.8 Mardia's Nonparametric Test for the Bivariate Two‐sample Problem
Exercises and Complements
7 Multivariate Regression Analysis
7.1 Introduction
7.2 Maximum‐likelihood Estimation
7.3 The General Linear Hypothesis
7.4 Design Matrices of Degenerate Rank
7.5 Multiple Correlation
7.6 Least‐squares Estimation
7.7 Discarding of Variables
Exercises and Complements
8 Graphical Models
8.1 Introduction
8.2 Graphs and Conditional Independence
8.3 Gaussian Graphical Models
8.4 Log‐linear Graphical Models
8.5 Directed and Mixed Graphs
Exercises and Complements
9 Principal Component Analysis
9.1 Introduction
9.2 Definition and Properties of Principal Components
9.3 Sampling Properties of Principal Components
9.4 Testing Hypotheses About Principal Components
9.5 Correspondence Analysis
9.6 Allometry – Measurement of Size and Shape
9.7 Discarding of Variables
9.8 Principal Component Regression
9.9 Projection Pursuit and Independent Component Analysis
9.10 PCA in High Dimensions
Exercises and Complements
10 Factor Analysis
10.1 Introduction
10.2 The Factor Model
10.3 Principal Factor Analysis
10.4 Maximum‐likelihood Factor Analysis
10.5 Goodness‐of‐fit Test
10.6 Rotation of Factors
10.7 Factor Scores
10.8 Relationships Between Factor Analysis and Principal Component Analysis
10.9 Analysis of Covariance Structures
Exercises and Complements
11 Canonical Correlation Analysis
11.1 Introduction
11.2 Mathematical Development
11.3 Qualitative Data and Dummy Variables
11.4 Qualitative and Quantitative Data
Exercises and Complements
12 Discriminant Analysis and Statistical Learning
12.1 Introduction
12.2 Bayes' Discriminant Rule
12.3 The Error Rate
12.4 Discrimination Using the Normal Distribution
12.5 Discarding of Variables
12.6 Fisher's Linear Discriminant Function
12.7 Nonparametric Distance‐based Methods
12.8 Classification Trees
12.9 Logistic Discrimination
12.10 Neural Networks
Exercises and Complements
Notes
13 Multivariate Analysis of Variance
13.1 Introduction
13.2 Formulation of Multivariate One‐way Classification
13.3 The Likelihood Ratio Principle
13.4 Testing Fixed Contrasts
13.5 Canonical Variables and A Test of Dimensionality
13.6 The Union Intersection Approach
13.7 Two‐way Classification
Exercises and Complements
Note
14 Cluster Analysis and Unsupervised Learning
14.1 Introduction
14.2 Probabilistic Membership Models
14.3 Parametric Mixture Models
14.4 Partitioning Methods
14.5 Hierarchical Methods
14.6 Distances and Similarities
14.7 Grouped Data
14.8 Mode Seeking
14.9 Measures of Agreement
Exercises and Complements
Note
15 Multidimensional Scaling
15.1 Introduction
15.2 Classical Solution
15.3 Duality Between Principal Coordinate Analysis and Principal Component Analysis
15.4 Optimal Properties of the Classical Solution and Goodness of Fit
15.5 Seriation
15.6 Nonmetric Methods
15.7 Goodness of Fit Measure: Procrustes Rotation
15.8 Multisample Problem and Canonical Variates
Exercises and Complements
16 High‐dimensional Data
16.1 Introduction
16.2 Shrinkage Methods in Regression
16.3 Principal Component Regression
16.4 Partial Least Squares Regression
16.5 Functional Data
Exercises and Complements
A Matrix Algebra
A.1 Introduction
A.2 Matrix Operations
A.3 Further Particular Matrices and Types of Matrices
A.4 Vector Spaces, Rank, and Linear Equations
A.5 Linear Transformations
A.6 Eigenvalues and Eigenvectors
A.7 Quadratic Forms and Definiteness
A.8 Generalized Inverse
A.9 Matrix Differentiation and Maximization Problems
A.10 Geometrical Ideas
B Univariate Statistics
B.1 Introduction
B.2 Normal Distribution
B.3 Chi‐squared Distribution
B.4 and Beta Variables
B.5 Distribution
B.6 Poisson Distribution
C R Commands and Data
C.1 Basic R Commands Related to Matrices
C.2 R Libraries and Commands Used in Exercises and Figures
C.3 Data Availability
D Tables
References and Author Index
Index
End User License Agreement
Chapter 1
Table 1.1 Data matrix with five students as objects, where is age in year...
Table 1.2 Marks in open‐ and closed‐book examination out of 100.
Table 1.3 Measurements (in cm) on three types of irises.
Table 1.4 Weights of cork deposits (in centigrams) for 28 trees in the fo...
Table 1.5 Sample means and standard deviations () for the two groups of ...
Table 1.6 Measures of multivariate skewness and kurtosis for each of th...
Chapter 2
Table 2.1 Jacobians of some transformations.
Chapter 3
Table 3.1 Selected examples (shown in Figure 3.1) for a variety of and ...
Table 3.2 Some examples of Archimedean copulas.
Chapter 5
Table 5.1 Extreme values to be excluded from robust estimates in the iris d...
Table 5.2 Estimates of location and scatter for each of four methods: maxim...
Chapter 6
Table 6.1 The measurements on the first and second adult sons in a sample o...
Table 6.2 Values of the statistics and for each of the iris species and...
Table 6.3 Data for the geological problem (in micrometers).
Chapter 7
Table 7.1 Correlation matrix for the physical properties of props.
Table 7.2 Correlation between the independent variables and for pitprop d...
Table 7.3 Variables selected in multiple regression for pitprop data.
Table 7.4 Variables selected in interdependence analysis for pitprop data....
Chapter 9
Table 9.1 Values of , defined by (9.14), – and their square root – for th...
Table 9.2 Eigenvectors and eigenvectors for the pitprop data.
Table 9.3 Open–closed‐book data in Example 9.3.2: standard errors for the l...
Table 9.4 Open–closed‐book data in Example 9.3.2: standard errors for the l...
Table 9.5 Cross‐tabulation of frequencies taken from the General Social Sur...
Table 9.6 Counts obtained from a random sample of size 1000 taken from tabl...
Chapter 10
Table 10.1 Principal factor solutions for the open/closed‐book data with ...
Table 10.2 Maximum‐likelihood factor solutions for the open/closed‐book dat...
Table 10.3 Correlation matrix for the applicant data
Table 10.4 Maximum‐likelihood factor solution of applicant data with fact...
Table 10.5 Maximum‐likelihood factor solution of applicant data with fact...
Chapter 11
Table 11.1 Social mobility contingency table (Glass (1954); see also Goodma...
Table 11.2 Data from 382 Hull University students ( denotes frequencies)....
Chapter 12
Table 12.1 Summary of the examples in this chapter using the iris data, inc...
Table 12.2 Percentage distribution of sentence endings in seven works of Pl...
Table 12.3 Mean scores and their standard errors from seven works of Plato....
Table 12.4 Illustrative data for travel claims.
Table 12.5 Summary of the Titanic data for the number of survivors/total nu...
Table 12.6 Summary of the estimated parameters and standard errors for a lo...
Table 12.7 Confusion matrix for logistic regression model of Titanic surviv...
Table 12.8 Flea‐beetle measurements, taken from Lubischew (1962).
Chapter 13
Table 13.1 Multivariate one‐way classification.
Table 13.2 Logarithms of multiple measurements on anteater skulls at three ...
Table 13.3 Matrices in
MANOVA
table for Reeve's data .
Table 13.4 Weight losses (in grams) for the first and second weeks for rats...
Table 13.5
MANOVA
table for the data in Table 13.4.
Chapter 14
Table 14.1 A subsample of six observations (and two variables) from the iri...
Table 14.2 Clusters for the data in Table 14.1.
Table 14.3 Mahalanobis distances between 10 island races of white‐toothed s...
Table 14.4 Single linkage procedure for shrew data.
Table 14.5 Relative frequencies of blood groups A
1
, A
2
, B, and O for four...
Table 14.6 Distances between two points and .
Table 14.7 Comparison of cluster allocations based on hill ‐climbing and mo...
Chapter 15
Table 15.1 Road distances in miles between 12 British towns.
Table 15.2 Percentage of times that the pairs of Morse code signals for two...
Table 15.3 Observed and fitted journey time distance matrices for Example 1...
Chapter 16
Table 16.1 Estimated regression coefficients using 16.33, with regressors...
Appendix A
Table A.1 Particular matrices and types of matrices (List 1). For List 2, se...
Table A.2 Basic matrix operations.
Table A.3 Particular types of matrices (List 2).
Table A.4 Rank of some matrices.
Table A.5 Basic concepts in ‐dimensional geometry.
Appendix C
Table C.1
R
commands for some basic matrix operations.
Table C.2 Using the
diag
command in R.
Table C.3
R
data sets used and associated library (see index for usage).
Table C.4 The following datasets are available on the website github.com/cha...
Appendix D
Table D.1 Upper critical values of the distribution.
Table D.2 Upper critical values of the distribution.
Table D.3 Upper critical values of the distribution ().
Table D.4 Upper critical values of the distribution ().
Table D.5 Upper critical values of the distribution ().
Table D.6 Upper critical values of the distribution ().
Table D.7 Using the
R
function
doubleWishart
(Turgeon, 2018) with , these a...
Chapter 1
Figure 1.1 Univariate representation of the cork data of Table 1.4.
Figure 1.2 Consecutive univariate representation. Source: Adapted from Pears...
Figure 1.3 Matrix of scatterplots for the cork data. The upper diagonal show...
Figure 1.4 A glyph representation of the cork data of Table 1.4.
Figure 1.5 Harmonic curves for simulated data with observations and vari...
Figure 1.6 Parallel coordinates plot for the cork data.
Chapter 2
Figure 2.1 Ellipses of equal probability density for the bivariate normal di...
Figure 2.2 Ellipses of equal probability density for the bivariate normal di...
Chapter 3
Figure 3.1 Contour plots of the quantiles for various choices of and in ...
Figure 3.2 Copulas combined with margins. (a) Gaussian (); (b) Clayton–Ma...
Chapter 5
Figure 5.1 Convex hull for the iris data. (
I. versicolor
variety) with me...
Chapter 7
Figure 7.1 Geometry of multiple regression model, .
Chapter 8
Figure 8.1 Graphical model for open/closed‐book data given in Table 1.2. The...
Figure 8.2 (a) Example of a graph with eight vertices; (b) decomposition int...
Figure 8.3 (a) Example of a graph with eight vertices; (b) decomposition int...
Figure 8.4 Proposed graphical model for turtle data.
Figure 8.5 This graph shows a graphical interaction model since the three tw...
Figure 8.6 Graphical model for the Copenhagen housing data.
Figure 8.7 (a) Decomposable graphical model for the vehicle accident data. (...
Figure 8.8 Example of directed acyclic graph (DAG).
Figure 8.9 Venn diagram depicting two events and (a) with and other pr...
Chapter 9
Figure 9.1 Scree graph for the Corsican pitprop data of Table 9.2.
Figure 9.2 The 88 individuals of the open/closed‐book data plotted on the fi...
Figure 9.3 Correlations between the variables and principal components for t...
Figure 9.4 Biplot (using ) for the open/closed‐book data.
Figure 9.5 Biplot (using ) using simulated data of Example 9.2.9.
Figure 9.6 Sampling distribution of the first two log eigenvalues for the op...
Figure 9.7 The row‐principal biplot and the column‐principal biplot for the ...
Figure 9.8 (a) Scatterplot of two of the iris data variables, with normalize...
Figure 9.9 (a) Scatterplot of two of the iris data variables, with direction...
Chapter 10
Figure 10.1 Plot of open/closed‐book factors before and after varimax rotati...
Chapter 12
Figure 12.1 Normal likelihoods with unequal means and unequal variances (fro...
Figure 12.2 Discrimination between two species of iris: a plot of the data p...
Figure 12.3 Discrimination between three species of iris using two variables...
Figure 12.4 (Mental health data). Discrimination between normal individuals ...
Figure 12.5 Classification regions (polygons obtained using the
R
library
sp
...
Figure 12.6 Classification regions for the first two variables of the iris d...
Figure 12.7 A possible classification tree for data in Table 12.4. The branc...
Figure 12.8 Splitting indices (some linearly transformed) of first two varia...
Figure 12.9 Scatter plot of the first two variables for iris data (
I. seto
...
Figure 12.10 Decision tree (created using
rpart
(Therneau and Atkinson, 2019...
Figure 12.11 (a) Classification tree and the corresponding partition (b) for...
Figure 12.12 Fitted probabilities for survival in the sinking of the Titanic...
Figure 12.13 Multilayer network with an input layer, one hidden layer, and t...
Figure 12.14 Multilayer perceptron boundaries (continuous line) and linear d...
Chapter 13
Figure 13.1 Canonical analysis of iris data, with approximate 99% confidence...
Figure 13.2 Plot of the 10 groups of shrews with respect to the first two ca...
Figure 13.3 Canonical analysis of image segmentation data used in Example 13...
Chapter 14
Figure 14.1 Scatter plot for the small subsample of six observations of iris...
Figure 14.2 Application of the ‐means algorithm, and
mclust
to the iris are...
Figure 14.3 Application of the fuzzy ‐means algorithm to the iris area data...
Figure 14.4 Dendrogram for shrew group means (single linkage) in Example 14....
Figure 14.5 Minimum spanning tree for the shrew data whose distances are giv...
Figure 14.6 Dendrogram for the shrew data (complete linkage).
Figure 14.7 Hierarchical clusters for grouped glass data. Groups are sequent...
Figure 14.8 Contour plots of the kernel density estimate for the transformed...
Figure 14.9 Clusters of countries based on five indicators, after selecting ...
Figure 14.10 World map for the 11 clusters of the dendogram of Figure 14.9. ...
Figure 14.11 Rand index showing the agreement between the complete linkage c...
Chapter 15
Figure 15.1 Classical solution for Morse code data in Table 15.2.
Figure 15.2 MDS solutions for the road data in Table 15.1, , original point...
Figure 15.3 Two‐dimensional MDS solution for Example 15.4.2. Train journey t...
Figure 15.4 Classical MDS solution in two dimensions using similarity matrix...
Figure 15.5 Two‐dimensional representation of Kendall's similarity matrix fo...
Figure 15.6 Seven regions in a country for Exercise 15.2.7.
Chapter 16
Figure 16.1 (a) Log of average nearest neighbor distances for various dimens...
Figure 16.2 Bias‐squared and variance for ridge regression estimates of , w...
Figure 16.3 Results from using LARS with simulated data with uncorrelated ex...
Figure 16.4 (a) Log of scree plot for yarn data; (b) Best subset selection f...
Figure 16.5 (a) Boxplots of scores for six sensory variables (); (b) Pairwi...
Figure 16.6 Height measurements for 39 boys (a) and 54 girls (b). Points sho...
Figure 16.7 Growth Data. (a) Estimated (Eq. 16.25; (b) Cumulative proporti...
Figure 16.8 Growth Data. The estimated regression function obtained using ...
Appendix A
Figure A.1 is the projection of onto the plane .
Figure A.2 Ellipsoid in the plane. Vectors given by and .
Cover
Table of Contents
Title Page
Copyright
Dedication
Epigraph
Preface to the Second Edition
Preface to the First Edition
Acknowledgments from First Edition
Notation, Abbreviations, and Key Ideas
Begin Reading
A Matrix Algebra
B Univariate Statistics
C R Commands and Data
D Tables
References and Author Index
Index
End User License Agreement
iii
iv
v
xvii
xix
xx
xxi
xxii
xxiii
xxv
xxvii
xxviii
xxix
xxx
xxxi
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
543
544
545
546
547
548
549
550
551
552
553
554
Second Edition
Kanti V. Mardia, John T. Kent, and Charles C. Taylor
This second edition first published 2024© 2024 John Wiley & Sons Ltd
Edition History© 1st Edition 1979 Academic Press
All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies. No part of this publication may be reproduced, stored in a retrieval system, or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording or otherwise, except as permitted by law. Advice on how to obtain permission to reuse material from this title is available at http://www.wiley.com/go/permissions.
The right of Kanti V. Mardia, John T. Kent, and Charles C. Taylor to be identified as the authors of this work have been asserted in accordance with law.
Registered OfficesJohn Wiley & Sons, Inc., 111 River Street, Hoboken, NJ 07030, USAJohn Wiley & Sons Ltd, The Atrium, Southern Gate, Chichester, West Sussex, PO19 8SQ, UK
For details of our global editorial offices, customer services, and more information about Wiley products visit us at www.wiley.com.
Wiley also publishes its books in a variety of electronic formats and by print‐on‐demand. Some content that appears in standard print versions of this book may not be available in other formats.
Trademarks: Wiley and the Wiley logo are trademarks or registered trademarks of John Wiley & Sons, Inc. and/or its affiliates in the United States and other countries and may not be used without written permission. All other trademarks are the property of their respective owners. John Wiley & Sons, Inc. is not associated with any product or vendor mentioned in this book.
Limit of Liability/Disclaimer of Warranty
In view of ongoing research, equipment modifications, changes in governmental regulations, and the constant flow of information relating to the use of experimental reagents, equipment, and devices, the reader is urged to review and evaluate the information provided in the package insert or instructions for each chemical, piece of equipment, reagent, or device for, among other things, any changes in the instructions or indication of usage and for added warnings and precautions. While the publisher and authors have used their best efforts in preparing this work, they make no representations or warranties with respect to the accuracy or completeness of the contents of this work and specifically disclaim all warranties, including without limitation any implied warranties of merchantability or fitness for a particular purpose. No warranty may be created or extended by sales representatives, written sales materials or promotional statements for this work. The fact that an organization, website, or product is referred to in this work as a citation and/or potential source of further information does not mean that the publisher and authors endorse the information or services the organization, website, or product may provide or recommendations it may make. This work is sold with the understanding that the publisher is not engaged in rendering professional services. The advice and strategies contained herein may not be suitable for your situation. You should consult with a specialist where appropriate. Further, readers should be aware that websites listed in this work may have changed or disappeared between when this work was written and when it is read. Neither the publisher nor authors shall be liable for any loss of profit or any other commercial damages, including but not limited to special, incidental, consequential, or other damages.
Library of Congress Cataloging‐in‐Publication Data
Names: Mardia, K. V., author. | Kent, J. T. (John T.), joint author. | Taylor, C. C (Charles C.), joint author.Title: Multivariate analysis / Kanti V. Mardia, John T. Kent, and Charles C. Taylor.Description: Second edition. | Hoboken, NJ : Wiley, 2024. | Series: Wiley series in probability and statistics | Includes bibliographical references and index.Identifiers: LCCN 2023002460 (print) | LCCN 2023002461 (ebook) | ISBN 9781118738023 (cloth) | ISBN 9781118892527 (adobe pdf) | ISBN 9781118892510 (epub)Subjects: LCSH: Multivariate analysis.Classification: LCC QA278 .M36 2023 (print) | LCC QA278 (ebook) | DDC 519.5/35–dc23/eng/20230531LC record available at https://lccn.loc.gov/2023002460LC ebook record available at https://lccn.loc.gov/2023002461
Cover Design: WileyCover Images: © Liyao Xie/Getty Images, Courtesy of Kanti Mardia
To my daughters Bela and Neeta
— with Jainness (Kanti V. Mardia)
To my son Edward and daughter Natalie
(John T. Kent)
To my wife Fiona and my children Mike, Anna, Ruth, and Kathryn
(Charles C. Taylor)
Everything is related with every other thing, and this relation involves the emergence of a relational quality. The qualities cannot be known a priori, though a good number of them can be deduced from certain fundamental characteristics.
— Jaina philosophy
The Jaina Philosophy of Non‐Absolutism by S. Mookerjee, q.v.
Mahalanobis (1957).
For over 40 years the first edition of this book (which was also translated into Persian) has been used by students to acquire a basic knowledge of the theory and methods of multivariate statistical analysis. The book has also served the wider statistical community to further their understanding of this field. Plans for the second edition started almost 20 years ago, and we have struggled with questions about which topics to add – something of a moving target in a field that has continued to evolve in this new era of artificial intelligence (AI) and “big data”. Since the first edition was published, multivariate analysis has been developed and extended in many directions. This new edition aims to bring the first edition up to date by substantial revision, rewriting, and additions, while seeking to maintain the overall length of the book. The basic approach has been maintained, namely a mathematical treatment of statistical methods for observations consisting of several measurements or characteristics of each subject and a study of their properties. The core topics, and the structure many of the chapters, have been retained.
Briefly, for those familiar with the first edition, the main changes (in addition to updating material in several places) are:
a new section giving Notation, Abbreviations, and Key Ideas used through the book;
a new chapter introducing some nonnormal distributions. This includes new sections on elliptical distributions and copulas;
a new chapter covering an introduction to
graphical models
;
a completely rewritten chapter that begins from discriminant analysis and extends to nonparametric methods, classification and regression trees, logistic discrimination, and multilayer perceptrons. These topics are commonly grouped into the heading of
supervised learning
;
the above chapter focuses on data in which group memberships are known, whereas “unsupervised learning” has more traditionally been known as cluster analysis, for which the current
Chapter 14
has also been substantially updated to reflect recent developments;
a new (final) chapter introduces some approaches to
high‐dimensional data
in which the number of variables may exceed the number of observations. This includes shrinkage methods in regression, principal components regression, partial least squares regression, and functional data analysis;
further development of discrete aspects, including log‐linear models, the EM algorithm for mixture models, and correspondence analysis for contingency tables.
As a consequence of the above new and extended/revised chapters and in order to save space, we have omitted some material from this edition:
the chapter on econometrics, since there are now dedicated books with an emphasis on statistical aspects (Maddala and Lahiri,
2009
); (Wooldridge,
2019
);
the chapter on directional statistics, since there are now related dedicated books by one of the authors (Dryden and Mardia,
2016
); (Mardia and Jupp,
2000
).
Further changes to this Edition, bringing many subjects up to date, include new graphical representations (Chapter 1), an introduction to the matrix normal distribution (Chapters 2 and 5), elliptical distributions and copulas (Chapter 3), robust estimators for location and dispersion (Chapter 5), a revision of correspondence analysis and biplots (Chapter 9), and projection pursuit and independent component analysis (Chapter 9).
The figures in the first edition have been redrawn in their original style, using the statistical package R. A new Appendix C contains some specific R (R Core Team, 2020) commands applicable to most of the matrix algebra used in the book. In addition, an online addendum to Appendix C contains the data files used in this book as well as the R commands used to obtain the calculations for the examples and figures. This public repository is at github.com/charlesctaylor/MVAdata-rcode. In many cases, we have chosen to use base R functions to mimic the equations used in the text in preference to more “black‐box” R functions. Note that intermediate steps in the calculation are generally rounded only for display purposes.
Multivariate analysis continues to be a research area of active development. We note that the Journal of Multivariate Analysis, in its 50th Anniversary Jubilee Edition (von Rosen and Kollo, 2022), has published a volume that describes the current state of the art and contains review papers. Beyond mainstream multivariate statistics, there have been developments in the applied sciences; one example in morphometrics is Bookstein (2018).
The first edition was published by Academic Press, and we are grateful to John Bibby for his contributions to that edition. For this edition, we thank the many readers who have offered their advice and suggestions. In particular, we would like to acknowledge the help of Susan Holmes for extensive discussions about a new structure as well as a draft of correspondence analysis material for Chapter 9.
We are extremely grateful to Wiley for their patience and help during the writing of the book, especially Helen Ramsey, Sharon Clutton, Richard Davies, Kathryn Sharples, Liz Wingett, Kelvin Matthews, Alison Oliver, Viktoria Hartl‐Vida, Ashley Alliano, Kimberly Monroe‐Hill, and Paul Sayer. Secretarial help at Leeds during the initial development was given by Christine Rutherford and Catherine Dobson.
Kanti would like to thank the Leverhulme Trust for an Emeritus Fellowship and Anna Grundy of the Trust for simplifying the administration process. Finally, he would like to express his sincere gratitude to his family for their continuous love, support, and tolerance.
We would be pleased to hear about any typographical or other errors in the text.
May 2022
Kanti V. Mardia
University of Leeds, Leeds, UK andUniversity of Oxford, Oxford, UK
John T. Kent
University of Leeds, Leeds, UK
Charles C. Taylor
University of Leeds, Leeds, UK
Multivariate Analysis deals with observations on more than one variable where there is some inherent interdependence between the variables. With several texts already available in this area, one may very well enquire of the authors as to the need for yet another book. Most of the available books fall into two categories, either theoretical or data analytic. The present book not only combines the two approaches but also emphasizes modern developments. The choice of material for the book has been guided by the need to give suitable matter for the beginner as well as illustrating some deeper aspects of the subject for the research worker. Practical examples are kept to the forefront, and, wherever feasible, each technique is motivated by such an example.
The book is aimed at final year undergraduates and postgraduate students in Mathematics/Statistics with sections suitable for practitioners and research workers. The book assumes a basic knowledge of Mathematical Statistics at undergraduate level. An elementary course on Linear Algebra is also assumed. In particular, we assume an exposure to Matrix Algebra to the level required to read Appendix A.
Broadly speaking, Chapters 1–6 and 12 can be described as containing direct extensions of univariate ideas and techniques. The remaining chapters concentrate on specifically multivariate problems that have no meaningful analogs in the univariate case. Chapter 1 is primarily concerned with giving exploratory analyses for multivariate data and briefly introduces some of the important techniques, tools, and diagrammatic representations. Chapter 2 introduces various distributions together with some fundamental results, whereas Chapter 3 concentrates exclusively on normal distribution theory. Chapters 4–6 deal with problems in inference. Chapter 7 [no longer included] gives an overview of Econometrics, while Principal Component Analysis, Factor Analysis, Canonical Correlation Analysis, and Discriminant Analysis are discussed from both theoretical and practical points of view in Chapters 8–11. Chapter 12 is on Multivariate Analysis of Variance, which can be better understood in terms of the techniques of previous chapters. The later chapters look into the presently developing techniques of Cluster Analysis, Multidimensional Scaling, and Directional Data [no longer included].
Each chapter concludes with a set of exercises. Solving these will not only enable the reader to understand the material better but will also serve to complement the chapter itself. In general, the questions have in‐built answers, but, where desirable, hints for the solution of theoretical problems are provided. Some of the numerical exercises are designed to be run on a computer, but as the main aim is on interpretation, the answers are provided. We found NAG routines and GLIM most useful, but nowadays any computer center will have some suitable statistics and matrix algebra routines.
There are three Appendices A, B, and C, which, respectively, provide a sufficient background of matrix algebra, a summary of univariate statistics, and some tables of critical values. The aim of Appendix A on Matrix Algebra is not only to provide a summary of results but also to give sufficient guidance to master these for students having little previous knowledge. Equations from Appendix A are referred to as (A.x.x) to distinguish them from (l.x.x), etc. Appendix A also includes a summary of results in n‐dimensional geometry that are used liberally in the book. Appendix B gives a summary of important univariate distributions.
The reference list is by no means exhaustive. Only directly relevant articles are quoted, and for a fuller bibliography, we refer the reader to Anderson et al. (1972) and Subrahmaniam and Subrahmaniam (1973). The reference list also serves as an author index. A subject index is provided.
The material in the book can be used in several different ways. For example, a one‐semester elementary course of 40 lectures could cover the following topics. Appendix A; Chapter 1 (Sections 1.1-1.7); Chapter 2 (Sections 2.1–2.5); Chapter 3 (Sections 3.4.1, 3.5, and 3.6.1, assuming results from previous sections, Definitions 3.7.1 and 3.7.2); Chapter 4 (Section 4.2.2); Chapter 5 (Sections 5.1, 5.2.1a, 5.2.1b, 5.2.2a, 5.2.2b, 5.3.2b, and 5.5); Chapter 8 (Sections 8.1, 8.2.1, 8.2.2, 8.2.5, 8.2.6, 8.4.3, and 8.7); Chapter 9 (Sections 9.1–9.3, 9.4 (without details), 9.5, 9.6, and 9.8); Chapter 10 (Sections 10.1 and 10.2); Chapter 11 (Sections –11.2.3, 11.3.1, and 11.6.1). Further material that can be introduced is Chapter 12 (Sections 12.1-12.3 and 12.6); Chapter 13 (Sections 13.1 and 13.3.1); Chapter 14 (Sections 14.1 and 14.2). This material has been covered in 40 lectures spread over two terms in different British universities. Alternatively, a one‐semester course with more emphasis on foundation rather than applications could be based on Appendix A and Chapters 1–5. Two‐semester courses could include all the chapters, excluding Chapters 7 and 15 on Econometrics and Directional Data, as well as the sections with asterisks. Mathematically orientated students may like to proceed to Chapter 2, omitting the data analytic ideas of Chapter 1.
Various new methods of presentation are utilized in the book. For instance, the data matrix is emphasized throughout, a density‐free approach is given for normal theory, the union intersection principle is used in testing as well as the likelihood ratio principle, and graphical methods are used in explanation. In view of the computer packages generally available, most of the numerical work is taken for granted, and therefore, except for a few particular cases, emphasis is not placed on numerical calculations. The style of presentation is generally kept descriptive except where rigor is found to be necessary for theoretical results, which are then put in the form of theorems. If any details of the proof of a theorem are felt tedious but simple, they are then relegated to the exercises.
Several important topics not usually found in multivariate texts are discussed in detail. Examples of such material include the complete chapters on Econometrics, Cluster Analysis, Multidimensional Scaling, and Directional Data. Further material is also included in parts of other chapters: methods of graphical presentation, measures of multivariate skewness and kurtosis, the singular multinormal distribution, various nonnormal distributions and families of distributions, a density‐free approach to normal distribution theory, Bayesian and robust estimators, a recent solution to the Fisher–Behrens problem, a test of multinormality, a nonparametric test, discarding of variables in regression, principal component analysis and discrimination analysis, correspondence analysis, allometry, the jack‐knifing method in discrimination, canonical analysis of qualitative and quantitative variables, and a test of dimensionality in MANOVA. It is hoped that coverage of these developments will be helpful for students as well as research workers.
There are various other topics that have not been touched upon partly because of lack of space as well as our own preferences, such as Control Theory, Multivariate Time Series, Latent Variable Models, Path Analysis, Growth Curves, Portfolio Analysis, and various Multivariate Designs.
In addition to various research papers, we have been influenced by particular texts in this area, especially Anderson (1958), Kendall (1975), Kshirsagar (1972), Morrison (1976), Press (1972), and Rao (1973). All these are recommended to the reader.
The authors would be most grateful to readers who draw their attention to any errors or obscurities in the book, or suggest other improvements.
January 1979
Kanti V. Mardia
John T. Kent
John C. Bibby
First of all, we wish to express our gratitude to the pioneers in this field. In particular, we should mention M. S. Bartlett, R. A. Fisher, H. Hotelling, D. G. Kendall, M. G. Kendall, P. C. Mahalanobis, C. R. Rao, S. N. Roy, W. S. Torgeson, and S. S. Wilks.
We are grateful to the authors and editors who have generously granted us permission to reproduce figures and tables.
We are also grateful to many of our colleagues for their valuable help and comments, in particular Martin Beale, Christopher Bingham, Lesley Butler, Richard Cormack, David Cox, Ian Curry, Peter Fisk, Allan Gordon, John Gower, Peter Harris, Chunni Khatri, Conrad Leser, Eric Okell, Ross Renner, David Salmond, Cyril Smith, and Peter Zemroch. We are also indebted to Joyce Snell for making various comments on an earlier draft of the book, which have led to considerable improvements. We should also express our gratitude to Rob Edwards for his help in various facets of the book, for calculations, proofreading, diagrams, etc.
Some of the questions are taken from examination papers in British universities, and we are grateful to various unnamed colleagues. Since the original sources of questions are difficult to trace, we apologize to any colleague who recognizes a question of their own.
The authors would like to thank their wives, Pavan Mardia, Susan Kent, and Zorina Bibby.
Finally, our thanks go to Barbara Forsyth and Margaret Richardson for typing a difficult manuscript with great skill.
KVM
JTK
JMB
Vectors
are viewed as column vectors and are represented using bold lower case letters. Round brackets are generally used when a vector is expressed in terms of its elements. For example, in which the th element or component is denoted . The transpose of is denoted , so is a row vector.
Matrices
are written using bold upper case letters, e.g. and . The matrix may be written as in which is the element of the matrix in row and column . If has rows and columns, then the th row of , written as a column vector, is
and the th column is written as
Hence, can be expressed in various forms,
We generally use square brackets when a matrix is expanded in terms of its elements.
Operations on a matrix include
–
transpose:
–
determinant:
–
inverse:
–
generalized inverse:
where for the final three operations, is assumed to be square, and for the inverse operation, is additionally assumed to be nonsingular. Different types of matrices are given in Tables A.1 and A.3. Table A.2 lists some further matrix operations.
In general, a random vector and a nonrandom vector are both indicated using a bold lower case letter, e.g. . Thus, the distinction between the two must be determined from the context. This convention is in contrast to the standard convention in statistics where upper case letters are used to denote random quantities, and lower case letters their observed values.
The reason for our convention is that bold upper case letters are generally used for a data matrix , both random and fixed.
In spite of the above convention, we very occasionally (e.g. parts of
Chapters 2
and
10
) use bold upper case letters for a random vector when it is important to distinguish between the random vector and a possible value .
The phrase “high‐dimensional data” often implies , whereas the phrase “big data” often just indicates that or is large.
Elements of an data matrix are generally written , where indices are used to label the observations, and indices are used to label the variables.
If the rows of a data matrix are normally distributed with mean and covariance matrix , and , the following notation is used to distinguish various population and sample quantities:
Parameter
Sample
Mean vector
Covariance matrix
Unbiased covariance matrix
Concentration matrix
Correlation matrix
The following notation is used for univariate and multivariate distributions. Appendix B summarizes the univariate distributions used in the Book.
cumulative distribution function/distribution function (d.f.)
probability density function (p.d.f.)
expectation
c.f.
characteristic function
d.f.
distribution function
Hotelling
multivariate normal distribution in ‐dimensions with mean (column vector of length ) and covariance matrix
variance–covariance matrix
correlation matrix
Wishart distribution
The terms variance matrix, covariance matrix, and variance–covariance matrix are synonymous.
Any symmetric matrix can (by the spectral decomposition theorem) be written as where is a diagonal matrix of eigenvalues of (which are real‐valued), i.e. , and is an orthogonal matrix whose columns are standardized eigenvectors, i.e. and . See Theorem A.6.8.
Using the above, we define the symmetric square root of a positive definite matrix by
If is an matrix of rank , then by the singular value decomposition, it can be written as where and are column orthonormal matrices, and is a diagonal matrix with positive elements. See Theorem A.6.8.
Table A.5 sets out the basic concepts in ‐dimensional geometry. In particular,
Length of a vector
Euclidean distance between and
Squared Mahalanobis distance – one of the most important distances in multivariate analysis, since it takes account of a covariance, i.e.
Table 14.6 gives a list of various distances.
approximately equal to
(conditionally) independent of
is distributed as
the set of elements that are members of but not
Euclidean distance between and
transpose of matrix
determinant of matrix
inverse of matrix
‐inverse (generalized inverse)
column vector of 1s
column vector or matrix of 0s
between‐groups sum of squares and products (SSP) matrix
beta variable
normalizing constant for beta distribution (note nonitalic font to distinguish from the above)
BLUE
best linear unbiased estimate
covariance between and
chi‐squared distribution with degrees of freedom
upper
α
critical value of chi‐squared distribution with degrees of freedom
c.f.
characteristic function
partial derivative – multivariate examples in Appendix A.9
distance matrix
squared Mahalanobis distance
d.f.
distribution function
Kronecker delta
diagonal elements of a square matrix (as column vector) or diagonal matrix created from a vector
(see above)
expectation
distribution with degrees of freedom and
upper
α
critical value of distribution with degrees of freedom and
cumulative distribution function
probability density function
gamma function
GLS
generalized least squares
centering matrix
identity matrix
ICA
independent component analysis
i.i.d.
independent and identically distributed
Jacobian of transformation (see
Table 2.1
)
concentration matrix ()
likelihood
log likelihood
LDA
linear discriminant analysis
logarithm to the base (natural logarithm)
LRT
likelihood ratio test
MANOVA
multivariate analysis of variance
MDS
multidimensional scaling
ML
maximum likelihood
m.l.e.
maximum likelihood estimate
mean (population) vector
multivariate normal distribution for ‐dimensions (usually omitted when )
OLS
ordinary least squares
(population) correlation matrix (sometimes a matrix of counts, e.g.
Section
9.5)
probability
PCA
principal component analysis
p.d.
positive definite
p.d.f.
probability density function
p.s.d.
positive semi definite
real numbers
correlation coefficient
sample correlation matrix
sample covariance matrix
unbiased sample covariance matrix
(population) covariance matrix
SLC
standardized linear combination
SSP
sums of squares and products
distribution with degrees of freedom
total SSP matrix
Hotelling statistic
trace
UIT
union intersection test
variance
variance‐covariance matrix of
within‐groups SSP matrix
Wishart distribution
data matrix
sample mean vector
Wilks' statistic
greatest root statistic
Multivariate analysis deals with data containing observations on two or more variables, each measured on a set of objects. For example, we may have the set of examination marks achieved by certain students, or the cork deposit in various directions of a set of trees, or flower measurements for different species of iris (see Tables 1.2, 1.4, and 1.3, respectively). Each of these data has a set of “variables” (the examination marks, trunk thickness, and flower measurements) and a set of “objects” (the students, trees, and flowers). In general, if there are objects, and variables, , the data contains pieces of information. These may be conveniently arranged using an “data matrix”, in which each row corresponds to an object, and each column corresponds to a variable. For instance, three variables on five “objects” (students) are shown as a data matrix in Table 1.1.
Note that all the variables need not be of the same type: in Table 1.1, is a “continuous” variable, is a discrete variable, and is a binary variable. Note also that attribute, characteristic, description, measurement, and response are synonyms for “variable”, whereas individual, observation, plot, reading, item, and unit can be used in place of “object”.
We may now illustrate various categories of multivariate technique.
Most univariate questions are capable of at least one multivariate generalization. For instance, using Table 1.2, we may ask, as an example, “What is the appropriate underlying parent distribution of examination marks on various papers of a set of students?” “What are the summary statistics?” “Are the differences between average marks on different papers significant?”, etc. These problems are direct generalizations of univariate problems, and their motivation is easy to grasp. See, for example, Chapters 2–7 and 13.
Table 1.1 Data matrix with five students as objects, where is age in years at entry to university, is marks out of 100 in an examination at the end of the first year, and is sex.
Variables
Objects
1
18.45
70
1
2
18.41
65
0
3
18.39
71
0
4
18.70
72
0
5
18.34
94
1
1 indicates male; 0 indicates female.
The data in Table 1.2, which were collected at the University of Hull in the early 1970s, formed part of an investigation into the merits of open‐book vs. closed‐book examinations. Marks (out of 100) were given for 88 students on each of five subjects; these observations were sorted (almost) according to the average. Initially, we may enquire as to the degree of dependence between performance on different papers taken by the same students. It may be useful, for counseling or other purposes, to have some idea of how final degree marks (“dependent” variables) are affected by previous examination results or by other variables such as age and sex (“explanatory” variables). This presents the so‐called regression problem, which is examined in Chapter 7.
Given examination marks on different topics (as in Table 1.2), the question arises of how to combine or average these marks in a suitable way. A straightforward method would use the simple arithmetic mean, but this procedure may not always be suitable. For instance, if the marks on some papers vary more than others, we may wish to weight them differently. This leads us to search for a linear combination (weighted sum) which is “optimal” in some sense. If all the examination papers fall in one group, then principal component analysis and factor analysis are two techniques that can help to answer such questions (see Chapters 9 and 10). In some situations, the papers may fall into more than one group – for instance, in Table 1.2, some examinations were “open book”, while others were “closed book”. In such situations, we may wish to investigate the use of linear combinations within each group separately. This leads to the method known as canonical correlation analysis, which is discussed in Chapter 11.
The idea of taking linear combinations is an important one in multivariate analysis, and we will return to it in Section 1.5.
Table 1.2 Marks in open‐ and closed‐book examination out of 100.
Mechanics (C)
Vectors (C)
Algebra (O)
Analysis (O)
Statistics (O)
77
82
67
67
81
63
78
80
70
81
75
73
71
66
81
55
72
63
70
68
63
63
65
70
63
53
61
72
64
73
51
67
65
65
68
59
70
68
62
56
62
60
58
62
70
64
72
60
62
45
52
64
60
63
54
55
67
59
62
44
50
50
64
55
63
65
63
58
56
37
31
55
60
57
73
60
64
56
54
40
44
69
53
53
53
42
69
61
55
45
62
46
61
57
45
31
49
62
63
62
44
61
52
62
46
49
41
61
49
64
12
58
61
63
67
49
53
49
62
47
54
49
56
47
53
54
53
46
59
44
44
56
55
61
36
18
44
50
57
81
46
52
65
50
35
32
45
49
57
64
30
69
50
52
45
46
49
53
59
37
40
27
54
61
61
31
42
48
54
68
36
59
51
45
51
56
40
56
54
35
46
56
57
49
32
45
42
55
56
40
42
60
54
49
33
40
63
53
54
25
23
55
59
53
44
48
48
49
51
37
41
63
49
46
34
46
52
53
41
40
46
61
46
38
41
40
57
51
52
31
49
49
45
48
39
22
58
53
56
41
35
60
47
54
33
48
56
49
42
32
31
57
50
54
34
17
53
57
43
51
49
57
47
39
26
59
50
47
15
46
37
56
49
28
45
40
43
48
21
61
35
35
41
51
50
38
44
54
47
24
43
43
38
34
49
39
46
46
32
43
62
44
36
22
42
48
38
41
44
33
34
42
50
47
29
18
51
40
56
30
35
36
46
48
29
59
53
37
22
19
41
41
43
30
33
31
52
37
27
40
17
51
52
35
31
34
30
50
47
36
46
40
47
29
17
10
46
36
47
39
46
37
45
15
30
30
34
43
46
18
13
51
50
25
31
49
50
38
23
09
18
32
31
45
40
08
42
48
26
40
23
38
36
48
15
30
24
43
33
25
03
09
51
47
40
07
51
43
17
22
15
40
43
23
18
15
38
39
28
17
05
30
44
36
18
12
30
32
35
21
05
26
15
20
20
00
40
21
09
14
O indicates open book, and C indicates closed book.
Table 1.3 gives three data matrices (or one data matrix if the species is coded as a variable). In each matrix, the “objects” are 50 irises of species Iris setosa, Iris versicolor, and Iris virginica, respectively. The “variables” are
=
sepal length,
=
sepal width,
=
petal length,
=
petal width.
The flowers of the first two iris species (I. setosa and I. versicolor) were taken from the same natural colony but the sample of the third iris species (I. virginica) is from a different colony; for more general details on the data, see Mardia (2023). If a new iris of unknown species has measurements , and , we may ask to which species it belongs. This presents the problem of discriminant analysis, which is discussed in Chapter 12. However, if we were presented with the 150 observations of Table 1.3 in an unclassified manner (say, before the three species were established), then the aim could have been to dissect the population into homogeneous groups. This problem is handled by cluster analysis (see Chapter 14).
Table 1.3 Measurements (in cm) on three types of irises.
Source: Fisher (1936) / John Wiley & Sons.
Iris setosa
Iris versicolor
Iris virginica
Sepal length
Sepal width
Petal length
Petal width
Sepal length
Sepal width
Petal length
Petal width
Sepal length
Sepal width
Petal length
Petal width
5.1
3.5
1.4
0.2
7.0
3.2
4.7
1.4
6.3
3.3
6.0
2.5
4.9
3.0
1.4
0.2
6.4
3.2
4.5
1.5
5.8
2.7
5.1
1.9
4.7
3.2
1.3
0.2
6.9
3.1
4.9
1.5
7.1
3.0
5.9
2.1
4.6
3.1
1.5
0.2
5.5
2.3
4.0
1.3
6.3
2.9
5.6
1.8
5.0
3.6
1.4
0.2