Structural Equations with Latent Variables - Kenneth A. Bollen - E-Book

Structural Equations with Latent Variables E-Book

Kenneth A. Bollen

4,4
147,99 €

oder
-100%
Sammeln Sie Punkte in unserem Gutscheinprogramm und kaufen Sie E-Books und Hörbücher mit bis zu 100% Rabatt.
Mehr erfahren.
Beschreibung

Analysis of Ordinal Categorical Data Alan Agresti Statistical Science Now has its first coordinated manual of methods for analyzing ordered categorical data. This book discusses specialized models that, unlike standard methods underlying nominal categorical data, efficiently use the information on ordering. It begins with an introduction to basic descriptive and inferential methods for categorical data, and then gives thorough coverage of the most current developments, such as loglinear and logit models for ordinal data. Special emphasis is placed on interpretation and application of methods and contains an integrated comparison of the available strategies for analyzing ordinal data. This is a case study work with illuminating examples taken from across the wide spectrum of ordinal categorical applications. 1984 (0 471-89055-3) 287 pp. Regression Diagnostics Identifying Influential Data and Sources of Collinearity David A. Belsley, Edwin Kuh and Roy E. Welsch This book provides the practicing statistician and econometrician with new tools for assessing the quality and reliability of regression estimates. Diagnostic techniques are developed that aid in the systematic location of data points that are either unusual or inordinately influential; measure the presence and intensity of collinear relations among the regression data and help to identify the variables involved in each; and pinpoint the estimated coefficients that are potentially most adversely affected. The primary emphasis of these contributions is on diagnostics, but suggestions for remedial action are given and illustrated. 1980 (0 471-05856-4) 292 pp. Applied Regression Analysis Second Edition Norman Draper and Harry Smith Featuring a significant expansion of material reflecting recent advances, here is a complete and up-to-date introduction to the fundamentals of regression analysis, focusing on understanding the latest concepts and applications of these methods. The authors thoroughly explore the fitting and checking of both linear and nonlinear regression models, using small or large data sets and pocket or high-speed computing equipment. Features added to this Second Edition include the practical implications of linear regression; the Durbin-Watson test for serial correlation; families of transformations; inverse, ridge, latent root and robust regression; and nonlinear growth models. Includes many new exercises and worked examples. 1981 (0 471-02995-5) 709 pp.

Sie lesen das E-Book in den Legimi-Apps auf:

Android
iOS
von Legimi
zertifizierten E-Readern

Seitenzahl: 842

Veröffentlichungsjahr: 2014

Bewertungen
4,4 (18 Bewertungen)
11
3
4
0
0
Mehr Informationen
Mehr Informationen
Legimi prüft nicht, ob Rezensionen von Nutzern stammen, die den betreffenden Titel tatsächlich gekauft oder gelesen/gehört haben. Wir entfernen aber gefälschte Rezensionen.



Contents

Preface

CHAPTER ONE: Introduction

HISTORICAL BACKGROUND

CHAPTER TWO: Model Notation, Covariances, and Path Analysis

MODEL NOTATION

COVARIANCE

PATH ANALYSIS

SUMMARY

CHAPTER THREE: Causality and Causal Models

NATURE OF CAUSALITY

ISOLATION

ASSOCIATION

DIRECTION OF CAUSATION

LIMITATIONS OF “CAUSAL” MODELING

SUMMARY

CHAPTER FOUR: Structural Equation Models with Observed Variables

MODEL SPECIFICATION

IMPLIED COVARIANCE MATRIX

IDENTIFICATION

ESTIMATION

FURTHER TOPICS

SUMMARY

APPENDIX 4A DERIVATION OF FML (y and x MULTINORMAL)

APPENDIX 4B DERIVATION OF FML (S WISHART DISTRIBUTION)

APPENDIX 4C NUMERICAL SOLUTIONS TO MINIMIZE FITTING FUNCTIONS

APPENDIX 4D ILLUSTRATIONS OF LISREL AND EQS PROGRAMS

CHAPTER FIVE: The Consequences of Measurement Error

UNIVARIATE CONSEQUENCES

BIVARIATE AND SIMPLE REGRESSION CONSEQUENCES

CONSEQUENCES IN MULTIPLE REGRESSION

CORRELATED ERRORS OF MEASUREMENT

CONSEQUENCES IN MULTIEQUATION SYSTEMS

SUMMARY

APPENDIX 5A ILLUSTRATIONS OF LISREL AND EQS PROGRAMS

CHAPTER SIX: Measurement Models: The Relation between Latent and Observed Variables

MEASUREMENT MODELS

VALIDITY

RELIABILITY

CAUSE INDICATORS

SUMMARY

APPENDIX 6A LISREL PROGRAM FOR THE MULTITRAIT-MULTIMETHOD EXAMPLE

CHAPTER SEVEN: Confirmatory Factor Analysis

EXPLORATORY AND CONFIRMATORY FACTOR ANALYSIS

MODEL SPECIFICATION

IMPLIED COVARIANCE MATRIX

IDENTIFICATION

ESTIMATION

MODEL EVALUATION

COMPARISON OF MODELS

RESPECIFICATION OF MODEL

EXTENSIONS

SUMMARY

APPENDIX 7A EXAMPLES OF PROGRAM LISTINGS

CHAPTER EIGHT: The General Model, Part I: Latent Variable and Measurement Models Combined

MODEL SPECIFICATION

IMPLIED COVARIANCE MATRIX

IDENTIFICATION

ESTIMATION AND MODEL EVALUATION

STANDARDIZED AND UNSTANDARDIZED COEFFICIENTS

MEANS AND EQUATION CONSTANTS

COMPARING GROUPS

MISSING VALUES

TOTAL, DIRECT, AND INDIRECT EFFECTS

SUMMARY

APPENDIX 8A ASYMPTOTIC VARIANCES OF EFFECTS

APPENDIX 8B LISTING OF THE LISREL VI PROGRAM FOR MISSING VALUE EXAMPLE

CHAPTER NINE: The General Model, Part II: Extensions

ALTERNATIVE NOTATIONS/REPRESENTATIONS

EQUALITY AND INEQUALITY CONSTRAINTS

QUADRATIC AND INTERACTION TERMS

INSTRUMENTAL-VARIABLE (IV) ESTIMATORS

DISTRIBUTIONAL ASSUMPTIONS

CATEGORICAL OBSERVED VARIABLES

SUMMARY

APPENDIX 9A LISREL PROGRAM FOR MODEL IN FIGURE 9.1(c)

APPENDIX A: Matrix Algebra Review

SCALARS, VECTORS, AND MATRICES

MATRIX OPERATIONS

APPENDIX B: Asymptotic Distribution Theory

CONVERGENCE IN PROBABILITY

CONVERGENCE IN DISTRIBUTIONS

References

Index

To Barbara and to my parents

Copyright © 1989 by John Wiley & Sons, Inc.

All rights reserved. Published simultaneously in Canada.

Reproduction or translation of any part of this work beyond that permitted by Section 107 or 108 of the 1976 United States Copyright Act without the permission of the copyright owner is unlawful. Requests for permission or further information should be addressed to the Permissions Department, John Wiley & Sons, Inc.

Library of Congress Cataloging in Publication Data:Bollen, Kenneth A.Structural equations with latent variables/Kenneth A. Bollen.p. cm.— (Wiley series in probability and mathematical statistics. Applied probability and statistics section, ISSN 0271-6356)“Wiley-Interscience publication.”Bibliography: p.Includes index.1. Social sciences—Statistical methods. 2. Latent variables.I. Title. II. Series.HA29.B732 1989300'.29—dc 1988-27272ISBN 0-471-01171-1CIP

Preface

Within the past decade the vocabulary of quantitative research in the social sciences has changed. Terms such as “LISREL,” “covariance structures,” “latent variables,” “multiple indicators,” and “path models” are commonplace. The structural equation models that lie behind these terms are a powerful generalization of earlier statistical approaches. They are changing researchers’ perspectives on statistical modeling and building bridges between the way social scientists think substantively and the way they analyze data.

We can view these models in several ways. They are regression equations with less restrictive assumptions that allow measurement error in the explanatory as well as the dependent variables. They consist of factor analyses that permit direct and indirect effects between factors. They routinely include multiple indicators and latent variables. In brief, these models encompass and extend regression, econometric, and factor analysis procedures.

The book provides a comprehensive introduction to the general structural equation system, commonly known as the “LISREL model.” One purpose of the book is to demonstrate the generality of this model. Rather than treating path analysis, recursive and nonrecursive models, classical econometrics, and confirmatory factor analysis as distinct and unique, I treat them as special cases of a common model. Another goal of the book is to emphasize the application of these techniques. Empirical examples appear throughout. To gain practice with the procedures, I encourage the reader to reestimate the examples, and then to devise and estimate new models. Several chapters contain some of the LISREL or EQS programs I used to obtain the results for the empirical examples. I have kept the examples as realistic as possible. This means that some of the initial specifications do not fit well. Through my experiences with students, colleagues, and in my own work, I frequently have found that the beginning model does not adequately describe the data. Respecification is often necessary. I note the difficulties this creates in proper interpretations of significance tests and the added importance of replication.

A final purpose is to emphasize the crucial role played by substantive expertise in most stages of the modeling process. Structural equation models are not very helpful if you have little idea about the subject matter. To begin the fitting process, the analysts must draw upon their knowledge to construct a multiequation system that specifies the relations between all latent variables, disturbances, and indicators. Furthermore they must turn to substantive information when respecifying models and when evaluating the final model. Empirical results can reveal that initial ideas are in error or they can suggest ways to modify a model, but they are given meaning only within the context of a substantively informed model.

Structural equation models can be presented in two ways. One is to start with the general model and then show its specializations to simpler models. The other is to begin with the simpler models and to build toward the general model. I have chosen the latter strategy. I start with the regression/econometric and factor analysis models and present them from the perspective of the general model. This has the advantage of gradually including new material while having types of models with which the reader is somewhat familiar. It also encourages viewing old techniques in a new light and shows the often unrealistic assumptions implicit in standard regression/econometric and factor analyses.

Specifically, I have organized the book as follows. Chapter 2 introduces several methodological tools. I present the model notation, covariances and covariance algebra, and a more detailed account of path analysis. Appendixes A and B at the end of the book provide reviews of matrix algebra and of asymptotic distribution theory. Chapter 3 addresses the issue of causality. Implicitly, the idea of causality pervades much of the structural equation writings. The meaning of causality is subject to much controversy. I raise some of the issues behind the controversy and present a structural equation perspective on the meaning of causality.

The regression/econometric models for observed variables are the subject of Chapter 4. Though many readers have experience with these, the covariance structure viewpoint will be new to many. The consequences of random measurement error in observed variable models is the topic of Chapter 5. The chapter shows why and when we should care about measurement error in observed variables.

Once we recognize that variables are measured with error, we need to consider the relation between the error-free variable and the observed variable. Chapter 6 is an examination of this relation. It introduces procedures for developing measures and explores the concepts of reliability and validity. Chapter 7 is on confirmatory factor analysis, which is used for estimating measurement models such as those in Chapter 6.

Finally, Chapters 8 and 9 provide the general structural equation model with latent variables. Chapter 8 emphasizes the “basics,” whereas Chapter 9 treats more advanced topics such as arbitrary distribution estimators and the treatment of categorical observed variables.

The main motivation for writing this book arose from my experiences teaching at the Interuniversity Consortium for Political and Social Research (ICPSR) Summer Training Program in Methodology at the University of Michigan (1980–1988). I could not find a text suitable for graduate students and professionals with training in different disciplines. Nor could I find a comprehensive introduction to these procedures. I have written the book for social scientists, market researchers, applied statisticians, and other analysts who plan to use structural equation or LISREL models. I assume that readers have prior exposure and experience with matrix algebra and regression analysis. A background in factor analysis is helpful but not essential. Jöreskog and Sörbom’s (1986) LISREL and Bender’s (1985) EQS are the two most popular structural equation software packages. I make frequent reference to them, but the ideas of the book extend beyond any specific program.

I have many people to thank for help in preparing this book. The Interuniversity Consortium for Political and Social Research (ICPSR) at the U niversity of Michigan (Ann Arbor) has made it possible for me to teach these techniques for the last nine years in their Summer Program in Quantitative Methods. Bob Hoyer started me there. Hank Heitowit and the staff of ICPSR have continued to make it an ideal teaching environment. A number of the hundreds of graduate students, professors, and other professionals who attended the courses provided general and specific comments to improve the book. Gerhard Arminger (University of Wuppertal), Jan de Leeuw (University of Leiden), Raymond Horton (Lehigh University), Frederick Lorenz (Iowa State University), John Fox (York University), Robert Stine (University of Pennsylvania), Boone Turchi (University of North Carolina), and members of the Statistical and Mathematical Sociology Group at the University of North Carolina provided valuable comments on several of the chapters. Barbara Entwisle Bollen read several drafts of most chapters, and her feedback and ideas are reflected throughout the book. Without her encouragement, I do not know when or if I would h ave completed the book.

Brenda Le Blanc of Dartmouth College and Priscilla Preston and Jenny Neville of the University of North Carolina (Chapel Hill) performed expert typing above and beyond the call of duty. Stephen Birdsall, Jack Kasarda, and Larry Levine smoothed the path several times as I moved the manuscript from Dartmouth College to the University of North Carolina. I thank the Committee for Scholarly Publications, Artistic Exhibitions, and Performances at the University of North Carolina, who provided financial support needed to complete the book.

KENNETH A. BOLLEN

Chapel Hill. North Carolina

CHAPTER ONE

Introduction

Most researchers applying statistics think in terms of modeling the individual observations. In multiple regression or ANOVA (analysis of variance), for instance, we learn that the regression coefficients or the error variance estimates derive from the minimization of the sum of squared differences of the predicted and observed dependent variable for each case. Residual analyses display discrepancies between fitted and observed values for every member of the sample.

The methods of this book demand a reorientation. The procedures emphasize covariances rather than cases.1 Instead of minimizing functions of observed and predicted individual values, we minimize the difference between the sample covariances and the covariances predicted by the model. The observed covariances minus the predicted covariances form the residuals. The fundamental hypothesis for these structural equation procedures is that the covariance matrix of the observed variables is a function of a set of parameters. If the model were correct and if we knew the parameters, the population covariance matrix would be exactly reproduced. Much of this book is about the equation that formalizes this fundamental hypothesis:

(1.1)

In (1.1), ∑ (sigma) is the population covariance matrix of observed variables, θ (theta) is a vector that contains the model parameters, and ∑ (θ) is the covariance matrix written as a function of θ. The simplicity of this equation is only surpassed by its generality. It provides a unified way of including many of the most widely used statistical techniques in the social sciences. Regression analysis, simultaneous equation systems, confirmatory factor analysis, canonical correlations, panel data analysis, ANOVA, analy-sis of covariance, and multiple indicator models are special cases of (1.1).

(1.2)

where VAR( ) and COV( ) refer to the population variances and covari- ances of the elements in parentheses. In (1.2) the left-hand side is 2, and the right-hand side is 2(6), with 0 containing γ, VAR(x), and VAR(ζ) as parameters. The equation implies that each element on the left-hand side equals its corresponding element on the right-hand side. For example, COV(x, y) γVAR(x) and VAR(γy2 VAR(x) + VAR(ζ). I could modify this example to create a multiple regression by adding explanatory variables, or I could add equations and other variables to make it a simultaneous equations system such as that developed in classical econo- metrics. Both cases can be represented as special cases of equation (1.1), as I show in Chapter 4.

(1.3)

where ϕ (phi) is the variance of the latent factor ξ. Here θ consists of three elements: ϕ, VAR(δX), and VAR(δ2). The covariance matrix of the observed variables is a function of these three parameters. I could add more indicators and more latent factors, allow for coefficients ("factor loadings") relating the observed variables to the factors, and allow correlated disturbances creating an extremely general factor analysis model. As Chapter 7 demonstrates, this is a special case of the covariance structure equation (1.1).

(1.4)

The term “structurai” stands for the assumption that the parameters are not just descriptive measures of association but rather that they reveal an invariant “causal” relation. I will have more to say about the meaning of “causality” with respect to these models in Chapter 3, but for now, let it suffice to say that the techniques do not “discover” causal relations. At best they show whether the causal assumptions embedded in a model match a sample of data. Also the models are for continuous latent and observed variables. The assumption of continuous observed variables is violated frequently in practice. In Chapter 9 I discuss the robustness of the standard procedures and the development of new ones for noncontinuous variables.

Structural equation models draw upon the rich traditions of several disciplines. I provide a brief description of their origins in the next section.

HISTORICAL BACKGROUND

Who invented general structural equation models? There is no simple answer to this question because many scholars have contributed to their development. The answer to this question is further complicated in that the models continue to unfold, becoming more general and more flexible. However, it is possible to outline various lines of research that have contributed to the evolution of these models.

My review is selective. More comprehensive discussions are available from the perspectives of sociology (Bielby and Hauser 1977), psychology (Bentler 1980; 1986), and economics (Goldberger 1972; Aigner et al. 1984). Two edited collections that represent the multidisciplinary origins of these techniques are the volumes by Goldberger and Duncan (1973) and Blalock ([1971] 1985). Other more recent collections are in Aigner and Goldberg (1977), Jöreskog and Wold (1982), the November 1982 issue of the Journal of Marketing Research, and the May–June 1983 issue of the Journal of Econometrics.

I begin by identifying three components present in today’s general structural equation models: (1) path analysis, (2) the conceptual synthesis of latent variable and measurement models, and (3) general estimation procedures. By tracing the rise of each component, we gain a better idea about the origins of these procedures.

Let me consider path analysis first. The biometrician Sewall Wright (1918, 1921, 1934, 1960) is its inventor. Three aspects of path analysis are the path diagram, the equations relating correlations or covariances to parameters, and the decomposition of effects. The first aspect, the path diagram, is a pictorial representation of a system of simultaneous equations. It shows the relation between all variables, including disturbances and errors. Figure 1.1 gives a path diagram for the last example of the previous section. It corresponds to the equations:

where ξ, δ1, and δ2 are uncorrelated with each other and with ξ. Straight single-headed arrows represent one-way causal influences from the variable at the arrow base to the variable to which the arrow points. The implicit coefficients of one for the effects of ξ on x1 and x2 are made explicit in the diagram.

Figure 1.1 Path Diagram Example

Using the path diagram, Wright proposed a set of rules for writing the equations that relate the correlations (or covariances) of variables to the model parameters; this constitutes the second aspect of path analysis. The equations are equivalent to covariance structure equations, an example of which appears in (1.4). He then proposed solving these equations for the unknown parameters, and substituting sample correlations or covariances for their population counterparts to obtain parameter estimates.

The third aspect of path analysis provides a means to distinguish direct, indirect, and total effects of one variable on another. The direct effects are those not mediated by any other variable; the indirect effects operate through at least one intervening variable, and the total effects is the sum of direct sind all indirect effects.4

With all these accomplishments it is surprising that social scientists and statisticians did not pay more attention to his work. As Ben tier (1986) documents, psychometrics only flirted (e.g., Dunlap and Cureton 1930; Englehart 1936) with Wright’s path analysis. Goldberger (1972) notes the neglect of econometricians and statisticians with a few exceptions (e.g., Fox 1958; Tukey 1954; Moran 1961; Dempster 1971). Wright’s work also was overlooked in sociology until the 1960s. Partially in reaction to work by Simon (1954), Tukey (1954), and Turner and Stevens (1959), sociologists such as Blalock (1961, 1963, 1964), Boudon (1965), and Duncan (1966) saw the potential of path analysis and related “partial-correlation” techniques as a means to analyze nonexperimental data. Following these works, and particularly following Duncan’s (1966) expository account, the late 1960s and early 1970s saw many applications of path analysis in the sociological journals. The rediscovery of path analysis in sociology diffused to political science and several other social science disciplines. Stimulated by work in sociology, Werts and Linn (1970) wrote an expository treatment of path analysis, but it was slow to catch on in psychology.

The next major boost to path analysis in the social sciences came when Jöreskog (1973), Keesing (1972), and Wiley (1973), who developed very general structural equation models, incorporated path diagrams and other features of path analysis into their presentations. Researchers know these techniques by the abbreviation of the JKW model (Bentler 1980), or more commonly as the LISREL model. The tremendous popularity of the LISREL model has facilitated the spread of path analysis. Path analysis has evolved over the years. Its present form has some elaboration in the symbols employed in path diagrams, has equations relating covariances to parameters that are derived with matrix operations rather than from “reading” the path diagram, and has a more refined and clearly defined decomposition of direct, indirect, and total effects (see, e.g., Duncan 1971; Alwin and Hauser 1975; Fox 1980; Graff and Schmidt 1982). But the contributions of Wright’s work are still clear.

In addition to path analysis, the conceptual synthesis of latent variable and measurement models was essential to contemporary structural equation techniques. The factor analysis tradition spawned by Spearman (1904) emphasized the relation of latent factors to observed variables. The central concern was on what we now call the measurement model. The structural relations between latent variables other than their correlation (or lack of correlation) were not examined. In econometrics the focus was the structural relation between observed variables with an occasional reference to error-in-the-variable situations.

Wright’s path analysis examples demonstrated that econometric-type models with variables measured with error could be identified and estimated. The conceptual synthesis of models containing structurally related latent variables and more elaborate measurement models was developed extensively in sociology during the 1960s and early 1970s. For instance, in 1963 Blalock argued that sociologists should use causal models containing both indicators and underlying variables to make inferences about the latent variables based on the covariances of the observed indicators. He suggested that observed variables can be causes or effects of latent variables or observed variables can directly affect each other. He contrasted this with the restrictive implicit assumptions of factor analysis where all indicators are viewed as effects of the latent variable. Duncan, Haller, and Portes (1968) developed a simultaneous equation model of peer influences on high school students’ ambitions. The model included two latent variables reciprocally related, multiple indicators of the latent variables, and several background characteristics that directly affected the latent variables. Heise (1969) and others applied path analysis to separate the stability of latent variables from the reliability of measures.

This and related work in sociology during the 1960s and early 1970s demonstrated the potential of synthesizing econometric-type models with latent rather than observed variables and psychometric-type measurement models with indicators linked to latent variables. But their approach was by way of examples; they did not establish a general model that could be applied to any specific problems. It awaited the work of Jöreskog (1973), Keesing (1972), and Wiley (1973) for a practical general model to be proposed. Their models had two parts. The first was a latent variable model that was similar to the simultaneous equation model of econometrics except that all variables were latent ones. The second part was the measurement model that showed indicators as effects of the latent variables as in factor analyses. Matrix expressions for these models were presented so that they could apply to numerous individual problems. Jöreskog and Sörbom’s LISREL programs were largely responsible for popularizing these structural equation models, as were the numerous publications and applications of Jöreskog (e.g., 1967, 1970, 1973, 1977, 1978) and his collaborators.

Bentler and Weeks (1980), McArdle and McDonald (1984), and others have proposed alternative representations of general structural equations. Though initially it seemed that these models were more general than the JKW model, most analysts now agree that both the new and old representations are capable of treating the range of linear models that typically occur in practice. I use what has come to be known as the “LISREL notation” throughout the book. To date, it is the most widely accepted representation. I will demonstrate ways to modify it to treat nonstandard applications in several of the chapters.

The last characteristic of the structural equation models are general estimation procedures. The early applications of path analysis by Wright and the sociologists influenced by his work used ad hoc estimation procedures to yield parameter estimates. There was little discussion of statistical inference and optimal ways of combining multiple estimates of a single parameter. Here work from econometrics and psychometrics proved indis- pensible. In econometrics the properties of estimators for structural equations with observed variables were well established (see, e.g., Goldberger 1964). In psychometrics the work of Lawley (1940), Anderson and Rubin (1956), and Jöreskog (1969) helped lay the foundations for hypothesis testing in factor analysis. Bock and Bargmann (1966) proposed an analysis of covariance structures to estimate the components of variance due to latent variables in multinormal observed variables. Jöreskog (1973) proposed a maximum likelihood estimator (based on the multinormality of the observed variables) for general structural equation models which is today the most widely used estimator. Jöreskog and Goldberger (1972) and Browne (1974, 1982, 1984) suggested generalized least squares (GLS) estimators that offer additional flexibility in the assumptions under which they apply. Browne (1982, 1984), for example, proposed estimators that assume arbitrary distributions or elliptical distributions for the observed variables. Bentler (1983) suggested estimators that treat higher-order product moments of the observed variables. He demonstrated that these moments can help identify model parameters that are not identified by the covariances and the gains in efficiency that may result. Muthén (1984, 1987), among others, has generalized these models to ordinal or limited observed variables.

Finally, this sketch of developments in structural equation models would not be complete without mentioning the computer software that has emerged. As I have already said, Jöreskog and Sörbom’s LISREL software perhaps has been the single greatest factor leading to the spread of these techniques throughout the social sciences. It is now entering its seventh version. Bentler’s (1985) EQS software has recently entered the field and also is likely to be widely used. McDonald (1980), Schoenberg (1987), and others have written programs with more limited circulations. The dual trend in software is that of providing more general and flexible models and programs that are more “user-friendly.” Examples of the former are that LISREL VII and EQS allow a variety of estimators including Arbitrary distribution estimators and Muthén’s (1987) LISCOMP allows ordinal or limited observed variables. At the same time there is a movement to allow programming through specifying equations (e.g., EQS and Jöreskog and Sörbom’s SIMPLIS) rather than matrices.

1As is clear from several places in the book, individual cases that are outliers can severely affect covariances and estimates of parameters. Thus, with these techniques, researchers still need to check for outliers. In addition, in many cases (e.g., regression models) the minimizations based on individuals and minimizations based on the predicted and observed covariance matrices lead to the same parameter estimates.

2 Given the symmetric nature of the covariance matrices only the lower half of these matrices is shown.

3Though an incredibly broad range of procedures falls under linear equations, those treated here are a special class of possible models of the general moment structure models (see Bentler, 1983).

41 review path analysis more fully in Chapters 2 and 8.

CHAPTER TWO

Model Notation, Covariances, and Path Analysis

Readers of this book are likely to have diverse backgrounds in statistics. There is a need to establish some common knowledge. I assume that readers have prior exposure to matrix algebra. Appendix A at the end of the book provides a summary of basic matrix algebra for those wishing to review it. Appendix B gives an overview of asymptotic distribution theory which I use in several chapters. This chapter discusses three basic tools essential to understanding structural equation models. They are model notation, covari­ances, and path analysis.

MODEL NOTATION

Jöreskog (1973, 1977), Wiley (1973), and Keesling (1972) developed the notation on which I rely. Jöreskog and Sörbom’s LISREL (LInear Structural RELationships) computer program popularized it, and many refer to it as the LISREL notation. I introduce the basic notation in this section and save the more specialized symbols for the later chapters where they are needed.

The full model consists of a system of structural equations. The equations contain random variables, structural parameters, and sometimes, nonrandom variables. The three types of random variables are latent, observed, and disturbance/error variables. The nonrandom variables are explanatory variables whose values remain the same in repeated random sampling (fixed or nonstochastic variables). These are less common than random explanatory variables.

The links between the variables are summarized in the structural parameters. The structural parameters are invariant constants that provide the “causal” relation between variables. The structural parameters may describe the causal link between unobserved variables, between observed variables, or between unobserved and observed variables. I further discuss the meaning of causality and the structural parameters in Chapter 3. The system of structural equations has two major subsystems: the latent variable model and the measurement model.

Latent Variable Model

Latent random variables represent unidimensional concepts in their purest form. Other terms for these are unobserved or unmeasured variables and factors. The observed variables or indicators of a latent variable contain random or systematic measurement errors, but the latent variable is free of these. Since all latent variables correspond to concepts, they are hypothetical variables. Concepts and latent variables, however, vary in their degree of abstractness. Intelligence, social class, power, and expectations are highly abstract latent variables that are central to many social science theories. Also important, but less abstract, are variables such as income, education, population size, and age. The latter type of latent variables are directly measurable, whereas the former are capable of being only indirectly measured. An example containing both types of latent variables is Emile Durkheim’s hypothesis of the inverse relationship between social cohesion and suicide. Social cohesion refers to group solidarity, a fairly abstract latent variable. Suicide is directly observable. But this direct-indirect demarcation becomes blurred when one considers that some suicides are disguised or misclassified as some other form of death. Thus the measurement of suicide may not be as direct as it initially appears. I make no distinction between directly and indirectly observable latent variables for the latent variable models. Analytically they may be treated the same. Chapter 6 provides further discussion of the nature of latent variables, their measurement, and their scales.

The latent variable model encompasses the structural equations that summarize the relationships between latent variables. Sometimes this part of the model is called the “structural equation” or “causal model.” I depart from this practice because it can be misleading. All equations in the model, both those for the latent variables and those for the measurement model, describe structural relationships. To apply structural to only the latent variable part of the full model suggests that the measurement model is not structural.

I use the relationship of political democracy to industrialization in developing countries to introduce the notation for latent variable models. International development researchers disagree about whether industrialization is positively associated with political democracy in Third World countries. The alternation between dictatorships and electoral regimes in some of these societies makes it difficult to discern whether any general association exists. Political democracy refers to the extent of political rights (e.g., fairness of elections) and political liberties (e.g., freedom of the press) in a country. Industrialization is the degree to which a society’s economy is characterized by mechanized manufacturing processes. It is some of the consequences of industrialization (e.g., societal wealth, an educated population, advances in living standards) that are thought to enhance the chances of democracy. However, to keep the model simple, I do not include these intervening variables. Suppose that I have three latent random variables: political democracy in 1965 and 1960, and industrialization in 1960. I assume that political democracy in 1965 is a function of 1960 political democracy and industrialization. The 1960 industrialization level also affects the 1960 political democracy level. Nothing is said about the determi­nants of industrialization that lie outside of the model. Industrialization is an (“independent”) latent variable and is symbolized as (xi). It is exogenous because its causes lie outside the model. The latent political democracy variables are they are determined by variables within the model. Each endogenous latent variable is represented by (eta). Political democracy in 1960 is represented as and 1965 democracy by . The latent endogenous variables are only partially explained by the model. The unexplained component is represented by (zeta) which is the random disturbance in the equation.

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!

Lesen Sie weiter in der vollständigen Ausgabe!