The CALIS Procedure

The LISMOD Model and Submodels

As a statistical model, the LISMOD modeling language is derived from the LISREL model proposed by Jöreskog and others (see Keesling 1972; Wiley 1973; Jöreskog 1973). But as a computer language, the LISMOD modeling language is quite different from the LISREL program. To maintain the consistence of specification syntax within the CALIS procedure, the LISMOD modeling language departs from the original LISREL programming language. In addition, to make the programming a little easier, some terminological changes from LISREL are made in LISMOD.

For brevity, models specified by the LISMOD modeling language are called LISMOD models, although you can also specify these LISMOD models by other general modeling languages that are supported in PROC CALIS.

The following descriptions of LISMOD models are basically the same as those of the original LISREL models. The main modifications are the names for the model matrices.

The LISMOD model is described by three component models. The first one is the structural equation model that describes the relationships among latent constructs or factors. The other two are measurement models that relate latent factors to manifest variables.

Structural Equation Model

The structural equation model for latent factors is

\[  \bm {\eta } = \balpha + \bbeta \bm {\eta } + \bGamma \bxi + \bzeta  \]

where: $\bm {\eta }$ is a random vector of $n_\eta $ endogenous latent factors $\bxi $ is a random vector of $n_\bxi $ exogenous latent factors $\bzeta $ is a random vector of errors $\balpha $ is a vector of intercepts $\bbeta $ is a matrix of regression coefficients of $\bm {\eta }$ variables on other $\bm {\eta }$ variables $\bGamma $ is a matrix of regression coefficients of $\bm {\eta }$ on $\bxi $

There are some assumptions in the structural equation model. To prevent a random variable in $\bm {\eta }$ from regressing directly on itself, the diagonal elements of $\bbeta $ are assumed to be zeros. Also, $(I-\bbeta )^{-1}$ is assumed to be nonsingular, and $\bzeta $ is uncorrelated with $\bxi $.

The covariance matrix of $\bzeta $ is denoted by $\bPsi $ and its expected value is a null vector. The covariance matrix of $\bxi $ is denoted by $\bPhi $ and its expected value is denoted by $\bkappa $.

Because variables in the structural equation model are not observed, to analyze the model these latent variables must somehow relate to the manifest variables. The measurement models, which are discussed in the subsequent sections, provide such relations.

Measurement Model for $\mi{y}$

\[  \mb{y} = \bnu _ y + \bLambda _ y\bm {\eta } + \bepsilon  \]

where: $\mb{y}$ is a random vector of $n_ y$ manifest variables $\bepsilon $ is a random vector of errors for $\mb{y}$ $\bnu _ y$ is a vector of intercepts for $\mb{y}$ $\bLambda _ y$ is a matrix of regression coefficients of $\mb{y}$ on $\bm {\eta }$

It is assumed that $\bepsilon $ is uncorrelated with either $\bm {\eta }$ or $\bzeta $. The covariance matrix of $\bepsilon $ is denoted by $\bTheta _ y$ and its expected value is the null vector.

Measurement Model for $\mi{x}$

\[  \mb{x} = \bnu _ x + \bLambda _ x\bxi + \bdelta  \]

where: $\mb{x}$ is a random vector of $n_ x$ manifest variables $\bdelta $ is a random vector of errors for $\mb{x}$ $\bnu _ x$ is a vector of intercepts for $\mb{x}$ $\bLambda _ x$ is a matrix of regression coefficients of $\mb{x}$ on $\bxi $

It is assumed that $\bdelta $ is uncorrelated with $\bxi $, $\bepsilon $, or $\bzeta $. The covariance matrix of $\bdelta $ is denoted by $\bTheta _ x$ and its expected value is a null vector.

Covariance and Mean Structures

Under the structural and measurement equations and the model assumptions, the covariance structures of the manifest variables $(\mb{y}^\prime ,\mb{x}^\prime )^\prime $ are expressed as

\[  \bSigma = \left( \begin{matrix}  \bLambda _ y (I - \bbeta )^{-1} (\bGamma \bPhi \bGamma ^\prime + \bPsi )(I - \bbeta ^\prime )^{-1} \bLambda _ y^\prime + \bTheta _ y   &  \bLambda _ y (I - \bbeta )^{-1}\bGamma \bPhi \bLambda _ x^\prime   \\ \bLambda _ x\bPhi \bGamma ^\prime (I - \bbeta ^\prime )^{-1}\bLambda _ y^\prime   &  \bLambda _ x\bPhi \bLambda _ x^\prime + \bTheta _ x   \\ \end{matrix} \right)  \]

The mean structures of the manifest variables $(\mb{y}^\prime ,\mb{x}^\prime )^\prime $ are expressed as

\[  \bmu = \left( \begin{matrix}  \bnu _ y + \bLambda _ y (I - \bbeta )^{-1} (\balpha + \bGamma \bkappa )   \\ \bnu _ x + \bLambda _ x \bkappa   \\ \end{matrix} \right)  \]

Model Matrices in the LISMOD Model

The parameters of the LISMOD model are elements in the model matrices, which are summarized as follows.

Matrix

Name

Description

Dimensions

Row Variables

Column Variables

$\balpha $

_ALPHA_

Intercepts for $\bm {\eta }$

$n_\eta \times 1$

$\bm {\eta }$
(ETAVAR=)

N/A

$\bbeta $

_BETA_

Effects of $\bm {\eta }$ on $\bm {\eta }$

$n_\eta \times n_\eta $

$\bm {\eta }$
(ETAVAR=)

$\bm {\eta }$
(ETAVAR=)

$\bGamma $

_GAMMA_

Effects of $\bxi $ on $\bm {\eta }$

$n_\eta \times n_\xi $

$\bm {\eta }$
(ETAVAR=)

$\bxi $
(XIVAR=)

$\bPsi $

_PSI_

Error covariance matrix for $\bm {\eta }$

$n_\eta \times n_\eta $

$\bm {\eta }$
(ETAVAR=)

$\bm {\eta }$
(ETAVAR=)

$\bPhi $

_PHI_

Covariance matrix for $\bxi $

$n_\xi \times n_\xi $

$\bxi $
(XIVAR=)

$\bxi $
(XIVAR=)

$\bkappa $

_KAPPA_

Mean vector for $\bxi $

$n_\xi \times 1$

$\bxi $
(XIVAR=)

N/A

$\bnu _ y$

_NUY_

Intercepts for $\mb{y}$

$n_ y \times 1$

$\mb{y}$
(YVAR=)

N/A

$\bLambda _ y$

_LAMBDAY_

Effects of $\bm {\eta }$ on $\mb{y}$

$n_ y \times n_\eta $

$\mb{y}$
(YVAR=)

$\bm {\eta }$
(ETAVAR=)

$\bTheta _ y$

_THETAY_

Error covariance matrix for $\mb{y}$

$n_ y \times n_ y$

$\mb{y}$
(YVAR=)

$\mb{y}$
(YVAR=)

$\bnu _ x$

_NUX_

Intercepts for $\mb{x}$

$n_ x \times 1$

$\mb{x}$
(XVAR=)

N/A

$\bLambda _ x$

_LAMBDAX_

Effects of $\bxi $ on $\mb{x}$

$n_ x \times n_\xi $

$\mb{x}$
(XVAR=)

$\bxi $
(XIVAR=)

$\bTheta _ x$

_THETAX_

Error covariance matrix for $\mb{x}$

$n_ x \times n_ x$

$\mb{x}$
(XVAR=)

$\mb{x}$
(XVAR=)

There are twelve model matrices in the LISMOD model. Not all of them are used in all situations. See the section LISMOD Submodels for details. In the preceding table, each model matrix is given a name in the column Name, followed by a brief description of the parameters in the matrix, the dimensions, and the row and column variables being referred to. In the second column of the table, the LISMOD matrix names are used in the MATRIX statements when specifying the LISMOD model. In the last two columns of the table, following the row or column variables is the variable list (for example, ETAVAR=, YVAR=, and so on) in parentheses. These lists are used in the LISMOD statement for defining variables.

Specification of the LISMOD Model

The LISMOD specification consists of two tasks. The first task is to define the variables in the model. The second task is to specify the parameters in the LISMOD model matrices.

Specifying Variables

The first task is accomplished in the LISMOD statement . In the LISMOD statement, you define the lists of variables of interest: YVAR=, XVAR=, ETAVAR=, and XIVAR= lists, respectively for the $\mb{y}$-variables, $\mb{x}$-variables, $\bm {\eta }$-variables, and the $\bxi $-variables. While you provide the names of variables in these lists, you also define implicitly the numbers of four types of variables: $n_ y$, $n_ x$, $n_\eta $, and $n_\xi $. The variables in the YVAR= and XVAR= lists are manifest variables and therefore must be present in the analyzed data set. The variables in the ETAVAR= and XIVAR= lists are latent factors, the names of which are assigned by the researcher to represent their roles in the substantive theory. After these lists are defined, the dimensions of the model matrices are also defined by the number of variables on various lists. In addition, the variable orders in the lists are referred to by the row and column variables of the model matrices.

Unlike the LINEQS model, in the LISMOD model you do not need to use the 'F' or 'f' prefix to denote factors in the ETAVAR= or XIVAR= list. You can use any valid SAS variable names for the factors, especially those names that reflect the nature of the factors. To avoid confusion with other names in the model, some general rules are recommended. See the section Naming Variables and Parameters for these general rules about naming variables and parameters.

Specifying Parameters in Model Matrices

The second task is accomplished by the MATRIX statements. In each MATRIX statement, you specify the model matrix by using the matrix names described in the previous table. Then you specify the parameters (free or fixed) in the locations of the model matrix. You can use as many MATRIX statements as needed for defining your model. But each model matrix can be specified only in one MATRIX statement, and each MATRIX statement is used for specifying only one model matrix.

An Example

In the section LISMOD Model, the LISMOD modeling language is used to specify the model described in the section A Structural Equation Example. In the LISMOD statement, you define four lists of variables, as shown in the following statement:

lismod
   yvar  = Anomie67 Powerless67 Anomie71 Powerless71,
   xvar  = Education SEI,
   etav  = Alien67 Alien71,
   xivar = SES;

Endogenous latent factors are specified in the ETAVAR= list. Exogenous latent factors are specified in the XIVAR= list. In this case, Alien67 and Alien71 are the $\bm {\eta }$-variables, and SES is the only $\bxi $-variable in the model. Manifest variables that are indicators of endogenous latent factors in $\bm {\eta }$ are specified in the YVAR= list. In this case, they are the Anomie and Powerless variables, measured at two different time points. Manifest variables that are indicators of exogenous latent factors in $\bxi $ are specified in the XVAR= list. In this case, they are the Education and the SEI variables. Implicitly, the dimensions of the model matrices are defined by these lists already; that is, $n_ y=4$, $n_ x=2$, $n_\eta =2$, and $n_\xi =1$.

The MATRIX statements are used to specify parameters in the model matrices. For example, in the following statement you define the _LAMBDAX_ ($\bLambda _ x$) matrix with two nonzero entries:

 matrix _LAMBDAX_ [1,1] = 1.0,
                  [2,1] = lambda;

The first parameter location is for [1,1], which is the effect of SES (the first variable in the XIVAR= list) on Education (the first element in the XVAR= list). A fixed value of 1.0 is specified there. The second parameter location is for [2,1], which is the effect of SES (the first variable in the XIVAR= list) on SEI (the second variable in the XVAR= list). A parameter named lambda without an initial value is specified there.

Another example is shown as follows:

matrix _THETAY_  [1,1] = theta1,
                 [2,2] = theta2,
                 [3,3] = theta1,
                 [4,4] = theta2,
                 [3,1] = theta5,
                 [4,2] = theta5;

In this MATRIX statement, the error variances and covariances (that is, the $\bTheta _ y$ matrix) for the $\mb{y}$-variables are specified. The diagonal elements of the _THETAY_ matrix are specified by parameters theta1, theta2, theta1, and theta2, respectively, for the four $\mb{y}$-variables Anomie67, Powerless67, Anomie71, and Powerless71. By using the same parameter name theta1, the error variances for Anomie67 and Anomie71 are implicitly constrained. Similarly, the error variances for Powerless67 and Powerless71 are also implicitly constrained. Two more parameter locations are specified. The error covariance between Anomie67 and Anomie71 and the error covariance between Powerless67 and Powerless71 are both represented by the parameter theta5. Again, this is an implicit constraint on the covariances. All other unspecified elements in the _THETAY_ matrix are treated as fixed zeros.

In this example, no parameters are specified for matrices _ALPHA_, _KAPPA_, _NUY_, or _NUX_. Therefore, mean structures are not modeled.

LISMOD Submodels

It is not necessary to specify all four lists of variables in the LISMOD statement. When some lists are unspecified in the LISMOD statement, PROC CALIS analyzes submodels derived logically from the specified lists of variables. For example, if only $\mb{y}$- and $\mb{x}$-variable lists are specified, the submodel being analyzed would be a multivariate regression model with manifest variables only. Not all combinations of lists lead to meaningful submodels, however. To determine whether and how a submodel (which is formed by a certain combination of variable lists) can be analyzed, the following three principles in the LISMOD modeling language are applied:

  • Submodels with at least one of the YVAR= and XVAR= lists are required.

  • Submodels that have an ETAVAR= list but no YVAR= list cannot be analyzed.

  • When a submodel has a YVAR= (or an XVAR=) list but without an ETAVAR= (or a XIVAR=) list, it is assumed that the set of $\mb{y}$-variables ($\mb{x}$-variables) is equivalent to the $\bm {\eta }$-variables ($\bxi $-variables). Hereafter, this principle is referred to as an equivalence interpretation.

Apparently, the third principle is the same as the situation where the latent factors $\bm {\eta }$ (or $\bxi $) are perfectly measured by the manifest variables $\mb{y}$ (or $\mb{x}$). That is, in such a perfect measurement model, $\bLambda _ y$ ($\bLambda _ x$) is an identity matrix and $\bTheta _ y$ ($\bTheta _ x$) and $\bnu _ y$ ($\bnu _ x$) are both null. This can be referred to as a perfect measurement interpretation. However, the equivalence interpretation stated in the last principle presumes that there are actually no measurement equations at all. This is important because under the equivalence interpretation, matrices $\bLambda _ y$ ($\bLambda _ x$), $\bTheta _ y$ ($\bTheta _ x$) and $\bnu _ y$ ($\bnu _ x$) are nonexistent rather than fixed quantities, which is assumed under the perfect measurement interpretation. Hence, the $\mb{x}$-variables are treated as exogenous variables with the equivalence interpretation, but they are still treated as endogenous with the perfect measurement interpretation. Ultimately, whether $\mb{x}$-variables are treated as exogenous or endogenous affects the default or automatic parameterization. See the section Default Parameters for more details.

By using these three principles, the models and submodels that PROC CALIS analyzes are summarized in the following table, followed by detailed descriptions of these models and submodels.

Presence of Lists

Description

Model Equations

Nonfixed Model Matrices

Presence of Both $\mb{x}$- and $\mb{y}$-variables

1

YVAR=, ETAVAR=,

Full model

$\mb{y} = \bnu _ y + \bLambda _ y\bm {\eta } + \bepsilon $

$\bnu _ y$,$\bLambda _ y$,$\bTheta _ y$

 

XVAR=, XIVAR=

 

$\mb{x} = \bnu _ x + \bLambda _ x\bxi + \bdelta $

$\bnu _ x$,$\bLambda _ x$,$\bTheta _ x$,$\bkappa $,$\bPhi $

     

$\bm {\eta } = \balpha + \bbeta \bm {\eta } + \bGamma \bxi + \bzeta $

$\balpha $ ,$\bbeta $,$\bGamma $,$\bPsi $

2

YVAR=,

Full model with

$\mb{x} = \bnu _ x + \bLambda _ x\bxi + \bdelta $

$\bnu _ x$,$\bLambda _ x$,$\bTheta _ x$,$\bkappa $,$\bPhi $

 

XVAR=, XIVAR=

$\mb{y} \equiv \bm {\eta }$

$\mb{y} = \balpha + \bbeta y + \bGamma \bxi + \bzeta $

$\balpha $,$\bbeta $,$\bGamma $,$\bPsi $

3

YVAR=, ETAVAR=,

Full model with

$\mb{y} = \bnu _ y + \bLambda _ y\bm {\eta } + \bepsilon $

$\bnu _ y$,$\bLambda _ y$,$\bTheta _ y$

 

XVAR=

$\mb{x} \equiv \bxi $

$\bm {\eta } = \balpha + \bbeta \bm {\eta } + \bGamma x + \bzeta $

$\balpha $ ,$\bbeta $,$\bGamma $,$\bPsi $,$\bkappa $,$\bPhi $

4

YVAR=,

Regression

$\mb{y} = \balpha + \bbeta y + \bGamma x + \bzeta $, or

$\balpha $,$\bbeta $,$\bGamma $,$\bPsi $,$\bkappa $,$\bPhi $

 

XVAR=

($\mb{y} \equiv \bm {\eta }$)

$(\mb{I} - \bbeta )^{-1} y = \balpha + \bGamma x + \bzeta $

 
   

($\mb{x} \equiv \bxi $)

   

Presence of $\mb{x}$-variables and Absence of $\mb{y}$-variables

5

XVAR=, XIVAR=

Factor model

$\mb{x} = \bnu _ x + \bLambda _ x\bxi + \bdelta $

$\bnu _ x$,$\bLambda _ x$,$\bTheta _ x$,$\bkappa $,$\bPhi $

   

for $\mb{x}$

   

6

XVAR=

$\mb{x}$-structures

 

$\bkappa $,$\bPhi $

   

($\mb{x} \equiv \bxi $)

   

Presence of $\mb{y}$-variables and Absence of $\mb{x}$-variables

7

YVAR=, ETAVAR=

Factor model

$\mb{y} = \bnu _ y + \bLambda _ y\bm {\eta } + \bepsilon $

$\bnu _ y$,$\bLambda _ y$,$\bTheta _ y$

   

for $\mb{y}$

$\bm {\eta } = \balpha + \bbeta \bm {\eta } + \bzeta $

$\balpha $,$\bbeta $,$\bPsi $

8

YVAR=

$\mb{y}$-structures

$\mb{y} = \balpha + \bbeta y + \bzeta $, or

$\balpha $,$\bbeta $,$\bPsi $

   

($\mb{y} \equiv \bm {\eta }$)

$(\mb{I} - \bbeta )^{-1} y = \balpha + \bzeta $

 

9

YVAR=, ETAVAR=,

Second-order

$\mb{y} = \bnu _ y + \bLambda _ y\bm {\eta } + \bepsilon $

$\bnu _ y$,$\bLambda _ y$,$\bTheta _ y$

 

XIVAR=

factor model

$\bm {\eta } = \balpha + \bbeta \bm {\eta } + \bGamma \bxi + \bzeta $

$\balpha $,$\bbeta $,$\bGamma $,$\bPsi $,$\bkappa $,$\bPhi $

10

YVAR=,

Factor model

$\mb{y} = \balpha + \bbeta y + \bGamma \bxi + \bzeta $, or

$\balpha $,$\bbeta $,$\bGamma $,$\bPsi $,$\bkappa $,$\bPhi $

 

XIVAR=

($\mb{y} \equiv \bm {\eta }$)

$(\mb{I} - \bbeta )^{-1} y = \balpha + \bGamma \bxi + \bzeta $

 

Models 1, 2, 3, and 4—Presence of Both $\mb{x}$- and $\mb{y}$-Variables

Submodels 1, 2, 3, and 4 are characterized by the presence of both $\mb{x}$- and $\mb{y}$-variables in the model. In fact, Model 1 is the full model with the presence of all four types of variables. All twelve model matrices are parameter matrices in this model.

Depending on the absence of the latent factor lists, manifest variables can replace the role of the latent factors in Models 2–4. For example, the absence of the ETAVAR= list in Model 2 means $\mb{y}$ is equivalent to $\bm {\eta }$ ($\mb{y} \equiv \bm {\eta }$). Consequently, you cannot, nor do you need to, use the MATRIX statement to specify parameters in the _LAMBDAY_, _THETAY_, or _NUY_ matrices under this model. Similarly, because $\mb{x}$ is equivalent to $\bxi $ ($\mb{x} \equiv \bxi $) in Model 3, you cannot, nor do you need to, use the MATRIX statement to specify the parameters in the _LAMBDAX_, _THETAX_, or _NUX_ matrices. In Model 4, $\mb{y}$ is equivalent to $\bm {\eta }$ ($\mb{y} \equiv \bm {\eta }$) and $\mb{x}$ is equivalent to $\bxi $ ($\mb{x} \equiv \bxi $). None of the six model matrices in the measurement equations are defined in the model. Matrices in which you can specify parameters by using the MATRIX statement are listed in the last column of the table.

Describing Model 4 as a regression model is a simplification. Because $\mb{y}$ can regress on itself in the model equation, the regression description is not totally accurate for Model 4. Nonetheless, if $\bbeta $ is a null matrix, the equation describes a multivariate regression model with outcome variables $\mb{y}$ and predictor variables $\mb{x}$. This model is the TYPE 2A model in LISREL VI (Jöreskog and Sörbom, 1985).

You should also be aware of the changes in meaning of the model matrices when there is an equivalence between latent factors and manifest variables. For example, in Model 4 the $\bPhi $ and $\bkappa $ are now the covariance matrix and mean vector, respectively, of manifest variables $\mb{x}$, while in Model 1 (the complete model) these matrices are of the latent factors $\bxi $.

Models 5 and 6 — Presence of $\mb{x}$-Variables and Absence of $\mb{y}$-Variables

Models 5 and 6 are characterized by the presence of the $\mb{x}$-variables and the absence of $\mb{y}$-variables.

Model 5 is simply a factor model for measured variables $\mb{x}$, with $\bLambda _ x$ representing the factor loading matrix, $\bTheta _ x$ the error covariance matrix, and $\bPhi $ the factor covariance matrix. If mean structures are modeled, $\bkappa $ represents the factor means and $\bnu _ x$ is the intercept vector. This is the TYPE 1 submodel in LISREL VI (Jöreskog and Sörbom, 1985).

Model 6 is a special case where there is no model equation. You specify the mean and covariance structures (in $\bkappa $ and $\bPhi $, respectively) for the manifest variables $\mb{x}$ directly. The $\mb{x}$-variables are treated as exogenous variables in this case. Because this submodel uses direct mean and covariance structures for measured variables, it can also be handled more easily by the MSTRUCT modeling language. See the MSTRUCT statement and the section The MSTRUCT Model for more details.

Note that because $\bm {\eta }$-variables cannot exist in the absence of $\mb{y}$-variables (see one of the three aforementioned principles for deriving submodels), adding the ETAVAR= list alone to these two submodels does not generate new submodels that can be analyzed by PROC CALIS.

Models 7, 8, 9 and 10—Presence of $\mb{y}$-Variables and Absence of $\mb{x}$-Variables

Models 7–10 are characterized by the presence of the $\mb{y}$-variables and the absence of $\mb{x}$-variables.

Model 7 is a factor model for $\mb{y}$-variables (TYPE 3B submodel in LISREL VI). It is similar to Model 5, but with regressions among latent factors allowed. When $\bbeta $ is null, Model 7 functions the same as Model 5. It becomes a factor model for $\mb{y}$-variables, with $\bLambda _ y$ representing the factor loading matrix, $\bTheta _ y$ the error covariance matrix, $\bPsi $ the factor covariance matrix, $\balpha $ the factor means, and $\bnu _ y$ the intercept vector.

Model 8 (TYPE 2B submodel in LISREL VI) is a model for studying the mean and covariance structures of $\mb{y}$-variables, with regression among $\mb{y}$-variables allowed. When $\bbeta $ is null, the mean structures of $\mb{y}$ are specified in $\balpha $ and the covariance structures are specified in $\bPsi $. This is similar to Model 6. However, there is an important distinction. In Model 6, the $\mb{x}$-variables are treated as exogenous (no model equation at all). But the $\mb{y}$-variables are treated as endogenous in Model 8 (with or without $\bbeta = 0$). Consequently, the default parameterization would be different for these two submodels. See the section Default Parameters for details about the default parameterization.

Model 9 represents a modified version of the second-order factor model for $\mb{y}$. It would be a standard second-order factor model when $\bbeta $ is null. This is the TYPE 3A submodel in LISREL VI. With $\bbeta $ being null, $\bm {\eta }$ represents the first-order factors and $\bxi $ represents the second-order factors. The first- and second-order factor loading matrices are $\bLambda _ y$ and $\bGamma $, respectively.

Model 10 is another form of factor model when $\bbeta $ is null, with factors represented by $\bxi $ and manifest variables represented by $\mb{y}$. However, if $\bbeta $ is indeed a null matrix in applications, you might want to use Model 5, in which the factor model specification is more direct and intuitive.

Default Parameters in the LISMOD Model

When a model matrix is defined in a LISMOD model, you can specify fixed values or free parameters for the elements of the matrix by the MATRIX statement. All other unspecified elements in the matrix are set by default. There are two types of default parameters for the LISMOD model matrices: one is free parameters; the other is fixed zeros.

The following sets of parameters are free parameters by default:

  • the diagonal elements of the _THETAX_, _THETAY_, and _PSI_ matrices; these elements represent the error variances in the model

  • all elements of the _PHI_ matrix; these elements represent the variances and covariance among exogenous variables in the model

  • all elements in the _NUX_ and _NUY_ vectors if the mean structures are modeled; these elements represent the intercepts of the observed variables

  • all elements in the _ALPHA_ vector if a YVAR= list is specified but an ETAVAR= list is not specified and the mean structures are modeled; these elements represent the intercepts of the $\mb{y}$-variables

  • all elements in the _KAPPA_ vector if an XVAR= list is specified but an XIVAR= list is not specified and the mean structures are modeled; these elements represent the means of the $\mb{x}$-variables

PROC CALIS names the default free parameters with the _Add prefix and a unique integer suffix. You can override the default free parameters by explicitly specifying them as free, constrained, or fixed parameter in the MATRIX statements for the matrices.

Parameters that are not default free parameters in the LISMOD model are fixed zeros by default. You can override almost all of these default fixed zeros of the LISMOD model by using the MATRIX statements for the matrices. The only set of default fixed zeros that you cannot override is the set of the diagonal elements of the _BETA_ matrix. These fixed zeros reflect a model restriction that precludes variables from having direct effects on themselves.