This section briefly introduces the basic notation for nonlinear regression models that applies in this chapter. Additional notation is introduced throughout as needed.
The vector of observed responses is denoted as . This vector is the realization of an random vector . The NLIN procedure assumes that the variance matrix of this random vector is . In other words, the observations have equal variance (are homoscedastic) and are uncorrelated. By defining the special variable
_WEIGHT_
in your NLIN programming statements, you can introduce heterogeneous variances. If a _WEIGHT_
variable is present, then , where is a diagonal matrix containing the values of the _WEIGHT_
variable.
The mean of the random vector is represented by a nonlinear model that depends on parameters and regressor (independent) variables :
In contrast to linear models, the number of regressor variables (k) does not necessarily equal the number of parameters (p) in the mean function . For example, the model fitted in the next subsection contains a single regressor and two parameters.
To represent the mean of the vector of observations, boldface notation is used in an obvious extension of the previous equation:
The vector , for example, is an vector of the values for the first regressor variables. The explicit dependence of the mean function on and/or the vectors is often omitted for brevity.
In summary, the stochastic structure of models fit with the NLIN procedure is mathematically captured by






Note that the residual variance is typically also unknown. Since it is not estimated in the same fashion as the other p parameters, it is often not counted in the number of parameters of the nonlinear regression. An estimate of is obtained after the model fit by the method of moments based on the residual sum of squares.
A matrix that plays an important role in fitting nonlinear regression models is the matrix of the first partial derivatives of the mean function with respect to the p model parameters. It is frequently denoted as
The use of the symbol —common in linear statistical modeling—is no accident here. The first derivative matrix plays a similar role in nonlinear regression to that of the matrix in a linear model. For example, the asymptotic variance of the nonlinear leastsquares estimators is proportional to , and projectiontype matrices in nonlinear regressions are based on . Also, fitting a nonlinear regression model can be cast as an iterative process where a nonlinear model is approximated by a series of linear models in which the derivative matrix is the regressor matrix. An important difference between linear and nonlinear models is that the derivatives in a linear model do not depend on any parameters (see previous subsection). In contrast, the derivative matrix is a function of at least one element of . It is this dependence that lies at the core of the fact that estimating the parameters in a nonlinear model cannot be accomplished in closed form, but it is an iterative process that commences with usersupplied starting values and attempts to continually improve on the parameter estimates.