'Singular Gradient Matrix' Errors and Nonlinear Regression in Axum.


Axum 12246 'Singular Gradient Matrix' Errors and Nonlinear Regression in Axum. Error regarding "singular gradient matrix"  Example. Problem in nls(EB ~ E  (a * T^2)/(1 + b * T), d..: singular gradient matrix Use traceback() to see the call stack The message 'Error in nls(..): Singular gradient matrix' Is returned by the 'Statistics > Nonlinear > Regression' dialog box and the nls() function when the problem matrix is nearly singular (or noninvertible). This error message is a side effect of the underlying nonlinear leastsquares algorithm, and is usually dependent on your chosen starting values. For your reference, there follows a discussion below on the theory behind the "singular gradient matrix" error message. The Nonlinear Regression dialog relies on the function 'nls()' for its main algorithm, and this is the function that is referred to in the discussion. It is explained below that the "singular gradient" error is usually eliminated by choosing different initial values for the model parameters, or if you have already tried this, it may be possible that your nonlinear model is singular in many points surrounding the optimal parameter values. If you have nonlinear parameters in your model, you might try selecting the 'Partial Linear Algorithm' box in the Options tab of the Nonlinear Regression dialog. You might also find the 'Graph > 2D Plots > Fit  Curvefit' plotting functions helpful in visualizing the fit of your dataset. The "singular gradient matrix" message in nls() nls() uses a method known as the GaussNewton method to minimise sums of squares of nonlinear functions. In the case of nls() the functions are just the residual function of a nonlinear model at certain parameter values and a given response vector. If r(x) is the vector of residuals (response  fitted values), then at each iteration the GaussNewton method moves from the current parameter values, x, along a direction d that is the solution to (1) J(x)^T J(x) d =  J(x)^T r(x) (I'm using ^T to denote the transpose) where J(x) is the Jacobian matrix of r(x). The next iterate is then x + a d, for some positive scalar a. The right hand side of (1) is the negative gradient of residuals, and if Newton's method were being used, then the next iterate would be in a direction d that solves (2) H(x) d =  J(x)^T r(x) where H(x) is the Hessian matrix of the sum of squares of the residuals r(x). H(x) has the form J(x)^T J(x) + B(x), where B(x) vanishes at points where r(x) = 0, so what GaussNewton is doing is using J(x)^T J(x) to approximate the Hessian matrix of the sum of squares. Now (1) is just the normal equations for the leastsquares problem (3) J(x) d = r(x) and leastsquares techniques are typically used to obtain d for reasons of numerical stability. Assuming there are more residuals than parameters, the GaussNewton method is guaranteed to converge to a local minimum of the sum of squares provided J(x) has full rank at each iterate. However, when J(x) is rankdeficient or nearly rankdeficient, the solution to (1) (and (3)) is no longer unique. You can still solve for the minimum norm solution, which is unique, but can cause computational problems because it involves a rank determination. Moreover, even if you were able to compute the exact minimum norm solution, the GaussNewton method would not be guaranteed to converge to a local minimum. What nls() means when it gives you the message "singular gradient matrix" is that the Jacobian matrix J(x) is nearly rank deficient. (I don't know why they call it a "gradient" matrix  as far as I know that terminology is not widely used.) When this happens, nls() actually gives up, and in Statistical Models in S, they tell you to try different initial estimates when this happens. There are nonlinear leastsquares problems with welldefined solutions that have rankdeficient or nearly rankdeficient Jacobians at many points. In fact this situation occurs not infrequently at the solution itself, so that using a different starting estimate will not help you with nls(). However, there are more robust methods for nonlinear leastsquares that routinely handle nearly rankdeficient or rank deficient Jacobians. In SPLUS, one such method is used by the function nlregb(). You can also solve the problem as an unconstrained minimization (in SPLUS, using ms() or nlminb()). Creating plots Axum All 7 en SPLUS
If you can't find a solution on the Knowledge Base then please contact us on the Technical Support Request Form or by email or by telephone on +44 (0) 203 695 7810
Adept Products
For the time being we are unable to offer the following product ranges although we are currently working hard to increase the number of products we can offer in the future. Please contact us to talk about alternative products that we may be able to offer you.