LAUSR.org creates dashboard-style pages of related content for over 1.5 million academic articles. Sign Up to like articles & get recommendations!

Faster model matrix crossproducts for large generalized linear models with discretized covariates

Photo by thinkmagically from unsplash

Wood et al. (J Am Stat Assoc 112(519):1199–1210, 2017 ) developed methods for fitting penalized regression spline based generalized additive models, with of the order of $$10^4$$ 10 4 coefficients,… Click to show full abstract

Wood et al. (J Am Stat Assoc 112(519):1199–1210, 2017 ) developed methods for fitting penalized regression spline based generalized additive models, with of the order of $$10^4$$ 10 4 coefficients, to up to $$10^8$$ 10 8 data. The methods offered two to three orders of magnitude reduction in computational cost relative to the most efficient previous methods. Part of the gain resulted from the development of a set of methods for efficiently computing model matrix products when model covariates each take only a discrete set of values substantially smaller than the sample size [generalizing an idea first appearing in Lang et al. (Stat Comput 24(2):223–238, 2014 )]. Covariates can always be rounded to achieve such discretization, and it should be noted that the covariate discretization is marginal. That is we do not rely on discretizing covariates jointly, which would typically require the use of very coarse discretization. The most expensive computation in model estimation is the formation of the matrix cross product $$\mathbf{X}^{\mathsf{T}}{\mathbf{WX}}$$ X T WX where $$\mathbf{X}$$ X is a model matrix and $${\mathbf{W}}$$ W a diagonal or tri-diagonal matrix. The purpose of this paper is to present a simple, novel and substantially more efficient approach to the computation of this cross product. The new method offers, for example, a 30 fold reduction in cross product computation time for the Black Smoke model dataset motivating Wood et al. ( 2017 ). Given this reduction in computational cost, the subsequent Cholesky decomposition of $$\mathbf{X}^{\mathsf{T}}{\mathbf{WX}}$$ X T WX and follow on computation of $$(\mathbf{X}^{\mathsf{T}}{\mathbf{WX}})^{-1}$$ ( X T WX ) - 1 become a more significant part of the computational burden, and we also discuss the choice of methods for improving their speed.

Keywords: cross product; model matrix; matrix; mathbf; model; computation

Journal Title: Statistics and Computing
Year Published: 2020

Link to full text (if available)


Share on Social Media:                               Sign Up to like & get
recommendations!

Related content

More Information              News              Social Media              Video              Recommended



                Click one of the above tabs to view related content.