SwePub
Sök i LIBRIS databas

  Utökad sökning

WFRF:(Allerbo Oskar 1985)
 

Sökning: WFRF:(Allerbo Oskar 1985) > Efficient training ...

Efficient training of interpretable, non-linear regression models

Allerbo, Oskar, 1985 (författare)
Gothenburg University,Göteborgs universitet,Institutionen för matematiska vetenskaper, Tillämpad matematik och statistik,Department of Mathematical Sciences, Applied Mathematics and Statistics
 (creator_code:org_t)
ISBN 9789180693370
2023
Engelska.
  • Doktorsavhandling (övrigt vetenskapligt/konstnärligt)
Abstract Ämnesord
Stäng  
  • Regression, the process of estimating functions from data, comes in many flavors. One of the most commonly used regression models is linear regression, which is computationally efficient and easy to interpret, but lacks in flexibility. Non-linear regression methods, such as kernel regression and artificial neural networks, tend to be much more flexible, but also harder to interpret and more difficult, and computationally heavy, to train. In the five papers of this thesis, different techniques for constructing regression models that combine flexibility with interpretability and computational efficiency, are investigated. In Papers I and II, sparsely regularized neural networks are used to obtain flexible, yet interpretable, models for additive modeling (Paper I) and dimensionality reduction (Paper II). Sparse regression, in the form of the elastic net, is also covered in Paper III, where the focus is on increased computational efficiency by replacing explicit regularization with iterative optimization and early stopping. In Paper IV, inspired by Jacobian regularization, we propose a computationally efficient method for bandwidth selection for kernel regression with the Gaussian kernel. Kernel regression is also the topic of Paper V, where we revisit efficient regularization through early stopping, by solving kernel regression iteratively. Using an iterative algorithm for kernel regression also enables changing the kernel during training, which we use to obtain a more flexible method, resembling the behavior of neural networks. In all five papers, the results are obtained by carefully selecting either the regularization strength or the bandwidth. Thus, in summary, this work contributes with new statistical methods for combining flexibility with interpretability and computational efficiency based on intelligent hyperparameter selection.

Ämnesord

NATURVETENSKAP  -- Matematik (hsv//swe)
NATURAL SCIENCES  -- Mathematics (hsv//eng)

Nyckelord

sparse regression
kernel regression
neural network regression
early stopping
bandwidth selection

Publikations- och innehållstyp

vet (ämneskategori)
dok (ämneskategori)

Hitta via bibliotek

Till lärosätets databas

Hitta mer i SwePub

Av författaren/redakt...
Allerbo, Oskar, ...
Om ämnet
NATURVETENSKAP
NATURVETENSKAP
och Matematik
Av lärosätet
Göteborgs universitet

Sök utanför SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy