SwePub
Sök i LIBRIS databas

  Utökad sökning

L773:2162 237X
 

Sökning: L773:2162 237X > Perceptron Theory C...

Perceptron Theory Can Predict the Accuracy of Neural Networks

Kleyko, Denis (författare)
RISE,Datavetenskap,University of California at Berkeley, USA
Rosato, Antonello (författare)
University of Rome “La Sapienza”, Italy
Frady, Edward Paxon (författare)
Intel Labs, USA
visa fler...
Panella, Massimo (författare)
University of Rome “La Sapienza”, Italy
Sommer, Friedrich T. (författare)
Intel Labs, USA; University of California at Berkeley, USA
visa färre...
 (creator_code:org_t)
2023
2023
Engelska.
Ingår i: IEEE Transactions on Neural Networks and Learning Systems. - 2162-237X .- 2162-2388.
  • Tidskriftsartikel (refereegranskat)
Abstract Ämnesord
Stäng  
  • Multilayer neural networks set the current state of the art for many technical classification problems. But, these networks are still, essentially, black boxes in terms of analyzing them and predicting their performance. Here, we develop a statistical theory for the one-layer perceptron and show that it can predict performances of a surprisingly large variety of neural networks with different architectures. A general theory of classification with perceptrons is developed by generalizing an existing theory for analyzing reservoir computing models and connectionist models for symbolic reasoning known as vector symbolic architectures. Our statistical theory offers three formulas leveraging the signal statistics with increasing detail. The formulas are analytically intractable, but can be evaluated numerically. The description level that captures maximum details requires stochastic sampling methods. Depending on the network model, the simpler formulas already yield high prediction accuracy. The quality of the theory predictions is assessed in three experimental settings, a memorization task for echo state networks (ESNs) from reservoir computing literature, a collection of classification datasets for shallow randomly connected networks, and the ImageNet dataset for deep convolutional neural networks. We find that the second description level of the perceptron theory can predict the performance of types of ESNs, which could not be described previously. Furthermore, the theory can predict deep multilayer neural networks by being applied to their output layer. While other methods for prediction of neural networks performance commonly require to train an estimator model, the proposed theory requires only the first two moments of the distribution of the postsynaptic sums in the output neurons. Moreover, the perceptron theory compares favorably to other methods that do not rely on training an estimator model.

Ämnesord

TEKNIK OCH TEKNOLOGIER  -- Elektroteknik och elektronik (hsv//swe)
ENGINEERING AND TECHNOLOGY  -- Electrical Engineering, Electronic Engineering, Information Engineering (hsv//eng)

Publikations- och innehållstyp

ref (ämneskategori)
art (ämneskategori)

Hitta via bibliotek

Till lärosätets databas

Hitta mer i SwePub

Av författaren/redakt...
Kleyko, Denis
Rosato, Antonell ...
Frady, Edward Pa ...
Panella, Massimo
Sommer, Friedric ...
Om ämnet
TEKNIK OCH TEKNOLOGIER
TEKNIK OCH TEKNO ...
och Elektroteknik oc ...
Artiklar i publikationen
IEEE Transaction ...
Av lärosätet
RISE

Sök utanför SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy