Sökning: id:"swepub:oai:DiVA.org:kth-326404" >
How are policy grad...
How are policy gradient methods affected by the limits of control?
-
- Ziemann, Ingvar (författare)
- KTH,Reglerteknik
-
- Tsiamis, Anastasios (författare)
- Swiss Fed Inst Technol, Automat Control Lab, Zurich, Switzerland.
-
- Sandberg, Henrik (författare)
- KTH,Reglerteknik
-
visa fler...
-
- Matni, Nikolai (författare)
- Univ Penn, Dept Elect & Syst Engn, Philadelphia, PA 19104 USA.
-
visa färre...
-
(creator_code:org_t)
- Institute of Electrical and Electronics Engineers (IEEE), 2022
- 2022
- Engelska.
-
Ingår i: 2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC). - : Institute of Electrical and Electronics Engineers (IEEE). ; , s. 5992-5999
- Relaterad länk:
-
https://urn.kb.se/re...
-
visa fler...
-
https://doi.org/10.1...
-
visa färre...
Abstract
Ämnesord
Stäng
- We study stochastic policy gradient methods from the perspective of control-theoretic limitations. Our main result is that ill-conditioned linear systems in the sense of Doyle inevitably lead to noisy gradient estimates. We also give an example of a class of stable systems in which policy gradient methods suffer from the curse of dimensionality. Finally, we show how our results extend to partially observed systems.
Ämnesord
- TEKNIK OCH TEKNOLOGIER -- Elektroteknik och elektronik -- Reglerteknik (hsv//swe)
- ENGINEERING AND TECHNOLOGY -- Electrical Engineering, Electronic Engineering, Information Engineering -- Control Engineering (hsv//eng)
Publikations- och innehållstyp
- ref (ämneskategori)
- kon (ämneskategori)