Efficient Q-learning modification aplied on active magnetic bearing control
The result's identifiers
Result code in IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F00216305%3A26210%2F04%3APU45638" target="_blank" >RIV/00216305:26210/04:PU45638 - isvavai.cz</a>
Result on the web
—
DOI - Digital Object Identifier
—
Alternative languages
Result language
angličtina
Original language name
Efficient Q-learning modification aplied on active magnetic bearing control
Original language description
The paper is focused on use of Q-learning for active magnetic bearing (AMB) control. Q-learning belongs to the reinforcement learning methods which are the part of real time machine learning approaches. The essence of proposed method is in separating theQ-learning into two phases ? prelearning phase, which use mathematical model of real system and tutorage phase, which works with the real system and is used for further improvement of Q-values found during the prelearning phase. Proposed learning methodd, used in prelearning phase, proved to be highly efficient. Controller based on Q-learning show better results (regarding the number of successful trials) than referential PID controller after only 1000 table passes. The control quality criterion results are comparable. The policy found by learning also shows high robustness against errors of system variables observations, even only very simple reinforcement function in shape of simple reduced penalty is used.
Czech name
Efficient Q-learning modification aplied on active magnetic bearing control
Czech description
The paper is focused on use of Q-learning for active magnetic bearing (AMB) control. Q-learning belongs to the reinforcement learning methods which are the part of real time machine learning approaches. The essence of proposed method is in separating theQ-learning into two phases ? prelearning phase, which use mathematical model of real system and tutorage phase, which works with the real system and is used for further improvement of Q-values found during the prelearning phase. Proposed learning methodd, used in prelearning phase, proved to be highly efficient. Controller based on Q-learning show better results (regarding the number of successful trials) than referential PID controller after only 1000 table passes. The control quality criterion results are comparable. The policy found by learning also shows high robustness against errors of system variables observations, even only very simple reinforcement function in shape of simple reduced penalty is used.
Classification
Type
J<sub>x</sub> - Unclassified - Peer-reviewed scientific article (Jimp, Jsc and Jost)
CEP classification
JD - Use of computers, robotics and its application
OECD FORD branch
—
Result continuities
Project
—
Continuities
Z - Vyzkumny zamer (s odkazem do CEZ)
Others
Publication year
2004
Confidentiality
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Data specific for result type
Name of the periodical
Inženýrská mechanika - Engineering Mechanics
ISSN
1210-2717
e-ISSN
—
Volume of the periodical
11/2004
Issue of the periodical within the volume
2
Country of publishing house
CZ - CZECH REPUBLIC
Number of pages
14
Pages from-to
83-96
UT code for WoS article
—
EID of the result in the Scopus database
—