Ye-feng YANG, Kai DENG, Ying-qi ZUO, et al. Parameter design and optimization of a flight attitude simulator system based on PILCO framework[J]. Optics and precision engineering, 2019, 27(11): 2365-2373.
DOI:
Ye-feng YANG, Kai DENG, Ying-qi ZUO, et al. Parameter design and optimization of a flight attitude simulator system based on PILCO framework[J]. Optics and precision engineering, 2019, 27(11): 2365-2373. DOI: 10.3788/OPE.20192711.2365.
Parameter design and optimization of a flight attitude simulator system based on PILCO framework
Proportional-integral-derivative (PID) controllers are widely used in flight control systems. However
it is often very cumbersome to adjust the parameters of a PID controller. In this study
we use Probabilistic Inference for Learning Control (PILCO) to optimize the parameters of a PID controller. As the first step
we develop a probabilistic dynamics model of the flight control system using input and output data. Next
the existing PID controller is evaluated using the policy evaluation method. Finally
the evaluated PID controller is optimized by policy update. The sampling frequency of the system is 100 Hz and the data acquisition time per round is 8 s. The optimized PID controller can achieve stable control post 10 rounds of offline training. Through PILCO optimization
the flight attitude simulator performed robustly in a fixed-point experiment
indicating that PILCO has tremendous potential in solving nonlinear control and parameter optimization problems.
GAO J ZH, JIA H G. Design of longitudinal control law for small fixed-wing UAV during auto landing[J]. Opt. Precision Eng. , 2016, 24(7):1799-1806. (in Chinese)
LI D, CHEN X J, XU ZH J. Gain adaptive sliding mode controller used for flight attitude control of MAV[J]. Opt. Precision Eng. , 2013, 21(5):1183-1191. (in Chinese)
YIN H, YANG Y F, ZHAO Y, Self-tuning controller design for a 2-DOF flight attitude simulator[J]. Electric Machines and Control , 2018, 22(4):105-112. (in Chinese)
BUȘONIUA L, BRUINB T D, TOLIÇC D , et al .. Reinforcement learning for control: performance, stability, and deep approximators[J]. Annual Reviews in Control , 2018(5):1-18.
RECHT B. A tour of reinforcement learning: the view from continuous control[Z/OL]. https://arxiv.org/pdf/1806.09460.pdf https://arxiv.org/pdf/1806.09460.pdf .[2018-09-09].
DONG L, GUANG-HONG Y. Model-free adaptive control design for nonlinear discrete-time processes with reinforcement learning techniques[J]. International Journal of Systems Science , 2018, 49(11):2298-2308.
LEVINE S. Reinforcement learning and control as probabilistic inference: tutorial and review[Z/OL]. https://arxiv.org/abs/1805.00909 https://arxiv.org/abs/1805.00909 .[2018-05-20].
ZHANG T. Research on Path Planning Method of Quadrotor UAV Based on Reinforcement Learning [D]. Harbin: Harbin Institute of Technology, 2018.(in Chinese)
KAELBLING L P, LITTMAN M L, MOORE A W. Reinforcement learning: a survey[J]. Artificial Intelligence Research, 1996, 4(1):237-285.
CHUA K, CALANDRA R, MCALLISTER R, et al .. Deep reinforcement learning in a handful of trials using probabilistic dynamics models[Z/OL]. https://arxiv.org/abs/1805.12114 https://arxiv.org/abs/1805.12114 .[2018-11-02].
DEISENROTH M, RASMUSSEN C. PILCO: A model-based and data-efficient approach to policy search[C]. International Conference on International Conference on Machine Learning. Omnipress , 2011. https://www.researchgate.net/publication/221345233_PILCO_A_Model-Based_and_Data-Efficient_Approach_to_Policy_Search
RICHARD S, ANDREW G. Reinforcement Learning: An Introduction [M]. Second Edition. London: The MIT Press, 2016:78-88.
DURRANT-WHYTE H, ROY N, ABBEEL P. Learning to control a low-cost manipulator using data-efficient reinforcement learning[C]. Robotics : Science and Systems Ⅶ . MIT Press , 2011. https://www.researchgate.net/publication/221344493_Learning_to_Control_a_Low-Cost_Manipulator_using_Data-Efficient_Reinforcement_Learning?_sg=o9yGsYC6qIAsC1CWW6DUbSVs8zFaNclJWJx0ptwVgZOkwd1FKS62ir30zS5JczbLhVa3rj6mS6kPt7vOi3TcUw
DEISENROTH M P. Efficient Reinforcement Learning using Gaussian Processes [D]. Karlsruhe: Karlsruhe Institute of Technology, 2015.
Adaptive optimization control method for overexposure of industrial camera
Design and parameter optimization of frictionless cylinder based on aerostatic bearing
Active disturbance rejection controller of fast steering mirror based on improved dragonfly algorithm
Application of improved artificial bee colony algorithm on chatter suppression in three-dimensional elliptical vibration cutting
Parameter optimization and direction recognition in angle measurement by laser self-mixing interference
Related Author
ZHUANG Jian
ZHOU Jun
LI Junzhong
LIAO Xiaobo
WU Wenlin
Yong CAI
Xingzhan LI
Shunshun LI
Related Institution
School of Manufacturing Science and Engineering, Key Laboratory of Testing Technology for Manufacturing Process, Minsitry of Education, Southwest University of Science and Technology
School of Mechanical Engineering, Xi’ an Jiaotong University, Xi’ an
Institute of Manufacturing Technology, China Academy of Engineering Physics
School of Manufacturing Science and Engineering, Southwest University of Science and Technology
Hunan Provincial Key Laboratory of High Efficiency and Precision Machining of Difficult-to-Cut Material, Intelligent Manufacturing Institute, Hunan University of Science and Technology