An Intelligent System for Parking Trailer in Presence of Fixed and Moving Obstacles Using Reinforcement Learning and Fuzzy Logic
-
1975
Downloads
-
3478
Views
Authors
M. Sharafi
- Islamic Azad University, Gonabad Branch
A. Zare
- Assistant Professor of Islamic Azad University, Gonabad Branch
A. V. Kamyad
- Full professor, Department of Mathematics, Ferdowsi University of Mashhad
Abstract
In examples of reinforcement learning where state space is continuous, it seems impossible to use reference tables to store value-action .In these problems a method is required for value estimation for each state-action pair .The inputs to this estimation system are (characteristics of)
state variables which reflect the status of agent in the environment .The system can be either linear of nonlinear .For each member in set of actions of an agent, there exists an estimation system which determines state value for the action .
On the other hand, in most real world problems, just as the state space is continuous, so is the action space for an agent .In these cases, fuzzy systems may provide a useful solution in selection of final action from action space .In this paper we intend to combine reinforcement learning algorithm with fuzzified actions and state space along with a linear estimation system into an intelligent systems for parking Trailers in cases where both state and action spaces are continuous .Finally, the successful performance of the proposed algorithm is shown through simulations on trailer parking problem .
Share and Cite
ISRP Style
M. Sharafi, A. Zare, A. V. Kamyad, An Intelligent System for Parking Trailer in Presence of Fixed and Moving Obstacles Using Reinforcement Learning and Fuzzy Logic, Journal of Mathematics and Computer Science, 2 (2011), no. 1, 141--149
AMA Style
Sharafi M., Zare A., Kamyad A. V., An Intelligent System for Parking Trailer in Presence of Fixed and Moving Obstacles Using Reinforcement Learning and Fuzzy Logic. J Math Comput SCI-JM. (2011); 2(1):141--149
Chicago/Turabian Style
Sharafi, M., Zare, A., Kamyad, A. V.. "An Intelligent System for Parking Trailer in Presence of Fixed and Moving Obstacles Using Reinforcement Learning and Fuzzy Logic." Journal of Mathematics and Computer Science, 2, no. 1 (2011): 141--149
Keywords
- Reinforcement Learning
- Fuzzy Systems
- Trailer Parking Problem
- SARSA Algorithm.
MSC
References
-
[1]
R. S. Sutton, A. G. Barto, Reinforcement Learning: An Introduction, M.I.T. Press, Cambridge (1998)
-
[2]
L. P. Kaelbling, M. L. Littman, A. W. Moore, Reinforcement learning: A survey, J. Artif. Intell. Res., 4 (1996), 237--287
-
[3]
D. P. Bertsekas, J. N. Tsitsiklis, Neuro-Dynamic Programming, Athena Scientific, Nashua (1996)
-
[4]
R. S. Sutton, Learning to predict by the methods of temporal difference, Mach. Learn., 3 (1988), 9--44
-
[5]
C. J. C. H. Watkins, P. Dayan, Q-learning, Mach. Learn., 8 (1992), 279--292
-
[6]
D. Vengerov, N. Bambos, H. R. Berenji, A fuzzy reinforcement learning approach to power control in wireless transmitters, IEEE Trans.Syst., Man, Cybern. B, Cybern., 35 (2005), 768--778
-
[7]
H. R. Beom, H. S. Cho, A sensor-based navigation for a mobile robot using fuzzy logic and reinforcement learning, IEEE Trans. Syst.,Man, Cybern., 25 (1995), 464--477
-
[8]
C. I. Connolly, Harmonic functions and collision probabilities, Int. J.Rob. Res., 16 (1997), 497--507
-
[9]
W. D. Smart, L. P. Kaelbling, Effective reinforcement learning for mobile robots, Proc. IEEE Int. Conf. Robot. Autom., 2002 (2002), 3404--3410
-
[10]
T. Kondo, K. Ito, A reinforcement learning with evolutionary state recruitment strategy for autonomous mobile robots control, Robot. Auton. Syst., 46 (2004), 111--124
-
[11]
M. Wiering, J. Schmidhuber, HQ-learning, Adapt. Behav., 6 (1997), 219--246
-
[12]
A. G. Barto, S. Mahadevan, Recent advances in hierarchical reinforcement learning, Discrete Event Dyn. Syst., 13 (2003), 41--77
-
[13]
R. S. Sutton, D. Precup, S. Singh, Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning, Artif. Intell., 112 (1999), 181--211
-
[14]
T. G. Dietterich, Hierarchical reinforcement learning with the MAXQ value function decomposition, J. Artificial Intelligence Res., 13 (2000), 227--303
-
[15]
G. Theocharous, Hierarchical learning and planning in partially observable Markov decision processes, Ph.D. dissertation (Michigan State Univ.), East Lansing (2002)
-
[16]
A. J. Smith, Applications of the self-organising map to reinforcement learning, Neural Netw., 15 (2002), 1107--1124
-
[17]
P. Y. Glorennec, L. Jouffe, Fuzzy Q-learning, Proc. 6th IEEE Int. Conf. Fuzzy Syst., 1997 (1997), 659--662
-
[18]
S. G. Tzafestas, G. G. Rigatos, Fuzzy reinforcement learning control for compliance tasks of robotic manipulators, IEEE Trans. Syst.,Man, Cybern. B, Cybern., 32 (2002), 107--113
-
[19]
M. J. Er, C. Deng, Online tuning of fuzzy inference systems using dynamic fuzzy Q-learning, IEEE Trans. Syst., Man, Cybern. B,Cybern., 34 (2004), 1478--1489
-
[20]
C. Chen, H. X. Li, D. Dong, Hybrid control for Robot Navigation: A hierarchical Q-learning algorithm, IEEE Robot. Autom. Mag., 15 (2008), 37--47
-
[21]
S. Whiteson, P. Stone, Evolutionary function approximation for reinforcement learning, J. Mach. Learn. Res., 7 (2006), 877--917
-
[22]
M. Kaya, R. Alhajj, A novel approach to multiagent reinforcement learning: Utilizing OLAP mining in the learning process, IEEE Trans.Syst., Man, Cybern. C, Appl. Rev., 35 (2005), 582--590
-
[23]
J. S. R. Jang, C. T. Sun, E. Mizutani, Neuro-Fuzzy and Soft Computing: a computational approach to learning and machine intelligence, Prentice-Hall, Englewood Cliffs (1997)
-
[24]
L.-X. Wang , A Course in fuzzy system and control, Prentice Hall, Upper Saddle River (1997)
-
[25]
Z. Moeini, V. Seyyedi Ghomshe, M. Teshne-lab, , 10th Iranian Fuzzy System Conference, (), 221--226