NEGOTIATION AGENT BEHAVIORS BASED ON REINFORCEMENT LEARNING APPROACHES AND FUZZY ARTMAP NEURAL NETWORKS
DOI:
https://doi.org/10.47839/ijc.7.3.532Keywords:
Intelligent behaviors, decision-making, reinforcement learning, fuzzy artmap neural network, field programmable gate arrayAbstract
Behaviors, in which the characters conciliatory, neutral, or aggressive define a ‘psychological’ aspect of human personality, play an important role for negotiation agent. Elsewhere, learning in negotiation is fundamental for understanding human behaviors and developing new concepts. In this paper, a negotiation strategy essentially based on such human personality behaviors is suggested for SISINE project which aims to develop innovative teaching methodology of negotiation skills. For this purpose, first, reinforcement learning (Q-learning and Sarsa-Learning) approaches are developed, analyzed, and compared in order to acquire the strategy negotiation behaviors. Second, a Fuzzy ArtMap Neural Network (FAMNN) is developed to acquire this strategy. Third, a Field Programmable Gate Array (FPGA) architecture is suggested for the FAMNN integration. The suggested strategy displays the ability to provide agents, through a basic buying strategy, with a first intelligence level in a social and cognitive system for learning negotiation strategies (human-agent and agent-agent).References
M. Wooldridge, An Introduction to MultiAgent Systems. John Wiley & Sons, England, 2002.
R. F. Bales, Interaction Process Analysis: A Method for the Study of Small Groups. Cambridge, Massach.: Addisson-Wesley, 1950.
J. Z. Rubin and B. R. Brown, The Social Psychology of Bargaining and Negotiation. NY Acad. Press, pp. 233-262, 1975.
D. Silver, R. Sutton, and M. Muller, Reinforcement Learning of Local Shape in the Game of Go, Int. Joint Conference on Artificial Intelligence, 2007, pp. 1053-1058.
O. Miglino, A. Di Ferdinando, A. Rega, and B. Benincasa, SISINE: Teaching negotiation through a multiplayer online role playing game, The 6th European Conference on E-Learning, Danemark, Oct. 04-05, 2007.
Orazio Miglino, The SISINE Project: Developing and E-Learning Platform for Educational Role-Playing Games, Technology-Enhanced Learning, European Research Consortium for Informatics and Mathematics (71) (2007), pp. 28-29.
R. Pfeifer and C. Scheier. Understanding Intelligence. MIT press, 1999.
A. Chohra, Embodied Cognitive Science, Intelligent Behavior Control, Machine Learning, Soft Computing, and FPGA Integration: Towards Fast, Cooperative and Adversarial Robot Team (RoboCup). Technical GMD Report, No. 136, Germany, June 2001.
D. Zeng and K. Sycara, Benefits of learning in negotiation, Proc. of the 14th National Conference on Artificial Intelligence (AAAI-97), Providence, RI, July 1997.
C. J. C. H. Watkins. Learning from Delayed Rewards. Ph.D Thesis, King’s College, 1989.
S. D. Whitehead. Reinforcement Learning for the Adaptive Control of Perception and Action. Technical Report 406, Rochester University, 1992.
R. S. Sutton and A. G. Barto. Reinforcement Learning. MIT press, 1998.
T. G. Dietterich, Hierarchical reinforcement learning with the MAXQ value function decomposition. Journal of Artificial Intelligence Research 13 (2000), pp. 227-303.
J. A. Anderson, An Introduction to Neural Networks, MIT Press, England, 1995.
D. W. Patterson, Artificial Neural Networks: Theory and Applications, Prentice-Hall, Simon & Schuster, Singapore, 1996.
S. Haykin, Neural Networks: A Comprehensive Foundation. 2nd Ed., Prentice-Hall, 1999.
Amine Chohra, Arash Bahrammirzaee, Kurosh Madani, Negotiation Agent Behaviors Based On Reinforcement Learning Approaches and Fuzzy ArtMap Neural Networks, Fifth Int. Conf. on Neural Networks and Artificial Intelligence (ICNNAI 2008), Minsk, Belarus, May 27-30, 2008, pp. 56-61.
T. W. Sandholm, Distributed Rational Decision Making. In Multiagent Systems: A Modern Introduction to Distributed Artificial Intelligence, MIT Press, pp. 201-258, 1999.
L.-J. Lin, Self-improving reactive agents based on reinforcement learning, planning and teaching, Machine Learning, 8 (1992) pp. 293-321.
C. F. Touzet, Neural reinforcement learning for behaviour synthesis, Robotics and Autonomous Sys. 22 (1997), pp. 251-281.
M. Coggan, “Exploration and Exploitation in Reinforcement Learning”, Research supervised by Prof. Doina Precup, CRA-W DMP Project at McGill University, 2004.
K. Takadama and H. Fujita, Lessons learned from comparison between Q-learning and Sarsa agents in bargaining game, North American Asso. for Computational Social and Organizational Science Conference 2004.
G. A. Carpenter, S. Grossberg and D. B. Rosen, Fuzzy ART: Fast stable learning and categorization of analog patterns by an adaptive resonance system, Neural Networks 4 (1991), pp. 759–771.
R.K. Awalt, Making ASIC/FPGA design, Integrated System Design (1999) pp. 22-28.
O. Azouaoui, A. Chohra, Soft computing based pattern classifiers for the obstacle avoidance behavior of Intelligent Autonomous Vehicles (IAV), Int. Journal of Applied Intelligence 16 (3) (2002), pp. 249-271.
Downloads
Published
How to Cite
Issue
Section
License
International Journal of Computing is an open access journal. Authors who publish with this journal agree to the following terms:• Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
• Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
• Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work.