A Comparative Study on the Performances of Q-Learning and Neural Q-Learning Agents toward Analysis of Emergence of Communication

Main Author: Takashi Sato and Fumiko Shirasaki, NIT, Okinawa College, Japan
Format: Article Journal
Bahasa: eng
Terbitan: , 2020
Subjects:
Online Access: https://zenodo.org/record/4309746
ctrlnum 4309746
fullrecord <?xml version="1.0"?> <dc schemaLocation="http://www.openarchives.org/OAI/2.0/oai_dc/ http://www.openarchives.org/OAI/2.0/oai_dc.xsd"><creator>Takashi Sato and Fumiko Shirasaki, NIT, Okinawa College, Japan</creator><date>2020-12-07</date><description>Abstract: In this paper, we suppose the gesture theory that is one theory on the origin of language, which tries to establish that speech originated from gestures. Based on the theory, we assume that &#x201C;actions&#x201D; having some purposes can be used as &#x201C;symbols&#x201D; in the communication through a learning process. The purpose of this study is to clarify what abilities of agents and what conditions are necessary to acquire usages of the actions as the symbols. To investigate them, we adopt a collision avoidance game and compare the performances of Q-learning agents with that of Neural Q-learning agents. In our simulation, we found that the Neural Q-learning agent&#x2019;s ability to reach the goal place is higher than the Q-learning agent&#x2019;s one. In contrast, the Neural Q-learning agent&#x2019;s ability to avoid collisions is lower than the Q-learning agent&#x2019;s one. If the inconsistencies in the learning data sets of the Neural Q-learning agent, however, can be resolved, the agent has enough potential to improve its ability for collision avoidance. Therefore, we conclude that the most suitable agent to analyze the emergence of communication is the Neural Q-learning agent who changed a feed forward type neural network into a recurrent type neural network that can resolve the inconsistencies in the learning data sets. Keywords: Q-learning, Neural Q-learning, Collision Avoidance Game, Reinforcement Learning Agents, Multi-Agent System.</description><description>Applied Science and Computer Science Publications</description><identifier>https://zenodo.org/record/4309746</identifier><identifier>10.5281/zenodo.4309746</identifier><identifier>oai:zenodo.org:4309746</identifier><language>eng</language><relation>doi:10.5281/zenodo.4309745</relation><rights>info:eu-repo/semantics/openAccess</rights><rights>https://creativecommons.org/licenses/by/4.0/legalcode</rights><source>Journal of Information and Communication Engineering(JICE) Volume 3(Issue 5) 128-135</source><subject>Applied Science and Computer Science Publications</subject><title>A Comparative Study on the Performances of Q-Learning and Neural Q-Learning Agents toward Analysis of Emergence of Communication</title><type>Journal:Article</type><type>Journal:Article</type><recordID>4309746</recordID></dc>
language eng
format Journal:Article
Journal
Journal:Journal
author Takashi Sato and Fumiko Shirasaki, NIT, Okinawa College, Japan
title A Comparative Study on the Performances of Q-Learning and Neural Q-Learning Agents toward Analysis of Emergence of Communication
publishDate 2020
topic Applied Science and Computer Science Publications
url https://zenodo.org/record/4309746
contents Abstract: In this paper, we suppose the gesture theory that is one theory on the origin of language, which tries to establish that speech originated from gestures. Based on the theory, we assume that “actions” having some purposes can be used as “symbols” in the communication through a learning process. The purpose of this study is to clarify what abilities of agents and what conditions are necessary to acquire usages of the actions as the symbols. To investigate them, we adopt a collision avoidance game and compare the performances of Q-learning agents with that of Neural Q-learning agents. In our simulation, we found that the Neural Q-learning agent’s ability to reach the goal place is higher than the Q-learning agent’s one. In contrast, the Neural Q-learning agent’s ability to avoid collisions is lower than the Q-learning agent’s one. If the inconsistencies in the learning data sets of the Neural Q-learning agent, however, can be resolved, the agent has enough potential to improve its ability for collision avoidance. Therefore, we conclude that the most suitable agent to analyze the emergence of communication is the Neural Q-learning agent who changed a feed forward type neural network into a recurrent type neural network that can resolve the inconsistencies in the learning data sets. Keywords: Q-learning, Neural Q-learning, Collision Avoidance Game, Reinforcement Learning Agents, Multi-Agent System.
Applied Science and Computer Science Publications
id IOS16997.4309746
institution ZAIN Publications
institution_id 7213
institution_type library:special
library
library Cognizance Journal of Multidisciplinary Studies
library_id 5267
collection Cognizance Journal of Multidisciplinary Studies
repository_id 16997
subject_area Multidisciplinary
city Stockholm
province INTERNASIONAL
shared_to_ipusnas_str 1
repoId IOS16997
first_indexed 2022-06-06T04:29:01Z
last_indexed 2022-06-06T04:29:01Z
recordtype dc
_version_ 1734902291074383872
score 17.608969