A parallel implementation of Q-learning based on communication with cache

Autores
Printista, Alicia Marcela; Errecalde, Marcelo Luis; Montoya, Cecilia Inés
Año de publicación
2002
Idioma
inglés
Tipo de recurso
artículo
Estado
versión publicada
Descripción
Q-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented.
Facultad de Informática
Materia
Ciencias Informáticas
Parallel programming
Redes de Comunicación de Computadores
Informática
Aprendizaje
communication based on cache
reinforcement learning
asynchronous dynamic programming
Nivel de accesibilidad
acceso abierto
Condiciones de uso
http://creativecommons.org/licenses/by-nc/3.0/
Repositorio
SEDICI (UNLP)
Institución
Universidad Nacional de La Plata
OAI Identificador
oai:sedici.unlp.edu.ar:10915/9432

id SEDICI_54db426bec83b3ebe0b705da882c3ec3
oai_identifier_str oai:sedici.unlp.edu.ar:10915/9432
network_acronym_str SEDICI
repository_id_str 1329
network_name_str SEDICI (UNLP)
spelling A parallel implementation of Q-learning based on communication with cachePrintista, Alicia MarcelaErrecalde, Marcelo LuisMontoya, Cecilia InésCiencias InformáticasParallel programmingRedes de Comunicación de ComputadoresInformáticaAprendizajecommunication based on cachereinforcement learningasynchronous dynamic programmingQ-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented.Facultad de Informática2002info:eu-repo/semantics/articleinfo:eu-repo/semantics/publishedVersionArticulohttp://purl.org/coar/resource_type/c_6501info:ar-repo/semantics/articuloapplication/pdfhttp://sedici.unlp.edu.ar/handle/10915/9432enginfo:eu-repo/semantics/altIdentifier/url/http://journal.info.unlp.edu.ar/wp-content/uploads/p41.pdfinfo:eu-repo/semantics/openAccesshttp://creativecommons.org/licenses/by-nc/3.0/Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0)reponame:SEDICI (UNLP)instname:Universidad Nacional de La Platainstacron:UNLP2025-09-29T10:50:40Zoai:sedici.unlp.edu.ar:10915/9432Institucionalhttp://sedici.unlp.edu.ar/Universidad públicaNo correspondehttp://sedici.unlp.edu.ar/oai/snrdalira@sedici.unlp.edu.arArgentinaNo correspondeNo correspondeNo correspondeopendoar:13292025-09-29 10:50:40.358SEDICI (UNLP) - Universidad Nacional de La Platafalse
dc.title.none.fl_str_mv A parallel implementation of Q-learning based on communication with cache
title A parallel implementation of Q-learning based on communication with cache
spellingShingle A parallel implementation of Q-learning based on communication with cache
Printista, Alicia Marcela
Ciencias Informáticas
Parallel programming
Redes de Comunicación de Computadores
Informática
Aprendizaje
communication based on cache
reinforcement learning
asynchronous dynamic programming
title_short A parallel implementation of Q-learning based on communication with cache
title_full A parallel implementation of Q-learning based on communication with cache
title_fullStr A parallel implementation of Q-learning based on communication with cache
title_full_unstemmed A parallel implementation of Q-learning based on communication with cache
title_sort A parallel implementation of Q-learning based on communication with cache
dc.creator.none.fl_str_mv Printista, Alicia Marcela
Errecalde, Marcelo Luis
Montoya, Cecilia Inés
author Printista, Alicia Marcela
author_facet Printista, Alicia Marcela
Errecalde, Marcelo Luis
Montoya, Cecilia Inés
author_role author
author2 Errecalde, Marcelo Luis
Montoya, Cecilia Inés
author2_role author
author
dc.subject.none.fl_str_mv Ciencias Informáticas
Parallel programming
Redes de Comunicación de Computadores
Informática
Aprendizaje
communication based on cache
reinforcement learning
asynchronous dynamic programming
topic Ciencias Informáticas
Parallel programming
Redes de Comunicación de Computadores
Informática
Aprendizaje
communication based on cache
reinforcement learning
asynchronous dynamic programming
dc.description.none.fl_txt_mv Q-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented.
Facultad de Informática
description Q-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented.
publishDate 2002
dc.date.none.fl_str_mv 2002
dc.type.none.fl_str_mv info:eu-repo/semantics/article
info:eu-repo/semantics/publishedVersion
Articulo
http://purl.org/coar/resource_type/c_6501
info:ar-repo/semantics/articulo
format article
status_str publishedVersion
dc.identifier.none.fl_str_mv http://sedici.unlp.edu.ar/handle/10915/9432
url http://sedici.unlp.edu.ar/handle/10915/9432
dc.language.none.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv info:eu-repo/semantics/altIdentifier/url/http://journal.info.unlp.edu.ar/wp-content/uploads/p41.pdf
dc.rights.none.fl_str_mv info:eu-repo/semantics/openAccess
http://creativecommons.org/licenses/by-nc/3.0/
Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0)
eu_rights_str_mv openAccess
rights_invalid_str_mv http://creativecommons.org/licenses/by-nc/3.0/
Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0)
dc.format.none.fl_str_mv application/pdf
dc.source.none.fl_str_mv reponame:SEDICI (UNLP)
instname:Universidad Nacional de La Plata
instacron:UNLP
reponame_str SEDICI (UNLP)
collection SEDICI (UNLP)
instname_str Universidad Nacional de La Plata
instacron_str UNLP
institution UNLP
repository.name.fl_str_mv SEDICI (UNLP) - Universidad Nacional de La Plata
repository.mail.fl_str_mv alira@sedici.unlp.edu.ar
_version_ 1844615757832388608
score 13.070432