A parallel implementation of Q-learning based on communication with cache
- Autores
- Printista, Alicia Marcela; Errecalde, Marcelo Luis; Montoya, Cecilia Inés
- Año de publicación
- 2002
- Idioma
- inglés
- Tipo de recurso
- artículo
- Estado
- versión publicada
- Descripción
- Q-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented.
Facultad de Informática - Materia
-
Ciencias Informáticas
Parallel programming
Redes de Comunicación de Computadores
Informática
Aprendizaje
communication based on cache
reinforcement learning
asynchronous dynamic programming - Nivel de accesibilidad
- acceso abierto
- Condiciones de uso
- http://creativecommons.org/licenses/by-nc/3.0/
- Repositorio
.jpg)
- Institución
- Universidad Nacional de La Plata
- OAI Identificador
- oai:sedici.unlp.edu.ar:10915/9432
Ver los metadatos del registro completo
| id |
SEDICI_54db426bec83b3ebe0b705da882c3ec3 |
|---|---|
| oai_identifier_str |
oai:sedici.unlp.edu.ar:10915/9432 |
| network_acronym_str |
SEDICI |
| repository_id_str |
1329 |
| network_name_str |
SEDICI (UNLP) |
| spelling |
A parallel implementation of Q-learning based on communication with cachePrintista, Alicia MarcelaErrecalde, Marcelo LuisMontoya, Cecilia InésCiencias InformáticasParallel programmingRedes de Comunicación de ComputadoresInformáticaAprendizajecommunication based on cachereinforcement learningasynchronous dynamic programmingQ-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented.Facultad de Informática2002info:eu-repo/semantics/articleinfo:eu-repo/semantics/publishedVersionArticulohttp://purl.org/coar/resource_type/c_6501info:ar-repo/semantics/articuloapplication/pdfhttp://sedici.unlp.edu.ar/handle/10915/9432enginfo:eu-repo/semantics/altIdentifier/url/http://journal.info.unlp.edu.ar/wp-content/uploads/p41.pdfinfo:eu-repo/semantics/openAccesshttp://creativecommons.org/licenses/by-nc/3.0/Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0)reponame:SEDICI (UNLP)instname:Universidad Nacional de La Platainstacron:UNLP2025-11-26T09:29:12Zoai:sedici.unlp.edu.ar:10915/9432Institucionalhttp://sedici.unlp.edu.ar/Universidad públicaNo correspondehttp://sedici.unlp.edu.ar/oai/snrdalira@sedici.unlp.edu.arArgentinaNo correspondeNo correspondeNo correspondeopendoar:13292025-11-26 09:29:12.688SEDICI (UNLP) - Universidad Nacional de La Platafalse |
| dc.title.none.fl_str_mv |
A parallel implementation of Q-learning based on communication with cache |
| title |
A parallel implementation of Q-learning based on communication with cache |
| spellingShingle |
A parallel implementation of Q-learning based on communication with cache Printista, Alicia Marcela Ciencias Informáticas Parallel programming Redes de Comunicación de Computadores Informática Aprendizaje communication based on cache reinforcement learning asynchronous dynamic programming |
| title_short |
A parallel implementation of Q-learning based on communication with cache |
| title_full |
A parallel implementation of Q-learning based on communication with cache |
| title_fullStr |
A parallel implementation of Q-learning based on communication with cache |
| title_full_unstemmed |
A parallel implementation of Q-learning based on communication with cache |
| title_sort |
A parallel implementation of Q-learning based on communication with cache |
| dc.creator.none.fl_str_mv |
Printista, Alicia Marcela Errecalde, Marcelo Luis Montoya, Cecilia Inés |
| author |
Printista, Alicia Marcela |
| author_facet |
Printista, Alicia Marcela Errecalde, Marcelo Luis Montoya, Cecilia Inés |
| author_role |
author |
| author2 |
Errecalde, Marcelo Luis Montoya, Cecilia Inés |
| author2_role |
author author |
| dc.subject.none.fl_str_mv |
Ciencias Informáticas Parallel programming Redes de Comunicación de Computadores Informática Aprendizaje communication based on cache reinforcement learning asynchronous dynamic programming |
| topic |
Ciencias Informáticas Parallel programming Redes de Comunicación de Computadores Informática Aprendizaje communication based on cache reinforcement learning asynchronous dynamic programming |
| dc.description.none.fl_txt_mv |
Q-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented. Facultad de Informática |
| description |
Q-Learning is a Reinforcement Learning method for solving sequential decision problems, where the utility of actions depends on a sequence of decisions and there exists uncertainty about the dynamics of the environment the agent is situated on. This general framework has allowed that Q-Learning and other Reinforcement Learning methods to be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others. Despite its interesting properties, Q-learning is a very slow method that requires a long period of training for learning an acceptable policy. In order to solve or at least reduce this problem, we propose a parallel implementation model of Q-learning using a tabular representation and via a communication scheme based on cache. This model is applied to a particular problem and the results obtained with different processor configurations are reported. A brief discussion about the properties and current limitations of our approach is finally presented. |
| publishDate |
2002 |
| dc.date.none.fl_str_mv |
2002 |
| dc.type.none.fl_str_mv |
info:eu-repo/semantics/article info:eu-repo/semantics/publishedVersion Articulo http://purl.org/coar/resource_type/c_6501 info:ar-repo/semantics/articulo |
| format |
article |
| status_str |
publishedVersion |
| dc.identifier.none.fl_str_mv |
http://sedici.unlp.edu.ar/handle/10915/9432 |
| url |
http://sedici.unlp.edu.ar/handle/10915/9432 |
| dc.language.none.fl_str_mv |
eng |
| language |
eng |
| dc.relation.none.fl_str_mv |
info:eu-repo/semantics/altIdentifier/url/http://journal.info.unlp.edu.ar/wp-content/uploads/p41.pdf |
| dc.rights.none.fl_str_mv |
info:eu-repo/semantics/openAccess http://creativecommons.org/licenses/by-nc/3.0/ Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0) |
| eu_rights_str_mv |
openAccess |
| rights_invalid_str_mv |
http://creativecommons.org/licenses/by-nc/3.0/ Creative Commons Attribution-NonCommercial 3.0 Unported (CC BY-NC 3.0) |
| dc.format.none.fl_str_mv |
application/pdf |
| dc.source.none.fl_str_mv |
reponame:SEDICI (UNLP) instname:Universidad Nacional de La Plata instacron:UNLP |
| reponame_str |
SEDICI (UNLP) |
| collection |
SEDICI (UNLP) |
| instname_str |
Universidad Nacional de La Plata |
| instacron_str |
UNLP |
| institution |
UNLP |
| repository.name.fl_str_mv |
SEDICI (UNLP) - Universidad Nacional de La Plata |
| repository.mail.fl_str_mv |
alira@sedici.unlp.edu.ar |
| _version_ |
1849875642378616832 |
| score |
13.011256 |