A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
- Autores
- Barsce, Juan Cruz; Palombarini, Jorge; Martínez, Ernesto
- Año de publicación
- 2019
- Idioma
- inglés
- Tipo de recurso
- documento de conferencia
- Estado
- versión publicada
- Descripción
- Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning.
Sociedad Argentina de Informática e Investigación Operativa - Materia
-
Ciencias Informáticas
Reinforcement learning
Hyper-parameter optimization
Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS) - Nivel de accesibilidad
- acceso abierto
- Condiciones de uso
- http://creativecommons.org/licenses/by-nc-sa/3.0/
- Repositorio
- Institución
- Universidad Nacional de La Plata
- OAI Identificador
- oai:sedici.unlp.edu.ar:10915/87851
Ver los metadatos del registro completo
id |
SEDICI_9605a4a9f5db5b2d72c4e06c19459149 |
---|---|
oai_identifier_str |
oai:sedici.unlp.edu.ar:10915/87851 |
network_acronym_str |
SEDICI |
repository_id_str |
1329 |
network_name_str |
SEDICI (UNLP) |
spelling |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement LearningBarsce, Juan CruzPalombarini, JorgeMartínez, ErnestoCiencias InformáticasReinforcement learningHyper-parameter optimizationBayesian optimization, Bayesian optimization of combinatorial structures (BOCS)Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning.Sociedad Argentina de Informática e Investigación Operativa2019-09info:eu-repo/semantics/conferenceObjectinfo:eu-repo/semantics/publishedVersionObjeto de conferenciahttp://purl.org/coar/resource_type/c_5794info:ar-repo/semantics/documentoDeConferenciaapplication/pdf32-38http://sedici.unlp.edu.ar/handle/10915/87851enginfo:eu-repo/semantics/altIdentifier/issn/2451-7585info:eu-repo/semantics/reference/hdl/10915/135049info:eu-repo/semantics/openAccesshttp://creativecommons.org/licenses/by-nc-sa/3.0/Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0)reponame:SEDICI (UNLP)instname:Universidad Nacional de La Platainstacron:UNLP2025-09-17T10:00:13Zoai:sedici.unlp.edu.ar:10915/87851Institucionalhttp://sedici.unlp.edu.ar/Universidad públicaNo correspondehttp://sedici.unlp.edu.ar/oai/snrdalira@sedici.unlp.edu.arArgentinaNo correspondeNo correspondeNo correspondeopendoar:13292025-09-17 10:00:13.924SEDICI (UNLP) - Universidad Nacional de La Platafalse |
dc.title.none.fl_str_mv |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
title |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
spellingShingle |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning Barsce, Juan Cruz Ciencias Informáticas Reinforcement learning Hyper-parameter optimization Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS) |
title_short |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
title_full |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
title_fullStr |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
title_full_unstemmed |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
title_sort |
A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning |
dc.creator.none.fl_str_mv |
Barsce, Juan Cruz Palombarini, Jorge Martínez, Ernesto |
author |
Barsce, Juan Cruz |
author_facet |
Barsce, Juan Cruz Palombarini, Jorge Martínez, Ernesto |
author_role |
author |
author2 |
Palombarini, Jorge Martínez, Ernesto |
author2_role |
author author |
dc.subject.none.fl_str_mv |
Ciencias Informáticas Reinforcement learning Hyper-parameter optimization Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS) |
topic |
Ciencias Informáticas Reinforcement learning Hyper-parameter optimization Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS) |
dc.description.none.fl_txt_mv |
Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning. Sociedad Argentina de Informática e Investigación Operativa |
description |
Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning. |
publishDate |
2019 |
dc.date.none.fl_str_mv |
2019-09 |
dc.type.none.fl_str_mv |
info:eu-repo/semantics/conferenceObject info:eu-repo/semantics/publishedVersion Objeto de conferencia http://purl.org/coar/resource_type/c_5794 info:ar-repo/semantics/documentoDeConferencia |
format |
conferenceObject |
status_str |
publishedVersion |
dc.identifier.none.fl_str_mv |
http://sedici.unlp.edu.ar/handle/10915/87851 |
url |
http://sedici.unlp.edu.ar/handle/10915/87851 |
dc.language.none.fl_str_mv |
eng |
language |
eng |
dc.relation.none.fl_str_mv |
info:eu-repo/semantics/altIdentifier/issn/2451-7585 info:eu-repo/semantics/reference/hdl/10915/135049 |
dc.rights.none.fl_str_mv |
info:eu-repo/semantics/openAccess http://creativecommons.org/licenses/by-nc-sa/3.0/ Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0) |
eu_rights_str_mv |
openAccess |
rights_invalid_str_mv |
http://creativecommons.org/licenses/by-nc-sa/3.0/ Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0) |
dc.format.none.fl_str_mv |
application/pdf 32-38 |
dc.source.none.fl_str_mv |
reponame:SEDICI (UNLP) instname:Universidad Nacional de La Plata instacron:UNLP |
reponame_str |
SEDICI (UNLP) |
collection |
SEDICI (UNLP) |
instname_str |
Universidad Nacional de La Plata |
instacron_str |
UNLP |
institution |
UNLP |
repository.name.fl_str_mv |
SEDICI (UNLP) - Universidad Nacional de La Plata |
repository.mail.fl_str_mv |
alira@sedici.unlp.edu.ar |
_version_ |
1843532492511903744 |
score |
13.000565 |