A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning

Autores
Barsce, Juan Cruz; Palombarini, Jorge; Martínez, Ernesto
Año de publicación
2019
Idioma
inglés
Tipo de recurso
documento de conferencia
Estado
versión publicada
Descripción
Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning.
Sociedad Argentina de Informática e Investigación Operativa
Materia
Ciencias Informáticas
Reinforcement learning
Hyper-parameter optimization
Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS)
Nivel de accesibilidad
acceso abierto
Condiciones de uso
http://creativecommons.org/licenses/by-nc-sa/3.0/
Repositorio
SEDICI (UNLP)
Institución
Universidad Nacional de La Plata
OAI Identificador
oai:sedici.unlp.edu.ar:10915/87851

id SEDICI_9605a4a9f5db5b2d72c4e06c19459149
oai_identifier_str oai:sedici.unlp.edu.ar:10915/87851
network_acronym_str SEDICI
repository_id_str 1329
network_name_str SEDICI (UNLP)
spelling A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement LearningBarsce, Juan CruzPalombarini, JorgeMartínez, ErnestoCiencias InformáticasReinforcement learningHyper-parameter optimizationBayesian optimization, Bayesian optimization of combinatorial structures (BOCS)Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning.Sociedad Argentina de Informática e Investigación Operativa2019-09info:eu-repo/semantics/conferenceObjectinfo:eu-repo/semantics/publishedVersionObjeto de conferenciahttp://purl.org/coar/resource_type/c_5794info:ar-repo/semantics/documentoDeConferenciaapplication/pdf32-38http://sedici.unlp.edu.ar/handle/10915/87851enginfo:eu-repo/semantics/altIdentifier/issn/2451-7585info:eu-repo/semantics/reference/hdl/10915/135049info:eu-repo/semantics/openAccesshttp://creativecommons.org/licenses/by-nc-sa/3.0/Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0)reponame:SEDICI (UNLP)instname:Universidad Nacional de La Platainstacron:UNLP2025-09-17T10:00:13Zoai:sedici.unlp.edu.ar:10915/87851Institucionalhttp://sedici.unlp.edu.ar/Universidad públicaNo correspondehttp://sedici.unlp.edu.ar/oai/snrdalira@sedici.unlp.edu.arArgentinaNo correspondeNo correspondeNo correspondeopendoar:13292025-09-17 10:00:13.924SEDICI (UNLP) - Universidad Nacional de La Platafalse
dc.title.none.fl_str_mv A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
title A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
spellingShingle A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
Barsce, Juan Cruz
Ciencias Informáticas
Reinforcement learning
Hyper-parameter optimization
Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS)
title_short A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
title_full A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
title_fullStr A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
title_full_unstemmed A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
title_sort A Hierarchical Two-tier Approach to Hyper-parameter Optimization in Reinforcement Learning
dc.creator.none.fl_str_mv Barsce, Juan Cruz
Palombarini, Jorge
Martínez, Ernesto
author Barsce, Juan Cruz
author_facet Barsce, Juan Cruz
Palombarini, Jorge
Martínez, Ernesto
author_role author
author2 Palombarini, Jorge
Martínez, Ernesto
author2_role author
author
dc.subject.none.fl_str_mv Ciencias Informáticas
Reinforcement learning
Hyper-parameter optimization
Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS)
topic Ciencias Informáticas
Reinforcement learning
Hyper-parameter optimization
Bayesian optimization, Bayesian optimization of combinatorial structures (BOCS)
dc.description.none.fl_txt_mv Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning.
Sociedad Argentina de Informática e Investigación Operativa
description Optimization of hyper-parameters in reinforcement learning (RL) algorithms is a key task, because they determine how the agent will learn its policy by interacting with its environment, and thus what data is gathered. In this work, an approach that uses Bayesian optimization to perform a two-step optimization is proposed: rst, categorical RL structure hyper-parameters are taken as binary variables and optimized with an acquisition function tailored for such variables. Then, at a lower level of abstraction, solution-level hyper-parameters are optimized by resorting to the expected improvement acquisition function, while using the best categorical hyper-parameters found in the optimization at the upper-level of abstraction. This two-tier approach is validated in a simulated control task. Results obtained are promising and open the way for more user-independent applications of reinforcement learning.
publishDate 2019
dc.date.none.fl_str_mv 2019-09
dc.type.none.fl_str_mv info:eu-repo/semantics/conferenceObject
info:eu-repo/semantics/publishedVersion
Objeto de conferencia
http://purl.org/coar/resource_type/c_5794
info:ar-repo/semantics/documentoDeConferencia
format conferenceObject
status_str publishedVersion
dc.identifier.none.fl_str_mv http://sedici.unlp.edu.ar/handle/10915/87851
url http://sedici.unlp.edu.ar/handle/10915/87851
dc.language.none.fl_str_mv eng
language eng
dc.relation.none.fl_str_mv info:eu-repo/semantics/altIdentifier/issn/2451-7585
info:eu-repo/semantics/reference/hdl/10915/135049
dc.rights.none.fl_str_mv info:eu-repo/semantics/openAccess
http://creativecommons.org/licenses/by-nc-sa/3.0/
Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0)
eu_rights_str_mv openAccess
rights_invalid_str_mv http://creativecommons.org/licenses/by-nc-sa/3.0/
Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0)
dc.format.none.fl_str_mv application/pdf
32-38
dc.source.none.fl_str_mv reponame:SEDICI (UNLP)
instname:Universidad Nacional de La Plata
instacron:UNLP
reponame_str SEDICI (UNLP)
collection SEDICI (UNLP)
instname_str Universidad Nacional de La Plata
instacron_str UNLP
institution UNLP
repository.name.fl_str_mv SEDICI (UNLP) - Universidad Nacional de La Plata
repository.mail.fl_str_mv alira@sedici.unlp.edu.ar
_version_ 1843532492511903744
score 13.000565