Closed-loop Rescheduling using Deep Reinforcement Learning
- Autores
- Palombarini, Jorge A.; Martínez, Ernesto C.
- Año de publicación
- 2019
- Idioma
- inglés
- Tipo de recurso
- documento de conferencia
- Estado
- versión publicada
- Descripción
- In this work, a novel approach for generating rescheduling knowledge which can be used in real-time for handling unforeseen events without extra deliberation is presented. For generating such control knowledge, the rescheduling task is modelled and solved as a closed-loop control problem by resorting to the integration of a schedule state simulator with a rescheduling agent that can learn successful schedule repairing policies directly from a variety of simulated transitions between schedule states, using as input readily available schedule color-rich Gantt chart images, and negligible prior knowledge. The generated knowledge is stored in a deep Q-network, which can be used as a computational tool in a closed-loop rescheduling control way that select repair actions to make progress towards a goal schedule state, without requiring to compute the rescheduling problem solution every time a disruptive event occurs and safely generalize control knowledge to unseen schedule states.
Sociedad Argentina de Informática e Investigación Operativa - Materia
-
Ciencias Informáticas
Control knowledge
Schedule state simulator
Computational tool - Nivel de accesibilidad
- acceso abierto
- Condiciones de uso
- http://creativecommons.org/licenses/by-sa/3.0/
- Repositorio
.jpg)
- Institución
- Universidad Nacional de La Plata
- OAI Identificador
- oai:sedici.unlp.edu.ar:10915/89513
Ver los metadatos del registro completo
| id |
SEDICI_9a4b2e0bf4e7f8f37a69e7d7e92be768 |
|---|---|
| oai_identifier_str |
oai:sedici.unlp.edu.ar:10915/89513 |
| network_acronym_str |
SEDICI |
| repository_id_str |
1329 |
| network_name_str |
SEDICI (UNLP) |
| spelling |
Closed-loop Rescheduling using Deep Reinforcement LearningPalombarini, Jorge A.Martínez, Ernesto C.Ciencias InformáticasControl knowledgeSchedule state simulatorComputational toolIn this work, a novel approach for generating rescheduling knowledge which can be used in real-time for handling unforeseen events without extra deliberation is presented. For generating such control knowledge, the rescheduling task is modelled and solved as a closed-loop control problem by resorting to the integration of a schedule state simulator with a rescheduling agent that can learn successful schedule repairing policies directly from a variety of simulated transitions between schedule states, using as input readily available schedule color-rich Gantt chart images, and negligible prior knowledge. The generated knowledge is stored in a deep Q-network, which can be used as a computational tool in a closed-loop rescheduling control way that select repair actions to make progress towards a goal schedule state, without requiring to compute the rescheduling problem solution every time a disruptive event occurs and safely generalize control knowledge to unseen schedule states.Sociedad Argentina de Informática e Investigación Operativa2019-09info:eu-repo/semantics/conferenceObjectinfo:eu-repo/semantics/publishedVersionResumenhttp://purl.org/coar/resource_type/c_5794info:ar-repo/semantics/documentoDeConferenciaapplication/pdf86http://sedici.unlp.edu.ar/handle/10915/89513enginfo:eu-repo/semantics/altIdentifier/issn/2618-3277info:eu-repo/semantics/openAccesshttp://creativecommons.org/licenses/by-sa/3.0/Creative Commons Attribution-ShareAlike 3.0 Unported (CC BY-SA 3.0)reponame:SEDICI (UNLP)instname:Universidad Nacional de La Platainstacron:UNLP2025-11-05T12:57:39Zoai:sedici.unlp.edu.ar:10915/89513Institucionalhttp://sedici.unlp.edu.ar/Universidad públicaNo correspondehttp://sedici.unlp.edu.ar/oai/snrdalira@sedici.unlp.edu.arArgentinaNo correspondeNo correspondeNo correspondeopendoar:13292025-11-05 12:57:39.994SEDICI (UNLP) - Universidad Nacional de La Platafalse |
| dc.title.none.fl_str_mv |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| title |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| spellingShingle |
Closed-loop Rescheduling using Deep Reinforcement Learning Palombarini, Jorge A. Ciencias Informáticas Control knowledge Schedule state simulator Computational tool |
| title_short |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| title_full |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| title_fullStr |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| title_full_unstemmed |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| title_sort |
Closed-loop Rescheduling using Deep Reinforcement Learning |
| dc.creator.none.fl_str_mv |
Palombarini, Jorge A. Martínez, Ernesto C. |
| author |
Palombarini, Jorge A. |
| author_facet |
Palombarini, Jorge A. Martínez, Ernesto C. |
| author_role |
author |
| author2 |
Martínez, Ernesto C. |
| author2_role |
author |
| dc.subject.none.fl_str_mv |
Ciencias Informáticas Control knowledge Schedule state simulator Computational tool |
| topic |
Ciencias Informáticas Control knowledge Schedule state simulator Computational tool |
| dc.description.none.fl_txt_mv |
In this work, a novel approach for generating rescheduling knowledge which can be used in real-time for handling unforeseen events without extra deliberation is presented. For generating such control knowledge, the rescheduling task is modelled and solved as a closed-loop control problem by resorting to the integration of a schedule state simulator with a rescheduling agent that can learn successful schedule repairing policies directly from a variety of simulated transitions between schedule states, using as input readily available schedule color-rich Gantt chart images, and negligible prior knowledge. The generated knowledge is stored in a deep Q-network, which can be used as a computational tool in a closed-loop rescheduling control way that select repair actions to make progress towards a goal schedule state, without requiring to compute the rescheduling problem solution every time a disruptive event occurs and safely generalize control knowledge to unseen schedule states. Sociedad Argentina de Informática e Investigación Operativa |
| description |
In this work, a novel approach for generating rescheduling knowledge which can be used in real-time for handling unforeseen events without extra deliberation is presented. For generating such control knowledge, the rescheduling task is modelled and solved as a closed-loop control problem by resorting to the integration of a schedule state simulator with a rescheduling agent that can learn successful schedule repairing policies directly from a variety of simulated transitions between schedule states, using as input readily available schedule color-rich Gantt chart images, and negligible prior knowledge. The generated knowledge is stored in a deep Q-network, which can be used as a computational tool in a closed-loop rescheduling control way that select repair actions to make progress towards a goal schedule state, without requiring to compute the rescheduling problem solution every time a disruptive event occurs and safely generalize control knowledge to unseen schedule states. |
| publishDate |
2019 |
| dc.date.none.fl_str_mv |
2019-09 |
| dc.type.none.fl_str_mv |
info:eu-repo/semantics/conferenceObject info:eu-repo/semantics/publishedVersion Resumen http://purl.org/coar/resource_type/c_5794 info:ar-repo/semantics/documentoDeConferencia |
| format |
conferenceObject |
| status_str |
publishedVersion |
| dc.identifier.none.fl_str_mv |
http://sedici.unlp.edu.ar/handle/10915/89513 |
| url |
http://sedici.unlp.edu.ar/handle/10915/89513 |
| dc.language.none.fl_str_mv |
eng |
| language |
eng |
| dc.relation.none.fl_str_mv |
info:eu-repo/semantics/altIdentifier/issn/2618-3277 |
| dc.rights.none.fl_str_mv |
info:eu-repo/semantics/openAccess http://creativecommons.org/licenses/by-sa/3.0/ Creative Commons Attribution-ShareAlike 3.0 Unported (CC BY-SA 3.0) |
| eu_rights_str_mv |
openAccess |
| rights_invalid_str_mv |
http://creativecommons.org/licenses/by-sa/3.0/ Creative Commons Attribution-ShareAlike 3.0 Unported (CC BY-SA 3.0) |
| dc.format.none.fl_str_mv |
application/pdf 86 |
| dc.source.none.fl_str_mv |
reponame:SEDICI (UNLP) instname:Universidad Nacional de La Plata instacron:UNLP |
| reponame_str |
SEDICI (UNLP) |
| collection |
SEDICI (UNLP) |
| instname_str |
Universidad Nacional de La Plata |
| instacron_str |
UNLP |
| institution |
UNLP |
| repository.name.fl_str_mv |
SEDICI (UNLP) - Universidad Nacional de La Plata |
| repository.mail.fl_str_mv |
alira@sedici.unlp.edu.ar |
| _version_ |
1847978618733461504 |
| score |
13.142177 |