• Classified by Topic • Classified by Publication Type • Sorted by Date • Sorted by First Author Last Name • Classified by Funding Source •
A Stitch in Time - Autonomous Model Management via Reinforcement Learning.
Elad
Liebman, Eric Zavesky, and Peter Stone.
In Proceedings of the 17th
International Conference on Autonomous Agents and Multiagent Systems (AAMAS), July 2018.
Concept drift - a change, either sudden or gradual, in the underlying properties of data - is one of the most prevalent challenges to maintaining high-performing learned models over time in autonomous systems. In the face of concept drift, one can hope that the old model is sufficiently representative of the new data despite the concept drift, one can discard the old data and retrain a new model with (often limited) new data, or one can use transfer learning methods to combine the old data with the new to create an updated model. Which of these three options is chosen affects not only near-term decisions, but also future needs to transfer or retrain. In this paper, we thus model response to concept drift as a sequential decision making problem and formally frame it as a Markov Decision Process. Our reinforcement learning approach to the problem shows promising results on one synthetic and two real-world datasets.
@InProceedings{AAMAS2018-eladlieb, author = {Elad Liebman and Eric Zavesky and Peter Stone}, title = {{A} {S}titch in {T}ime - {A}utonomous {M}odel {M}anagement via {R}einforcement {L}earning}, booktitle = {Proceedings of the 17th International Conference on Autonomous Agents and Multiagent Systems (AAMAS)}, location = {Stockholm, Sweden}, month = {July}, year = {2018}, abstract = { Concept drift - a change, either sudden or gradual, in the underlying properties of data - is one of the most prevalent challenges to maintaining high-performing learned models over time in autonomous systems. In the face of concept drift, one can hope that the old model is sufficiently representative of the new data despite the concept drift, one can discard the old data and retrain a new model with (often limited) new data, or one can use transfer learning methods to combine the old data with the new to create an updated model. Which of these three options is chosen affects not only near-term decisions, but also future needs to transfer or retrain. In this paper, we thus model response to concept drift as a sequential decision making problem and formally frame it as a Markov Decision Process. Our reinforcement learning approach to the problem shows promising results on one synthetic and two real-world datasets. }, }
Generated by bib2html.pl (written by Patrick Riley ) on Sun Nov 24, 2024 20:24:54