Paper ID | MLSP-26.1 | ||
Paper Title | Introducing Deep Reinforcement Learning to NLU Ranking Tasks | ||
Authors | Ge Yu, Emre Barut, Chengwei Su, Amazon Inc, United States | ||
Session | MLSP-26: Reinforcement Learning 2 | ||
Location | Gather.Town | ||
Session Time: | Thursday, 10 June, 13:00 - 13:45 | ||
Presentation Time: | Thursday, 10 June, 13:00 - 13:45 | ||
Presentation | Poster | ||
Topic | Machine Learning for Signal Processing: [MLR-REI] Reinforcement learning | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Natural language understanding (NLU) models in production systems rely heavily on human annotated data, which involves an expensive, time-consuming, and error-prone process. Moreover, the model release process requires offline supervised learning and human in-the-loop. Together, these factors prolong the model update cycle and result in sub-standard model performance in situations where usage behavior is non-stationary. In this paper, we address these issues with a deep reinforcement learning approach that ranks suggestions from multiple experts in an online fashion. Our proposed method removes the reliance on annotated data, and can effectively adapt to recent changes in the data distribution. The efficiency of the new approach is demonstrated through simulation experiments using logged data from voice-based virtual assistants. Our results show that our algorithm, without any reliance on annotation, over-performs offline supervised learning methods. |