Symbol Guided Hindsight Priors for Reward Learning from Human Preferences

This paper was accepted at the “Human in the Loop Learning Workshop” at NeurIPS 2022.
Specification of reward functions for Reinforcement Learning is a challenging task which is bypassed by the framework of Preference Based Learning methods which instead learn from preference labels on trajectory queries. These methods, however, still suffer from high requirements of preference labels and often would still achieve low reward recovery. We present the PRIOR framework that alleviates the issues of impractical number of queries to humans as well as poor reward recovery through computing priors…Apple Machine Learning Research