Search Results for author: June Yang

Found 1 papers, 1 papers with code

Aligning Language Models with Offline Learning from Human Feedback

1 code implementation23 Aug 2023 Jian Hu, Li Tao, June Yang, Chandler Zhou

Learning from human preferences is crucial for language models (LMs) to effectively cater to human needs and societal values.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.