Ziegler, Daniel M. (2019). Fine-Tuning Language Models from Human Preferences. arxiv.org(англ.). doi:10.48550/arXiv.1909.08593.
Warnell, Garrett (25 апреля 2018). Deep TAMER: Interactive Agent Shaping in High-Dimensional State Spaces. Proceedings of the AAAI Conference on Artificial Intelligence(англ.). 32 (1). doi:10.1609/aaai.v32i1.11485.
Bai, Yuntao (2022). Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback. arxiv.org(англ.). doi:10.48550/arXiv.2204.05862.
Ouyang, Long (2022). Training language models to follow instructions with human feedback. arxiv.org(англ.). doi:10.48550/arXiv.2203.02155.
Glaese, Amelia (2022). Improving alignment of dialogue agents via targeted human judgements. arxiv.org(англ.). doi:10.48550/arXiv.2209.14375.