返回到 Reinforcement Learning from Human Feedback
DeepLearning.AI

Reinforcement Learning from Human Feedback

Large language models (LLMs) are trained on human-generated text, but additional methods are needed to align an LLM with human values and preferences. Reinforcement Learning from Human Feedback (RLHF) is currently the main method for aligning LLMs with human values and preferences. RLHF is also used for further tuning a base LLM to align with values and preferences that are specific to your use case. In this course, you will gain a conceptual understanding of the RLHF training process, and then practice applying RLHF to tune an LLM. You will: 1. Explore the two datasets that are used in RLHF training: the “preference” and “prompt” datasets. 2. Use the open source Google Cloud Pipeline Components Library, to fine-tune the Llama 2 model with RLHF. 3. Assess the tuned LLM against the original base model by comparing loss curves and using the “Side-by-Side (SxS)” method.

状态:LLM Application
状态:Model Evaluation
中级项目小时

精选评论

AA

5.0评论日期:Jun 18, 2025

better to be expanded a bit, but overall, it is super course

ME

4.0评论日期:Jan 11, 2025

Overall worth a shot. Not in depth but good overview

所有审阅

显示:6/6

Ahmad Alsharef
5.0
评论日期:Jun 19, 2025
Neil
5.0
评论日期:Aug 17, 2025
sajjad shahali
5.0
评论日期:May 14, 2025
Fady Ashraf Sulaiman
4.0
评论日期:Dec 12, 2024
Manideep Reddy Enugala
4.0
评论日期:Jan 12, 2025
Alessandro Varriale
4.0
评论日期:Aug 28, 2024