AI for Healthcare
Dev & Doc
Understanding Reinforcement Learning from Human Feedback (RLHF)
0:00
-42:01

Understanding Reinforcement Learning from Human Feedback (RLHF)

Aligning AI models for healthcare
Transcript

No transcript...

You can also listen to the episode on YouTube DevAndDoc and Spotify DevAndDoc

🎙️ Podcast Highlights

  • Exploring the alignment of AI

  • A deep dive into RLHF and AI training methods, shedding light on cutting-edge advancements.

  • Understanding the intricacies of reinforcement learning in AI development.

  • Discussing the critical role of preference models - how rewarding models correctly is pivotal for success.

  • Addressing the challenges of exploiting reward functions and model degradation, with solutions to fix it.

Thank you for listening,

Zeljko

0 Comments
AI for Healthcare
Dev & Doc
Our podcast is structured as a collection of conversations between doctors and developers exploring the potential of AI in healthcare. Only the last couple of episodes are on Substack, you can find the older ones on YouTube or Spotify - https://www.youtube.com/@DevAndDoc