AI for Healthcare
Dev & Doc
Understanding Reinforcement Learning from Human Feedback (RLHF)
0:00
Current time: 0:00 / Total time: -42:01
-42:01

Understanding Reinforcement Learning from Human Feedback (RLHF)

Aligning AI models for healthcare

You can also listen to the episode on YouTube DevAndDoc and Spotify DevAndDoc

🎙️ Podcast Highlights

  • Exploring the alignment of AI

  • A deep dive into RLHF and AI training methods, shedding light on cutting-edge advancements.

  • Understanding the intricacies of reinforcement learning in AI development.

  • Discussing the critical role of preference models - how rewarding models correctly is pivotal for success.

  • Addressing the challenges of exploiting reward functions and model degradation, with solutions to fix it.

Thank you for listening,

Zeljko

Discussion about this podcast

AI for Healthcare
Dev & Doc
Our podcast is structured as a collection of conversations between doctors and developers exploring the potential of AI in healthcare. Only the last couple of episodes are on Substack, you can find the older ones on YouTube or Spotify - https://www.youtube.com/@DevAndDoc