Main Deep Reinforcement Learning with Python: RLHF for Chatbots and Large Language Models

Deep Reinforcement Learning with Python: RLHF for Chatbots and Large Language Models

5.0 / 5.0
0 comments
Deep Reinforcement Learning with Python, Second Edition Gain a theoretical understanding to the most popular libraries in deep reinforcement learning (deep RL). This new edition focuses on the latest advances in deep RL using a learn-by-coding approach, allowing readers to assimilate and replicate the latest research in this field. New agent environments ranging from games, and robotics to finance are explained to help you try different ways to apply reinforcement learning. A chapter on multi-agent reinforcement learning covers how multiple agents compete, while another chapter focuses on the widely used deep RL algorithm, proximal policy optimization (PPO). You’ll see how reinforcement learning with human feedback (RLHF) has been used by chatbots, built using Large Language Models, e.g. ChatGPT to improve conversational capabilities. You’ll also review the steps for using the code on multiple cloud systems and deploying models on platforms such as Hugging Face Hub. The code is in Jupyter Notebook, which canbe run on Google Colab, and other similar deep learning cloud platforms, allowing you to tailor the code to your own needs. Whether it’s for applications in gaming, robotics, or Generative AI, Deep Reinforcement Learning with Python will help keep you ahead of the curve. What You’ll Learn Explore Python-based RL libraries, including StableBaselines3 and CleanRL Work with diverse RL environments like Gymnasium, Pybullet, and Unity ML Understand instruction finetuning of Large Language Models using RLHF and PPO Study training and optimization techniques using HuggingFace, Weights and Biases, and Optuna
Request Code : ZLIBIO4352623
Categories:
Year:
2024
Edition:
2nd ed., Second
Publisher:
Apress
Language:
English
ISBN 13:
9798868802720
ISBN:
9798868802720
This book is not available due to the complaint of the copyright holder.

Comments of this book

There are no comments yet.