<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Reinforcement Learning on Baam's Techlog</title><link>https://baampark.github.io/tags/reinforcement-learning/</link><description>Recent content in Reinforcement Learning on Baam's Techlog</description><generator>Hugo -- 0.128.0</generator><language>en-us</language><lastBuildDate>Sat, 04 Oct 2025 15:32:28 -0400</lastBuildDate><atom:link href="https://baampark.github.io/tags/reinforcement-learning/index.xml" rel="self" type="application/rss+xml"/><item><title>From Policy Gradient to GRPO: Policy Optimization for LLM Training</title><link>https://baampark.github.io/posts/2025-10-04_rl_policy_optimization/</link><pubDate>Sat, 04 Oct 2025 15:32:28 -0400</pubDate><guid>https://baampark.github.io/posts/2025-10-04_rl_policy_optimization/</guid><description>You’ve probably heard that DeepSeek R1 was fine-tuned using reinforcement learning, specifically an algorithm called Generalized Reparameterized Policy Optimization (GRPO). DeepSeek research team demostrated that reinforcement learning (RL) without any supervised fine-tuning can teach LLMs to reason, and this drew widespread interest and scrutiny across academia. In my previous blog post, Mathmatical Foundation from Markov to Deep Q-learning, we dabbled in Q-learning, which is value-based (off-policy) RL where the agent learns value (\(Q\) or \(V\)) and derives its policy \(\pi\) from the value.</description></item><item><title>Mathmatical Foundation from Markov to Deep Q-learning</title><link>https://baampark.github.io/posts/2025-02-23_rl_math/</link><pubDate>Sun, 23 Feb 2025 15:04:51 -0500</pubDate><guid>https://baampark.github.io/posts/2025-02-23_rl_math/</guid><description>When I first started studying reinforcement learning, I was intimidated by the amount of mathematical background required to understand even the basic concepts. Terms like “Markov property,” “Bellman equation,” and “Q-learning” felt abstract and overwhelming. In this blog post, we will walk through these foundations step by step, starting from probability basics and building up toward deep reinforcement learning. Specifically, we will cover: 1) Markov decision process (MDP) 2) Value function, 3) Q-learning, and 4) Deep Q-learning (DQN).</description></item></channel></rss>