Preprint
Brief Report

This version is not peer-reviewed.

A Brief Tutorial on Reinforcement Learning: From MDP to DDPG

Submitted:

19 February 2026

Posted:

20 February 2026

You are already at the latest version

Abstract
Following the evolution from theoretical foundations to advanced deep learning algorithms, a coherent overview of reinforcement learning (RL) is proposed in this tutorial. We begin with the mathematical formalization of sequential decision-making via Markov decision processes (MDPs). In MDP, Bellman equation and Bellman optimality equation play important roles, they provide for policy evaluation and the fundamental condition for optimal behavior, respectively. The movement from these equations to practical algorithms is explored, starting with model-based dynamic programming and progressing to model-free temporal-difference (TD) learning. As a pivotal model-free algorithm, Q-learning directly implements the Bellman optimality equation through sampling. To handle high-dimensional state spaces, function approximation and deep reinforcement learning emerge, exemplified by Deep Q-Networks (DQN). Thereafter, actor-critic methods address the challenge of continuous action spaces. As a typical actor-critic scheme, the deep deterministic policy gradient (DDPG) algorithm is illustrated in detail on how it adapts the principles of optimality to continuous control by maintaining separate actor and critic networks. Finally, the tutorial concludes with a unified perspective, observing the development of RL as a logical progression from defining optimality conditions to developing scalable solution algorithms. Furthermore, future directions are summarized.
Keywords: 
;  

1. Introduction: How Do Machines Learn to Make Decisions?

Imagine teaching a robot to walk—you cannot specify the movement of every muscle in detail, but you can tell it “you’ll get a reward for walking well, and lose points for falling.” Through repeated attempts, the robot gradually figures out a stable walking method. This is the core idea of reinforcement learning (RL), i.e., an agent learns an optimal decision-making policy through interaction with the environment and trial-and-error [1].
RL has made groundbreaking progress over the past decade, from AlphaGo defeating human Go champions to agents surpassing humans in complex games. Behind these achievements lies a series of sophisticated mathematical frameworks and algorithmic advancements. This tutorial guides you from the most fundamental Markov decision process (MDP), step by step, to explore its core equations and algorithm evolution, up to understanding advanced algorithms like the deep deterministic policy gradient (DDP), unveiling the mystery behind intelligent decision-making systems. The logistic structure of the tutorial is illustrated in Figure 1.

2. The Mathematical Foundation of Decision-Making: MDP

In this section, we first introduce the Markov property, and then give the definition of MDP. Finally, the policy and return are explained.
The Markov property is the core assumption of reinforcement learning, it shows that the future state depends only on the current state and not on past states. Mathematically,
P ( S t + 1 | S t ) = P ( S t + 1 | S t , S t - 1 , , S 0 )
where P is the probability, and S t is the state at time t
This memoryless property greatly simplifies the modeling of decision problems.
Formally, a standard MDP consists of five key elements.
(1)
State space ( S ): The set of all possible situations of the environment.
(2)
Action space ( A ): The set of operations the agent can perform.
(3)
Transition probability ( P ): The probability distribution of state transitions after taking an action.
(4)
Reward function ( R ): The environment’s immediate feedback for a state-action pair.
(5)
Discount factor (γ): Measures the present value of future rewards (0 ≤ γ ≤ 1).
Remarks:The discount factor is similar to the concept of “present value” in economics—100 $ tomorrow is worth less than 100 $ today, and the agent also values immediate rewards more.
A policy π is the agent’s decision rule, specifying what action to take in each state. The agent’s goal is to maximize the cumulative discounted return as
G t = R t + 1 + γ R t + 2 + γ ² R t + 3 +
where G t is referred to as the return at time t.

3. From Policy Evaluation to Optimal Policy: The Evolution of the Bellman Equation

In this section, we first give the definitions of state-value function and action-value function, respectively. And then the Bellman equation and Bellman optimality equation are discussed successively. Finally, the dynamical programming methods to solve Bellman optimality equation under known model are given.
To evaluate the quality of a policy, RL introduces two core concepts as follows.
1) State-value function V π ( s ) : The expected return starting from state s and following policy π.
V π ( s ) = E π [ G t | S t = s ]
2) Action-value function Q π ( s , a ) : The expected return after taking action a in state s and thereafter following policy π.
Q π ( s , a ) = E π [ G t | S t = s , A t = a ]
The value functions satisfy an important recursive relationship—the Bellman Equation:
V π ( s ) = Σ a π ( a | s ) Σ s P ( s | s , a ) [ R ( s , a , s ) + γ V π ( s ) ]
This equation reveals a profound insight: the value of the current state equals the immediate reward plus the discounted value of the successor state.
Remarks: The Bellman equation allows us to evaluate the quality of any given policy π, but it does not tell us how to find a better policy.
The ultimate goal of RL is to find the optimal policy π * . This leads to one of the most central equations in RL theory, i.e., the Bellman optimality equation.
1) Optimality equation for the state-value function:
V * ( s ) = m a x a Σ s P ( s | s , a ) [ R ( s , a , s ) + γ V * ( s ) ]
2) Optimality equation for the action-value function:
Q * ( s , a ) = Σ s P ( s | s , a ) [ R ( s , a , s ) + γ m a x a Q * ( s , a ) ]
Remarks: The Bellman optimality equation no longer depends on a specific policy π, but directly defines the self-consistent condition that the optimal value functions must satisfy. It tells us: the optimal state value equals the value corresponding to the action that yields the maximum expected return; the optimal action value equals the immediate reward plus the maximum optimal action value of the successor state.
Remarks: Imagine finding the shortest path between cities. The Bellman equation tells you “how long it will take to reach the destination along the current route”; the Bellman optimality equation tells you “the shortest possible time from the current location to the destination, and which direction to go next.”
Remarks: From the Bellman equation to the Bellman optimality equation, we complete the conceptual leap from policy evaluation (analyzing a given policy) to policy optimization (finding the optimal policy). This equation is the theoretical foundation for all subsequent optimal control algorithms.
When the environment’s dynamics model (P and R) are known, we can solve the Bellman optimality equation using dynamic programming methods [2] as follows.
1)Policy iteration: Alternates between policy evaluation (solving the Bellman equation) and policy improvement.
2)Value iteration: Directly iterates the Bellman optimality equation:
V k + 1 ( s ) = m a x a Σ s P ( s | s , a ) [ R ( s , a , s ) + γ V k ( s ) ]
Remarks: The value iteration algorithm essentially repeatedly applies the Bellman optimality equation until the value function converges. This verifies that the Bellman optimality equation is not only descriptive but also constructive—it directly provides a solution algorithm.

4. Model-Free Learning: From Theory to Practice

In realistic circumstances, the environment model is often unknown. In this section, the model-free scenario is discussed. Monte Carlo (MC) methods and temporal difference (TD) learning are given at the beginning.
1) Monte Carlo methods: Estimate the value function using returns from complete trajectories.
V ( S t ) V ( S t ) + α [ G t V ( S t ) ]
2) Temporal difference (TD) learning: Combines ideas from Monte Carlo and dynamic planning, using estimated values for updates.
V ( S t ) V ( S t ) + α [ R t + 1 + γ V ( S t + 1 ) V ( S t ) ]
where the TD error δ t = R t + 1 + γ V ( S t + 1 ) V ( S t ) measures the discrepancy between prediction and actual outcome.
Q-learning is a milestone algorithm in RL history [3]. As a practical breakthrough from the Bellman optimality equation. Its update rule derives directly from the Bellman optimality equation as
Q ( S t , A t ) Q ( S t , A t ) + α [ R t + 1 + γ m a x a Q ( S t + 1 , a ) Q ( S t , A t ) ]
Remarks:Comparing the Q-learning update formula with the Bellman optimality equation, you’ll find that Q-learning is approximately solving the optimal equation through sampling. m a x a Q ( S t + 1 , a ) corresponds to m a x a Q * ( s , a ) in the optimal equation, and the entire update moves towards the optimal Q-value.
Remarks: Q learning is an off-policy learning, i.e., the agent learns the optimal action-value function to find the optimal policy, while potentially exploring other behaviors. This allows it to directly estimate the optimal Q-function without depending on the currently executed policy.
When the state space is huge or continuous, tabular methods become infeasible. The function approximation scheme emerges, that uses parameterized functions (e.g., neural networks) to represent value functions or policies as
Q ( s , a ; θ ) Q π ( s , a )
where θ are learnable parameters. This paves the way for combining deep learning with reinforcement learning.

5. The Rise of Deep Reinforcement Learning

Proposed by DeepMind in 2015, the deep Q-networks (DQN) achieves human-level performance on Atari games [4]. The main innovations include
1) Experience replay: Stores transition samples (s,a,r,s’) and randomly samples from them to break data correlations.
2) Target network: Uses a separate target network to compute the TD target, stabilizing the learning process.
3) End-to-end learning: Learns control policies directly from raw pixel input.
The DQN loss function essentially minimizes the error of the Bellman optimality equation as
L ( θ ) = E [ ( r + γ m a x a Q ( s , a ; θ ) Q ( s , a ; θ ) ) ² ]
where θ are the target network parameters.
Remarks: DQN can be viewed as a practical implementation that approximates the Q-function with a deep neural network and solves the Bellman optimality equation using stochastic gradient descent.

6. The Deep Deterministic Policy Gradient (DDPG) Algorithm

Many practical problems (e.g., robot control, autonomous driving) require decision-making in continuous action spaces [6]. Discrete-action methods like DQN face the curse of dimensionality, i.e., if each dimension has m possible values, a d-dimensional action space has m d possible actions. More importantly, the m a x a Q ( s , a ) operation in DQN becomes computationally intractable in continuous spaces.
Unlike value-based methods, policy gradient (PG) methods directly parameterize the policy π(a|s;θ) and optimize the expected return [5]. The fundamental theorem is
θ J ( θ ) = E π [ θ l o g π ( A t | S t ; θ ) Q π ( S t , A t ) ]
Remarks: Increase the probability of actions that lead to high returns and decrease the probability of actions that lead to low returns.
The REINFORCE is a typical MC-PG method. The sampling efficiency of REINFORCE is low. Therefore, by combing value-based and policy-based methods, the DDPG emerges.
The DDPG is an actor-critic algorithm [7] designed for continuous control, cleverly avoiding the maximization problem. The characteristic of DDPG is
1) Deterministic policy:  a = μ ( s | θ μ ) directly outputs a deterministic action.
2) Actor-critic architecture:
actor network (policy network): μ ( s | θ μ ) , responsible for selecting actions.
critic network (value function network): Q ( s , a | θ Q ) , evaluates the quality of actions.
3) Soft update mechanism: Target network parameters slowly track the online network parameters.
DDPG’s update rules embody the adaptation of Bellman optimality ideas to continuous spaces.
1) Critic update (based on the continuous form of the Bellman optimality equation):
y = r + γ Q ( s , μ ( s | θ μ ) | θ Q )
Note that the target policy network μ’ replaces the max operation here, as the deterministic policy directly gives the optimal action selection.
3) Actor update(improving the policy along the gradient of the Q-function):
θ μ J E [ a Q ( s , a | θ Q ) θ μ μ ( s | θ μ ) ]
The actor network is trained to output actions that maximize the Q-value, which is essentially the “policy improvement” step in continuous action spaces.
The Pseudocode of DDPG is listed as
Initialize actor network μ ( s | θ μ ) and critic network Q ( s , a | θ Q )
Initialize corresponding target networks μ’ and Q’
Initialize experience replay buffer R
for each time step do
Select action a = μ ( s | θ μ ) + exploration noise N 0
Execute a, observe reward r and new state s’
Store transition (s,a,r,s’) in R
Randomly sample a minibatch from R
   Update critic (based on modified Bellman optimality equation)
y i = r i + γ Q ( s i , μ ( s i | θ μ ) | θ Q )
Minimize L = 1 N Σ i ( y i - Q ( s i , a i | θ Q ) ) ²
   Update actor (approximate policy improvement)
θ μ J 1 N Σ i a Q ( s , a | θ Q ) | s = s i , a = μ ( s i ) θ μ μ ( s | θ μ ) | s = s i
   Soft update target networks
θ Q τ θ Q + ( 1 - τ ) θ Q
θ μ τ θ μ + ( 1 - τ ) θ μ
end for
Remarks: DDPG can be seen as an approximate method for solving the Bellman optimality equation in continuous spaces. The critic learns the optimal Q-function, the actor learns the corresponding optimal policy, and they are optimized alternately, similar to policy iteration in dynamic programming.
The advantages and challenges of DDPG is outlined as
First, we give the advantages.
1) Efficiently handles continuous action spaces, avoiding the curse of dimensionality.
2) Relatively high sample efficiency.
3) Can learn deterministic policies, suitable for tasks requiring precise control.
However, there are challenges.
1) Sensitive to hyperparameters.
2) Exploration efficiency issues (requires carefully designed exploration noise).
3) May converge to local optima.

7. Frontier Developments and a Unified Theoretical Perspective

In this section, we give the logistical evolution and future directions.
Following DDPG, researchers proposed various improved algorithms, all revolving around the core of solving the Bellman optimality equation more stably. Some of them are
1) TD3 (twin delayed deep deterministic policy gradient): Addresses overestimation via double Q-learning, a correction for errors in the max operator of the Bellman optimality equation.
2) SAC(soft actor-critic): Introduces a maximum entropy framework, modifying the Bellman equation to encourage exploration [8].
3) PPO(proximal policy optimization): Stabilizes training by constraining policy updates, ensuring monotonic policy improvement [9].
From MDP to the DDPG, the development of reinforcement learning presents a clear logical thread. Table 1 shows this complete framework of theoretical evolution and algorithmic development.
We conclude the key turning points in the logical evolution.
1) From evaluation to optimization, i.e., Bellman equation to Bellman optimality equation.
2) From known to unknown, i.e., dynamic programming (model known) to model-free learning.
3) From discrete to continuous, i.e., tabular methods to function approximation to deep neural networks.
4) From value function to policy, i.e., value-based methods to policy gradient to actor-critic.
5) From theory to practice, i.e., theoretical equations to practical algorithms to engineering implementation.
Potential future directions of RL are listed as follows.
1)Theoretical deepening: Deeper understanding of the generalization capabilities and convergence properties of deep reinforcement learning.
2)Algorithmic efficiency: Improving sample efficiency, reducing the need for environment interaction.
3) Safety and reliability: Ensuring safety during exploration.
4) Multi-task learning: Enabling transfer of knowledge and skills.

8. Conclusion: The Science of Intelligent Decision-Making Within a Unified Framework

From the mathematical formalization of MDPs, to the establishment of the Bellman equation, and further to the theoretical breakthrough of the Bellman optimality equation, RL has built a complete theoretical system for decision-making. Q-learning transforms the optimality equation into a practical algorithm, deep neural networks solve the problem of high-dimensional representation, and algorithms like DDPG extend it to the domain of continuous control.
This development history reveals a profound scientific methodology. That is from clear problem definition (i.e., MDP), to establishing optimality conditions (i.e., Bellman optimality equation), to developing practical solution algorithms (i.e., from dynamic programming to deep reinforcement learning). Each layer of progress is built upon a solid theoretical foundation while simultaneously pushing the theory further.
Despite fast development, the core of RL revolves around the fundamental problem of how to make optimal decisions in uncertain environments. The Bellman optimality equation continues to guide the development direction of new algorithms. With computing power improvement and theoretical refinement, intelligent systems based on these principles play key roles in more complex decision-making tasks.
The Bellman equation not only became the cornerstone of RL but also opened a new era of enabling machines to learn autonomous decision-making. The path of exploration from theoretical equations to practical algorithms continues, and each step brings us closer to truly intelligent machine decision systems.

References

  1. Sutton, R. S., & Barto, A. G. (2018). Reinforcement Learning: An Introduction (2nd ed.). MIT Press.
  2. Bellman, R. (1957). Dynamic Programming. Princeton University Press.
  3. Watkins, C. J. C. H., & Dayan, P. (1992). Q-learning. Machine Learning, 8(3-4), 279–292. [CrossRef]
  4. Mnih, V., Kavukcuoglu, K., Silver, et al., (2015). Human-level control through deep reinforcement learning. Nature, 518(7540), 529–533. [CrossRef]
  5. Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., & Wierstra, D. (2016). Continuous control with deep reinforcement learning. International Conference on Learning Representations (ICLR).
  6. Silver, D., Lever, G., Heess, N., Degris, T., Wierstra, D., & Riedmiller, M. (2014). Deterministic policy gradient algorithms. International Conference on Machine Learning (ICML).
  7. Fujimoto, S., van Hoof, H., & Meger, D. (2018). Addressing function approximation error in actor-critic methods. International Conference on Machine Learning (ICML).
  8. Haarnoja, T., Zhou, A., Abbeel, P., & Levine, S. (2018). Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. International Conference on Machine Learning (ICML).
  9. Schulman, J., Wolski, F., Dhariwal, P., Radford, A., & Klimov, O. (2017). Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347.
Figure 1. Logistic structure of the tutorial.
Figure 1. Logistic structure of the tutorial.
Preprints 199612 g001
Table 1. Correspondence between Theoretical Foundations and Algorithm Evolution.
Table 1. Correspondence between Theoretical Foundations and Algorithm Evolution.
Theoretical Stage Core Concept Mathematical Expression Representat-ive Algorithms Applicable Scenarios
Problem Definition Markov Decision Process ( S , A , P , R , γ ) - Decision problem formulation
Policy Evaluation Bellman Equation V π ( s ) = Σ a π ( a | s ) Σ s P ( s | s , a ) [ R + γ V π ( s ) ] Policy evaluation algorithms Analyzing a given policy
Optimal Condition Bellman Optimality Equation V ( s ) = m a x a Σ s P ( s | s , a ) [ R + γ V ( s ) ] Theoretical benchmark Defining optimal policy standard
Model-Based Solution Dynamic Programming V k + 1 ( s ) = m a x a Σ s P ( s | s , a ) [ R + γ V k ( s ) ] Value iteration, Policy iteration Environments with known model
Model-Free Learning Temporal Difference Learning Q ( s , a ) Q ( s , a ) + α [ R + γ Q ( s , a ) - Q ( s , a ) ] SARSA、Q-learning Environments with unknown model
High-Dimensional Extension Function Approximation Q ( s , a ; θ ) Q π ( s , a ) Linear function approximation Large state spaces
Deep Integration Deep Q-Learning L ( θ ) =   E [ ( r + γ m a x a Q ( s , a ; θ ) Q ( s , a ; θ ) ) ² ] DQN and its variants High-dimensional inputs like images
Policy Optimization Policy Gradient θ J ( θ ) = E [ θ l o g π ( a | s ; θ ) Q π ( s , a ) ] REINFORCE Direct policy optimization
Integrated Methods Actor-Critic actor: θ J E [ θ μ ( s ) a Q ( s , a ) ] A3C、DDPG Continuous action spaces
Continuous Control Deterministic Policy Gradient a = μ ( s | θ μ ) DDPG、TD3、SAC Robot control, etc.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated