From 04818f770bcc19e3d639f7c1c0951c8ff6ba2ced Mon Sep 17 00:00:00 2001 From: phong vu <54299527+phongvu009@users.noreply.github.com> Date: Wed, 8 Oct 2025 08:42:26 -0500 Subject: [PATCH] Update deep-q-network.mdx I am adding a note that we are using Same network Not 2 networks yet. So we know that this will cause issue for training. And then to solve this issue we will use Target Network. On the next chapter The Deep Q-Learning Algorithm, the picture of pseudocode is showing that we are using 2 networks. It may cause confusion for beginner . Please add some part saying we are using One network first . Thanks ! --- units/en/unit3/deep-q-network.mdx | 2 ++ 1 file changed, 2 insertions(+) diff --git a/units/en/unit3/deep-q-network.mdx b/units/en/unit3/deep-q-network.mdx index 50cd4f2b..22bd3c4a 100644 --- a/units/en/unit3/deep-q-network.mdx +++ b/units/en/unit3/deep-q-network.mdx @@ -39,3 +39,5 @@ Finally, we have a couple of fully connected layers that output a Q-value for ea Deep Q Network So, we see that Deep Q-Learning uses a neural network to approximate, given a state, the different Q-values for each possible action at that state. Now let's study the Deep Q-Learning algorithm. + +NOTE: We are using same network to calculate predicting (Q_current) and targeting (Q_target). This will cause some issues of Unstable training since the target is moving every time network’s parameters is updated.