You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for the implementation and wonderful work. I was trying to run the IDDPG with a simple tag and trying to observe the result. Can you help me understand how the Fig-3 is plotted in the paper? Is it that you check the entropy, or is there any other term I'm missing here?
Also, I have another question: what should we do to observe the training progression on the terminal instead of saving it? Can you also provide a command to monitor the training process?
Thanks,
Kailash
The text was updated successfully, but these errors were encountered:
I'm sorry to trouble you again. The problem I encountered was not smoothness, but rather the non convergence of rewards, and the values of "Action Loss, Value Loss, Entropy" also fluctuated greatly. I haven't made any changes to the parameters, I just modified the expression of success, which shouldn't involve any policy update issues. I really want to restore your code. Do you still want to save the original code data?
Hi,
Thanks for the implementation and wonderful work. I was trying to run the IDDPG with a simple tag and trying to observe the result. Can you help me understand how the Fig-3 is plotted in the paper? Is it that you check the entropy, or is there any other term I'm missing here?
Also, I have another question: what should we do to observe the training progression on the terminal instead of saving it? Can you also provide a command to monitor the training process?
Thanks,
Kailash
The text was updated successfully, but these errors were encountered: