-
Notifications
You must be signed in to change notification settings - Fork 28
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Change the example of model #8
Comments
Hi @iamysy! I suggest that if there is a sudden drop in Reward, you can restart the training a few times to get the effect similar to the paper:) For the second question: If you want to train cylinders using the SAC algorithm, you can compare these two files: launch_multiprocessing_traning_cylinder.py and launch_multiprocessing_traning_square.py. You need to put the cylinder changes into launch_multiprocessing_traning_square.py of the square cylinder:
In fact, these are some changes to the jets in the OpenFoam case, which will become clearer if you look further. Good luck:) |
Thank you! @1900360 |
Hello, I am still doing similar work recently, I would like to ask you, if I want to break the limit of 100 trajectory, where do I need to change it? I have tried to change
I hope my question doesn't seem stupid, and please give me your advice. |
Hi @iamysy ! Well, I think you still have the following parameter unchanged, which is the key to determine the total number of steps each episode runs in each environment: I don't think you need to save the best episode during DRL training. You can use DRL test function(that is no exploration). The desired episode can be obtained from testing by using the training convergence policy. Thanks for your reminding, I will update this function later. Also, if you want to save the entire environment calculation file, it is not recommended to do so under environment_tianshou file. After all, there is no way to compare the effects of training (such as reward) between each environments. I think it's appropriate to need to operate under this file. But this is related to tianshou platform, please refer to it in tutorials. |
Your advice is very useful to me! I found that I only modified one |
Hello, your work is great. Thank you for your inspiration.
I had some problems during the operation.(I run under the docker environment.)
First of all, in the cylinder of example, Even after 200 rounds, the value of the reward is still unstable, and even falling off a cliff.
Secondly, I want to change the reinforcement learning model of the cylindrical example to SAC. I changed it on the basis of the square column example, but the reward is unchanged.
May I ask you about these two questions?
Thank you again.
The text was updated successfully, but these errors were encountered: