Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can't reproduce test results with FinRL after restarting Jupyter Notebook Kernel #1264

Open
Bottagandalf opened this issue Aug 3, 2024 · 3 comments
Assignees
Labels
good first issue Good for newcomers

Comments

@Bottagandalf
Copy link

Hi everyone!

Ive been doing some research for my masters thesis on reinforcement learning using FinRL.
I'm mainly using the Copy_of_FinRL_Ensemble_StockTrading_ICAIF_2020.ipynb file and tweaking the parameters slightly.
Unfortunately, I'm having issues getting reproducible results. I already set Seeds and my results are reproducible until I restart the Jupyter Notebook Kernel. From there, everything is completely mixed up again.

I looked this issue up on the web and came accross Hashseeds. I tried different methods implementing them but none of it changed the outcome. At this point I think the issue must be somewhere else, but I have no idea what to look out for anymore.

Thanks in advance for your help!

@zhumingpassional
Copy link
Collaborator

stability in RL is an important issue, which is not well solved now.

@zhumingpassional zhumingpassional added the good first issue Good for newcomers label Aug 4, 2024
@zhumingpassional zhumingpassional self-assigned this Aug 4, 2024
@Bottagandalf
Copy link
Author

stability in RL is an important issue, which is not well solved now.

Yeah, that's something I already came across 😢
I was just thinking as I could narrow it down to the Kernel restart, maybe theres something I can do that I didnt think about yet.
Would be cool to be able to compare all my results to each other. If thats not possible, I might just do smaller chunks of tests for each parameter only compare them

@arun-dezerv
Copy link

I have tried this multiple times and I have not faced issues with the reproducibility of the test set. Why does seed matter for the testing set ? It matters only for model training I think. Please let me know if I am wrong

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
good first issue Good for newcomers
Projects
None yet
Development

No branches or pull requests

3 participants