-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ndv dont share mi clr but still lock per bootstrap #63
base: master
Are you sure you want to change the base?
Ndv dont share mi clr but still lock per bootstrap #63
Conversation
Codecov Report
@@ Coverage Diff @@
## master #63 +/- ##
=========================================
- Coverage 70.54% 70.44% -0.1%
=========================================
Files 18 18
Lines 1480 1482 +2
=========================================
Hits 1044 1044
- Misses 436 438 +2
Continue to review full report at Codecov.
|
👍 |
I tried this on NYU HPC. I submitted the following code into the system using
When I ran this code using the original code on the master branch, everything ran fine and the results looked fine. But when I switched to the
...
|
@nickdeveaux any ideas why i'm getting that error? |
Has anybody else tried this? Does it for for anyone else? I am still getting the same error on NYU HPC. This time I was working on the InfereCLaDR branch and I put in the same changes that you did into bbsr_tfa_runner.py manually, and I still got the same error. |
@kostyat @dayanne-castro
Calculating Mi and CLR and sending it to workers was sending a large amount of data to each worker per bootstrap. For example, for a 60k gene by 150 sample input file, the mi and clr matrices summed to .6 GB, and ended up being 1.6 GB of data once they were pickled. This was sent to 70 workers across 20 bootstraps on the cluster, leading to a massive (>10x) slowdown.
Now, each worker calculates mi and clr independently, and needs to wait for a new special key (bootstrap %idx) before moving forward