-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Run the pipeline on a small / medium / large genome #11
Comments
Avocado is always failing at the moment... |
what's the error message? |
(let's leave comments here, for future tracking) François Langelier 2:35 PM (22 minutes ago) |
Potential solution: https://plumbr.eu/outofmemoryerror/gc-overhead-limit-exceeded |
If you can run it with jvisualvm attached you will have more idea of what What JVM parameters you are currently using ? -tony On Mon, Mar 23, 2015 at 3:13 PM, David Lauzon [email protected]
|
Currently, it is the default settings. I think there is no enough memory allocated to the spark workers. Francois will is describing the steps-by-steps to reproduce the problem. |
@flangelier I think I encountered the same problem as you with Avocado! Change the values of the two lines below in the file bin/avocado-submit. 4g is probably more than what you have available in memory and the JVM can't start.
See: https://github.com/bigdatagenomics/avocado/blob/master/bin/avocado-submit#L56-58 |
With updated versions of the BDGenomics pipeline (snap, adam, avocado), use a small / medium / large genome to validate the new images and the orchestration scripts. We'll compare these results with the results from S. Bonami and the BDGenomics papers.
References
TODO: Find the data used in the Snap / Adam / Avocado papers.
The text was updated successfully, but these errors were encountered: