-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Two questions related to the Fig1 in the paper #48
Comments
Yes. The size of bloom filters is determined by the expected false-positive rate and the length of the largest sequence in a block.
|
Sorry for the misunderstanding regarding "rows" and "columns," thank you so much! |
After a round of filtering, some multiple reads are assigned to fewer species (even one, making them uniquely matched reads), so another round is needed to use the new statistics to recompute detected reference genomes. However, there's no need to use many rounds, two or three is enough. Like the EM algorithm, each round of filtration improves the result, but it will soon reach the plateau. In R3, the grey diamond criterion is not used cause it is slow, and there's no need according to my observation.
Each block contains data from different genome chunks. In the figure, chunks of a genome are next to each other for simplicity. Actually, k-mers of all chunks are sorted, and divided into multiple blocks. You can also read the COBS paper.
They might be different or the same.
Yes.
It is. |
Hi! I have two questions about Fig 1 in the paper.
The first one is about the indexing part. In Block 1, the number of columns is n-1. What exactly does this n-1 refer to? Does the n-1 (number of columns) change depending on different blocks?
The other question is about the filtering step in the Profiling part. Why should KMCP have three filtering steps? Why not just use the second filter step (the most rigorous round) so that KMCP can only have one round filtering step?
These are the questions that I am concerned about. Hope you can reply as soon as possible. Thank you!
The text was updated successfully, but these errors were encountered: