duplicate reads #500
Unanswered
INFINITY1993
asked this question in
Q&A
Replies: 1 comment 3 replies
-
Many metagenoics libraries are prepared with a small amplification step. Therefore there might be some reads that are artificial duplicates. This step removes them it also sorts and compresses the reads so that the dastq gz file is up to 30% smaller, according to the claim. Do you have a reason for nod-removing the deduplicaded reads? In my experience the deduplicaded reads are only an tiny fraction. But if you have low biomass samples they might be higher. I suggest you to see also the docs of the tool in question, clumpify |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, dear developer,
I noticed that you integrated reads deduplication in the QC step. What I wanna ask is that is it really necessary to do deduplication? what are the benefits or drawbacks? Thanks in advance!
Best regards!
Beta Was this translation helpful? Give feedback.
All reactions