Hi,
A collaborator of mine is working on a new software and we while we were
doing a sanity check to compare the base-level coverage from BAM files and
bigWig files generated from his software we realized that by default bases
corresponding to a 'D' on the CIGAR string get counted when reading
coverage from BAM files.
That is:
[1] ?1.2.0?
You can see that this is default elsewhere, for example:
extractAlignmentRangesOnReference
function (cigar, pos = 1L, drop.D.ranges = FALSE, f = NULL)
{
if (!isTRUEorFALSE(drop.D.ranges))
stop("'drop.D.ranges' must be TRUE or FALSE")
if (drop.D.ranges) {
ops <- c("M", "=", "X", "I")
}
else {
ops <- c("M", "=", "X", "I", "D")
}
cigarRangesAlongReferenceSpace(cigar, flag = NULL, pos = pos,
f = f, ops = ops, drop.empty.ranges = FALSE, reduce.ranges = TRUE)
}
<environment: namespace:GenomicAlignments>
What is the rationale for setting `drop.D.ranges` by default to FALSE?
Thanks,
Leo