Description Usage Arguments Value Author(s)
Split the bins into big and small chunks. A big chunk represents all the bins used to normalize bins of this chunk. Small chunk represents the bins to analyze by one job on the cluster. While the size of the small chunks is not important and can be adjusted to fit the cluster, the big chunk size will impact on the efficiency of the normalization (the bigger the better).
1 2 |
bins.df |
a data.frame with the bins definition (one row per bin). E.g. created by 'fragment.genome.hg19'. |
bg.chunk.size |
the number of bins in a big chunk. |
sm.chunk.size |
the number of bins in a small chunk. |
large.chr.chunks |
should the big chunks be made of a few large genomic sub-regions ? Default is false. Normalization is faster (but a bit less efficient) than when using random bins. Recommended when dealing with a large number of bins. |
an updated data.frame with new columns 'sm.chunk', 'bg.chunk' and 'bin' with the small chunk ID, big chunk ID and bin definition.
Jean Monlong
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.