R/basic_functions.R

Defines functions correct_rounded round_precision test_exposureAffected find_affected_PIDs compare_exposures melt_exposures merge_exposures add_as_fist_to_list split_exposures_by_subgroups shapiro_if_possible make_subgroups_df annotate_intermut_dist_cohort annotate_intermut_dist_PID attribute_nucleotide_exchanges translate_to_1kG translate_to_hg19 stderrmean_over_present stderrmean sum_over_list_of_df sd_over_present average_over_present normalize_df_per_dim repeat_df deriveSigInd_df relateSigs disambiguateVector uniquify compare_expousre_sets compare_sets vplayout cosineMatchDist cosineDist

Documented in add_as_fist_to_list annotate_intermut_dist_cohort annotate_intermut_dist_PID attribute_nucleotide_exchanges average_over_present compare_exposures compare_expousre_sets compare_sets correct_rounded cosineDist cosineMatchDist deriveSigInd_df disambiguateVector find_affected_PIDs make_subgroups_df melt_exposures merge_exposures normalize_df_per_dim relateSigs repeat_df round_precision sd_over_present shapiro_if_possible split_exposures_by_subgroups stderrmean stderrmean_over_present sum_over_list_of_df test_exposureAffected translate_to_1kG translate_to_hg19

# Copyright © 2014-2019  The YAPSA package contributors
# This file is part of the YAPSA package. The YAPSA package is licenced under
# GPL-3

#' Compute the cosine distance of two vectors
#'
#' @param a,b Numerical vectors of same length
#' @return The scalar product of the two input vectors divided by the
#'        product of the norms of the two input vectors
#' 
#' @examples
#' ## 1. Orthogonal vectors:
#' cosineDist(c(1,0),c(0,1))
#' ## 2. Non-orthogonal vectors:
#' cosineDist(c(1,0),c(1,1))
#' ## Compare trigonometry:
#' 1-cos(pi/4)
#' 
#' @export
#' 
cosineDist <- function(a,b){
    return(1 - sum(a * b) / ( sqrt(sum(a * a)) * sqrt(sum(b * b)) )) 
}


#' Compute an altered cosine distance of two vectors
#'
#' This is an altered cosine distance: it first reduced the dimension of the
#' two input vectors to only those coordinates where both have non-zero entries.
#' The cosine similarity is then computed on these reduced vectors, i.e. on a
#' sub-vector space.
#'
#' @param a,b Numerical vectors of same length
#' @return The scalar product of the reduced input vectors divided by the
#'        product of the norms of the two reduced input vectors
#' 
#' @examples
#' ## 1. Orthogonal vectors:
#' cosineMatchDist(c(1,0),c(0,1))
#' ## 2. Non-orthogonal vectors:
#' cosineMatchDist(c(1,0),c(1,1))
#' 
#' @export
#' 
cosineMatchDist <- function(a, b){
    choice_ind <- intersect(which(a != 0), which(b != 0))
    if(length(choice_ind) == 0) return(1)
    return(cosineDist(a[choice_ind], b[choice_ind]))
}


#' @import grid
#' 
vplayout <- function(x, y) viewport(layout.pos.row = x, layout.pos.col = y)


#' Compare two sets of signatures by cosine distance
#'
#' Compare two sets of signatures, stored in numerical data frames
#' \code{W1} and \code{W2}, by computing the column-wise cosine distance
#'
#' @param in_df_small,in_df_big Numerical data frames \code{W1} and 
#'  \code{W2}, ideally the bigger one first, both with \code{n} rows
#'  and \code{l1} and \code{l2} columns, \code{n} being the number of
#'  features and \code{l1} and \code{l2} being the respective numbers
#'  of signatures of \code{W1} and \code{W2}
#' @param in_distance A function which computes the distance measure, default
#'  is \code{\link{cosineDist}}
#'  
#' @return A list with entries
#'  \code{distance},
#'  \code{hierarchy_small} and
#'  \code{hierarchy_big}.
#' \itemize{
#'  \item \code{distance}:
#'    A numerical data frame with the cosine distances between the
#'    columns of \code{W1}, indexing the rows, and \code{W2}, indexing
#'    the columns
#'  \item \code{hierarchy_small}:
#'    A data frame carrying the information of ranked similarity between
#'    the signatures in \code{W2} with the signatures in \code{W1}
#'  \item \code{hierarchy_big}:
#'    A data frame carrying the information of ranked similarity between
#'    the signatures in \code{W1} with the signatures in \code{W2}
#' }
#' 
#' @seealso \code{\link{cosineDist}}
#' 
#' @examples
#' sig_1_df <- data.frame(matrix(c(1,0,0,0,0,1,0,0,0,0,1,0),ncol=3))
#' names(sig_1_df) <- paste0("B",seq_len(dim(sig_1_df)[2]))
#' sig_2_df <- data.frame(matrix(c(1,1,0,0,0,0,1,1),ncol=2))
#' compare_sets(sig_1_df,sig_2_df)
#' 
#' @export
#' 
compare_sets <- function(in_df_small,in_df_big,
                            in_distance = cosineDist) {
    sig_names_small <- colnames(in_df_small)
    sig_names_big <- colnames(in_df_big)
    number_of_sigs_small <- dim(in_df_small)[2]
    number_of_sigs_big <- dim(in_df_big)[2]
    number_of_features_small <- dim(in_df_small)[1]
    number_of_features_big <- dim(in_df_big)[1]
        if(number_of_features_small != number_of_features_big) {
            cat("compare_sets: Error: features mismatch")
            return()
    }
    temp_ind <- match(rownames(in_df_small),rownames(in_df_big))
    df_big <- in_df_big[temp_ind,]
    df_small <- in_df_small
    distance_df <- as.data.frame(
        matrix(rep(1,number_of_sigs_small*number_of_sigs_big),
            ncol=number_of_sigs_big))
    rownames(distance_df) <- sig_names_small
    colnames(distance_df) <- sig_names_big
    for (i in 1:number_of_sigs_small) {
        for (j in 1:number_of_sigs_big) {
        distance_df[i,j] <- in_distance(df_small[,i],df_big[,j])
        }
    }  
    hierarchy_small_df <- 
        t(apply(distance_df,1,
            function(l) return(colnames(distance_df)[order(l)])))
    hierarchy_big_df <- 
        t(apply(distance_df,2,
            function(l) return(rownames(distance_df)[order(l)])))
    return(list(distance=distance_df,
            hierarchy_small=hierarchy_small_df,
            hierarchy_big=hierarchy_big_df))
}

#' Compare two sets of exposures by cosine distance
#'
#' Compare two sets of exposures, stored in numerical data frames
#' \code{H1} and \code{H2}, by computing the row-wise cosine distance
#'
#' @param in_df_small,in_df_big Numerical data frames \code{H1} and 
#'  \code{H2}, ideally the bigger one first, both with \code{l} rows
#'  and \code{m1} and \code{m2} columns, \code{l} being the number of
#'  signatures and \code{m1} and \code{m2} being the respective numbers
#'  of samples or patient identefier of \code{H1} and \code{H2}
#' @param in_distance A function which computes the distance measure, default
#'  is \code{\link{cosineDist}}
#'  
#' @return A list with entries
#'  \code{distance},
#'  \code{hierarchy_small} and
#'  \code{hierarchy_big}.
#' \itemize{
#'  \item \code{distance}:
#'    A numerical data frame with the cosine distances between the
#'    columns of \code{H1}, indexing the rows, and \code{H2}, indexing
#'    the columns
#'  \item \code{hierarchy_small}:
#'    A data frame carrying the information of ranked similarity between
#'    the signatures in \code{H2} with the signatures in \code{H1}
#'  \item \code{hierarchy_big}:
#'    A data frame carrying the information of ranked similarity between
#'    the signatures in \code{H1} with the signatures in \code{H2}
#' }
#' 
#' @seealso \code{\link{cosineDist}}
#' 
#' @examples
#' sig_1_df <- data.frame(matrix(c(1,0,0,0,0,1,0,0,0,0,1,0),ncol=3))
#' names(sig_1_df) <- paste0("B",seq_len(dim(sig_1_df)[2]))
#' sig_2_df <- data.frame(matrix(c(1,1,0,0,0,0,1,1),ncol=2))
#' compare_expousre_sets(sig_1_df,sig_2_df)
#' 
#' @export
#' 
compare_expousre_sets <-  function(in_df_small,in_df_big,
                                    in_distance = cosineDist) {
    sig_names_small <- colnames(in_df_small)
    sig_names_big <- colnames(in_df_big)
    number_of_sigs_small <- dim(in_df_small)[2]
    number_of_sigs_big <- dim(in_df_big)[2]
    number_of_samples_small <- dim(in_df_small)[1]
    number_of_samples_big <- dim(in_df_big)[1]
    if(number_of_samples_small != number_of_samples_big) {
        cat("compare_sets: Error: features mismatch")
        return()
    }
    temp_ind <- match(rownames(in_df_small),rownames(in_df_big))
    df_big <- in_df_big[temp_ind,]
    df_small <- in_df_small
    distance_df <- as.data.frame(
        matrix(rep(1,number_of_sigs_small*number_of_sigs_big),
                ncol=number_of_sigs_big))
    rownames(distance_df) <- sig_names_small
    colnames(distance_df) <- sig_names_big
    for (i in 1:number_of_sigs_small) {
        for (j in 1:number_of_sigs_big) {
            distance_df[i,j] <- in_distance(df_small[,i],df_big[,j])
        }
    }  
    hierarchy_small_df <- 
        t(apply(distance_df,1,function(l) 
            return(colnames(distance_df)[order(l)])))
    hierarchy_big_df <- 
        t(apply(distance_df,2,function(l)
            return(rownames(distance_df)[order(l)])))
    return(list(distance=distance_df,
                hierarchy_small=hierarchy_small_df,
                hierarchy_big=hierarchy_big_df))
}

uniquify<-function(x) if (length(x) == 1) x else sprintf("%s_%02d", x, 
                                                        seq_along(x))


#' Disambiguate a vector
#' 
#' Add numbered suffixes to redundant entries in a vector
#'
#' @param in_vector Input vector
#'
#' @return The disambiguated vector.
#' @export
#'
#' @examples
#'  NULL
#' 
disambiguateVector <- function(in_vector){
    ave(in_vector, in_vector, FUN = uniquify)
}


#' Make unique assignments between sets of signatures
#' 
#' Make unique assignments between a set of given signatures and a set of new
#' signatures.
#'
#' @param querySigs The signatures to compare to (given signatures).
#' @param subjectSigs The signatures to be compared (new signatures).
#'
#' @return A list of comparison vectors
#' @export
#'
#' @seealso \code{\link{compare_sets}}
#' @seealso \code{\link{disambiguateVector}}
#' 
#' @examples
#'  NULL
#' 
relateSigs <- function(querySigs, subjectSigs){
    compareList <- compare_sets(querySigs, subjectSigs)
    queryMatchVector <- compareList$hierarchy_small[, 1]
    subjectMatchVector <- compareList$hierarchy_big[, 1]
    subjectUniqueVector <- disambiguateVector(subjectMatchVector)
    uniqueInd <- 
        which(queryMatchVector[subjectMatchVector] == names(subjectMatchVector))
    subjectMatchUniqueVector <- subjectMatchVector
    subjectMatchUniqueVector[-uniqueInd] <- 
    paste0(subjectUniqueVector[-uniqueInd], "_mm")
    uniqueMatchInd <- 
        which(queryMatchVector[subjectMatchVector] == 
            names(subjectMatchVector) | 
            subjectMatchVector == subjectUniqueVector)
    subjectUniqueMatchVector <- subjectMatchVector
    subjectUniqueMatchVector[-uniqueMatchInd] <- 
        paste0(subjectUniqueVector[-uniqueMatchInd], "_mm")
    distVector <- 
        unlist(lapply(seq_along(subjectMatchVector), function(currentInd) 
    compareList$distance[subjectMatchVector[currentInd],
        names(subjectMatchVector)[currentInd]]))
    return(list(match = subjectMatchVector,
                unique = subjectUniqueVector,
                matchUnique = subjectMatchUniqueVector,
                uniqueMatch = subjectUniqueMatchVector,
                dist = distVector))
}


#' Derive a signature_indices_df object
#'
#' Derive a data frame of type signature_indices_df (additional information for
#' a set of signatures) from a set of given signatures for a set of new
#' signatures.
#'
#' @param querySigs The signatures to compare to (given signatures).
#' @param subjectSigs The signatures to be compared (new signatures).
#'   Alternatively this may be a complex object of type list and contain data
#'   from different deconvolutions, each of which having a set of signaturen to
#'   be compared.
#' @param querySigInd The object of type signature_indices_df (additional
#'   informatio for a set of signatures) belonging to the set of known
#'   signatures.
#' @param in_sort Whether to sort or not
#'
#' @return An object of type signature_indices_df (additional informatio for a
#'   set of signatures) belonging to the set of new signatures.
#' @export
#'
#' @seealso \code{\link{relateSigs}}
#'
#' @examples
#'  NULL
#' 
deriveSigInd_df <- function(querySigs, subjectSigs, 
                            querySigInd = NULL,
                            in_sort = FALSE){
    if(inherits(subjectSigs, "list")){
        if(all(unlist(lapply(subjectSigs, function(current_NMFlist){
            "signatures" %in% names(current_NMFlist)})))){
    outListsList <- lapply(subjectSigs, function(current_NMFlist){
        temp_list <- deriveSigInd_df(querySigs,
                                    current_NMFlist$signatures,
                                    querySigInd,in_sort)
        current_NMFlist$out_sig_ind_df <- temp_list$df
        if(in_sort){
            current_NMFlist$exposures <- 
                current_NMFlist$exposures[temp_list$ind,]
            current_NMFlist$norm_exposures <- 
                current_NMFlist$norm_exposures[temp_list$ind,]
            current_NMFlist$signatures <- 
                current_NMFlist$signatures[, temp_list$ind]
        }
        return(current_NMFlist)
    })
    return(outListsList)
    } else {
        cat("YAPSA:::deriveSigInd_df::error: wrong input type. Abort.\n")
    }
    } else if(inherits(subjectSigs, "data.frame")){
        reorder_ind = NULL
        relateList <- relateSigs(querySigs, subjectSigs)
        sigInd_df <- data.frame(sig = colnames(subjectSigs),
                                index = seq(dim(subjectSigs)[2]),
                                colour = rainbow(dim(subjectSigs)[2]),
                                process = relateList$uniqueMatch,
                                match = relateList$uniqueMatch,
                                dist = relateList$dist)
    if(!is.null(querySigInd)){
        if("sig" %in% names(querySigInd)){
            match_ind <- match(relateList$uniqueMatch, querySigInd$sig)
        if("process" %in% names(querySigInd)){
            sigInd_df$process = querySigInd$process[match_ind]
            na_ind <- which(is.na(sigInd_df$process))
            sigInd_df$process <- as.character(sigInd_df$process)
            sigInd_df$process[na_ind] <- "nonunique"
        }
        if("colour" %in% names(querySigInd)){
            sigInd_df$colour = querySigInd$colour[match_ind]
            na_ind <- which(is.na(sigInd_df$colour))
            valeurVector <- 
                floor(seq(from = 0, to = length(na_ind) - 1)*100/length(na_ind))
            sigInd_df$colour <- as.character(sigInd_df$colour)
            sigInd_df$colour[na_ind] <- paste0("grey",valeurVector)
        }
        if("index" %in% names(querySigInd)){
            sigInd_df$index = querySigInd$index[match_ind]
            na_ind <- which(is.na(sigInd_df$index))
            sigInd_df$index <- as.numeric(as.character(sigInd_df$index))
            #tempIndex <- as.numeric(as.character(sigInd_df$index))
            tempMax <- max(sigInd_df$index, na.rm = TRUE)
            sigInd_df$index[na_ind] <- 
                seq(from = tempMax + 1, to = tempMax + length(na_ind))
            if(in_sort){
                reorder_ind <- order(as.numeric(as.character(sigInd_df$index)))
                sigInd_df <- sigInd_df[reorder_ind,]
            }
        }
    }
}
    return(list(df = sigInd_df,
                ind = reorder_ind))
    }
}


#' Create a data frame with default values
#'
#' @param in_value Default entry to be repeated in the data frame
#' @param in_rows,in_cols Dimensions of the data frame to be created
#' @return The created data frame
#'
#' @examples
#' ## 1. Initialize with numeric value:
#' repeat_df(1,2,3)
#' ## 2. Initialize with NA value:
#' repeat_df(NA,3,2)
#' ## 3. Initialize with character:
#' repeat_df("a",4,3)
#'
#' @export
#' 
repeat_df <- function(in_value,in_rows,in_cols) {
    return(as.data.frame(matrix(rep(in_value,in_cols*in_rows),ncol=in_cols)))
}


#' Useful functions on data frames
#'
#' \code{normalize_df_per_dim}: Normalization is carried out by dividing by
#' \code{rowSums} or \code{colSums}; for rows with \code{rowSums=0} or columns
#' with \code{colSums=0}, the normalization is left out.
#'
#' @param in_df Data frame to be normalized
#' @param in_dimension Dimension along which the operation will be carried out
#' @return The normalized numerical data frame (\code{normalize_df_per_dim})
#'
#' @examples
#' test_df <- data.frame(matrix(c(1,2,3,0,5,2,3,4,0,6,0,0,0,0,0,4,5,6,0,7),
#'                              ncol=4))
#' ## 1. Normalize over rows:
#' normalize_df_per_dim(test_df,1)
#' ## 2. Normalize over columns:
#' normalize_df_per_dim(test_df,2)
#'
#' @export
#' 
normalize_df_per_dim <- function(in_df,in_dimension) {
    out_df <- repeat_df(in_value = 0,in_rows = dim(in_df)[1],
                        in_cols = dim(in_df)[2])
    if(in_dimension==1) {
        choice_ind <- which(rowSums(in_df)>0)
        if(length(choice_ind) > 0){
            out_df[choice_ind,] <- in_df[choice_ind,]/rowSums(in_df)[choice_ind]
    } else {
        warning("YAPSA:::normalize_df_per_dim::warning: input is all zero.\n",
            "Return input as output.\n")
        out_df <- in_df
    }
    } else if (in_dimension==2) {
    t_df <- t(in_df)
    choice_ind <- which(rowSums(t_df)>0)
    if(length(choice_ind) > 0){
        temp_df <- repeat_df(in_value = 0,in_rows = dim(in_df)[2],
                                in_cols = dim(in_df)[1])
        temp_df[choice_ind,] <- t_df[choice_ind,]/rowSums(t_df)[choice_ind]
        out_df <- as.data.frame(t(temp_df))
    } else {
        warning("YAPSA:::normalize_df_per_dim::warning: input is all zero.\n",
            "Return input as output.\n")
        out_df <- in_df
    }
    } else {
        return(NULL)
    }
    colnames(out_df) <- colnames(in_df)
    rownames(out_df) <- rownames(in_df)
    return(out_df)
}


#' Average a data frame over a specified dimension
#'
#' \code{average_over_present}: 
#' If averaging over columns, zero rows (i.e. those with \code{rowSums=0})
#' are left out, if averaging over rows, zero columns (i.e. those with
#' \code{colSums=0}) are left out.
#'
#' @return A vector of the means (\code{average_over_present})
#' 
#' @examples
#' test_df <- data.frame(matrix(c(1,2,3,0,5,2,3,4,0,6,0,0,0,0,0,4,5,6,0,7),
#'                              ncol=4))
#' ## 1. Average over non-zero rows:
#' average_over_present(test_df,1)
#' ## 2. Average over non-zero columns:
#' average_over_present(test_df,2)
#' 
#' @export
#' @rdname normalize_df_per_dim
#' 
average_over_present <- function(in_df,in_dimension) {
    out_vector <- c()
    if(in_dimension==2) {
        row_ind <- which(rowSums(in_df)>0)
        reduced_df <- in_df[row_ind,]
    if (length(row_ind)==0) {
        return(NULL)
    } else if (length(row_ind)==1) {
        out_vector <- reduced_df
    } else {
        out_vector <- apply(reduced_df,2,mean)      
    }
    } else if (in_dimension==1) {
        col_ind <- which(colSums(in_df)>0)
        reduced_df <- in_df[,col_ind]
        if (length(col_ind)==0) {
            return(NULL)
    } else if (length(col_ind)==1) {
        out_vector <- reduced_df
    } else {
        out_vector <- apply(reduced_df,1,mean)      
    }
    } else {
        return(NULL)
    }
    return(out_vector)
}


#' Standard deviation of a data frame over a specified dimension
#'
#' \code{sd_over_present}: 
#' If computing the standard deviation over columns, zero rows
#' (i.e. those with \code{rowSums=0}) are left out, if computing
#' the standard deviation over rows, zero columns (i.e. those with
#' \code{colSums=0}) are left out.
#'
#' @return A vector of the standard deviations (\code{sd_over_present})
#' 
#' @examples
#' test_df <- data.frame(matrix(c(1,2,3,0,5,2,3,4,0,6,0,0,0,0,0,4,5,6,0,7),
#'                              ncol=4))
#' ## 1. Compute standard deviation over non-zero rows:
#' sd_over_present(test_df,1)
#' ## 2. Compute standard deviation over non-zero columns:
#' sd_over_present(test_df,2)
#' 
#' @export
#' @rdname normalize_df_per_dim
#' 
sd_over_present <- function(in_df,in_dimension) {
    out_vector <- c()
    if(in_dimension==2) {
        row_ind <- which(rowSums(in_df)>0)
        reduced_df <- in_df[row_ind,]
        if (length(row_ind)==0) {
            return(NULL)
    } else if (length(row_ind)==1) {
        out_vector <- rep(0,dim(in_df)[1])
    } else {
        out_vector <- apply(reduced_df,2,sd)      
    }
    } else if (in_dimension==1) {
        col_ind <- which(colSums(in_df)>0)
        reduced_df <- in_df[,col_ind]
        if (length(col_ind)==0) {
            return(NULL)
    } else if (length(col_ind)==1) {
        out_vector <- rep(0,dim(in_df)[2])
    } else {
        out_vector <- apply(reduced_df,1,sd)      
    }
    } else {
        return(NULL)
    }
    return(out_vector)
}


#' Elementwise sum over a list of (numerical) data frames
#'
#' @param in_df_list
#'  List of (numerical) data frames
#' @return A numerical data frame with the same dimensions as the entries of
#'  \code{in_df_list} with elementwise sums
#' 
#' @examples
#' A <- data.frame(matrix(c(1,1,1,2,2,2),ncol=2))
#' B <- data.frame(matrix(c(3,3,3,4,4,4),ncol=2))
#' df_list <- list(A=A,B=B)
#' sum_over_list_of_df(df_list)
#' 
#' @export
#' 
sum_over_list_of_df <- function(in_df_list) {
    sum_df <- repeat_df(0,dim(in_df_list[[1]])[1],dim(in_df_list[[1]])[2])
    for (i in seq_len(length(in_df_list))) {
        sum_df <- sum_df + in_df_list[[i]]
    }
    colnames(sum_df) <- colnames(in_df_list[[1]])
    rownames(sum_df) <- rownames(in_df_list[[1]])
    return(sum_df)
}


#' Compute the standard error of the mean
#'
#' This function returns the standard deviation of an input numerical vector 
#' divided by the square root of the length of the input vector
#' 
#' @param x
#'  A numerical vector
#'  
#' @return
#'  Standard deviation of an input numerical vector divided by the square
#'  root of the length of the input vector
#'
#' @examples
#' A <- c(1,2,3)
#' sd(A)
#' stderrmean(A)
#'  
#' @export
#'  
stderrmean <- function(x) sqrt(var(x,na.rm=TRUE)/length(na.omit(x)))


#' Standard error of the mean of a data frame over a specified dimension
#'
#' \code{stderrmean_over_present}: 
#' If computing the standard error of the mean over columns, zero rows
#' (i.e. those with \code{rowSums=0}) are left out, if computing the
#' standard error of the mean over rows, zero columns (i.e. those with
#' \code{colSums=0}) are left out. Uses the function \code{\link{stderrmean}}
#'
#' @return A vector of the standard errors of the mean 
#'  (\code{stderrmean_over_present})
#' 
#' @seealso \code{\link{stderrmean}}
#' 
#' @examples
#' test_df <- data.frame(matrix(c(1,2,3,0,5,2,3,4,0,6,0,0,0,0,0,4,5,6,0,7),
#'                              ncol=4))
#' ## 1. Compute standard deviation over non-zero rows:
#' stderrmean_over_present(test_df,1)
#' ## 2. Compute standard deviation over non-zero columns:
#' stderrmean_over_present(test_df,2)
#' 
#' @export
#' @rdname normalize_df_per_dim
#' 
stderrmean_over_present <- function(in_df,in_dimension) {
    out_vector <- c()
    if(in_dimension==2) {
        row_ind <- which(rowSums(in_df)>0)
        reduced_df <- in_df[row_ind,]
    if (length(row_ind)==0) {
        return(NULL)
    } else if (length(row_ind)==1) {
        out_vector <- rep(0,dim(in_df)[1])
    } else {
        out_vector <- apply(reduced_df,2,stderrmean)      
    }
    } else if (in_dimension==1) {
        col_ind <- which(colSums(in_df)>0)
        reduced_df <- in_df[,col_ind]
    if (length(col_ind)==0) {
        return(NULL)
    } else if (length(col_ind)==1) {
        out_vector <- rep(0,dim(in_df)[2])
    } else {
        out_vector <- apply(reduced_df,1,stderrmean)      
    }
    } else {
        return(NULL)
    }
    return(out_vector)
}


#' Translate chromosome names to the hg19 naming convention
#'
#' \code{translate_to_hg19}: In hg19 naming convention, chromosome names start
#' with the prefix \emph{chr} and the gonosomes are called \emph{X} and
#' \emph{Y}. If data analysis is performed e.g. with
#' \code{\link[BSgenome.Hsapiens.UCSC.hg19]{BSgenome.Hsapiens.UCSC.hg19}}, this
#' naming convention is needed. The inverse transform is done with
#' \code{\link{translate_to_1kG}}.
#'
#' @param in_dat GRanges object, VRanges object or data frame which carries one
#'   column with chromosome information to be reformatted.
#' @param in_CHROM.field String indicating which column of \code{in_dat} carries
#'   the chromosome information
#' @param in_verbose Whether verbose or not.
#'
#' @return GRanges object, VRanges object or data frame identical to
#'   \code{in_dat}, but with the names in the chromosome column replaced (if
#'   dealing with data frames) or alternatively the seqlevels replaced (if
#'   dealing with GRanges or VRanges objects).
#'
#' @examples
#' test_df <- data.frame(CHROM=c(1,2,23,24),POS=c(100,120000000,300000,25000),
#'                       dummy=c("a","b","c","d"))
#' hg19_df <- translate_to_hg19(test_df, in_CHROM.field = "CHROM")
#' hg19_df
#'
#' @importFrom GenomeInfoDb seqlevels seqlevels<-
#' @export
#' 
translate_to_hg19 <- function(in_dat,
                            in_CHROM.field="CHROM",
                            in_verbose = FALSE) {
    out_dat <- in_dat
    if(inherits(in_dat, "GRanges")) {
        seqlevels(out_dat) <- gsub("23", "X", seqlevels(out_dat))
        seqlevels(out_dat) <- gsub("24", "Y", seqlevels(out_dat))
        seqlevels(out_dat) <- paste0("chr", seqlevels(out_dat))
    } else if(inherits(in_dat, "data.frame") & 
        in_CHROM.field %in% names(in_dat)){
    chrom_ind <- which(names(out_dat)==in_CHROM.field)
    #names(out_dat)[chrom_ind] <- "chr"
    out_dat[,chrom_ind] <- as.character(out_dat[,chrom_ind])
    out_dat[,chrom_ind] <- gsub("23","X",out_dat[,chrom_ind])
    out_dat[,chrom_ind] <- gsub("24","Y",out_dat[,chrom_ind])
    logical_chr_vector <- grepl("^chr.*",out_dat[,chrom_ind])
    if(!(all(logical_chr_vector))) {
        replace_ind <- which(!logical_chr_vector)
        out_dat[replace_ind,chrom_ind] <- 
            paste0("chr",out_dat[replace_ind,chrom_ind])
    }
    } else {
        if(in_verbose) cat("YAPSA:::translate_to_hg19::error: input not of",
            " suitable type. Return original object.\n")
    }
    return(out_dat)
}


#' Translate chromosome names to the 1kG naming convention
#'
#' \code{translate_to_1kG}: In 1kG, i.e. 1000 genomes naming convention, 
#' chromosome names have no prefix \emph{chr} and the gonosomes are called 
#' \emph{23} for \emph{X} and \emph{24} for \emph{Y}. If data analysis is 
#' performed e.g. with \code{hs37d5.fa}, this naming convention is needed. The
#' inverse transform is done with \code{\link{translate_to_hg19}}.
#'
#' @examples
#' test_df <- data.frame(CHROM=c(1,2,23,24),POS=c(100,120000000,300000,25000),
#'                       dummy=c("a","b","c","d"))
#' hg19_df <- translate_to_hg19(test_df, in_CHROM.field = "CHROM")
#' onekG_df <- translate_to_1kG(hg19_df, in_CHROM.field = "CHROM")
#' onekG_df
#' 
#' @importFrom GenomeInfoDb seqlevels seqlevels<-
#' @export
#' @rdname translate_to_hg19
#' 
translate_to_1kG <- function(in_dat,
                            in_CHROM.field = "chr",
                            in_verbose = FALSE) {
    out_dat <- in_dat
    # account for input data type
    if(inherits(in_dat, "GRanges")) {
        seqlevels(out_dat) <- gsub("chr", "", seqlevels(out_dat))
        seqlevels(out_dat) <- gsub("X", "23", seqlevels(out_dat))
        seqlevels(out_dat) <- gsub("Y", "24", seqlevels(out_dat))
    } else if(inherits(in_dat, "data.frame") & 
        in_CHROM.field %in% names(in_dat)){
        chrom_ind <- which(names(out_dat)==in_CHROM.field)
        #names(out_dat)[chrom_ind] <- "CHROM"
        out_dat[,chrom_ind] <- as.character(out_dat[,chrom_ind])
        out_dat[,chrom_ind] <- gsub("X","23",out_dat[,chrom_ind])
        out_dat[,chrom_ind] <- gsub("Y","24",out_dat[,chrom_ind])
        logical_chr_vector <- grepl("^[0-9]+$",out_dat[,chrom_ind])
        if(!(all(logical_chr_vector))) {
            replace_ind <- which(!logical_chr_vector)
            out_dat[replace_ind,chrom_ind] <- gsub("chr","",
                                                out_dat[replace_ind,chrom_ind])
    }
    } else {
        if(in_verbose) cat("YAPSA:::translate_to_1kG::error: input not of",
            " suitable type. Return original object.\n")
    }
        return(out_dat)
}


#' Attribute the nucleotide exchange for an SNV
#'
#' SNVs are grouped into 6 different categories (12/2 as reverse complements are
#' summed over). This function defines the attribution.
#'
#' @param in_dat VRanges object or data frame which carries one column for the
#'   reference base and one column for the variant base
#' @param in_REF.field String indicating which column of \code{in_dat} carries
#'   the reference base if dealing with data frames
#' @param in_ALT.field String indicating which column of \code{in_dat} carries
#'   the variant base if dealing with data frames
#' @param in_verbose Whether verbose or not.
#'
#' @return A character vector with as many rows as there are in \code{in_dat}
#'   which can be annotated (i.e. appended) to the input data frame.
#'
#' @examples
#' test_df <- data.frame(
#'  CHROM=c(1,1,1,2,2,2,3,3,3,4,4,4,5,5),
#'  POS=c(1,2,3,4,5,6,1,2,3,4,5,6,7,8),
#'  REF=c("C","C","C","T","T","T","A","A","A","G","G","G","N","A"),
#'  ALT=c("A","G","T","A","C","G","C","G","T","A","C","T","A","N"))
#' test_df$change <- attribute_nucleotide_exchanges(
#'  test_df,in_REF.field = "REF",in_ALT.field = "ALT")
#' test_df
#'
#' @export
#' 
attribute_nucleotide_exchanges <- function(in_dat, in_REF.field = "REF",
                                            in_ALT.field = "ALT",
                                            in_verbose = FALSE) {
    # account for input data type
    if(inherits(in_dat, "VRanges")) {
        choice_column_vector <- c("seqnames", "start", "ref", "alt")
    in_dat <- as.data.frame(in_dat)
    colum_names <- intersect(names(in_dat),choice_column_vector)
    in_dat <- in_dat[,colum_names]
    names(in_dat) <- gsub("ref", in_REF.field, names(in_dat))
    names(in_dat) <- gsub("alt", in_ALT.field, names(in_dat))
    }
    if(!(inherits(in_dat, "data.frame"))){
        if(in_verbose) cat(
            "YAPSA:::attribute_nucleotide_exchanges::error:input",
            " data is of wrong type.\n")
    return(NULL)
    }
    name_list <- names(in_dat)
    ## exception handling for input fields
    if(tolower(in_REF.field) %in% tolower(name_list)) {
        if(in_verbose) cat(
            "YAPSA:::attribute_nucleotide_exchanges::in_REF.field",
            " found. Retrieving REF information.\n")
    REF_ind <- min(which(tolower(name_list)==tolower(in_REF.field)))
    } else {
        if(in_verbose) cat("YAPSA:::attribute_nucleotide_exchanges::error: ",
            "in_REF.field not found. Return NULL.\n")
    return(NULL)
    }
    if(tolower(in_ALT.field) %in% tolower(name_list)) {
        if(in_verbose) cat(
            "YAPSA:::attribute_nucleotide_exchanges::in_ALT.field",
            " found. Retrieving ALT information.\n")
    ALT_ind <- min(which(tolower(name_list)==tolower(in_ALT.field)))
    } else {
        if(in_verbose) cat("YAPSA:::attribute_nucleotide_exchanges::error: ",
            "in_ALT.field not found. Return NULL.\n")
    return(NULL)
    }
    ## adapt nomenclature for nucleotide exchanges
    complement <- c('A' = 'T', 'C' = 'G','G' = 'C', 'T' = 'A', 'N' = 'N')
    change_vec <- rep("dummy",dim(in_dat)[1])
    sel <- which(in_dat[,REF_ind] == 'C' | in_dat[,REF_ind] == 'T')
    change_vec[sel] <- 
        paste0(as.character(in_dat[sel,REF_ind]),
                as.character(in_dat[sel,ALT_ind]))
    sel <- which(in_dat[,REF_ind] == 'A' | in_dat[,REF_ind] == 'G')
    change_vec[sel] <- 
        paste0(complement[as.character(in_dat[sel,REF_ind])],
                complement[as.character(in_dat[sel,ALT_ind])])  
    change_vec <- factor(change_vec, 
                            levels = c("CA", "CG", "CT", "TA", "TC", "TG"))
    return(change_vec)
}


#' Annotate the intermutation distance of variants per PID
#'
#' The function annotates the intermutational distance to a PID wide data frame
#' by applying \code{\link[circlize]{rainfallTransform}} to every
#' chromosome-specific subfraction of the PID wide data.
#'
#' @param in_dat VRanges object or data frame which carries (at least) one
#'   column for the chromosome and one column for the position.
#' @param in_CHROM.field String indicating which column of \code{in_dat} carries
#'   the chromosome information if dealing with data frames.
#' @param in_POS.field String indicating which column of \code{in_dat} carries
#'   the position information if dealing with data frames.
#' @param in_mode String passed to \code{\link[circlize]{rainfallTransform}}
#'   indicating which method to choose for the computation of the
#'   intermutational distance.
#' @param in_verbose Whether verbose or not.
#'
#' @return VRanges object or data frame identical to \code{in_dat}, but with the
#'   intermutation distance annotated as an additional column on the right named
#'   \code{dist}.
#'
#' @seealso \code{\link{annotate_intermut_dist_cohort}}
#' @seealso \code{\link[circlize]{rainfallTransform}}
#'
#' @examples
#' test_df <- data.frame(
#'  CHROM=c(1,1,1,2,2,2,3,3,3,4,4,4,5,5),
#'  POS=c(1,2,4,4,6,9,1,4,8,10,20,40,100,200),
#'  REF=c("C","C","C","T","T","T","A","A","A","G","G","G","N","A"),
#'  ALT=c("A","G","T","A","C","G","C","G","T","A","C","T","A","N"))
#' min_dist_df <- annotate_intermut_dist_PID(test_df,in_CHROM.field="CHROM",
#'                                           in_POS.field="POS",
#'                                           in_mode="min")
#' max_dist_df <- annotate_intermut_dist_PID(test_df,in_CHROM.field="CHROM",
#'                                           in_POS.field="POS",
#'                                           in_mode="max")
#' min_dist_df
#' max_dist_df
#'
#' @importFrom circlize rainfallTransform
#' @export
#' 
annotate_intermut_dist_PID <- function(in_dat, in_CHROM.field = "CHROM",
                                        in_POS.field = "POS", in_mode = "min",
                                        in_verbose = FALSE){
    out_dat <- in_dat
    # account for input data type
    if(inherits(in_dat, "VRanges")) {
        in_dat$dist1 <- c(1e10, as.numeric(width(gaps(in_dat)))[-1])
        in_dat$dist2 <- c(as.numeric(width(gaps(in_dat)))[-1], 1e10)
        out_dat$dist <- apply(mcols(in_dat)[,c("dist1","dist2")], 1, min)
    } else if(inherits(in_dat, "data.frame")){
        CHROM_ind <- which(names(in_dat)==in_CHROM.field)
        POS_ind <- which(names(in_dat)==in_POS.field)
        inflate_df <- data.frame()
        for(my_chrom in unique(in_dat[,CHROM_ind])) {
            my_df <- in_dat[which(in_dat[,CHROM_ind]==my_chrom),]
            region_df <- data.frame(start=my_df[,POS_ind],end=my_df[,POS_ind])
        if(dim(my_df)[1]<2) {
            temp_df <- region_df
            temp_df$dist <- c(100000000)
        } else {
            temp_df <- circlize::rainfallTransform(region_df, mode = in_mode)
        }
        temp_df$CHROM <- my_chrom
        inflate_df <- rbind(inflate_df,temp_df)
    }
    out_dat$dist <- inflate_df[,dim(inflate_df)[2]-1]
    } else {
        if(in_verbose) cat("YAPSA:::annotate_intermut_dist_PID::error: input",
                " data is neither of type VRanges nor data.frame")
        return(NULL)
    }
    return(out_dat)
}


# rainfallTransformVR <- function(in_vr){
#   in_vr$dist1 <- c(1e10, as.numeric(width(gaps(in_vr)))[-1])
#   in_vr$dist2 <- c(as.numeric(width(gaps(in_vr)))[-1], 1e10)
#   return(apply(mcols(in_vr)[,c("dist1","dist2")], 1, min))
# }


#' Annotate the intermutation distance of variants cohort-wide
#'
#' The function annotates the intermutational distance to a cohort wide data
#' frame by applying \code{\link{annotate_intermut_dist_PID}} to every
#' PID-specific subfraction of the cohort wide data. Note that
#' \code{\link{annotate_intermut_dist_PID}} calls
#' \code{\link[circlize]{rainfallTransform}}. If the PID information is missing,
#' \code{\link{annotate_intermut_dist_PID}} is called directly for the whole
#' input.
#'
#' @param in_dat VRanges object, VRangesList, data frame or list of data frames
#'   which carries (at least) one column for the chromosome and one column for
#'   the position. Optionally, a column to specify the PID can be provided.
#' @param in_CHROM.field String indicating which column of \code{in_df} carries
#'   the chromosome information
#' @param in_POS.field String indicating which column of \code{in_df} carries
#'   the position information
#' @param in_PID.field String indicating which column of \code{in_df} carries
#'   the PID information
#' @param in_mode String passed through \code{\link{annotate_intermut_dist_PID}}
#'   to \code{\link[circlize]{rainfallTransform}} indicating which method to
#'   choose for the computation of the intermutational distance.
#' @param in_verbose Whether verbose or not.
#'
#' @return VRanges object, VRangesList, data frame or list of data frames
#'   identical to \code{in_df} (reordered by \code{in_PID.field}), but with the
#'   intermutation distance annotated as an additional column on the right named
#'   \code{dist}.
#'
#' @seealso \code{\link{annotate_intermut_dist_PID}}
#' @seealso \code{\link[circlize]{rainfallTransform}}
#'
#' @examples
#' test_df <- data.frame(CHROM=c(1,1,1,2,2,2,3,3,3,4,4,4,5,5),
#'                       POS=c(1,2,4,4,6,9,1,4,8,10,20,40,100,200),
#'                       REF=c("C","C","C","T","T","T","A",
#'                             "A","A","G","G","G","N","A"),
#'                       ALT=c("A","G","T","A","C","G","C",
#'                             "G","T","A","C","T","A","N"),
#'                       PID=c(1,1,1,2,2,2,1,1,2,2,2,1,1,2))
#' test_df <- test_df[order(test_df$PID,test_df$CHROM,test_df$POS),]
#' min_dist_df <-
#'   annotate_intermut_dist_cohort(test_df,in_CHROM.field="CHROM",
#'                                 in_POS.field="POS", in_PID.field="PID",
#'                                 in_mode="min")
#' max_dist_df <-
#'   annotate_intermut_dist_cohort(test_df,in_CHROM.field="CHROM",
#'                                 in_POS.field="POS", in_PID.field="PID",
#'                                 in_mode="max")
#' min_dist_df
#' max_dist_df
#'
#' @export
#' 
annotate_intermut_dist_cohort <- function(in_dat, in_CHROM.field = "CHROM",
                                        in_POS.field = "POS", 
                                        in_PID.field = NULL, 
                                        in_mode = "min", in_verbose = FALSE){
    # account for input data type
    if(inherits(in_dat, "data.frame")) {
        if(!is.null(in_PID.field)) {
            #PID_ind <- which(names(in_dat)==in_PID.field)
            in_dat[,in_PID.field] <- as.character(in_dat[,in_PID.field])
            dat_list <- split(in_dat, f = in_dat[,in_PID.field])
            out_list <- lapply(seq_along(dat_list),function(current_ind) {
            temp_out <- annotate_intermut_dist_PID(dat_list[[current_ind]],
                                            in_CHROM.field = in_CHROM.field,
                                            in_POS.field = in_POS.field,
                                            in_mode = in_mode, 
                                            in_verbose = in_verbose)
            current_name <- names(dat_list)[current_ind]
            if(is.null(current_name)) current_name <- current_ind
            temp_out[,in_PID.field] <- current_name
            return(temp_out)
        })
        out_dat <- do.call("rbind", out_list)
        } else {
            if(in_verbose) 
                cat("YAPSA:::annotate_intermut_dist_cohort::warning:",
                "no PID variable, thus no split. Simply calling ",
                "annotate_intermut_dist_PID.\n")
            out_dat <- annotate_intermut_dist_PID(in_dat,
                                            in_CHROM.field = in_CHROM.field,
                                            in_POS.field = in_POS.field,
                                            in_mode = in_mode, 
                                            in_verbose = in_verbose)
        }
    } else if(inherits(in_dat, "list")){
        if(all(unlist(lapply(in_dat, function(current_item){
            inherits(current_item, "data.frame")})))){
        if(is.null(in_PID.field)) in_PID.field <- "PID"
            out_list <- lapply(seq_along(in_dat),function(current_ind) {
            temp_out <- annotate_intermut_dist_PID(in_dat[[current_ind]],
                                            in_CHROM.field = in_CHROM.field,
                                            in_POS.field = in_POS.field,
                                            in_mode = in_mode, 
                                            in_verbose = in_verbose)
        current_name <- names(in_dat)[current_ind]
        if(is.null(current_name)) current_name <- current_ind
            temp_out[,in_PID.field] <- current_name
            return(temp_out)
        })
    names(out_list) <- names(in_dat)
    out_dat <- out_list
    }
    } else if(inherits(in_dat, "VRanges")){
    if(!is.null(in_PID.field)) {
        if(in_verbose) cat("YAPSA:::annotate_intermut_dist_cohort: input ",
                        "VRanges, PID variable found.\n")
        mcols(in_dat)[,in_PID.field] <- 
        as.character(mcols(in_dat)[,in_PID.field])
        dat_list <- split(in_dat, mcols(in_dat)[,in_PID.field])
        out_list <- lapply(seq_along(dat_list),function(current_ind) {
            temp_out <- annotate_intermut_dist_PID(dat_list[[current_ind]],
                                            in_CHROM.field = in_CHROM.field,
                                            in_POS.field = in_POS.field,
                                            in_mode = in_mode, 
                                            in_verbose = in_verbose)
            current_name <- names(dat_list)[current_ind]
            if(is.null(current_name)) current_name <- current_ind
            mcols(temp_out)[,in_PID.field] <- current_name
            return(temp_out)
            })
        out_dat <- unlist(VRangesList(out_list))
    } else {
        if(in_verbose) cat("YAPSA:::annotate_intermut_dist_cohort::warning:",
                            "no PID variable, thus no split. Simply calling ",
                            "annotate_intermut_dist_PID.\n")
        out_dat <- annotate_intermut_dist_PID(in_dat,
                                            in_CHROM.field = in_CHROM.field,
                                            in_POS.field = in_POS.field,
                                            in_mode = in_mode, 
                                            in_verbose = in_verbose)
    }
    } else if(inherits(in_dat, "VRangesList")){
        out_list <- lapply(seq_along(in_dat),function(current_ind) {
            temp_out <- annotate_intermut_dist_PID(in_dat[[current_ind]],
                                            in_CHROM.field = in_CHROM.field,
                                            in_POS.field = in_POS.field,
                                            in_mode = in_mode, 
                                            in_verbose = in_verbose)
        current_name <- names(in_dat)[current_ind]
        if(is.null(current_name)) current_name <- current_ind
        mcols(temp_out)[,in_PID.field] <- current_name
        return(temp_out)
        })
    names(out_list) <- names(in_dat)
    out_dat <- VRangesList(out_list)
    } else {
        if(in_verbose) cat("YAPSA:::annotate_intermut_dist_cohort::error:",
                            "input of wrong type, return NULL.\n")
        return(NULL)
    }
    return(out_dat)
}


#' Make a custom data structure for subgroups
#'
#' Creates a data frame carrying the subgroup information and the order in which
#' the PIDs have to be displayed. Calls \code{\link{aggregate}} on
#' \code{in_vcf_like_df}.
#'
#' @param in_vcf_like_df vcf-like data frame with point mutation calls
#' @param in_exposures_df Data frame with the signature exposures
#' @param in_palette Palette for colour attribution to the subgroups if nun-NULL
#' @param in_subgroup.field String indicating which column of
#'   \code{in_vcf_like_df} carries the subgroup information
#' @param in_PID.field String indicating which column of \code{in_vcf_like_df}
#'   and of \code{in_exposures_df} carries the PID information
#' @param in_verbose Whether verbose or not.
#'
#' @return subgroups_df: A data frame carrying the subgroup and rank
#'   information.
#'
#' @examples
#'  data(lymphoma_test)
#'  data(lymphoma_cohort_LCD_results)
#'  choice_ind <- (names(lymphoma_Nature2013_COSMIC_cutoff_exposures_df)
#'                 %in% unique(lymphoma_test_df$PID))
#'  lymphoma_test_exposures_df <-
#'    lymphoma_Nature2013_COSMIC_cutoff_exposures_df[,choice_ind]
#'  make_subgroups_df(lymphoma_test_df,lymphoma_test_exposures_df)
#'
#' @seealso \code{\link{aggregate}}
#'
#' @export
#' 
make_subgroups_df <- function(in_vcf_like_df,
                                in_exposures_df = NULL,
                                in_palette = NULL,
                                in_subgroup.field="SUBGROUP",
                                in_PID.field="PID", in_verbose = FALSE){
    # account for input data type
    if(!(inherits(in_vcf_like_df, "VRanges")) & 
        !(inherits(in_vcf_like_df, "data.frame"))){
        if(in_verbose) 
            cat("YAPSA:::make_subgroups_df::warning: Input is neither ",
                "a VRanges object nor a data frame. Return NULL.\n")
        return(NULL)
    }
    if(inherits(in_vcf_like_df, "VRanges")) {
        choice_column_vector <- c("seqnames", "start", "ref", "alt",
                                    in_subgroup.field, in_PID.field)
        in_vcf_like_df <- as.data.frame(in_vcf_like_df)
        colum_names <- intersect(names(in_vcf_like_df),choice_column_vector)
        in_vcf_like_df <- in_vcf_like_df[,colum_names]
        names(in_vcf_like_df) <- 
            gsub("seqnames", "CHROM", names(in_vcf_like_df))
        names(in_vcf_like_df) <- 
            gsub("start", "POS", names(in_vcf_like_df))
        names(in_vcf_like_df) <- gsub("ref", "REF", names(in_vcf_like_df))
        names(in_vcf_like_df) <- gsub("alt", "ALT", names(in_vcf_like_df))
    }
    ## 1. rename the colnames in in_vcf_like_df if necessary to be able to run
    ## the aggregate command later
    subgroup_ind <- which(names(in_vcf_like_df)==in_subgroup.field)
    names(in_vcf_like_df)[subgroup_ind] <- "SUBGROUP"
    PID_ind <- which(names(in_vcf_like_df)==in_PID.field)
    names(in_vcf_like_df)[PID_ind] <- "PID"
    ## 2. start work
    out_subgroups_df <- aggregate(SUBGROUP~PID,data=in_vcf_like_df,
                                    function(l) return(l[1]))
    if(!is.null(in_exposures_df)) 
        this_sum_df <- data.frame(sum=apply(in_exposures_df,2,sum))
    else {
        this_sum_df <- as.data.frame(table(in_vcf_like_df[,PID_ind]))
        rownames(this_sum_df) <- this_sum_df[,1]
        names(this_sum_df)[2] <- "sum"
        this_sum_df[,1] <- NULL
    }
    out_subgroups_df <- merge(out_subgroups_df, this_sum_df,
                            by.x=in_PID.field,by.y=0)
    max_total_count <- max(out_subgroups_df$sum)
    out_subgroups_df$compl_sum <- max_total_count - out_subgroups_df$sum
    temp_ind <- order(out_subgroups_df$SUBGROUP,out_subgroups_df$compl_sum)
    out_subgroups_df$index <- order(temp_ind)
    names(out_subgroups_df)[2] <- "subgroup"
    if(is.null(in_palette)){
        in_palette <- rainbow(length(unique(out_subgroups_df$subgroup)))
    }
    out_subgroups_df$col <- in_palette[factor(out_subgroups_df$subgroup)]
    out_subgroups_df$subgroup <- as.character(out_subgroups_df$subgroup)
    return(out_subgroups_df)
}


#' Wrapper for Shapiro test but allow for all identical values
#' 
#' @param in_vector
#'  Numerical vector the Shapiro-Wilk test is computed on
#' 
#' @return p-value of the Shapiro-Wilk test, zero if all entries in the input
#'  vector \code{in_vector} are identical.
#' 
#' @seealso \code{\link[stats]{shapiro.test}}
#' 
#' @examples
#'  shapiro_if_possible(runif(100,min=2,max=4))
#'  shapiro_if_possible(rnorm(100,mean=5,sd=3))
#'  shapiro_if_possible(rep(4.3,100))
#'  shapiro_if_possible(c("Hello","World"))
#' 
#' @export
#' 
shapiro_if_possible <- function(in_vector){
    if(is.numeric(in_vector)){
        if(length(unique(in_vector)) > 1){
            return(shapiro.test(in_vector)$p.value)
        } else {
            return(0.0)
        }
    } else {
        cat("YAPSA:::shapiro_if_possible::error: Non-numeric input")
        return(NULL)
    }
}


#' Split an exposures data frame by subgroups
#'
#' If a cohort consists of different subgroups, this function enables to split
#' the data frame storing the signature exposures into a list of data frames
#' with signature exposures, one per subgroup. This functionality is needed for
#' \code{\link{stat_test_subgroups}} and \code{\link{stat_plot_subgroups}}
#'
#' @param in_exposures_df Numerical data frame of the exposures (i.e.
#'   contributions of the different signatures to the number of point mutations
#'   per PID)
#' @param in_subgroups_df Data frame indicating which PID belongs to which
#'   subgroup
#' @param in_subgroups.field Name indicating which column in
#'   \code{in_subgroups_df} contains the subgroup information
#' @param in_PID.field Name indicating which column in \code{in_subgroups_df}
#'   contains the PID information
#'
#' @return List of data frames with the subgroup specific signature exposures.
#'
#' @seealso \code{\link{stat_test_subgroups}}
#' @seealso \code{\link{stat_plot_subgroups}}
#'
#' @examples
#'  NULL
#'
#' @export
#' 
split_exposures_by_subgroups <- function(in_exposures_df,in_subgroups_df,
                                        in_subgroups.field="subgroup",
                                        in_PID.field="PID"){
    subgroup_index <- which(names(in_subgroups_df)==in_subgroups.field)
    PID_index <- which(names(in_subgroups_df)==in_PID.field)
    # split in_exposures_df
    subgroups_list <- split(in_subgroups_df,in_subgroups_df[,subgroup_index])
    exposures_list <- lapply(subgroups_list, FUN=function(x) {
        PID_vector <- x[,PID_index]
        choice_ind <- which(names(in_exposures_df) %in% PID_vector)
        if(length(choice_ind)>1){
            out_df <- in_exposures_df[,choice_ind]
        } else {
            out_df <- data.frame(temp=in_exposures_df[,choice_ind])
            names(out_df)[1] <- PID_vector[1]
        }
        return(out_df)
    })
    return(exposures_list)
}



#' Add an element as first entry to a list
#' 
#' Works for all types of lists and inputs
#' 
#' @param in_list
#'  List to which an element is to be added
#' @param in_element
#'  Element to be added
#' 
#' @return List with input element as first entry.
#' 
#' @examples
#'  NULL
#' 
#' @export
#' 
add_as_fist_to_list <- function(in_list,in_element){
    out_list <- in_list
    number_of_elements <- length(in_list)
    out_list[[number_of_elements+1]] <- in_element
    my_ind <- c(number_of_elements+1,seq_len(number_of_elements))
    return(out_list[my_ind])
}


#' Merge exposure data frames
#'
#' Merges with the special feature of preserving the signatures and signature
#' order.
#'
#' @param in_exposures_list List of data frames (carrying information on
#'   exposures).
#' @param in_signatures_df Data frame \code{W} in which the columns represent
#'   the signatures.
#'
#' @return A data frame with the merged exposures.
#'
#' @examples
#'  NULL
#'
#' @export
#' 
merge_exposures <- function(in_exposures_list,
                            in_signatures_df){
    exposures_list <- lapply(in_exposures_list,FUN=function(current_df) {
        current_df$sig <- rownames(current_df)
        return(current_df)
    })
    expousre_list_no_null <- exposures_list[!sapply(exposures_list, is.null)] 
    #Filter for NA
    exposures_df <- Reduce(function(...) merge(..., by="sig",
                                            all=TRUE, sort=FALSE),
                                            expousre_list_no_null)
    row_order <- order(match(exposures_df$sig,names(in_signatures_df)))
    exposures_df <- exposures_df[row_order,]
    rownames(exposures_df) <- exposures_df$sig
    exposures_df$sig <- NULL
    exposures_df[is.na(exposures_df)] <- 0
    return(exposures_df)
}


#' Generically melts exposure data frames
#' 
#' Melt an exposure data frame with signatures as ID variables.
#' 
#' @param in_df
#'  Numeric data frame with exposures.
#' 
#' @return A data frame with the molten exposures.
#' 
#' @examples
#'  NULL
#' 
#' @export
#' 
melt_exposures <- function(in_df){
    in_df$sig <- rownames(in_df)
    in_df_melt <- melt(in_df,id.vars = "sig")
    names(in_df_melt) <- gsub("variable","PID",names(in_df_melt))
    return(in_df_melt)
}


#' Compares alternative exposures
#'
#' Compares exposures computed by two alternative approaches for the same cohort
#'
#' @param in_exposures1_df Numeric data frame with exposures, ideally the
#'   smaller exposure data is supplied first.
#' @param in_exposures2_df Numeric data frame with exposures, ideally the bigger
#'   exposure data is supplied second.
#' @param deselect_flag Wehther signatures absent in both exposure data frames
#'   should be removed.
#'
#' @return A list with entries \code{merge_df}, \code{all_cor.coeff},
#'   \code{all_p.value}, \code{cor.coeff_vector}, \code{p.value_vector},
#'   \code{all_cor.test}, and \code{cor.test_list}. \itemize{ \item
#'   \code{merge_df}: Merged molten input exposure data frames \item
#'   \code{all_cor.coeff}: Pearson correlation coefficient for all data points,
#'   i.e. taken all signatures together \item \code{all_p.value}: P-value of the
#'   Pearson test for all data points, i.e. taken all signatures together \item
#'   \code{cor.coeff_vector}: A vector of Pearson correlation coefficients
#'   evaluated for every signature independently \item \code{p.value_vector}: A
#'   vector of p-values of the Pearson tests evaluated for every signature
#'   independently \item \code{all_cor.test}: A data structure as returned by
#'   \code{\link{cor.test}} for all data points, i.e. taken all signatures
#'   together \item \code{cor.test_list}: A list of data structures as returned
#'   by \code{\link{cor.test}}, but evaluated for every signature independently
#'   }
#'
#' @examples
#'  NULL
#'
#' @export
#' 
compare_exposures <- function(in_exposures1_df,
                                in_exposures2_df,
                                deselect_flag=TRUE){
    signatures1 <- rownames(in_exposures1_df)
    signatures2 <- rownames(in_exposures2_df)
    common_signatures <- intersect(signatures1,signatures2)
    if(length(common_signatures) == 0) return(NULL)
    common_signatures <- union(signatures2,signatures1)
    exposures1_df_melt <- melt_exposures(in_exposures1_df)
    names(exposures1_df_melt) <- gsub("value","exposures1",
                                names(exposures1_df_melt))
    exposures2_df_melt <- melt_exposures(in_exposures2_df)
    names(exposures2_df_melt) <- gsub("value","exposures2",
                                names(exposures2_df_melt))
    merge_df <- merge(exposures1_df_melt,exposures2_df_melt,by=c("sig","PID"),
                                sort=FALSE,all=TRUE)
    merge_df$exposures1[which(!is.finite(merge_df$exposures1))] <- 0
    merge_df$exposures2[which(!is.finite(merge_df$exposures2))] <- 0
    if(deselect_flag){
        deselect_ind <- which(merge_df$exposures1==0 & merge_df$exposures2==0)
        merge_df <- merge_df[-deselect_ind,]    
    }
    all_cor.test <- cor.test(merge_df$exposures1,merge_df$exposures2)
    cor.test_list <- lapply(
        split(merge_df,f = merge_df$sig),
        FUN=function(current_df){
        levels1 <- length(unique(current_df$exposures1))
        levels2 <- length(unique(current_df$exposures2))
        if(levels1 > 1 & levels2 > 1) {
            out.test <- cor.test(current_df$exposures1,current_df$exposures2)
        } else {
            out.test <- NULL       
        } 
    })
    cor.coeff_vector <- unlist(lapply(
        cor.test_list,
        function(current_test){
            if(is.null(current_test)) return(0)
            return(as.numeric(current_test$estimate))
    }))
    p.value_vector <- unlist(lapply(
    cor.test_list,
    function(current_test){
        if(is.null(current_test)) return(1)
            return(as.numeric(current_test$p.value))
        }))
    match_ind <- match(common_signatures,names(cor.test_list))
    cor.test_list <- cor.test_list[match_ind]
    cor.coeff_vector <- cor.coeff_vector[match_ind]
    p.value_vector <- p.value_vector[match_ind]
    return(list(merge_df=merge_df,
                all_cor.coeff=all_cor.test$estimate,
                all_p.value=all_cor.test$p.value,
                cor.coeff_vector=cor.coeff_vector,
                p.value_vector=p.value_vector,
                all_cor.test=all_cor.test,
                cor.test_list=cor.test_list))
}


#' Find samples affected
#'
#' Find samples affected by SNVs in a certain pathway
#'
#' @param in_gene_list List of genes in the pathway of interest.
#' @param in_gene_vector Character vector for genes annotated to SNVs as in
#'   \code{vcf_like_df}.
#' @param in_PID_vector Character vector for sample names annotated to SNVs as
#'   in \code{vcf_like_df}.
#'
#' @return A character vector of the names of the affected samples
#'
#' @examples
#'  NULL
#'
#' @export
#' 
find_affected_PIDs <- function(in_gene_list,
                                in_gene_vector,
                                in_PID_vector){
    choice_gene_pattern <- paste0(in_gene_list,collapse="|")
    choice_mut_ind <- grep(choice_gene_pattern,in_gene_vector)
    affected_PIDs <- unique(in_PID_vector[choice_mut_ind])
    return(affected_PIDs)
}


#' Test significance of association
#'
#' Test significance of association between a vector of exposures and a
#' selection of samples, e.g. those affected by mutations in a pathway as
#' returned by \code{\link{find_affected_PIDs}}
#'
#' @param in_exposure_vector Named vector of a phenotype (e.g. exposures to a
#'   specific signature)
#' @param in_affected_PIDs Character vector of samples affected by some
#'   criterion, e.g. mutations in a pathway as returned by
#'   \code{\link{find_affected_PIDs}}
#' @param in_mutation_label If non-NULL, prefix to the mutation status (x-axis
#'   label) in the produced boxplot
#' @param in_exposure_label If non-NULL, prefix to the exposures (y-axis label)
#'   in the produced boxplot
#'
#' @return A list with entries: \itemize{ \item \code{current_kruskal}: Kruskal
#'   test object from testing phenotype against affection \item
#'   \code{current_boxplot}: Boxplot of phenotype against affection }
#'
#' @examples
#'  NULL
#'
#' @importFrom ggbeeswarm geom_beeswarm
#' @export
#' 
test_exposureAffected <- function(in_exposure_vector,
                                    in_affected_PIDs,
                                    in_mutation_label=NULL,
                                    in_exposure_label=NULL){
    .e <- environment()
    factor_vector <- rep("wt",length(in_exposure_vector)) 
    names(factor_vector) <- names(in_exposure_vector)
    factor_vector[in_affected_PIDs] <- "mut"
    if(!is.null(in_mutation_label)) factor_vector <- 
        paste0(in_mutation_label,"-", factor_vector)
    factor_vector <- factor(factor_vector)
    test_df <- data.frame(
        exposure=as.numeric(in_exposure_vector),mut_stat=factor_vector)
    current_kruskal <- kruskal.test(exposure~mut_stat,data=test_df)
    current_boxplot <- ggplot(environment = .e, data = test_df) +  
        geom_boxplot(aes_string(x = "mut_stat", y = "exposure")) +
        geom_beeswarm(aes_string(x = "mut_stat", y = "exposure", 
                                col = "mut_stat")) +
        theme(legend.position="none")
    if(!is.null(in_exposure_label)){
        current_boxplot <- current_boxplot + ylab(paste0(in_exposure_label,
                                            "exposure"))
    }
    return(list(kruskal=current_kruskal,
                boxplot=current_boxplot))
}


#' Round to a defined precision
#'
#' This function is an extension with regard to the function \code{\link{round}}
#' from base R as it allows not only digits as precision, but can also round to
#' a user-specified precision. The interval in which the rounding operation is
#' to be carried out also can be specified by the user (default is the unit
#' interval). Alternatively, breaks can be provided.
#'
#' @param x Vector to be rounded
#' @param breaks The breaks used for rounding. Default NULL
#' @param in_precision Precition default 0.05
#' @param in_interval Interval needs to be larger than the precision value
#'
#' @return A list with two entries: \itemize{ \item \code{values}: the rounded
#'   vector \item \code{breaks}: the breaks used for rounding }
#'
#' @export
#'
#' @examples
#'  NULL
#' 
round_precision <- function(x, breaks = NULL, in_precision = 0.05, 
                            in_interval = c(0,1)){
    if(is.null(breaks)){
        if(is.null(in_interval)) in_interval <- c(min(x), max(x))
        if(any(x < in_interval[1] | x > in_interval[2])) {
            warning("YAPSA:::round_precision::warning: ",
                "input is not contained in specified interval.\n")
        return(x)
    }
    if((in_interval[2] - in_interval[1]) < in_precision) {
        warning("YAPSA:::round_precision::warning: ",
                "specified interval is smaller than specified precision.\n")
        return(x)
    }
    breaks <- seq(in_interval[1], in_interval[2], in_precision)
    }
    out_vec <- unlist(lapply(x, function(y) breaks[which.min(abs(y - breaks))]))
    return(list(values = out_vec,
                breaks = breaks))
}


#' Readjust the vector to it's original norm after rounding
#' 
#' After use of the function \code{\link{round_precision}} the norm of the 
#' input vector may have been altered by the rounding procedure. This 
#' function restores the norm by altering only the largest entry in the 
#' rounded vector (in order to create the least possible relative error).
#'
#' @param x vector to be rounded
#' @param in_interval Interval
#'
#' @return The adapted form of the input vector \code{x}.
#' 
#' @export
#'
#' @examples
#'  NULL
#' 
correct_rounded <- function(x, in_interval = c(0,1)){
    if(sum(x) == in_interval[2]) return(x)
    max_ind <- which.max(x)
    x[max_ind] <- in_interval[2] - sum(x[-max_ind])
    return(x)
}

Try the YAPSA package in your browser

Any scripts or data that you put into this service are public.

YAPSA documentation built on Nov. 8, 2020, 4:59 p.m.