R/03_GLCMtexturalFeatures25Dmerged.R

Defines functions glcmTexturalFeatures25Dmerged

glcmTexturalFeatures25Dmerged <- function(imgObj,n_grey){

  # compute number of non-NA voxels

  nVoxel <- dim(imgObj)[1]*dim(imgObj)[2]*dim(imgObj)[3] - sum(is.na(imgObj))
  grayLevlesMax <- max(imgObj,na.rm = T)
  ### compute Gray Levels Cooccurrence Matrices

  G_list_0 <- list()
  G_list_45 <- list()
  G_list_90 <- list()
  G_list_135 <- list()

  #Compute grey level cooccurrence matrices for 4 different directions within each slice
  for (i in 1:dim(imgObj)[3]){
    if(length(imgObj[,,i])*.005 >= sum(!is.na(imgObj[,,i]))) next
    imgObj[,,i] <- round(imgObj[,,i])
    #if(length(table(unique(imgObj[,,i])))<n_grey) n_grey <- length(table(unique(imgObj[,,i])))-1
    if (min(imgObj[,,i],na.rm = T) < 0) {imgObj[,,i] <- imgObj[,,i] + abs(min(imgObj[,,i],na.rm = T))}
    G_list_0[[i]] <- as.matrix(glcm(imgObj[,,i], angle = 0, normalize = F,verbose=F,n_grey = n_grey))
    G_list_45[[i]] <- as.matrix(glcm(imgObj[,,i], angle = 45, normalize = F,verbose=F,n_grey = n_grey))
    G_list_90[[i]] <- as.matrix(glcm(imgObj[,,i], angle = 90, normalize = F,verbose=F,n_grey = n_grey))
    G_list_135[[i]] <- as.matrix(glcm(imgObj[,,i], angle = 135, normalize = F,verbose=F,n_grey = n_grey))
  }
  
  
  #elimino gli elementi NULL della lista
  if(length(G_list_0)!=0){
    if(length(which(sapply(G_list_0, is.null)))!=0){
      G_list_0 = G_list_0[-which(sapply(G_list_0, is.null))]
    }
  }
  
  #elimino gli elementi NULL della lista
  if(length(G_list_45)!=0){
    if(length(which(sapply(G_list_45, is.null)))!=0){
      G_list_45 = G_list_45[-which(sapply(G_list_45, is.null))]
    }
  }
  
  #elimino gli elementi NULL della lista
  if(length(G_list_90)!=0){
    if(length(which(sapply(G_list_90, is.null)))!=0){
      G_list_90 = G_list_90[-which(sapply(G_list_90, is.null))]
    }
  }
  
  #elimino gli elementi NULL della lista
  if(length(G_list_135)!=0){
    if(length(which(sapply(G_list_135, is.null)))!=0){
      G_list_135 = G_list_135[-which(sapply(G_list_135, is.null))]
    }
  }

  sumtot <- list()
  ### DIRECTION 0

  matrix.df <- ldply(G_list_0, data.table::melt, varnames=c("row", "col"))
  sumtot[[1]] <- acast(matrix.df, row ~ col, sum)
  ### DIRECTION 45
  matrix.df <- ldply(G_list_45, data.table::melt, varnames=c("row", "col"))
  sumtot[[2]] <- acast(matrix.df, row ~ col, sum)
  ### DIRECTION 90
  matrix.df <- ldply(G_list_90, data.table::melt, varnames=c("row", "col"))
  sumtot[[3]] <- acast(matrix.df, row ~ col, sum)
  ### DIRECTION 135
  matrix.df <- ldply(G_list_135, data.table::melt, varnames=c("row", "col"))
  sumtot[[4]] <- acast(matrix.df, row ~ col, sum)
  
  ## Tutte insieme
  matrix.df <- ldply(sumtot, data.table::melt, varnames=c("row", "col"))
  sumtottutte <- acast(matrix.df, row ~ col, sum)
  

  #Initialise data table for storing GLCM features; I have added a few
  featNames <- c("F_cm.2.5Dmerged.joint.max", "F_cm.2.5Dmerged.joint.avg", "F_cm.2.5Dmerged.joint.var", "F_cm.2.5Dmerged.joint.entr",
                 "F_cm.2.5Dmerged.diff.avg", "F_cm.2.5Dmerged.diff.var", "F_cm.2.5Dmerged.diff.entr",
                 "F_cm.2.5Dmerged.sum.avg", "F_cm.2.5Dmerged.sum.var", "F_cm.2.5Dmerged.sum.entr", "F_cm.2.5Dmerged.energy","F_cm.2.5Dmerged.contrast","F_cm.2.5Dmerged.dissimilarity",
                 "F_cm.2.5Dmerged.inv.diff","F_cm.2.5Dmerged.inv.diff.norm","F_cm.2.5Dmerged.inv.diff.mom","F_cm.2.5Dmerged.inv.diff.mom.norm","F_cm.2.5Dmerged.inv.var",
                 "F_cm.2.5Dmerged.corr","F_cm.2.5Dmerged.auto.corr","F_cm.2.5Dmerged.clust.tend","F_cm.2.5Dmerged.clust.shade","F_cm.2.5Dmerged.clust.prom",
                 "F_cm.2.5Dmerged.info.corr.1","F_cm.2.5Dmerged.info.corr.2")
  F_cm <- data.table(matrix(NA, nrow=1, ncol=length(featNames)))
  colnames(F_cm) <- featNames


    Ng <- ncol(sumtottutte)

    #Convert matrix to data table
    df.G    <- data.table(sumtottutte)

    #Add row grey level intensity
    df.G$i <- as.numeric(row.names(sumtottutte))

    #Convert from wide to long table. This is the preferred format for data tables and data frames
    df.G   <- data.table::melt(df.G, id.vars="i", variable.name="j", value.name="n", variable.factor=FALSE)

    #Convert j from string to numeric
    df.G$j <- as.numeric(df.G$j)

    #Remove combinations with 0 counts
    df.G   <- df.G[n>0,]

    #Convert Grey level coccurrence matrix to joint probability
    df.p_ij <- df.G[,.(p_ij=n/sum(df.G$n)), by=.(i,j)] #joint probability
    df.p_i  <- df.p_ij[,.(p_i=sum(p_ij)), by=i]        #marginal probability over columns
    df.p_j  <- df.p_ij[,.(p_j=sum(p_ij)), by=j]        #marginal probability over rows

    #Diagonal probabilities (p(i-j))
    #First, we create a new column k which contains the absolute value of i-j.
    #Second, we sum the joint probability where k is the same.
    #This can written as one line by chaining the operations.
    df.p_imj <- copy(df.p_ij)
    df.p_imj <- df.p_imj[,"k":=abs(i-j)][,.(p_imj=sum(p_ij)), by=k]

    #Cross-diagonal probabilities (p(i+j))
    #Again, we first create a new column k which contains i+j
    #Second, we sum the probability where k is the same.
    #This is written in one line by chaining the operations.
    df.p_ipj <- copy(df.p_ij)
    df.p_ipj <- df.p_ipj[,"k":=i+j][,.(p_ipj=sum(p_ij)), by=k]

    #Merger of df.p_ij, df.p_i and df.p_j
    df.p_ij <- merge(x=df.p_ij, y=df.p_i, by="i")
    df.p_ij <- merge(x=df.p_ij, y=df.p_j, by="j")

    #Thus we have five probability matrices
    #Joint probability:          df.p_ij with probabilities p_ij, p_i and p_j, and indices i, j
    #Marginal probability:       df.p_i with probability p_i, and index i
    #Marginal probability:       df.p_j with probability p_j, and index j
    #Diagonal probability:       df.p_imj with probability p_imj and index k
    #Cross-diagonal probability: df.p_ipj with probability p_ipj and index k

    #Calculate features

    #Joint maximum
    F_cm$F_cm.2.5Dmerged.joint.max        <- max(df.p_ij$p_ij)

    #Joint average
    F_cm$F_cm.2.5Dmerged.joint.avg        <- sum(df.p_ij$i * df.p_ij$p_ij)

    #Joint variance
    mu <- sum(df.p_ij$i * df.p_ij$p_ij)
    F_cm$F_cm.2.5Dmerged.joint.var         <- sum((df.p_ij$i-mu)^2 * df.p_ij$p_ij)

    #Joint entropy
    F_cm$F_cm.2.5Dmerged.joint.entr        <- - sum(df.p_ij$p_ij * log2(df.p_ij$p_ij))

    #Difference average
    F_cm$F_cm.2.5Dmerged.diff.avg          <- sum((df.p_imj$k) * df.p_imj$p_imj)

    #Difference variance
    mu <- sum((df.p_imj$k) * df.p_imj$p_imj)
    F_cm$F_cm.2.5Dmerged.diff.var          <- sum((df.p_imj$k-mu)^2 * df.p_imj$p_imj)

    #Difference entropy
    F_cm$F_cm.2.5Dmerged.diff.entr         <- - sum(df.p_imj$p_imj * log2(df.p_imj$p_imj))

    #Sum average
    F_cm$F_cm.2.5Dmerged.sum.avg           <- sum(df.p_ipj$k * df.p_ipj$p_ipj)

    #Sum variance
    mu <- sum(df.p_ipj$k * df.p_ipj$p_ipj)
    F_cm$F_cm.2.5Dmerged.sum.var           <- sum((df.p_ipj$k-mu)^2 * df.p_ipj$p_ipj)

    #Sum entropy
    F_cm$F_cm.2.5Dmerged.sum.entr          <- - sum(df.p_ipj$p_ipj * log2(df.p_ipj$p_ipj))

    #Angular second moment
    F_cm$F_cm.2.5Dmerged.energy          <- sum(df.p_ij$p_ij^2)

    #Contrast
    F_cm$F_cm.2.5Dmerged.contrast          <- sum((df.p_ij$i - df.p_ij$j)^2 * df.p_ij$p_ij)

    #Dissimilarity
    F_cm$F_cm.2.5Dmerged.dissimilarity          <- sum(abs(df.p_ij$i - df.p_ij$j) * df.p_ij$p_ij)

    #Inverse difference
    F_cm$F_cm.2.5Dmerged.inv.diff         <- sum( df.p_ij$p_ij/(1+abs(df.p_ij$i - df.p_ij$j)) )

    #Inverse difference normalized
    D <- abs(df.p_ij$i - df.p_ij$j)/grayLevlesMax
    F_cm$F_cm.2.5Dmerged.inv.diff.norm         <- sum(df.p_ij$p_ij / (1+D) )

    #Inverse difference moment
    F_cm$F_cm.2.5Dmerged.inv.diff.mom          <- sum(df.p_ij$p_ij/(1+(df.p_ij$i - df.p_ij$j)^2))

    #Inverse difference moment normalized
    DD <- ((df.p_ij$i - df.p_ij$j)/grayLevlesMax)^2
    F_cm$F_cm.2.5Dmerged.inv.diff.mom.norm         <- sum(df.p_ij$p_ij / (1+DD))

    #Inverse variance
    df.jmorethani <- df.p_ij[which(df.p_ij$j>df.p_ij$i),]
    F_cm$F_cm.2.5Dmerged.inv.var          <- 2*sum(df.jmorethani$p_ij/(df.jmorethani$i-df.jmorethani$j)^2)

    #Correlation
    mu.i <- sum(df.p_i$i * df.p_i$p_i)
    sigma.i <- sqrt(sum((df.p_i$i - mu.i)^2 * df.p_i$p_i))
    F_cm$F_cm.2.5Dmerged.corr <- (1/sigma.i)^2 * sum((df.p_ij$i - mu.i) * (df.p_ij$j - mu.i) * df.p_ij$p_ij)

    #Autocorrelation
    F_cm$F_cm.2.5Dmerged.auto.corr <- sum(df.p_ij$i * df.p_ij$j * df.p_ij$p_ij)

    #Cluster tendency
    F_cm$F_cm.2.5Dmerged.clust.tend <- sum((df.p_ij$i + df.p_ij$j -2*mu.i)^2 * df.p_ij$p_ij)

    #Cluster shade
    F_cm$F_cm.2.5Dmerged.clust.shade <- sum((df.p_ij$i + df.p_ij$j -2*mu.i)^3 * df.p_ij$p_ij)

    #Cluster prominence
    F_cm$F_cm.2.5Dmerged.clust.prom <- sum((df.p_ij$i + df.p_ij$j -2*mu.i)^4 * df.p_ij$p_ij)

    #First measure of information correlation
    HXY <- - sum(df.p_ij$p_ij * log2(df.p_ij$p_ij))
    HX <- - sum(df.p_i$p_i * log2(df.p_i$p_i))
    HXY1 <- - sum(df.p_ij$p_ij * log2(df.p_ij$p_i * df.p_ij$p_j))
    F_cm$F_cm.2.5Dmerged.info.corr.1 <- (HXY - HXY1) / HX

    #Second measure of information correlation
    #HXY2 <- - sum(df.p_ij$p_i * df.p_ij$p_j * log2(df.p_ij$p_i * df.p_ij$p_j))
    # if (HXY > HXY2 ) F_cm$F_cm.info.corr.2[iter] <- 0
    # else if (HXY <= HXY2 ) F_cm$F_cm.info.corr.2[iter] <- sqrt(1-exp(-2*(HXY2 - HXY)))

    HXY2 <- - sum(df.p_ij$p_i * df.p_ij$p_j * log2(df.p_ij$p_i * df.p_ij$p_j))
    F_cm$F_cm.2.5Dmerged.info.corr.2 <- sqrt(1-exp(-2*(abs(HXY2 - HXY))))

  return(F_cm)
}
kbolab/moddicom documentation built on Nov. 29, 2020, 9:11 p.m.