#The goal of this function is to calculate the LoD/LoQ of the data provided in the data frame.
#The function returns a new data frame containing the value of the LoD/LoQ
#' @export
#' @import minpack.lm
# DON'T TOUCH THIS ONE! THIS IS FROM 3.3.11
nonlinear_quantlim_first <- function(datain){
#Need to rename variables as needed:
names(datain)[names(datain) == 'CONCENTRATION'] <- 'C'
names(datain)[names(datain) == 'INTENSITY'] <- 'I'
##Coarse:
B <- 500
BB <- 200
NMAX = 30
Npoints <- 30
Maxit = 3
datain <- datain[!is.na(datain$I) & !is.na(datain$C),]
datain <- datain[!is.infinite(datain$I) & !is.infinite(datain$C),]
datain <- datain[order(datain$C),]
tmp_nob <- subset(datain,datain$C >0)
tmp_all <- datain
tmp_blank <-subset(datain,datain$C == 0)
#Calculate the value of the noise:
#Use the zero concentration to calculate the LOD:
noise = mean(tmp_blank$I)
var_noise = var(tmp_blank$I)
pb <- txtProgressBar(min = 0, max = 1, initial = 0, char = "%",
width = 40, title, label, style = 1, file = "")
#Need to change this to account for limited number of samples (4):
#The confidence interval for future observations of normal observations with unknown mean and variance:
#t(alpha/2,dof = n-1)*s*sqrt(1+1/n)
n_blank = length(unique(tmp_blank$I))
fac_low = qt(1-0.25/2,n_blank - 1)*sqrt(1+1/n_blank)
fac = qt(1-0.05/2,n_blank - 1)*sqrt(1+1/n_blank)
fac_high = qt(1-0.01/2,n_blank - 1)*sqrt(1+1/n_blank)
#2.575
low_noise = noise - fac* sqrt(var_noise)
up_noise_low = noise + fac_low * sqrt(var_noise)
up_noise = noise + fac* sqrt(var_noise)
up_noise_high = noise + fac_high*sqrt(var_noise)
unique_c = sort(unique(tmp_all$C)); var_v <- rep(0.0, length(unique_c))
weights <- rep(0.0, length(tmp_all$C));
weights_nob <- rep(0.0, length(tmp_nob$C));
#Calculate variance for all concentrations:
ii = 1
for (j in unique_c){
data_f <- subset(tmp_all, C == j)
var_v[ii] <- var(data_f$I)#/mean(data_f$log2Int)**2
ii = ii +1
}
#Log scale discretization:
xaxis_orig_2 <- exp(c( seq( from = log(10+0), to = log(1+max(unique_c)), by = log(1+max(unique_c))/Npoints ))) -10 #0.250 to go fast here
xaxis_orig_2 <- unique(sort(c(xaxis_orig_2,unique_c)))
###Loop here to make sure that the discretization is fine enough##
x_new <- NULL
for(it in 1:Maxit){
if(!is.null(x_new)) xaxis_orig_2 <- c(xaxis_orig_2,x_new)
y.loessm04 <- loess(y ~ x, span=0.6, data.frame(x=log(1+unique_c), y=log(1+var_v)))
y.predic_log <- predict(y.loessm04, data.frame(x=log(1+xaxis_orig_2)))
y.predic_log_unique <- predict(y.loessm04, data.frame(x=log(1+unique_c)))
#var_v_s_log <- exp(y.predic_log) -1
#Can use this instead of the origianl variance for everything
var_v_s_log <- pmax(exp(y.predic_log) -1, rep(1.0,length(y.predic_log)))
var_v_s_log_unique <- pmax(exp(y.predic_log_unique) -1, rep(1.0,length(y.predic_log_unique)))
##
#In the following, consider that the smoothed variance is that from the log:
var_v_s <- var_v_s_log
var_v_s_unique <- var_v_s_log_unique
##
if(1){# Full bootstrap
outB <- matrix(NA_real_, nrow=B, ncol=length(xaxis_orig_2))
outBB_pred <- matrix(NA_real_, nrow=B*BB, ncol=length(xaxis_orig_2))
change_B <- rep(NA, B)
for (j in 1:B){ #Number of first boostrap samples j
setTxtProgressBar(pb, j/B, title = NULL, label = NULL)
#if(j %% 10 == 0) print(paste("Boot=",j))
lin.blank_B <- NULL
tmpB <- tmp_all[sample(1:nrow(tmp_all), replace=TRUE),] #Pick ** observations with replacement among all that are available.
#Blank samples are included
weights = rep(0,length(tmpB$C) )
for (kk in 1:length(tmpB$C)){
weights[kk] = 1/var_v_s_unique[which( unique_c == tmpB$C[kk])]
}
noise_B = mean(sample(tmp_blank$I,length(tmp_blank$I),replace=TRUE)) #Mean of resampled noise (= mean of noise)
ii = 0;
if(1) while(ii < NMAX){#Number of ii trials for bilinear fit < NMAX
ii = ii + 1
{
change = median(tmpB$C)*runif(1)*0.25
slope=median(tmpB$I)/median(tmpB$C)*runif(1)
sink("NULL");
#Set intercept at noise and solve for the slope and change
fit.blank_B <- NULL
fit.blank_B <- tryCatch({nlsLM( I ~ .bilinear_LOD(C , noise_B, slope, change),data=tmpB, trace = TRUE,start=c(slope=slope, change=change), weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
}
out_loop = 0
if(!is.null(fit.blank_B)){ #Converges but cannot have a real threshold here anyway
if( summary(fit.blank_B)$coefficient[2] < min(tmpB$C) ){
fit.blank_B <- NULL
out_loop =1
}
}
if(!is.null(fit.blank_B) && out_loop == 0){ #Boostrap again to see whether bilinear fit is real:
change_BB <- rep(NA,NMAX)
#print(paste('Change=',summary(fit.blank_B)$coefficient[2]))
bb = 0;
while(bb < NMAX){#Number of second boostrap samples bb < NMAX
bb = bb + 1
iii = 0
while(iii < NMAX){#Number of iii trials for convergence of bilinear
iii = iii + 1
tmpBB <- tmpB[sample(1:nrow(tmpB), replace=TRUE),]
change = median(tmpBB$C)*runif(1)*0.25
slope=median(tmpBB$I)/median(tmpBB$C)*runif(1)
weightsB = rep(0,length(tmpB$C) )
for (kk in 1:length(tmpBB$C)){
weightsB[kk] = 1/var_v_s_unique[which( unique_c == tmpBB$C[kk])]
}
#Need to also bootstrap for the value of the mean:
#Pick with replacement blank samples:
noise_BB = noise
sink("NULL");
fit.blank_BB <- NULL
fit.blank_BB <- tryCatch({nlsLM( I ~ .bilinear_LOD(C , noise_BB, slope, change),data=tmpBB, trace = TRUE,start=c(slope=slope, change=change), weights = weightsB,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
if(!is.null(fit.blank_BB)){
change_BB[bb] = summary(fit.blank_BB)$coefficient[2]
}else{
change_BB[bb] = NA
}
if(!is.null(fit.blank_BB)) break
} #Number of iii trials for convergence of bilinear
#if(is.null(fit.blank_BB)) print(paste('No convergence for coefficient ',j,bb))
}#Number of second boostrap samples bb < NMAX
CI_change <- quantile(change_BB,probs=c(0.05,0.95),na.rm= TRUE) #95% Confidence interval for the value of change
#Ensure that the 95% confidence interval is included inside the concentration range:
if(is.na(CI_change[1]) || is.na(CI_change[2])){
fit.blank_B <- NULL
out_loop =1
}
if(!is.na(CI_change[1]) && !is.na(CI_change[2])) if(CI_change[1] < min(tmp_all$C) | CI_change[2] > max(tmp_all$C)){
fit.blank_B <- NULL
out_loop =1
}else{
#fit.blank_B <- NULL#print('Acceptable fit')
}
} #Boostrap again to see whether bilinear fit is real
if(out_loop == 1) break #Bilinear fit converges but CI not acceptable
if(!is.null(fit.blank_B)) break #Could never find a converged bilinear fit
} #Number of ii trials for bilinear fit < NMAX
#fit.blank_B <- NULL
if(is.null(fit.blank_B)){ #Do linear fit:
ll = 0
while(ll < NMAX){
ll = ll + 1
slope = median(tmpB$I)/median(tmpB$C)*runif(1)
intercept = noise*runif(1)
sink("NULL");
lin.blank_B <- tryCatch({nlsLM( I ~ .linear(C , intercept, slope),data=tmpB, trace = TRUE,start=c(intercept=intercept, slope=slope), weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
if(!is.null(lin.blank_B)) break
}
} #Do linear fit if is.null(fit.blank_B)
#Store the curve fits obtained via bootstrap with bilinear and linear:
if (!is.null(fit.blank_B)){
outB[j,] <- .bilinear_LOD(xaxis_orig_2, noise_B, summary(fit.blank_B)$coefficient[1] , summary(fit.blank_B)$coefficient[2])
change_B[j] <- summary(fit.blank_B)$coefficient[2]
} else{
if(!is.null(lin.blank_B)){#If linear fit, change = 0 anyway
outB[j,] <- .linear(xaxis_orig_2,summary(lin.blank_B)$coefficient[1] , summary(lin.blank_B)$coefficient[2] )
change_B[j] <- 0
}
else{
outB[j,] <- rep(NA, length(xaxis_orig_2))
change_B[j] <- NA
}
}
for (jj in 1:BB){#Calculate predictions
outBB_pred[(j-1)*BB +jj,] <- outB[j,] + rnorm( length(xaxis_orig_2),0,sqrt(var_v_s_log))
}
} # Number of first bootstrap samples j <=B
#Calculate the variance of the fits:
var_bilinear <- apply(outB, 2, var,na.rm = TRUE)
mean_bilinear <- apply(outB, 2, mean,na.rm = TRUE)
#Calculate confidence interval based on quantiles:
# Do not use: based on normal distribution
##lower_CI = mean_bilinear - 1.96* sqrt(var_bilinear)
##upper_CI = mean_bilinear + 1.96* sqrt(var_bilinear)
mean_pred <- apply(outBB_pred, 2, mean,na.rm = TRUE)
var_pred <- apply(outBB_pred, 2, var,na.rm = TRUE)
lower_Q = apply(outB, 2, quantile, probs=c(0.05) ,na.rm = TRUE)
upper_Q = apply(outB, 2, quantile, probs=c(0.95) ,na.rm = TRUE)
lower_Q_pred = apply(outBB_pred, 2, quantile, probs=c(0.05) ,na.rm = TRUE)
upper_Q_pred = apply(outBB_pred, 2, quantile, probs=c(0.95) ,na.rm = TRUE)
}#Full bootstrap method
LOD_pred_mean_low = 0; LOD_pred_high =0;
#Calculate the LOD/LOQ from the prediction interval:
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred = LOD_pred_mean
y_LOD_pred = up_noise
} else{
LOD_pred = 0
y_LOD_pred = up_noise
}
#Calculate the LOD with the upper-upper and upper-lower limits of the noise (Calculated to make sure that we have a large enough resolution)
i_before = which(diff(sign( up_noise_low - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_low - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_low - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise_high - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_high - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_high - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
i_before = which(diff(sign( up_noise_low - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_low - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_low - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise_high - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_high - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_high - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred = x_inter
y_LOQ_pred = up_noise
} else{
LOQ_pred = 0
y_LOQ_pred = up_noise
}
if(length(LOD_pred) > 1) print('multiple intersection between fit and upper bound of noise, picking first')
LOQ_pred=LOQ_pred[1]
i_before = which(diff(sign( up_noise_low - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_low - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_low - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise_high - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_high - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_high - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_low - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_low - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred_low = x_inter
} else{
LOQ_pred_low = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise_high - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise_high - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred_high = x_inter
} else{
LOQ_pred_high = 0
}
#Want to require at least one point between the various LOD's
found_low_LOD = 0; found_high_LOD = 0; x_low = -7; x_high = -7
found_low_LOQ = 1; found_high_LOQ = 1;
for( x in xaxis_orig_2 ){
if( x < LOD_pred_mean_high && x > LOD_pred){
found_high_LOD =1 ;x_high = x;
}
if( x > LOD_pred_mean_low && x < LOD_pred){
found_low_LOD =1 ; x_low = x;
}
if( x < LOQ_pred_high && x > LOQ_pred){
found_high_LOQ =1 ;x_high = x;
}
if( x > LOQ_pred_low && x < LOQ_pred){
found_low_LOQ =1 ; x_low = x;
}
}
x_new <- NULL
if(found_low_LOD == 0){
x_new = c(x_new, 0.5*(LOD_pred_low + LOD_pred))
}
if(found_high_LOD == 0){
x_new = c(x_new, 0.5*(LOD_pred_high + LOD_pred))
}
if(found_low_LOQ== 0){
x_new = c(x_new, 0.5*(LOQ_pred_low + LOQ_pred))
}
if(found_high_LOQ == 0){
x_new = c(x_new, 0.5*(LOQ_pred_high + LOQ_pred))
}
if(is.null(x_new)) break;
}
#if(is.null(x_new)) break
return(list(CONCENTRATION = xaxis_orig_2, MEAN=mean_bilinear,LOW= lower_Q_pred, UP = upper_Q_pred, LOD= rep(LOD_pred, length(upper_Q_pred)), LOQ = rep(LOQ_pred, length(upper_Q_pred)),
NAME = rep(datain$NAME[1], length(upper_Q_pred)),
METHOD = rep("NONLINEAR", length(upper_Q_pred))
))
}
# DON'T TOUCH THIS ONE! THIS IS THE LEGACY FUNCTION
nonlinear_quantlim <- function(datain, alpha = 0.05, Npoints = 100, Nbootstrap = 500){
switch(Sys.info()[['sysname']],
Windows = {null_output <- "NUL"},
Linux = {null_outpur <- "/dev/null"},
Darwin = {null_output <- "/dev/null"})
#Need to rename variables as needed:
names(datain)[names(datain) == 'CONCENTRATION'] <- 'C'
names(datain)[names(datain) == 'INTENSITY'] <- 'I'
#percentile of the prediction interval considered
if(missing(alpha)){
alpha = 5/100;
}
if( alpha >= 1 | alpha <= 0){
print("incorrect specified value for alpha, 0 < alpha < 1")
return(NULL)
}
#Number of boostrap samples
if(missing(Nbootstrap)){
B <- 500
} else B <= Nbootstrap
#Number of points for the discretization interval
if(missing(Npoints)){
Npoints = 100
}
#Number of bootstrap samples for the prediction inteval for the changepoint
#Large values can make calculations very expensive
B1 <- 30
#Number of prediction samples to generate
BB <- 200
#Number of trials for convergence of every curvefit algorithm
NMAX = 30
datain <- datain[!is.na(datain$I) & !is.na(datain$C),]
datain <- datain[!is.infinite(datain$I) & !is.infinite(datain$C),]
datain <- datain[order(datain$C),]
tmp_nob <- subset(datain,datain$C >0)
tmp_all <- datain
tmp_blank <-subset(datain,datain$C == 0)
#Calculate the value of the noise:
#Use the zero concentration to calculate the LOD:
noise = mean(tmp_blank$I)
var_noise = var(tmp_blank$I)
pb <- txtProgressBar(min = 0, max = 1, initial = 0, char = "%",
width = 40, title, label, style = 1, file = "")
n_blank = length(unique(tmp_blank$I))
if(nrow(tmp_blank) <= 1 || var_noise <= 0){
print("Not enough blank samples!!!")
return(NULL)
}
fac = qt(1-alpha,n_blank - 1)*sqrt(1+1/n_blank)
#upper bound of noise prediction interval
up_noise = noise + fac* sqrt(var_noise)
unique_c = sort(unique(tmp_all$C)); var_v <- rep(0.0, length(unique_c))
weights <- rep(0.0, length(tmp_all$C));
weights_nob <- rep(0.0, length(tmp_nob$C));
#Calculate variance for all concentrations keeping NA values:
ii = 1
for (j in unique_c){
data_f <- subset(tmp_all, C == j)
var_v[ii] <- var(data_f$I)
ii = ii +1
}
#Log scale discretization:
xaxis_orig_2 <- exp(c( seq( from = log(10+0), to = log(1+max(unique_c)), by = log(1+max(unique_c))/Npoints ))) -10 #0.250 to go fast here
xaxis_orig_2 <- unique(sort(c(xaxis_orig_2,unique_c)))
#Instead simply create a piecewise linear approximation:
var_v_lin = approx(unique_c[!is.na(var_v)], var_v[!is.na(var_v)], xout = xaxis_orig_2)$y
var_v_lin_unique = approx(unique_c[!is.na(var_v)], var_v[!is.na(var_v)], xout = unique_c)$y
##
#In the following, consider that the smoothed variance is that from the log:
var_v_s <- var_v_lin
var_v_s_unique <- var_v_lin_unique
var_v_s_log <- var_v_s
var_v_s_log_unique <- var_v_s_unique
##
if(1){# Full bootstrap
outB <- matrix(NA_real_, nrow=B, ncol=length(xaxis_orig_2))
outBB_pred <- matrix(NA_real_, nrow=B*BB, ncol=length(xaxis_orig_2))
change_B <- rep(NA, B)
set.seed(123)
for (j in 1:B){ #Number of first boostrap samples j
setTxtProgressBar(pb, j/B, title = NULL, label = NULL)
lin.blank_B <- NULL
tmpB <- tmp_all[sample(1:nrow(tmp_all), replace=TRUE),] #Pick ** observations with replacement among all that are available.
#Blank samples are included
weights = rep(0,length(tmpB$C) )
for (kk in 1:length(tmpB$C)){
weights[kk] = 1/var_v_s_unique[which( unique_c == tmpB$C[kk])]
}
noise_B = mean(sample(tmp_blank$I,length(tmp_blank$I),replace=TRUE)) #Mean of resampled noise (= mean of noise)
ii = 0;
if(1) while(ii < NMAX){#Number of ii trials for bilinear fit < NMAX
ii = ii + 1
{
change = median(tmpB$C)*runif(1)*0.25
slope=median(tmpB$I)/median(tmpB$C)*runif(1)
sink(null_output);
#Set intercept at noise and solve for the slope and change
fit.blank_B <- NULL
fit.blank_B <- tryCatch({nlsLM( I ~ .bilinear_LOD(C , noise_B, slope, change),data=tmpB, trace = TRUE,start=c(slope=slope, change=change), weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
}
out_loop = 0
if(!is.null(fit.blank_B)){ #Converges but cannot have a real threshold here anyway
if( summary(fit.blank_B)$coefficient[2] < min(tmpB$C) ){
fit.blank_B <- NULL
out_loop =1
}
}
if(!is.null(fit.blank_B) && out_loop == 0){ #Boostrap again to see whether bilinear fit is real:
change_BB <- rep(NA,B1)
bb = 0;
while(bb < B1){#Number of second boostrap samples bb < NMAX
bb = bb + 1
iii = 0
while(iii < NMAX){#Number of iii trials for convergence of bilinear
iii = iii + 1
tmpBB <- tmpB[sample(1:nrow(tmpB), replace=TRUE),]
change = median(tmpBB$C)*runif(1)*0.25
slope=median(tmpBB$I)/median(tmpBB$C)*runif(1)
weightsB = rep(0,length(tmpB$C) )
for (kk in 1:length(tmpBB$C)){
weightsB[kk] = 1/var_v_s_unique[which( unique_c == tmpBB$C[kk])]
}
#Need to also bootstrap for the value of the mean:
#Pick with replacement blank samples:
noise_BB = noise
sink(null_output);
fit.blank_BB <- NULL
fit.blank_BB <- tryCatch({nlsLM( I ~ .bilinear_LOD(C , noise_BB, slope, change),data=tmpBB, trace = TRUE,start=c(slope=slope, change=change), weights = weightsB,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
if(!is.null(fit.blank_BB)){
change_BB[bb] = summary(fit.blank_BB)$coefficient[2]
}else{
change_BB[bb] = NA
}
if(!is.null(fit.blank_BB)) break
} #Number of iii trials for convergence of bilinear
#if(sum(is.na(change_BB))>10 && mean(change_BB, na.rm = TRUE) < 0){ out_loop = 1; break; }
}#Number of second boostrap samples bb < B1
# #print(change_BB[order(change_BB)])
# CI_change <- quantile(change_BB,probs=c(0.1),na.rm= TRUE)
#
# #Ensure that the 95% confidence interval is included inside the concentration range:
# if(is.na(CI_change[1] && out_loop == 0) ){
# fit.blank_B <- NULL
# out_loop =1
# }
#
#
# if(!is.na(CI_change[1]) && out_loop == 0) if(CI_change[1] < min(tmp_all$C)){
# fit.blank_B <- NULL
# out_loop =1
# }else{
# #fit.blank_B <- NULL#print('Acceptable fit')
# }
#
CI_change <- quantile(change_BB,probs=c(0.05,1-0.05),na.rm= TRUE) #90% Confidence interval for the value of change
#Ensure that the 90% confidence interval is included inside the concentration range:
if(is.na(CI_change[1]) || is.na(CI_change[2])){
fit.blank_B <- NULL
out_loop =1
}
if(!is.na(CI_change[1]) && !is.na(CI_change[2])) if(CI_change[1] < min(tmp_all$C) | CI_change[2] > max(tmp_all$C)){
fit.blank_B <- NULL
out_loop =1
}else{
#fit.blank_B <- NULL#print('Acceptable fit')
}
} #Boostrap again to see whether bilinear fit is real
if(out_loop == 1) break #Bilinear fit converges but CI not acceptable
if(!is.null(fit.blank_B)) break #Could never find a converged bilinear fit
} #Number of ii trials for bilinear fit < NMAX
#fit.blank_B <- NULL
if(is.null(fit.blank_B)){ #Do linear fit:
ll = 0
while(ll < NMAX){
ll = ll + 1
slope = median(tmpB$I)/median(tmpB$C)*runif(1)
intercept = noise*runif(1)
sink(null_output);
lin.blank_B <- tryCatch({nlsLM( I ~ .linear(C , intercept, slope),data=tmpB, trace = TRUE,start=c(intercept=intercept, slope=slope), weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
if(!is.null(lin.blank_B)) break
}
} #Do linear fit if is.null(fit.blank_B)
#Store the curve fits obtained via bootstrap with bilinear and linear:
if (!is.null(fit.blank_B)){
outB[j,] <- .bilinear_LOD(xaxis_orig_2, noise_B, summary(fit.blank_B)$coefficient[1] , summary(fit.blank_B)$coefficient[2])
change_B[j] <- summary(fit.blank_B)$coefficient[2]
} else{
if(!is.null(lin.blank_B)){#If linear fit, change = 0 anyway
outB[j,] <- .linear(xaxis_orig_2,summary(lin.blank_B)$coefficient[1] , summary(lin.blank_B)$coefficient[2] )
change_B[j] <- 0
}
else{
outB[j,] <- rep(NA, length(xaxis_orig_2))
change_B[j] <- NA
}
}
for (jj in 1:BB){#Calculate predictions
outBB_pred[(j-1)*BB +jj,] <- outB[j,] + rnorm( length(xaxis_orig_2),0,sqrt(var_v_s_log))
}
} # Number of first bootstrap samples j <=B
#Calculate the variance of the fits:
var_bilinear <- apply(outB, 2, var,na.rm = TRUE)
mean_bilinear <- apply(outB, 2, mean,na.rm = TRUE)
mean_pred <- apply(outBB_pred, 2, mean,na.rm = TRUE)
var_pred <- apply(outBB_pred, 2, var,na.rm = TRUE)
lower_Q_pred = apply(outBB_pred, 2, quantile, probs=c(alpha) ,na.rm = TRUE)
upper_Q_pred = apply(outBB_pred, 2, quantile, probs=c(1 - alpha) ,na.rm = TRUE)
}#Full bootstrap method
LOD_pred_mean_low = 0; LOD_pred_high =0;
#Calculate the LOD/LOQ from the prediction interval:
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred = LOD_pred_mean
y_LOD_pred = up_noise
} else{
LOD_pred = 0
y_LOD_pred = up_noise
}
#Calculate the LOD with the upper-upper and upper-lower limits of the noise (Calculated to make sure that we have a large enough resolution)
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred = x_inter
y_LOQ_pred = up_noise
} else{
LOQ_pred = 0
y_LOQ_pred = up_noise
}
if(length(LOD_pred) > 1){ print('multiple intersection between fit and upper bound of noise, picking first')
LOQ_pred=LOQ_pred[1]}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred_low = x_inter
} else{
LOQ_pred_low = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred_high = x_inter
} else{
LOQ_pred_high = 0
}
#Calculate slope and intercept above the LOD/LOQ:
slope_lin <- NA
intercept_lin <- NA
data_linear <- datain[datain$C > LOQ_pred,]
ii = 0
while(ii < NMAX){#Number of ii trials for bilinear fit < NMAX
ii = ii + 1
{
intercept = data_linear$I[1]*runif(1)
slope=median(data_linear$I)/median(data_linear$C)*runif(1)
weights = rep(0,length(data_linear$C) )
for (kk in 1:length(data_linear$C)){
weights[kk] = 1/var_v_s[which( unique_c == data_linear$C[kk])]
}
sink(null_output);
fit.blank_lin <- NULL
fit.blank_lin <- tryCatch({nlsLM( I ~ .linear(C , intercept, slope),data=data_linear, trace = TRUE,start=c(slope=slope, intercept = intercept), weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
}
if(!is.null(fit.blank_lin)) break
}#ii < NMAX
if(!is.null(fit.blank_lin)){
slope_lin <- summary(fit.blank_lin)$coefficients[[1]]
intercept_lin <- summary(fit.blank_lin)$coefficients[[2]]
}
if(is.null(fit.blank_lin)) print("Linear assay characterization above LOD did not converge")
return(
data.frame(as.data.frame(list(CONCENTRATION = xaxis_orig_2, MEAN=mean_bilinear,LOW= lower_Q_pred, UP = upper_Q_pred, LOB= rep(LOD_pred, length(upper_Q_pred)), LOD = rep(LOQ_pred, length(upper_Q_pred)),
SLOPE =slope_lin , INTERCEPT =intercept_lin, NAME = rep(datain$NAME[1], length(upper_Q_pred)),
METHOD = rep("NONLINEAR", length(upper_Q_pred))
)
))
)
}
nonlinear_quantlim_modded <- function(datain, alpha = 0.05, Npoints = 100, Nbootstrap = 500){
switch(Sys.info()[['sysname']],
Windows = {null_output <- "NULL"},
Linux = {null_output <- "/dev/null"},
Darwin = {null_output <- "/dev/null"})
#Need to rename variables as needed:
names(datain)[names(datain) == 'CONCENTRATION'] <- 'C'
names(datain)[names(datain) == 'INTENSITY'] <- 'I'
#percentile of the prediction interval considered
if(missing(alpha)){
alpha = 5/100;
}
if( alpha >= 1 | alpha <= 0){
print("incorrect specified value for alpha, 0 < alpha < 1")
return(NULL)
}
#Number of boostrap samples
if(missing(Nbootstrap)){
B <- 500
} else B <= Nbootstrap
#Number of points for the discretization interval
if(missing(Npoints)){
Npoints = 100
}
#Number of bootstrap samples for the prediction inteval for the changepoint
#Large values can make calculations very expensive
B1 <- 30
#Number of prediction samples to generate
BB <- 200
#Number of trials for convergence of every curvefit algorithm
NMAX = 30
datain <- datain[!is.na(datain$I) & !is.na(datain$C),]
datain <- datain[!is.infinite(datain$I) & !is.infinite(datain$C),]
datain <- datain[order(datain$C),]
tmp_blank <- subset(datain,datain$C == 0)
#datain <- subset(datain,datain$I >0 & datain$C >0) # PATCH pt3(?) TO SEE IF MODEL WILL CONVERGE WHEN THERE'S NOT MANY POINTS with I>0
tmp_nob <- subset(datain,datain$C >0)
tmp_all <- datain
#Calculate the value of the noise:
#Use the zero concentration to calculate the LOD:
noise = mean(tmp_blank$I)
var_noise = var(tmp_blank$I)
#print(paste("Variance of noise =", var_noise))
pb <- txtProgressBar(min = 0, max = 1, initial = 0, char = "%",
width = 40, title, label, style = 1, file = "")
## must be unique intensity values!!!
n_blank = length(unique(tmp_blank$I))
#if(nrow(tmp_blank) <= 1 || var_noise <= 0){
# print("Not enough blank samples or variance is <=0!")
# return(NULL)
#}
unique_c = sort(unique(tmp_all$C)); # list unique concentration points
var_v <- rep(0.0, length(unique_c)) # intialize vector to hold variance of each point
weights <- rep(0.0, length(tmp_all$C)); # initialize vector to hold weights for each concentration point
weights_nob <- rep(0.0, length(tmp_nob$C)); # initialize vector to hold weights for each point, minus blank points
#Calculate variance for all concentrations keeping NA values:
ii = 1
for (j in unique_c){
data_f <- subset(tmp_all, C == j)
var_v[ii] <- var(data_f$I)
ii = ii +1
}
## BLANK VARIANCE ZERO PATCH
iii = 1
for (j in unique_c){
data_f <- subset(tmp_all, C == j)
## patch to account for when the blank has values of 0
if (min(data_f$I) > 0){
nonzeroC <- iii # save the first nonzero concentration point
n_blank <- length(unique(data_f$I)) # set the number of blanks to this
noise <- mean(data_f$I) # set the noise to the mean of this nonzero point
var_blank <- var(data_f$I)
var_noise <- var_blank
tmp_blank <- data_f
break
}
iii = iii +1
}
fac = qt(1-alpha,n_blank - 1)*sqrt(1+1/n_blank) ## qt() requires n_blank >= 2
#upper bound of noise prediction interval
up_noise = noise + fac* sqrt(var_noise) ## ok this is a problem... the noise threshold is way too high
#Log scale discretization:
xaxis_orig_2 <- exp(c( seq( from = log(10+0), to = log(1+max(unique_c)), by = log(1+max(unique_c))/Npoints ))) -10 #0.250 to go fast here
xaxis_orig_2 <- unique(sort(c(xaxis_orig_2,unique_c)))
#Instead simply create a piecewise linear approximation:
var_v_lin = approx(unique_c[!is.na(var_v)], var_v[!is.na(var_v)], xout = xaxis_orig_2)$y
var_v_lin_unique = approx(unique_c[!is.na(var_v)], var_v[!is.na(var_v)], xout = unique_c)$y
##
#In the following, consider that the smoothed variance is that from the log:
var_v_s <- var_v_lin
var_v_s_unique <- var_v_lin_unique
var_v_s_log <- var_v_s
var_v_s_log_unique <- var_v_s_unique
##
if(1){# Full bootstrap
outB <- matrix(NA_real_, nrow=B, ncol=length(xaxis_orig_2))
outBB_pred <- matrix(NA_real_, nrow=B*BB, ncol=length(xaxis_orig_2))
change_B <- rep(NA, B)
set.seed(123)
for (j in 1:B){ #Number of first boostrap samples j
setTxtProgressBar(pb, j/B, title = NULL, label = NULL)
lin.blank_B <- NULL
tmpB <- tmp_all[sample(1:nrow(tmp_all), replace=TRUE),] #Pick ** observations with replacement among all that are available.
#Blank samples are included
weights = rep(0,length(tmpB$C) )
for (kk in 1:length(tmpB$C)){
weights[kk] = 1/var_v_s_unique[which( unique_c == tmpB$C[kk])]
}
## dirty fix for when var(I) of point C is 0, resulting in weight 1/0=Inf
#weights[is.infinite(weights)] <- min(weights[is.finite(weights)])
#weights[is.infinite(weights)] <- max(weights[is.finite(weights)])
weights[is.infinite(weights)] <- 0
noise_B = mean(sample(tmp_blank$I,length(tmp_blank$I),replace=TRUE)) #Mean of resampled noise (= mean of noise)
ii = 0;
if(1) while(ii < NMAX){#Number of ii trials for bilinear fit < NMAX
ii = ii + 1
{
change = median(tmpB$C)*runif(1)*0.25
slope=median(tmpB$I)/median(tmpB$C)*runif(1)
sink(null_output);
#Set intercept at noise and solve for the slope and change
fit.blank_B <- NULL
fit.blank_B <- tryCatch({nlsLM( I ~ .bilinear_LOD(C , noise_B, slope, change),
data=tmpB,
trace = TRUE,
start=c(slope=slope, change=change),
weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))},
error = function(e) {NULL})
sink();
}
out_loop = 0
if(!is.null(fit.blank_B)){ #Converges but cannot have a real threshold here anyway
if( summary(fit.blank_B)$coefficient[2] < min(tmpB$C) ){
fit.blank_B <- NULL
out_loop =1
}
}
if(!is.null(fit.blank_B) && out_loop == 0){ #Boostrap again to see whether bilinear fit is real:
change_BB <- rep(NA,B1)
bb = 0;
while(bb < B1){#Number of second boostrap samples bb < NMAX
bb = bb + 1
iii = 0
while(iii < NMAX){#Number of iii trials for convergence of bilinear
iii = iii + 1
tmpBB <- tmpB[sample(1:nrow(tmpB), replace=TRUE),]
change = median(tmpBB$C)*runif(1)*0.25
slope = median(tmpBB$I)/median(tmpBB$C)*runif(1)
weightsB = rep(0,length(tmpB$C) )
for (kk in 1:length(tmpBB$C)){
weightsB[kk] = 1/var_v_s_unique[which( unique_c == tmpBB$C[kk])]
}
## dirty fix for when var(I) of point C is 0, resulting in weight 1/0=Inf
#weightsB[is.infinite(weightsB)] <- min(weightsB[is.finite(weightsB)])
#weightsB[is.infinite(weightsB)] <- max(weightsB[is.finite(weightsB)])
weightsB[is.infinite(weightsB)] <- 0
#Need to also bootstrap for the value of the mean:
#Pick with replacement blank samples:
noise_BB = noise
sink(null_output);
fit.blank_BB <- NULL
fit.blank_BB <- tryCatch({nlsLM( I ~ .bilinear_LOD(C , noise_BB, slope, change),
data=tmpBB,
trace = TRUE,
start=c(slope=slope, change=change),
weights = weightsB,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))},
error = function(e) {NULL})
sink();
if(!is.null(fit.blank_BB)){
change_BB[bb] = summary(fit.blank_BB)$coefficient[2]
}else{
change_BB[bb] = NA
}
if(!is.null(fit.blank_BB)) break
} #Number of iii trials for convergence of bilinear
#if(sum(is.na(change_BB))>10 && mean(change_BB, na.rm = TRUE) < 0){ out_loop = 1; break; }
}#Number of second boostrap samples bb < B1
# #print(change_BB[order(change_BB)])
# CI_change <- quantile(change_BB,probs=c(0.1),na.rm= TRUE)
#
# #Ensure that the 95% confidence interval is included inside the concentration range:
# if(is.na(CI_change[1] && out_loop == 0) ){
# fit.blank_B <- NULL
# out_loop =1
# }
#
#
# if(!is.na(CI_change[1]) && out_loop == 0) if(CI_change[1] < min(tmp_all$C)){
# fit.blank_B <- NULL
# out_loop =1
# }else{
# #fit.blank_B <- NULL#print('Acceptable fit')
# }
#
CI_change <- quantile(change_BB,probs=c(0.05,1-0.05),na.rm= TRUE) #90% Confidence interval for the value of change
#Ensure that the 90% confidence interval is included inside the concentration range:
if(is.na(CI_change[1]) || is.na(CI_change[2])){
fit.blank_B <- NULL
out_loop =1
}
if(!is.na(CI_change[1]) && !is.na(CI_change[2])) if(CI_change[1] < min(tmp_all$C) | CI_change[2] > max(tmp_all$C)){
fit.blank_B <- NULL
out_loop =1
}else{
#fit.blank_B <- NULL#print('Acceptable fit')
}
} #Boostrap again to see whether bilinear fit is real
if(out_loop == 1) break #Bilinear fit converges but CI not acceptable
if(!is.null(fit.blank_B)) break #Could never find a converged bilinear fit
} #Number of ii trials for bilinear fit < NMAX
#fit.blank_B <- NULL
if(is.null(fit.blank_B)){ #Do linear fit:
ll = 0
while(ll < NMAX){
ll = ll + 1
slope = median(tmpB$I)/median(tmpB$C)*runif(1)
intercept = noise*runif(1)
sink(null_output);
lin.blank_B <- tryCatch({nlsLM( I ~ .linear(C , intercept, slope),
data=tmpB,
trace = TRUE,
start=c(intercept=intercept, slope=slope),
weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))},
error = function(e) {NULL})
sink();
if(!is.null(lin.blank_B)) break
}
} #Do linear fit if is.null(fit.blank_B)
#Store the curve fits obtained via bootstrap with bilinear and linear:
if (!is.null(fit.blank_B)){
outB[j,] <- .bilinear_LOD(xaxis_orig_2, noise_B, summary(fit.blank_B)$coefficient[1] , summary(fit.blank_B)$coefficient[2])
change_B[j] <- summary(fit.blank_B)$coefficient[2]
} else{
if(!is.null(lin.blank_B)){#If linear fit, change = 0 anyway
outB[j,] <- .linear(xaxis_orig_2,summary(lin.blank_B)$coefficient[1] , summary(lin.blank_B)$coefficient[2] )
change_B[j] <- 0
}
else{
outB[j,] <- rep(NA, length(xaxis_orig_2))
change_B[j] <- NA
}
}
for (jj in 1:BB){#Calculate predictions
outBB_pred[(j-1)*BB +jj,] <- outB[j,] + rnorm( length(xaxis_orig_2),0,sqrt(var_v_s_log))
}
} # Number of first bootstrap samples j <=B
#Calculate the variance of the fits:
var_bilinear <- apply(outB, 2, var,na.rm = TRUE)
mean_bilinear <- apply(outB, 2, mean,na.rm = TRUE)
mean_pred <- apply(outBB_pred, 2, mean,na.rm = TRUE)
var_pred <- apply(outBB_pred, 2, var,na.rm = TRUE)
lower_Q_pred = apply(outBB_pred, 2, quantile, probs=c(alpha) ,na.rm = TRUE)
upper_Q_pred = apply(outBB_pred, 2, quantile, probs=c(1 - alpha) ,na.rm = TRUE)
}#Full bootstrap method
LOD_pred_mean_low = 0; LOD_pred_high =0;
#Calculate the LOD/LOQ from the prediction interval:
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred = LOD_pred_mean
y_LOD_pred = up_noise
} else{
LOD_pred = 0
y_LOD_pred = up_noise
}
#Calculate the LOD with the upper-upper and upper-lower limits of the noise (Calculated to make sure that we have a large enough resolution)
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred = x_inter
y_LOQ_pred = up_noise
} else{
LOQ_pred = 0
y_LOQ_pred = up_noise
}
if(length(LOD_pred) > 1){ print('multiple intersection between fit and upper bound of noise, picking first')
LOQ_pred=LOQ_pred[1]}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_low = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_low = LOD_pred_mean_low
y_LOD_pred_low = up_noise
} else{
LOD_pred_low = 0
y_LOD_pred_low = up_noise
}
i_before = which(diff(sign( up_noise - mean_bilinear ))!=0) #before sign change
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - mean_bilinear)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - mean_bilinear)[i_after]
#Linear interpollation to find where the function changes sign:
LOD_pred_mean_high = x1 - f1*(x2-x1)/(f2-f1)
LOD_pred_high = LOD_pred_mean_high
} else{
LOD_pred_mean_high = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred_low = x_inter
} else{
LOQ_pred_low = 0
}
#Do a linear fit to find the intersection:
i_before = which(diff(sign(up_noise - lower_Q_pred))!=0)
if(length(i_before)>0){
i_after = i_before+1
x1 = xaxis_orig_2[i_before]; f1 = (up_noise - lower_Q_pred)[i_before]
x2 = xaxis_orig_2[i_after]; f2 = (up_noise - lower_Q_pred)[i_after]
x_inter = x1 - f1*(x2-x1)/(f2-f1)
LOQ_pred_high = x_inter
} else{
LOQ_pred_high = 0
}
#Calculate slope and intercept above the LOD/LOQ:
slope_lin <- NA
intercept_lin <- NA
data_linear <- datain[datain$C > LOQ_pred,]
ii = 0
while(ii < NMAX){#Number of ii trials for bilinear fit < NMAX
ii = ii + 1
{
intercept = data_linear$I[1]*runif(1)
slope=median(data_linear$I)/median(data_linear$C)*runif(1)
weights = rep(0,length(data_linear$C) )
for (kk in 1:length(data_linear$C)){
weights[kk] = 1/var_v_s[which( unique_c == data_linear$C[kk])]
}
sink(null_output);
fit.blank_lin <- NULL
fit.blank_lin <- tryCatch({nlsLM( I ~ .linear(C , intercept, slope),data=data_linear, trace = TRUE,start=c(slope=slope, intercept = intercept), weights = weights,
control = nls.lm.control(nprint=1,ftol = sqrt(.Machine$double.eps)/2, maxiter = 50))}, error = function(e) {NULL}
)
sink();
}
if(!is.null(fit.blank_lin)) break
}#ii < NMAX
if(!is.null(fit.blank_lin)){
slope_lin <- summary(fit.blank_lin)$coefficients[[1]]
intercept_lin <- summary(fit.blank_lin)$coefficients[[2]]
}
if(is.null(fit.blank_lin)) print("Linear assay characterization above LOD did not converge")
return(
data.frame(as.data.frame(list(CONCENTRATION = xaxis_orig_2,
MEAN=mean_bilinear,
LOW= lower_Q_pred,
UP = upper_Q_pred,
LOB= rep(LOD_pred, length(upper_Q_pred)),
LOD = rep(LOQ_pred, length(upper_Q_pred)),
SLOPE =slope_lin ,
INTERCEPT =intercept_lin,
NAME = rep(datain$NAME[1], length(upper_Q_pred)),
METHOD = rep("NONLINEAR", length(upper_Q_pred)))))
)
}
#rm(list = ls())
library(tidyverse)
library(gplots)
library(lme4)
library(ggrepel)
library(reshape)
library(reshape2)
library(data.table)
library(Rcpp)
library(survival)
library(limma)
library(marray)
library(preprocessCore)
library(MSnbase)
#library(MSstats, lib.loc= "/net/maccoss/vol6/home/lpino/R/3.3.0/libs")
library(minpack.lm)
library(dplyr)
#curve.df <- read.csv("C:/Users/Lindsay/Documents/proj/MSstats-patch/MSstats-patch/dev/calibration_data_norm.csv", header= TRUE, stringsAsFactors = FALSE)
#curve.df <- read.csv("C:/Users/Lindsay/Documents/proj/MSstats-patch/MSstats-patch/dev/TEST_df.csv", header= TRUE, stringsAsFactors = FALSE)
#curve.df <- read.csv("C:/Users/Lindsay/Documents/proj/dia-quant/data/2016analyses/20160329lkp_CalibrationCurves_quantlib.elib.peptides.searle2msstats.csv", header=TRUE, stringsAsFactors=FALSE)
#curve.df <- read.csv("C:/Users/lpino/Documents/proj/MSstats-patch/dev/UPS_water_curve_encyclopedia.elib.peptides.MELTED.csv", header=TRUE, stringsAsFactors=FALSE)
curve.df <- read.csv("C:/Users/lpino/Documents/proj/dia-quant/data/20170326_YeastQuant_CalibrationCurve_201705analysis.elib.peptides.MELTED.csv", header=TRUE, stringsAsFactors=FALSE)
#peptide <- "LPPGLLANFTLLR" #nonlinear
#peptide <- "FVGTPEVNQTTLYQR" #linear
#peptide <- "BLANKVARIANCENONZERO"
#peptide <- "BLANKVARIANCEZERO"
#peptide <- "ALALGSSTVMMGGMLAGTTESPGEYFYK" # gives null result on cluster
#peptide <- "VSLPSVPSNK"
peptide <- sample(unique(curve.df$NAME),size=1,replace=FALSE)
subset.df <- curve.df %>% filter(NAME == peptide)
subset.df <- na.omit(subset.df)
boxplot(formula = INTENSITY ~ CONCENTRATION, data=subset.df)
testout.vanilla <- nonlinear_quantlim(subset.df)
testout.modded <- nonlinear_quantlim_modded(subset.df)
plot_quantlim(spikeindata = subset.df, quantlim_out = testout.modded, dir_output=getwd())
# calculate LOD/LOQ for each peptide, storing plots in a designated directory
peptide.batch <- unique(curve.df$NAME)
date <- Sys.Date()
fo <- paste(getwd(), "/proj/MSstats-patch/dev/", date, ".txt", sep="")
counter <- 1 # inititalize counter for "for" loop below
for (peptide in peptide.batch){
time <- Sys.time()
print(paste("counter=",counter)) # sanity check
print(time)
print(peptide)
counter <- counter + 1 # increment counter to ensure update
#df_in contains data for peptide i
df_in <- curve.df %>% filter(NAME == peptide)
#Count the number of blank samples for the peptide (with a non NA intensity) [commented out 'and with different values']
df_blank = df_in %>% filter(CONCENTRATION == 0 & !is.na(INTENSITY)) #%>% distinct(INTENSITY)
#n_blank = number of "acceptable" blank samples:
n_blank = nrow(df_blank)
print(paste("n_blank=",n_blank))
if(n_blank <= 1) {next}
df_out <- nonlinear_quantlim_modded(df_in)
try(plot_quantlim(spikeindata = df_in, quantlim_out = df_out, dir_output=paste(getwd(),'/proj/MSstats-patch/dev/',date,sep="")))
# write the nonlinear_quantlim() results to an outfile for more downstream processing
if(counter == 1){
write.table(df_out, file=fo, append=FALSE, col.names=TRUE)
} else {
try(write.table(df_out, file=fo, append=TRUE, col.names=FALSE))
}
print(df_out)
}
for (peptide in peptide.batch){
df_in <- curve.df %>% filter(NAME == peptide)
plot(x=df_in$CONCENTRATION,y=df_in$INTENSITY)
fo <- paste(getwd(), "/proj/MSstats-patch/dev/afed/", peptide, ".jpg", sep="")
dev.copy(jpeg,filename=fo);
dev.off ();
}
print(df_out)
close(fo)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.