inst/doc/Keras.R

## ----setup, include = FALSE---------------------------------------------------
knitr::opts_chunk$set(
  collapse = TRUE,
  comment = "#>",
  fig.width=5, fig.height=5 ,fig.align="center", eval=FALSE
)
fpath <- ""

## -----------------------------------------------------------------------------
#  library(keras)
#  library(condvis2)
#  library(MASS)
#  set.seed(123)

## -----------------------------------------------------------------------------
#  # Training features
#  Pima.training <- Pima.tr[,1:7]
#  # Testing features
#  Pima.testing <- Pima.te[,1:7]
#  
#  # Scale the data
#  Pima.training <-as.matrix(scale(Pima.training))
#  means <- attr(Pima.training,"scaled:center")
#  sds<- attr(Pima.training,"scaled:scale")
#  Pima.testing <- as.matrix(scale(Pima.testing, center=means, scale=sds))
#  
#  # One hot encode training target values
#  Pima.trainLabels <- to_categorical(as.numeric(Pima.tr[,8]) -1)[, 2]
#  # One hot encode test target values
#  Pima.testLabels <- to_categorical(as.numeric(Pima.te[,8]) -1)[, 2]
#  
#  # Create dataframes for Condvis
#  dtf <- data.frame(Pima.training)
#  dtf$Pima.trainLabels <- Pima.tr[,8]
#  
#  dtf.te <- data.frame(Pima.testing)
#  dtf.te$Pima.testLabels <- Pima.te[,8]

## -----------------------------------------------------------------------------
#  
#  model <- keras_model_sequential() # Add layers to the model
#  model %>%
#    layer_dense(units = 8, activation = 'tanh', input_shape = c(7)) %>%
#    layer_dense(units = 1, activation = 'sigmoid')
#  
#  # Print a summary of a model
#  summary(model)
#  
#  
#  # Compile the model
#  model %>% compile(
#    loss = 'binary_crossentropy',
#    optimizer = 'adam',
#    metrics = 'accuracy'
#  )
#  
#  # Fit the model
#  history <-model %>% fit(Pima.training, Pima.trainLabels,
#                epochs = 500,
#                batch_size = 50,
#                    validation_split = 0.2,
#                class_weight = as.list(c("0" = 1, "1"=3))
#  )

## -----------------------------------------------------------------------------
#  kresponse <- "Pima.testLabels"
#  kpreds <- setdiff(names(dtf.te),kresponse)
#  CVpredict(model, dtf.te[1:10,], response=kresponse, predictors=kpreds)
#  CVpredict(model, dtf.te[1:10,], response=kresponse, predictors=kpreds, ptype="prob")
#  CVpredict(model, dtf.te[1:10,], response=kresponse, predictors=kpreds, ptype="probmatrix")

## -----------------------------------------------------------------------------
#  mean(CVpredict(model, dtf.te, response=kresponse, predictors=kpreds) == dtf.te$Pima.testLabels)

## -----------------------------------------------------------------------------
#  
#  fit.lda <- lda(Pima.trainLabels~., data = dtf)
#  mean(CVpredict(fit.lda, dtf.te) == dtf.te$Pima.testLabels)
#  

## ----eval=F-------------------------------------------------------------------
#  kresponse <- "Pima.trainLabels"
#  kArgs1 <-  list(response=kresponse,predictors=kpreds)
#  condvis(dtf, list(model.keras = model, model.lda = fit.lda), sectionvars = c("bmi", "glu"), response="Pima.trainLabels",predictArgs = list(kArgs1, NULL), pointColor = "Pima.trainLabels")
#  

## ----echo=FALSE, out.width='100%', eval=T-------------------------------------
knitr::include_graphics(paste0(fpath, "Pima.png"))

## -----------------------------------------------------------------------------
#  boston_housing <- dataset_boston_housing()
#  c(train_data, train_labels) %<-% boston_housing$train
#  c(test_data, test_labels) %<-% boston_housing$test
#  
#  # Normalize training data
#  train_data <- scale(train_data)
#  
#  # Use means and standard deviations from training set to normalize test set
#  col_means_train <- attr(train_data, "scaled:center")
#  col_stddevs_train <- attr(train_data, "scaled:scale")
#  test_data <- scale(test_data, center = col_means_train, scale = col_stddevs_train)
#  

## -----------------------------------------------------------------------------
#  build_model <- function() {
#  
#    model <- keras_model_sequential() %>%
#      layer_dense(units = 64, activation = "relu",
#                  input_shape = dim(train_data)[2]) %>%
#      layer_dense(units = 64, activation = "relu") %>%
#      layer_dense(units = 1)
#  
#    model %>% compile(
#      loss = "mse",
#      optimizer = optimizer_rmsprop(),
#      metrics = list("mean_absolute_error")
#    )
#  
#    model
#  }
#  
#  model <- build_model()
#  model %>% summary()
#  
#  
#  # Display training progress by printing a single dot for each completed epoch.
#  print_dot_callback <- callback_lambda(
#    on_epoch_end = function(epoch, logs) {
#      if (epoch %% 80 == 0) cat("\n")
#      cat(".")
#    }
#  )
#  
#  epochs <- 500
#  
#  # Fit the model
#  early_stop <- callback_early_stopping(monitor = "val_loss", patience = 20)
#  
#  model <- build_model()
#  history <- model %>% fit(
#    train_data,
#    train_labels,
#    epochs = epochs,
#    validation_split = 0.2,
#    verbose = 0,
#    callbacks = list(early_stop, print_dot_callback)
#  )
#  

## -----------------------------------------------------------------------------
#  
#  column_names <- c('CRIM', 'ZN', 'INDUS', 'CHAS', 'NOX', 'RM', 'AGE',
#                    'DIS', 'RAD', 'TAX', 'PTRATIO', 'B', 'LSTAT')
#  train_df <- data.frame(train_data)
#  colnames(train_df) <- column_names
#  train_df$medv <- as.numeric(train_labels)
#  
#  test_df <- data.frame(test_data)
#  colnames(test_df) <- column_names
#  test_df$medv <- as.numeric(test_labels)
#  kpreds <- column_names
#  kresponse <- "medv"

## -----------------------------------------------------------------------------
#  suppressMessages(library(mgcv))
#  gam.model = gam(medv ~ s(LSTAT) + s(RM) + s(CRIM), data=train_df)
#  
#  suppressMessages(library(randomForest))
#  rf.model <- randomForest(formula = medv ~ ., data = train_df)
#  

## -----------------------------------------------------------------------------
#  
#  mean((test_labels - CVpredict(model, test_df, response=kresponse, predictors=kpreds))^2)
#  mean((test_labels - CVpredict(gam.model, test_df))^2)
#  mean((test_labels - CVpredict(rf.model, test_df))^2, na.rm=TRUE)

## ----eval = FALSE-------------------------------------------------------------
#  
#  kArgs <-  list(response=kresponse,predictors=kpreds)
#  condvis(train_df, list(gam = gam.model, rf = rf.model, kerasmodel = model), sectionvars = c("LSTAT", "RM"),predictArgs = list(NULL, NULL, kArgs) )
#  

## ----echo=FALSE, out.width='100%', eval=T-------------------------------------
knitr::include_graphics(paste0(fpath, "Boston.png"))

Try the condvis2 package in your browser

Any scripts or data that you put into this service are public.

condvis2 documentation built on Sept. 14, 2022, 5:06 p.m.