This package is for training large data random competing risks forests. See https://github.com/jatherrien/largeRCRF for reporting issues.
Find a file
2019-07-05 12:08:03 -07:00
data Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00
inst/java Update steps for obtaining Java code 2019-07-03 09:35:45 -07:00
java Update steps for obtaining Java code 2019-07-03 09:35:45 -07:00
man Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00
R Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00
tests Version 1.0.1; cleaned up documentation. 2019-06-30 15:07:29 -07:00
vignettes Small update to vignette 2019-07-05 12:08:03 -07:00
.gitignore Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00
.Rbuildignore Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00
COPYING Version 1.0 release; update README. 2019-06-28 11:15:53 -07:00
DESCRIPTION Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00
NAMESPACE Slight cleanup 2019-06-30 15:41:33 -07:00
README.md Include wihs data and vignette into package 2019-07-05 11:56:53 -07:00

README

This R package is used to train random competing risks forests, ideally for large data. It's based heavily off of randomForestSRC, although there are some differences.

This package is not yet on CRAN, so in the meantime to install it use the devtools package and run the following command:

R> devtools::install_git("https://github.com/jatherrien/largeRCRF.git")

If you care about vignettes and have the packages available to build them you can include build_vignettes = TRUE as a parameter in the command above.

System Requirements

You need:

  • R version 3.4.0 or greater
  • The rJava package version 0.9-9 or greater
  • A Java runtime version 1.8 or greater

Troubleshooting

I get an OutOfMemoryException error but I have plenty of RAM

largeRCRF makes use of the Java virtual machine, which unfortunately restricts itself by default to a quarter of your system memory. You can override the default by including before loading largeRCRF or any other rJava based package the following line:

R> options(java.parameters <- c("-Xmx13G", "-Xms13G"))

with 13G replaced with a little less than your available system memory.

I get an OutOfMemoryException error and I'm short on RAM

Obviously if you're short on RAM there is a limit on how large of a dataset you can train, but there are some techniques you can use to limit how much largeRCRF needs.

  • If your training dataset is large you might not want both R and largeRCRF to have their own separate copies (limitations due to Java require largeRCRF have its own copy). When specifying the data parameter into train, instead provide an environment containing one object called data which is the dataset. largeRCRF will delete that variable after importing it into the Java environment.

Example:

R> data.env <- new.env()
R> data.env$data <- trainingData
R> rm(trainingData)
R> model <- train(..., data=data.env, ...)
  • Each core that is training trees requires its own memory; you can try limiting largeRCRF to train only one tree at a time by specifiying cores=1.

  • By default largeRCRF keeps the entire forest loaded in memory during training, when in practice only the trees being trained on need to be loaded. You can specify savePath to give a directory for largeRCRF to save trees in during training, which will allow to largeRCRF to conserve memory for only those trees being currently trained.

Training stalls immediately at 0 trees and the CPU is idle

This issue has been observed before on one particular system (and only on that system) but it's not clear what causes it. It would be appreciated if you could report this bug to joelt@sfu.ca and give your operating system and the version of Java installed (the entire output of java --version).

As a workaround, this issue seems to occur randomly; so try restarting your code to see if it runs.