Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

Handling java.lang.OutOfMemoryError when writing to Excel from R

Tags:

r

xlsx

The xlsx package can be used to read and write Excel spreadsheets from R. Unfortunately, even for moderately large spreadsheets, java.lang.OutOfMemoryError can occur. In particular,

Error in .jcall("RJavaTools", "Ljava/lang/Object;", "invokeMethod", cl, :
java.lang.OutOfMemoryError: Java heap space

Error in .jcall("RJavaTools", "Ljava/lang/Object;", "newInstance", .jfindClass(class), :
java.lang.OutOfMemoryError: GC overhead limit exceeded

(Other related exceptions are also possible but rarer.)

A similar question was asked regarding this error when reading spreadsheets.

Importing a big xlsx file into R?

The main advantage of using Excel spreadsheets as a data storage medium over CSV is that you can store multiple sheets in the same file, so here we consider a list of data frames to be written one data frame per worksheet. This example dataset contains 40 data frames, each with two columns of up to 200k rows. It is designed to be big enough to be problematic, but you can change the size by altering n_sheets and n_rows.

library(xlsx) set.seed(19790801) n_sheets <- 40 the_data <- replicate(   n_sheets,   {     n_rows <- sample(2e5, 1)     data.frame(       x = runif(n_rows),       y = sample(letters, n_rows, replace = TRUE)     )   },   simplify = FALSE ) names(the_data) <- paste("Sheet", seq_len(n_sheets)) 

The natural method of writing this to file is to create a workbook using createWorkbook, then loop over each data frame calling createSheet and addDataFrame. Finally the workbook can be written to file using saveWorkbook. I've added messages to the loop to make it easier to see where it falls over.

wb <- createWorkbook()   for(i in seq_along(the_data)) {   message("Creating sheet", i)   sheet <- createSheet(wb, sheetName = names(the_data)[i])   message("Adding data frame", i)   addDataFrame(the_data[[i]], sheet) } saveWorkbook(wb, "test.xlsx")   

Running this in 64-bit on a machine with 8GB RAM, it throws the GC overhead limit exceeded error while running addDataFrame for the first time.

How do I write large datasets to Excel spreadsheets using xlsx?

like image 239
Richie Cotton Avatar asked Feb 21 '14 14:02

Richie Cotton


People also ask

How do I fix Java Lang Outofmemory?

1) An easy way to solve OutOfMemoryError in java is to increase the maximum heap size by using JVM options "-Xmx512M", this will immediately solve your OutOfMemoryError.

What causes Java Lang OutOfMemoryError?

OutOfMemoryError exception. Usually, this error is thrown when there is insufficient space to allocate an object in the Java heap. In this case, The garbage collector cannot make space available to accommodate a new object, and the heap cannot be expanded further.


2 Answers

This is a known issue: http://code.google.com/p/rexcel/issues/detail?id=33

While unresolved, the issue page links to a solution by Gabor Grothendieck suggesting that the heap size should be increased by setting the java.parameters option before the rJava package is loaded. (rJava is a dependency of xlsx.)

options(java.parameters = "-Xmx1000m") 

The value 1000 is the number of megabytes of RAM to allow for the Java heap; it can be replaced with any value you like. My experiments with this suggest that bigger values are better, and you can happily use your full RAM entitlement. For example, I got the best results using:

options(java.parameters = "-Xmx8000m") 

on the machine with 8GB RAM.

A further improvement can be obtained by requesting a garbage collection in each iteration of the loop. As noted by @gjabel, R garbage collection can be performed using gc(). We can define a Java garbage collection function that calls the Java System.gc() method:

jgc <- function() {   .jcall("java/lang/System", method = "gc") }     

Then the loop can be updated to:

for(i in seq_along(the_data)) {   gc()   jgc()   message("Creating sheet", i)   sheet <- createSheet(wb, sheetName = names(the_data)[i])   message("Adding data frame", i)   addDataFrame(the_data[[i]], sheet) } 

With both these code fixes, the code ran as far as i = 29 before throwing an error.

One technique that I tried unsuccessfully was to use write.xlsx2 to write the contents to file at each iteration. This was slower than the other code, and it fell over on the 10th iteration (but at least part of the contents were written to file).

for(i in seq_along(the_data)) {   message("Writing sheet", i)   write.xlsx2(     the_data[[i]],      "test.xlsx",      sheetName = names(the_data)[i],      append    = i > 1   ) } 
like image 86
Richie Cotton Avatar answered Nov 06 '22 14:11

Richie Cotton


Building on @richie-cotton answer, I found adding gc() to the jgc function kept the CPU usage low.

jgc <- function() {   gc()   .jcall("java/lang/System", method = "gc") }     

My previous for loop still struggled with the original jgc function, but with extra command, I no longer run into GC overhead limit exceeded error message.

like image 29
guyabel Avatar answered Nov 06 '22 14:11

guyabel