data1.dl.r2 = vector()
for (i in 1:100) {
if (i==1) {
data1.hex = as.h2o(data1)
} else {
data1.hex = nextdata
}
data1.dl = h2o.deeplearning (x=2:1000,y=1,training_frame=data1.hex,nfolds=5,activation="Tanh",hidden=30,seed=5,reproducible=TRUE)
data1.dl.pred = h2o.predict(data1.dl,data1.hex)
data1.dl.r2[i] = sum((as.matrix(data1.dl.pred)-mean(as.matrix(data1.hex[,1])))^2)/
sum((as.matrix(data1.hex[,1])-mean(as.matrix(data1.hex[,1])))^2) # R-squared
prevdata = as.matrix(data1.hex)
nextpred = as.matrix(h2o.predict(data1.dl,as.h2o(data0[i,])))
colnames(nextpred) = "response"
nextdata = as.h2o(rbind(prevdata,cbind(nextpred,data0[i,-1])))
print(i)
}
This is my code with a dataset (data1) of 100 observations and 1000 features. When I run this, it gave me an error message at 50~60th iteration "
Error in .h2o.doSafeREST(h2oRestApiVersion = h2oRestApiVersion, urlSuffix = page, :
ERROR MESSAGE:
Total input file size of 87.5 KB is much larger than total cluster memory of Zero , please use either a larger cluster or smaller data.
When I run 'h20.init()', it tells me that the total cluster memory is zero.
H2O cluster total nodes: 1
H2O cluster total memory: 0.00 GB
H2O cluster total cores: 8
H2O cluster allowed cores: 8
H2O cluster healthy: TRUE
So I wonder why cluster total memory is zero, and why it didn't go wrong at earlier iterations.
You need to restart your H2O cluster.
Try h2o.cluster().shutdown()
and then h2o.init()
.
You can also explicitly set the memory allocated to H2O by h2o.init(min_mem_size_GB=8)
, which depends upon how much memory your machine has of course.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With