Hi all,
I am trying to download 1 year of AGERA5 data on 30 polygon of size (1280m*1280m). With my script I am already able to download Sentinel 2 and Sentinel 1 data. In the log file from AGERA5 job, I have the following error: “OpenEO batch job failed: Exception during Spark execution: Java heap space”. From what I understand, this error can be solved by customizing the batch job resources, through job_options. It seems that I should try to increase : "driver-memoryOverhead"
or "driver-memory"
. However, I find it difficult to assess, how to correctly set this job_options. I had similar issue with S2 and setting the job_options eventually worked:
job_options = {
"executor-memory": "3G",
"executor-memoryOverhead": "10G", #default 2G
"executor-cores": 2,
"task-cpus": 1,
"executor-request-cores": "400m",
"max-executors": "100",
"driver-memory": "12G",
"driver-memoryOverhead": "10G",
"driver-cores": 5,
"udf-dependency-archives":[],
"logging-threshold": "info"
}
I would like to have a better understanding on how to set this job_options. Has someone already run Batch Job with AGERA5 on large period of data ? Is it a problem that I can solve by modifying job options ?
Thanks in advance,
Iris