I have an ETL flow through talend
and there:
- Read the zipped files from a remote server with a job.
- Take this files unzipes them and parse them into HDFS with a job. Inside the job exists a schema check so if something is not
My problem is that TAC server stopes the execution because of this error:
Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at org.talend.fileprocess.TOSDelimitedReader$ColumnBuffer4Joiner.saveCharInJoiner(TOSDelimitedReader.java:503) at org.talend.fileprocess.TOSDelimitedReader.joinAndRead(TOSDelimitedReader.java:261) at org.talend.fileprocess.TOSDelimitedReader.readRecord_SplitField(TOSDelimitedReader.java:148) at org.talend.fileprocess.TOSDelimitedReader.readRecord(TOSDelimitedReader.java:125) ....
Is there any option to avoid and handle this error automatically? There are only few files which cause this error but I want to find a solution for further similar situation.