Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hello
We have a spark streaming Talend job that consumes events in json format from kafka and writes to hive. The input is a large json with 500+ attributes and ending up in 64K byte limit on the method generated for the subjob.
I understand the best way to work around this is to split the subjob but with the streaming job, that is not possible. Are there any suggestions/pointers to work around this?
We have the following flexibility, if any of this helps..
Thanks
Radhika
Hi
Take a look at these KB articles about this Java 65535 bytes limit error.
https://community.talend.com/s/article/Exceeding-the-Java-bytes-limit-1Z1UZ
https://community.talend.com/s/article/Building-a-Job-with-one-tExtractPositionalFields-component-fails-with-the-error-The-code-of-method-is-exceeding-the-bytes-limit-17gnl
https://community.talend.com/s/article/tMSSqlInput-Process-Map-String-Object-is-exceeding-the-bytes-limit-InMpE
The workaround is optimizing the Job to reduce the size of the final generated code of a subjob. Try the following:
In your case, I think option 1 may be a solution that can be tried.
Regards
Shong