Do not input private or sensitive data. View Qlik Privacy & Cookie Policy.
Skip to main content

Announcements
Join us in Bucharest on Sept 18th for Qlik's AI Reality Tour! Register Now
cancel
Showing results for 
Search instead for 
Did you mean: 
Anonymous
Not applicable

What is the efficient way to parse large xml around 2 GB in Talend 6.4

Hi,

I have around 2 Gb of multi schema XML file. Currently I am parsing XML file using ParseUtils.parseTo_Document in tjavaflex for converting string to document and using tXML Map for mapping. But for large file it is taking much time around 40 mins to parse and extract the data based on xml tags.

So another approach I have tried is to split XML file into multiple chunks and use the same process further (ParseUtils.parseTo_Document)  but still it is taking much time as we are converting string to document while parsing.

Can anyone please suggest what would be the best and fastest way to parse and extract data from XML

 

Thanks,

Sanket

Labels (2)
3 Replies
Anonymous
Not applicable
Author

An XML document can be larger in two ways: (1) in its proper XML form, requiring more storage space and bandwidth; and (2) in its compiled, in-memory form, requiring more computing resources. Sometimes, XML deserves this reputation.

Anonymous
Not applicable
Author

But can you provide solution to parse 2 gb of multi loop XML in faster way in Talend

manodwhb
Champion II
Champion II

@sm , you need to spilt that xml into small files and do the required things down the line.