Do not input private or sensitive data. View Qlik Privacy & Cookie Policy.
Skip to main content

Announcements
Join us to spark ideas for how to put the latest capabilities into action. Register here!
cancel
Showing results for 
Search instead for 
Did you mean: 
Anonymous
Not applicable

[resolved] tStatCatcher/tFlowMeter for file iterations

Hi,
If I use the stat catcher and/or flowmeter, and I am iterating through a set if input files, will the counts be available for each file or will the count be for the combined total for all files that are looped through?
Thanks in advance,
Labels (2)
13 Replies
Anonymous
Not applicable
Author

Hi Shong,
thanks for your reply.
1) Delete the 'oncomponentok' link, see picture 1.

That's the way I did it at first, but then the RowMeterCatcher fires only after the subjob "Browse Files" finishes; as a consequence, the "Get Filename" mapper reads the file name as it is after subjob termination, i.e., the last file that was processed, and outputs this single filename for each line of statistics. So for instance instead of getting
file1 100 lines
file2 200 lines
file3 300 lines
I get
file3 100 lines
file3 200 lines
file3 300 lines
2) The tOracleOutputBulk and tOracleBulkExec components must be used together. see picture 2.

I don't understand this one too well. Wouldn't adding another OracleOutputBulk step introduce an additional, redundant, intermediate csv file? Currently the job fills up a csv file with the ouput from several input binary files then load the csv file in one go using sqlldr. This seems to work quite well as long as I don't introduce the statistics steps. Is there a better way to do that under Talend? Are the multithreading issues related to this architecture?
Thanks a lot for your help,
Chris
Anonymous
Not applicable
Author

Hi again Shong,
do you have any input regarding the points above? Your help is definitely appreciated!
Thanks,
Chris
Anonymous
Not applicable
Author

Hi there again,
sorry to insist but does anyone has any insight on the issue at hand? Is there a flaw in TOS multithreading support or should the job be designed differently?
Thanks to all for your help,
Chris
Anonymous
Not applicable
Author

Hello Chris
Sorry, I miss this topic yesterday.
The processing is done in parallel (cf the "Iterate x5" on Browse_Dirs

As I known, there was a bug on the parallel iterate link. So, don't check 'enable parallel execution' option and try if the problem still exsit.
as a consequence, the "Get Filename" mapper reads the file name as it is after subjob termination,

Yes, using the 'oncomponentok' link after tOracleOutput if you want to get the current file name on tMap and output it.
I don't understand this one too well. Wouldn't adding another OracleOutputBulk step introduce an additional, redundant, intermediate csv file?

Yes, you are right. If the file exists, you just need a tOracleBulkExec to read it and bulk insert records into db.
Best regards

shong