Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hi,
I have a very massive table whose record count will be more than 10 billion.
1. I wanted to read the data from input table for first 2 billion, process it and write to output table.
Then fetch next 2 billion and do the same process and write to output table.
how could I design my job so that it loops to fetch 2 billion at a time. This design should also consider restartability at record level if it errors(checkpoint at each row if this is the best approach).
2. Can I run as a seperate process each 2 billion records?
Thanks
Hi,
Thanks for the suggestion.
I have used table to store restart and record split details which will be input to the child job.
And I have created parallel flow creating multiple job reading from the same table.
Thanks.
Hi,
Thanks for the suggestion.
I have used table to store restart and record split details which will be input to the child job.
And I have created parallel flow creating multiple job reading from the same table.
Thanks.
Hi TRF,
Like tCheckpoint customised component, is there any customised component that support parallel execution as well as restartability at job-level?
Thanks,
Revathy.