Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
I have a number of files in 1 folder that have slightly different schemas. I want to standardise the schema before processing all rows as flow. I will need to extend the design to include at least 2 more schema designs not shown here.
I'm having trouble designing the job to unite the flows into 1 while avoiding loops (tUnite doesn't work) so tHash components were suggested but my design as seen below only outputs the rows from 1 of the 4 tHashOut components (whichever is listed as linked in the tHashInput).
What I've tried
While this reference guide shows 2 HashOuts its not clear to me why the configuration works or how I would extend it to include more tHashOutputs
Anyone able to advise me ?
The way to do this is create a mini subjob at the beginning of your job. In that subjob add a tFixedFlowInput with the schema you require for your files and connect that to a tHashOutput. Set the tFixedFlowInput to produce 0 rows. This will initialise the tHashOutput.
Now in all of your other tHashOutputs tick the "Link with a tHashOutput" and select the name of the pre-initialised tHashOutput.
When you want to read from the tHashOutputs connect your tHashInput to the pre-initialised one.
That is it.
The way to do this is create a mini subjob at the beginning of your job. In that subjob add a tFixedFlowInput with the schema you require for your files and connect that to a tHashOutput. Set the tFixedFlowInput to produce 0 rows. This will initialise the tHashOutput.
Now in all of your other tHashOutputs tick the "Link with a tHashOutput" and select the name of the pre-initialised tHashOutput.
When you want to read from the tHashOutputs connect your tHashInput to the pre-initialised one.
That is it.