Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hi,
I am trying to find out the best practice in order to read a file, transform some of its values and then load the transformed data in a hive table. My problem is that the components available only allow entire file loading in Hive and not row by row data loading. Any ideas on how to do it ???
The only solution I have found this far is to store the output file in HDFS and then create a job, which will import the HDFS file in HIVE. However this seems too slow.... Additionally there used to exist a tHiveOutput component which would be perfect for what I am trying to do, but it is not available in the standard jobs any more. Any ideas ???
Thank,
Myrto
Hello,
tHiveOutput only exists in spark, it is not a standard DI component.
Here comes tHiveCreateTable component which is running in the Standard Job framework. And you could use tHDFSPut component to load large-scale files.
Best regards
Sabrina
Hello,
tHiveOutput only exists in spark, it is not a standard DI component.
Here comes tHiveCreateTable component which is running in the Standard Job framework. And you could use tHDFSPut component to load large-scale files.
Best regards
Sabrina