Do not input private or sensitive data. View Qlik Privacy & Cookie Policy.
Skip to main content

Announcements
Qlik Open Lakehouse is Now Generally Available! Discover the key highlights and partner resources here.
cancel
Showing results for 
Search instead for 
Did you mean: 
Anonymous
Not applicable

Big Data Spark Job - Load data into Hive

I am creating big data spark job and want to load data into dynamic partitioned hive tables.

 

Which component I can use to load data into hive and what would be the workflow?

Labels (3)
3 Replies
Anonymous
Not applicable
Author

tHDFSConnection--> HIVE-->tHiveRow--->tFileInputDelimiter---->tHDFSOutput

 

0683p000009LuUI.pngHive load data

tHiveRow configuration 

0683p000009LuAm.pngtHiveRow1

 

tHiveRow2

0683p000009LuUR.png

 

tFileInput Delimiter

 

0683p000009Ltxv.png

tHDFSoutput

 

0683p000009LuUb.png

1.tHDFSConnection set your connection to your Hadoop cluster

2.set connection to hive in HIVE component

3.tHiveRow drop the table if present

4.tHiveRow create external table table

5. tFileInutDelimiter load the data to HDFS external table location

6. tHDFSoutput to load data to external table path  

Anonymous
Not applicable
Author

Which version you are using? With 6.2.1, all these tHive component is not available. Only tHiveConfiguration, tHiveInput and tHiveOutput is there.

 

Data is loaded using spark job but facing problem when the Hive table is dynamically partitioned.

 

Thanks. 

Anonymous
Not applicable
Author

I am using Talend open studio for Big data Version TOS_BD-20150508_1414-V5.6.2 . try to Download those components from https://exchange.talend.com/.