Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hi All,
Is there an example of a talend job that loads data into a parquet file and then extracts the data from the parquet file to the target table? I am aware of the component used tFileInputParquet and tFileOutputParquet
Do we need to have HDFS connectivity for this, or can we do it locally on the server or local folder path? Please help with this.
Hello,
Please have a look at this KB article about:
https://community.talend.com/s/article/Using-Parquet-format-best-practices-feZEm
Note: If you are using a Windows platform, make sure Hadoop Winutils and Microsoft Visual C++ 2010 Service Pack 1 Redistributable Package MFC Security Update are installed before using this component. The two links also give the installation information.
Hope it helps.
Best regards
Sabrina