Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hi All,
I have 2 questions which I am looking in multiple documents but could not find concrete answers:
I appreciate your quick response. Based on that I am in need to check about the feasibility of using Talend in our project.
Thank you @Thomas Dye. So you mean if I use Talend Stitch, I can directly insert the data from MySQL to AWS S3 without any file? I would check meanwhile.
"But the other alternative of using like Spark Job" - Can I get any references for this for inspiration. I could not find it.
Also if I use tDBInput_1 -->tFileOutputParquet_1
Would this work? Would it insert the data directly from Server without using any spark jobs?
Hi Sireesha, You have 2 possibilities. First look at Talend Stitch. This tool is designed to move data in bulk and supports both your endpoints. If you can't use Stitch for some reason, you may try to create a Spark Job that reads MySQL and writes Parquet. You will want to add as many executors as you have hardware, to foster performance. This also assumes you have a Talend Big Data License. Stitch is the preferred solution. It is easy to get a trial license.
Thank you @Thomas Dye. So you mean if I use Talend Stitch, I can directly insert the data from MySQL to AWS S3 without any file? I would check meanwhile.
"But the other alternative of using like Spark Job" - Can I get any references for this for inspiration. I could not find it.
Also if I use tDBInput_1 -->tFileOutputParquet_1
Would this work? Would it insert the data directly from Server without using any spark jobs?