Do not input private or sensitive data. View Qlik Privacy & Cookie Policy.
Skip to main content

Announcements
Independent validation for trusted, AI-ready data integration. See why IDC named Qlik a Leader: Read the Excerpt!
cancel
Showing results for 
Search instead for 
Did you mean: 
Sireesha_Chapa
Contributor
Contributor

Migrate data from My SQL to AWS S3 without any Delimited files

Hi All,

I have 2 questions which I am looking in multiple documents but could not find concrete answers:

  1. I would like to know if there are any components that would migrate directly from My SQL to S3 without using any delimited file.
  2. Even if it should be written to any file, can I directly write the data from My SQL to Parquet and then to AWS S3 ? I could see many references where one writes into CSV and from CSV it would be written into S3. So, I should know without using CSV, is there any architecture in Talend that transfers to Parquet from My SQL directly and then from Parquet to S3.

I appreciate your quick response. Based on that I am in need to check about the feasibility of using Talend in our project.

Labels (2)
1 Solution

Accepted Solutions
Sireesha_Chapa
Contributor
Contributor
Author

Thank you @Thomas Dye​. So you mean if I use Talend Stitch, I can directly insert the data from MySQL to AWS S3 without any file? I would check meanwhile.

 

"But the other alternative of using like Spark Job" - Can I get any references for this for inspiration. I could not find it.

 

Also if I use tDBInput_1 -->tFileOutputParquet_1

Would this work? Would it insert the data directly from Server without using any spark jobs?

View solution in original post

2 Replies
Anonymous
Not applicable

Hi Sireesha, You have 2 possibilities. First look at Talend Stitch. This tool is designed to move data in bulk and supports both your endpoints. If you can't use Stitch for some reason, you may try to create a Spark Job that reads MySQL and writes Parquet. You will want to add as many executors as you have hardware, to foster performance. This also assumes you have a Talend Big Data License. Stitch is the preferred solution. It is easy to get a trial license.

Sireesha_Chapa
Contributor
Contributor
Author

Thank you @Thomas Dye​. So you mean if I use Talend Stitch, I can directly insert the data from MySQL to AWS S3 without any file? I would check meanwhile.

 

"But the other alternative of using like Spark Job" - Can I get any references for this for inspiration. I could not find it.

 

Also if I use tDBInput_1 -->tFileOutputParquet_1

Would this work? Would it insert the data directly from Server without using any spark jobs?