Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hi,
Why it takes around 11 minutes to dump 2.8 million rows from one database table to other table on Azure Sql server?
Do anyone have solution to speed up the performance on Azure Server?
Thanks,
Dhara
Hi Dhara,
the MySQL input component I work with has an advanced setting "enable stream". I do not know about Azure SQL server, maybe this setting is there, too?
Are you using a bulk load component? If you are using an output component with inserts, did you consider to use and ajust the 'batch size' and 'commit every' settings? Have you thought about 'enable parallel execution'?
If you are sending the rows through a tMap, you might want to ajust the 'max buffer size' advanced setting.
Best regards,
Thomas
Hi Thomas,
I am only using tMSSqlInput and tMSSqlOutput component. I have tried 'batch size' and 'commit every' settings but still performance is low. I am only executing single job.
Thanks,
Dhara
Check increasing JVM size. Also remove if you have used any tLogRow component within any subjob.
Hi,
I believe you are using tDBInput and tDBOutput components to load the data. You can do two changes to make it faster.
a) Currently you must be using a single tDBInput Component to extract the data. You can use a tparallelize option to run parallel operations where each tDBInput will fetch data from certain partitions. Please make sure that you are providing enough memory also for this operation. This way, we can increase the throughput at reading stage.
b) In writing stage, could you please use tDBBulk components instead of normal output component? This will increase the processing at output stages.
Warm Regards,
Nikhil Thampi
Please appreciate our Talend community members by giving Kudos for sharing their time for your query. If your query is answered, please mark the topic as resolved 🙂
Hi,
I am getting the following error:
Thanks,
Dhara
Hi Dhara,
Could you please share your job details screenshots so that Talend Community members will get more clarity?
Warm Regards,
Nikhil Thampi
Please appreciate our Talend community members by giving Kudos for sharing their time for your query. If your query is answered, please mark the topic as resolved 🙂
Hi,
I have set Additional JDBC parameter = "encrypt=true;trustServerCertificate=false;hostNameInCertificate=*.database.windows.net;loginTimeout=30;" as I am dumping data on Azure Server. Following is my Talend Job:
Thanks,
Dhara
Hi Dhara,
relating to the error "file ... could not be opened": Did this job run in Talend Studio on your local PC or on a remote jobserver? Does the file "mssql_data.txt" exist in the directory D:\<something> (on your PC or the host that runs Talend jobserver)?
Best regards,
Thomas