<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Replicating huge databases in Qlik Replicate</title>
    <link>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2039711#M5048</link>
    <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.qlik.com/t5/user/viewprofilepage/user-id/193288"&gt;@guilherme-matte&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;Thanks for you reaching out.&lt;/P&gt;
&lt;P&gt;Not sure what's the source/target database type and Replicate version, in general the suggestions are:&lt;/P&gt;
&lt;P&gt;1. We may utilize filters to 'cut' the massive size data from source to smaller size to transfer to target side; eg each part contains 1 year history data only. or (2)&lt;/P&gt;
&lt;P&gt;2.&amp;nbsp;&lt;SPAN&gt;Using&amp;nbsp;&lt;A title="Parallel Load" href="https://help.qlik.com/en-US/replicate/November2022/Content/Global_Common/Content/SharedEMReplicate/Customize%20Tasks/Parallel_Load.htm" target="_blank" rel="nofollow noopener noreferrer"&gt;Parallel Load&lt;/A&gt;&amp;nbsp;to speed up the transfer if network bandwidth is sufficient.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;3. Run Full Load ONLY task to pass the&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;unchanged&amp;nbsp;old data to target by dedicated task(s), or by different filter conditions in the single task; &lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;4. We may transfer unchanged data to different temporary&amp;nbsp;table(s) in target in parallel, then merge the temporary table(s) into target table in&amp;nbsp;off-peak time, before the CDC task startup.&amp;nbsp;&lt;SPAN&gt;Transfer the 'history' data prior to 'change' data to avoid the UPDATE/DELETE cannot find the target rows in target side database/tables.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;5. If possible please use partition tables (eg 1 partition for 1 year data) in target DB for easy management and better performance&lt;/SPAN&gt;&lt;SPAN&gt;, also it has Primary Key/Unique Index/Unique Key etc to make sure no full table scanning during CDC stage otherwise latency builds up.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;6.&lt;/SPAN&gt;&lt;SPAN&gt;&amp;nbsp;We'd like to suggest PS team engaged as this not an easy performance tuning job, and need to solve the various issues during the long time running.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Hope this helps.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Regards,&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;John.&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Mon, 20 Feb 2023 07:47:33 GMT</pubDate>
    <dc:creator>john_wang</dc:creator>
    <dc:date>2023-02-20T07:47:33Z</dc:date>
    <item>
      <title>Replicating huge databases</title>
      <link>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2039591#M5044</link>
      <description>&lt;P&gt;Hello guys!&lt;/P&gt;
&lt;P&gt;More of a general question today..&lt;/P&gt;
&lt;P&gt;Does Qlik Replicate have issues replicating HUGE databases? We have a client who wants to replicate one db that they even have issues querying due its massive size on premises.&lt;/P&gt;
&lt;P&gt;Does Replicate have a limit in this regard or we should be fine?&lt;/P&gt;
&lt;P&gt;Thank you as always for the collaboration!&lt;/P&gt;
&lt;P&gt;Kind regards!&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sun, 19 Feb 2023 22:34:36 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2039591#M5044</guid>
      <dc:creator>guilherme-matte</dc:creator>
      <dc:date>2023-02-19T22:34:36Z</dc:date>
    </item>
    <item>
      <title>Re: Replicating huge databases</title>
      <link>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2039711#M5048</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.qlik.com/t5/user/viewprofilepage/user-id/193288"&gt;@guilherme-matte&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;Thanks for you reaching out.&lt;/P&gt;
&lt;P&gt;Not sure what's the source/target database type and Replicate version, in general the suggestions are:&lt;/P&gt;
&lt;P&gt;1. We may utilize filters to 'cut' the massive size data from source to smaller size to transfer to target side; eg each part contains 1 year history data only. or (2)&lt;/P&gt;
&lt;P&gt;2.&amp;nbsp;&lt;SPAN&gt;Using&amp;nbsp;&lt;A title="Parallel Load" href="https://help.qlik.com/en-US/replicate/November2022/Content/Global_Common/Content/SharedEMReplicate/Customize%20Tasks/Parallel_Load.htm" target="_blank" rel="nofollow noopener noreferrer"&gt;Parallel Load&lt;/A&gt;&amp;nbsp;to speed up the transfer if network bandwidth is sufficient.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;3. Run Full Load ONLY task to pass the&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;unchanged&amp;nbsp;old data to target by dedicated task(s), or by different filter conditions in the single task; &lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;4. We may transfer unchanged data to different temporary&amp;nbsp;table(s) in target in parallel, then merge the temporary table(s) into target table in&amp;nbsp;off-peak time, before the CDC task startup.&amp;nbsp;&lt;SPAN&gt;Transfer the 'history' data prior to 'change' data to avoid the UPDATE/DELETE cannot find the target rows in target side database/tables.&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;5. If possible please use partition tables (eg 1 partition for 1 year data) in target DB for easy management and better performance&lt;/SPAN&gt;&lt;SPAN&gt;, also it has Primary Key/Unique Index/Unique Key etc to make sure no full table scanning during CDC stage otherwise latency builds up.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;6.&lt;/SPAN&gt;&lt;SPAN&gt;&amp;nbsp;We'd like to suggest PS team engaged as this not an easy performance tuning job, and need to solve the various issues during the long time running.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Hope this helps.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Regards,&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;John.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 20 Feb 2023 07:47:33 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2039711#M5048</guid>
      <dc:creator>john_wang</dc:creator>
      <dc:date>2023-02-20T07:47:33Z</dc:date>
    </item>
    <item>
      <title>Re: Replicating huge databases</title>
      <link>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2040097#M5057</link>
      <description>&lt;P&gt;Hello John!&lt;/P&gt;
&lt;P&gt;Thank you as always for your help.&lt;/P&gt;
&lt;P&gt;I will get more information about the steps you mentioned (parallel load, partitions,&amp;nbsp; etc...) and also consider engaging the PS team as well after I get some more info.&lt;/P&gt;
&lt;P&gt;In general, what would Qlik be able to handle without many issues? I mean, what would be considered a really big database that should require some extra tunning and usually which are the sizes that Qlik would handle without many problems? I know it might depend on extra&amp;nbsp; factors, but its just to have an idea since HUGE databases, as in the question, is a bit subjective.&lt;/P&gt;
&lt;P&gt;Cheers!&lt;/P&gt;</description>
      <pubDate>Tue, 21 Feb 2023 03:00:15 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2040097#M5057</guid>
      <dc:creator>guilherme-matte</dc:creator>
      <dc:date>2023-02-21T03:00:15Z</dc:date>
    </item>
    <item>
      <title>Re: Replicating huge databases</title>
      <link>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2040236#M5062</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.qlik.com/t5/user/viewprofilepage/user-id/193288"&gt;@guilherme-matte&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;In general we need a configuration tuning if the data is huge however it's hard to tell a number , that depends on hardware/OS settings/network throughputs/database type etc factors, just as you said.&lt;/P&gt;
&lt;P&gt;Best Regards,&lt;/P&gt;
&lt;P&gt;John.&lt;/P&gt;</description>
      <pubDate>Tue, 21 Feb 2023 09:10:06 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Qlik-Replicate/Replicating-huge-databases/m-p/2040236#M5062</guid>
      <dc:creator>john_wang</dc:creator>
      <dc:date>2023-02-21T09:10:06Z</dc:date>
    </item>
  </channel>
</rss>

