<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Job hangs after start (LogStream to DataBricks) in Qlik Replicate</title>
    <link>https://community.qlik.com/t5/Qlik-Replicate/Job-hangs-after-start-LogStream-to-DataBricks/m-p/2495404#M13312</link>
    <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.qlik.com/t5/user/viewprofilepage/user-id/127219"&gt;@TomaszRomanowski&lt;/a&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Could you please open a support ticket to analyze further.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Regards,&lt;/P&gt;
&lt;P&gt;Suresh&lt;/P&gt;</description>
    <pubDate>Sun, 01 Dec 2024 10:11:58 GMT</pubDate>
    <dc:creator>sureshkumar</dc:creator>
    <dc:date>2024-12-01T10:11:58Z</dc:date>
    <item>
      <title>Job hangs after start (LogStream to DataBricks)</title>
      <link>https://community.qlik.com/t5/Qlik-Replicate/Job-hangs-after-start-LogStream-to-DataBricks/m-p/2494879#M13286</link>
      <description>&lt;P&gt;&lt;STRONG&gt;Qlik replicate version&lt;/STRONG&gt;: 2023.11.0.468&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;Source&lt;/STRONG&gt;: MSSQL with LogStream (data rotation 3 days, max data 300GB).&lt;BR /&gt;&lt;STRONG&gt;Target&lt;/STRONG&gt;: DataBricks Delta&lt;/P&gt;
&lt;P&gt;Job (MSSQL -&amp;gt; LogStream) tables:&lt;BR /&gt;9 typical tables + 1 table with XML column with max lob size 20MB.&lt;/P&gt;
&lt;P&gt;Job (LogStream -&amp;gt; DataBricks) tables:&lt;BR /&gt;Only one table with XML column with max lob size 20MB.&lt;/P&gt;
&lt;P&gt;Current Status:&lt;BR /&gt;In DataBricks I have data of this tables but 32 hours behind real-time.&lt;BR /&gt;So, I run task without full load just to point date (32 hours in the past)&lt;BR /&gt;In Logstream I have 3 days rotation data (currently about 80GB data).&lt;/P&gt;
&lt;P&gt;So when I run task I use option:&lt;BR /&gt;Advanced Run Options -&amp;gt; Table are already loaded. Start processing changes from -&amp;gt; Data and Time&lt;BR /&gt;I put there past date (32h ago).&lt;/P&gt;
&lt;P&gt;Configuration on job (LogStream -&amp;gt; DataBricks):&lt;BR /&gt;- max lob size = 20480&lt;BR /&gt;- stream_buffers_number = 5 &lt;BR /&gt;- stream_buffer_size = 1024&lt;BR /&gt;- Apply Conflicts insert -&amp;gt; „Duplicate key when applying Insert” = UPDATE &lt;BR /&gt;- one extra column that is calculated.&lt;BR /&gt;- Total transactions memory size exceeds (MB) = 2048&lt;BR /&gt;- target max file uload = 2000 (2GB)&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;The task starts, sends some data to the target, and &lt;STRONG&gt;then hangs indefinitely.&lt;/STRONG&gt;&lt;BR /&gt;There is no bigger cpu or disk usage.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;STRONG&gt;Logs show the following key messages:&lt;/STRONG&gt;&lt;/P&gt;
&lt;P&gt;[SORTER ]T: Reading from source is paused (sorter_transaction.c:76)&lt;BR /&gt;... &lt;BR /&gt;[SORTER_STORAGE ]T: Memory size needed 2147684472, Memory size limit: 2147483648 (transaction_storage.c:5380)&lt;BR /&gt;... &lt;BR /&gt;[SORTER_STORAGE ]T: Swap committed transaction to free memory, transaction index 10 (transaction_storage.c:5398)&lt;BR /&gt;[SORTER_STORAGE ]T: Transaction (Type 'Commited Transactions storage', Id 00000000000000000000000000000bae, Events # 3) is moving to file 'D:\qlik\Replicate\data\tasks\MSSQL_LS_DBRICKS/sorter/ars_swap_tr_00000000000000000001.tswp' (transaction_storage.c:5482)&lt;BR /&gt;... &lt;BR /&gt;[SORTER_STORAGE ]T: Memory size needed 2094508184, Memory size limit: 2147483648 (transaction_storage.c:5420)&lt;BR /&gt;...&lt;BR /&gt;[SORTER ]T: Reading from source is paused (sorter_transaction.c:76)&lt;BR /&gt;...&lt;BR /&gt;[FILE_FACTORY ]T: Source 'D:\qlik\Replicate\data\tasks\MSSQL_LS_DBRICKS\cloud\bulk\CDC00000001.csv.gz' exists (type = 2), size is 7129552 bytes (at_universal_fs_object.c:658)&lt;BR /&gt;[FILE_FACTORY ]T: uploading file 'D:\qlik\Replicate\data\tasks\MSSQL_LS_DBRICKS\cloud\bulk\CDC00000001.csv.gz' to '/staging/attrep_changes32432562243S4234D/CDC00000001.csv.gz' (AttAdls2FileFactory.java:52)&lt;BR /&gt;[FILE_FACTORY ]T: upload done (AttAdls2FileFactory.java:55)&lt;BR /&gt;...&lt;BR /&gt;[FILE_FACTORY ]T: upload of file &amp;lt;D:\qlik\Replicate\data\tasks\MSSQL_LS_DBRICKS\cloud\bulk\CDC00000001.&lt;BR /&gt;[TARGET_APPLY ]T: Data is copied to attrep_changes table (cloud_bulk.c:1245)&lt;BR /&gt;[TARGET_APPLY ]T: cloud_bulk_start_applying - refreshing net changes table (cloud_bulk.c:1527)&lt;BR /&gt;[TARGET_APPLY ]T: Refresh table - owner: qlik, table: attrep_changes32432562243S4234D (databricks_imp.c:607)&lt;BR /&gt;[TARGET_APPLY ]T: Execute statement: REFRESH TABLE `qlik`.`attrep_changes32432562243S4234D` (databricks_imp.c:612)&lt;BR /&gt;.. &lt;BR /&gt;[TARGET_APPLY ]T: Start applying of 'UPDATE (3)' events for table 'dbo'.'TABLE_XML' (1). (bulk_apply.c:2903) &lt;BR /&gt;[AT_GLOBAL ]T: Bulk update statement:&lt;BR /&gt;[TARGET_APPLY ]T: Going to run update statement, from seq 1 to seq 3. 'MERGE INTO `qlik`.&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;Then in the loop:&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 0, Joined tr. # 259 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 1, Joined tr. # 252 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 2, Joined tr. # 251 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 3, Joined tr. # 242 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 4, Joined tr. # 246 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 5, Joined tr. # 307 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 6, Joined tr. # 370 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 7, Joined tr. # 370 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 8, Joined tr. # 362 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 9, Joined tr. # 318 (transaction_storage.c:2298)&lt;BR /&gt;[SORTER_STORAGE ]T: Forwarded counters. Tr. index 10, Joined tr. # 191 (transaction_storage.c:2298)&lt;/P&gt;
&lt;P&gt;There is one time:&lt;BR /&gt;[IO ]T: rep_net_server_select: Server poll timeout. failed (at_repnet.c:489)&lt;/P&gt;
&lt;P&gt;After this I have still: "T: Forwarded counters. Tr. index ...."&lt;/P&gt;
&lt;P&gt;In sorter folder, Qlik Replicate created some Swap files (few with data and few with 0 data) &lt;BR /&gt;After a while, I tried to stop Task but it was for a long time in stopping mode. So I restarted QR service to regain control.&lt;/P&gt;
&lt;P&gt;It seems that somehow Qlik Replicate was not able to read data from LogStream (with back in time). &lt;BR /&gt;It reads some part of data and then hangs.&lt;/P&gt;
&lt;P&gt;&lt;BR /&gt;&lt;U&gt;&lt;STRONG&gt;Questions:&lt;/STRONG&gt;&lt;/U&gt; &lt;BR /&gt;1. Why did replication to the target stall despite no errors on the Databricks side?&lt;/P&gt;
&lt;P&gt;2. Are the following log messages indicative of the root cause?&lt;BR /&gt;"Reading from source is paused"&lt;BR /&gt;"rep_net_server_select: Server poll timeout. failed"&lt;/P&gt;
&lt;P&gt;3.Could the issue stem from:&lt;BR /&gt;-Memory constraints on the sorter?&lt;BR /&gt;-LogStream not delivering data correctly when processing historical changes?&lt;/P&gt;
&lt;P&gt;For me interesting messages are:&lt;BR /&gt;Reading from source is paused &lt;BR /&gt;rep_net_server_select: Server poll timeout. failed&lt;/P&gt;</description>
      <pubDate>Wed, 27 Nov 2024 10:28:27 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Qlik-Replicate/Job-hangs-after-start-LogStream-to-DataBricks/m-p/2494879#M13286</guid>
      <dc:creator>TomaszRomanowski</dc:creator>
      <dc:date>2024-11-27T10:28:27Z</dc:date>
    </item>
    <item>
      <title>Re: Job hangs after start (LogStream to DataBricks)</title>
      <link>https://community.qlik.com/t5/Qlik-Replicate/Job-hangs-after-start-LogStream-to-DataBricks/m-p/2495404#M13312</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.qlik.com/t5/user/viewprofilepage/user-id/127219"&gt;@TomaszRomanowski&lt;/a&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Could you please open a support ticket to analyze further.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Regards,&lt;/P&gt;
&lt;P&gt;Suresh&lt;/P&gt;</description>
      <pubDate>Sun, 01 Dec 2024 10:11:58 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Qlik-Replicate/Job-hangs-after-start-LogStream-to-DataBricks/m-p/2495404#M13312</guid>
      <dc:creator>sureshkumar</dc:creator>
      <dc:date>2024-12-01T10:11:58Z</dc:date>
    </item>
  </channel>
</rss>

