<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: ProcessingSpeed Up Processing From tOracleInput to tHDFSOutput in Talend Studio</title>
    <link>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296243#M68927</link>
    <description>What if in case I want to load the data from Oracle in memory first and then do processing on it?
&lt;BR /&gt;If I use sqoop then I will have to get data to HDFS first then read from there. There will 2 I/O operations involved here. In case I use tOracleInput, then data will come in memory. I will do direct processing on it and then load it in HDFS.
&lt;BR /&gt;
&lt;BR /&gt;Which according to you is the better approach?</description>
    <pubDate>Mon, 12 Dec 2016 05:18:09 GMT</pubDate>
    <dc:creator>Anonymous</dc:creator>
    <dc:date>2016-12-12T05:18:09Z</dc:date>
    <item>
      <title>ProcessingSpeed Up Processing From tOracleInput to tHDFSOutput</title>
      <link>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296241#M68925</link>
      <description>&lt;P&gt;I am running a job that is pulling data from an Oracle DB from a remote server, and I am trying to push that table into HDFS. The best I am getting is 4100 rows per second, and there is total of 53 million rows. I have six tables like that.&amp;nbsp;&lt;BR /&gt;I have set JVM setting&amp;nbsp;&lt;BR /&gt;-xms&amp;nbsp;16GB | -xmx 32GB&lt;BR /&gt;What Can I do to increase the performance,&amp;nbsp;at this rate, I will have all the data loaded in over 12 hours?&amp;nbsp;&lt;BR /&gt;Thanks,&lt;/P&gt;</description>
      <pubDate>Sat, 16 Nov 2024 10:13:15 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296241#M68925</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-11-16T10:13:15Z</dc:date>
    </item>
    <item>
      <title>Re: ProcessingSpeed Up Processing From tOracleInput to tHDFSOutput</title>
      <link>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296242#M68926</link>
      <description>Hi,&lt;BR /&gt;Usually, we&amp;nbsp;&lt;FONT size="1"&gt;&lt;FONT face="Verdana," helvetica=""&gt;are using the tsqoopimport to load the data to HDFS from&amp;nbsp;&lt;/FONT&gt;&lt;/FONT&gt;&lt;FONT size="2"&gt;&lt;FONT face="noto," helvetica=""&gt;a relational database management system (RDBMS), MySql, Oracle..&lt;/FONT&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT size="2"&gt;&lt;FONT face="noto," helvetica=""&gt;Please take a look at component reference about:&lt;A href="https://help.talend.com/search/all?query=tSqoopImport" target="_blank" rel="nofollow noopener noreferrer"&gt;TalendHelpCenter:tSqoopImport&lt;/A&gt;.&lt;/FONT&gt;&lt;/FONT&gt;&lt;BR /&gt;&lt;FONT size="2"&gt;&lt;FONT face="noto," helvetica=""&gt;Best regards&lt;/FONT&gt;&lt;/FONT&gt;&lt;BR /&gt;[font=noto, Helvetica, Arial, sans-serif]&lt;FONT size="2"&gt;Sabrina&lt;/FONT&gt;[/font]</description>
      <pubDate>Fri, 09 Dec 2016 10:21:20 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296242#M68926</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2016-12-09T10:21:20Z</dc:date>
    </item>
    <item>
      <title>Re: ProcessingSpeed Up Processing From tOracleInput to tHDFSOutput</title>
      <link>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296243#M68927</link>
      <description>What if in case I want to load the data from Oracle in memory first and then do processing on it?
&lt;BR /&gt;If I use sqoop then I will have to get data to HDFS first then read from there. There will 2 I/O operations involved here. In case I use tOracleInput, then data will come in memory. I will do direct processing on it and then load it in HDFS.
&lt;BR /&gt;
&lt;BR /&gt;Which according to you is the better approach?</description>
      <pubDate>Mon, 12 Dec 2016 05:18:09 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/ProcessingSpeed-Up-Processing-From-tOracleInput-to-tHDFSOutput/m-p/2296243#M68927</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2016-12-12T05:18:09Z</dc:date>
    </item>
  </channel>
</rss>

