<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Kafka - Talend - Teradata Huge data processing in Talend Studio</title>
    <link>https://community.qlik.com/t5/Talend-Studio/Kafka-Talend-Teradata-Huge-data-processing/m-p/2326675#M96131</link>
    <description>&lt;P&gt;first write in to file and then load file into table.&lt;/P&gt;</description>
    <pubDate>Thu, 18 Feb 2021 19:09:30 GMT</pubDate>
    <dc:creator>sagu</dc:creator>
    <dc:date>2021-02-18T19:09:30Z</dc:date>
    <item>
      <title>Kafka - Talend - Teradata Huge data processing</title>
      <link>https://community.qlik.com/t5/Talend-Studio/Kafka-Talend-Teradata-Huge-data-processing/m-p/2326674#M96130</link>
      <description>&lt;P&gt;I am trying to read huge amount of data about 1Million+ messages from message streaming service(Kafka). My current methodology is tKafkaInput &amp;gt; tExtractJsonFields &amp;gt; tMap &amp;gt;tTeradataOutput.&lt;/P&gt;&lt;P&gt;When I am running my job, I am getting the following error&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;B&gt;Exception in thread "main" java.lang.OutOfMemoryError: Java heap space&lt;/B&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Could someone please suggest on what can be done to avoid this issue and handle unexpected huge data of 5Million and upwards&lt;/P&gt;</description>
      <pubDate>Sat, 16 Nov 2024 00:35:03 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/Kafka-Talend-Teradata-Huge-data-processing/m-p/2326674#M96130</guid>
      <dc:creator>STelkar1613587356</dc:creator>
      <dc:date>2024-11-16T00:35:03Z</dc:date>
    </item>
    <item>
      <title>Re: Kafka - Talend - Teradata Huge data processing</title>
      <link>https://community.qlik.com/t5/Talend-Studio/Kafka-Talend-Teradata-Huge-data-processing/m-p/2326675#M96131</link>
      <description>&lt;P&gt;first write in to file and then load file into table.&lt;/P&gt;</description>
      <pubDate>Thu, 18 Feb 2021 19:09:30 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/Kafka-Talend-Teradata-Huge-data-processing/m-p/2326675#M96131</guid>
      <dc:creator>sagu</dc:creator>
      <dc:date>2021-02-18T19:09:30Z</dc:date>
    </item>
  </channel>
</rss>

