<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: How to load data into HDFS using Spark streaming or Batch job with reference to data in Talend Studio</title>
    <link>https://community.qlik.com/t5/Talend-Studio/How-to-load-data-into-HDFS-using-Spark-streaming-or-Batch-job/m-p/2372464#M135300</link>
    <description>&lt;P&gt;Hello,&lt;/P&gt; 
&lt;P&gt;So far, tflowtoIterate is available in Standard ETL only. &amp;nbsp;&lt;/P&gt; 
&lt;P&gt;Here is a KB article about:&lt;A title="https://community.talend.com/t5/Architecture-Best-Practices-and/Spark-Dynamic-Context/ta-p/33038" href="https://community.qlik.com/s/article/ka03p0000006EJjAAM" target="_self"&gt;https://community.talend.com/t5/Architecture-Best-Practices-and/Spark-Dynamic-Context/ta-p/33038&lt;/A&gt;.&lt;/P&gt; 
&lt;P&gt;Hope it will help.&lt;/P&gt; 
&lt;P&gt;Best regards&lt;/P&gt; 
&lt;P&gt;Sabrina&lt;/P&gt;</description>
    <pubDate>Wed, 26 Sep 2018 09:01:26 GMT</pubDate>
    <dc:creator>Anonymous</dc:creator>
    <dc:date>2018-09-26T09:01:26Z</dc:date>
    <item>
      <title>How to load data into HDFS using Spark streaming or Batch job with reference to data</title>
      <link>https://community.qlik.com/t5/Talend-Studio/How-to-load-data-into-HDFS-using-Spark-streaming-or-Batch-job/m-p/2372463#M135299</link>
      <description>&lt;P&gt;Hi All,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Usecase:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have a data in coming form a file.&lt;/P&gt;&lt;P&gt;For example&lt;/P&gt;&lt;P&gt;RawDataA, A&lt;/P&gt;&lt;P&gt;RawDataB, B&lt;/P&gt;&lt;P&gt;RawDataC, C&lt;/P&gt;&lt;P&gt;.&lt;/P&gt;&lt;P&gt;.&lt;/P&gt;&lt;P&gt;RawDataZ, Z&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Now I wanted to store "RawDataX" in corresponding X value location&lt;/P&gt;&lt;P&gt;/X/RawDataX&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Note:&lt;/P&gt;&lt;P&gt;I don't want to create 26 tFileOutputDelimited in job&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is there any possible way where i can use single tFileOutputDelimited for all records&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Heads up&lt;/P&gt;&lt;P&gt;In DI, we can use tFlowtoIterate and context variable in tFileOutputDelimited to generate above requirement&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Can anyone give some ideas how to implement same thing in spark or map-reduce job ?&lt;/P&gt;</description>
      <pubDate>Sat, 16 Nov 2024 08:08:14 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/How-to-load-data-into-HDFS-using-Spark-streaming-or-Batch-job/m-p/2372463#M135299</guid>
      <dc:creator>Bluemoon</dc:creator>
      <dc:date>2024-11-16T08:08:14Z</dc:date>
    </item>
    <item>
      <title>Re: How to load data into HDFS using Spark streaming or Batch job with reference to data</title>
      <link>https://community.qlik.com/t5/Talend-Studio/How-to-load-data-into-HDFS-using-Spark-streaming-or-Batch-job/m-p/2372464#M135300</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt; 
&lt;P&gt;So far, tflowtoIterate is available in Standard ETL only. &amp;nbsp;&lt;/P&gt; 
&lt;P&gt;Here is a KB article about:&lt;A title="https://community.talend.com/t5/Architecture-Best-Practices-and/Spark-Dynamic-Context/ta-p/33038" href="https://community.qlik.com/s/article/ka03p0000006EJjAAM" target="_self"&gt;https://community.talend.com/t5/Architecture-Best-Practices-and/Spark-Dynamic-Context/ta-p/33038&lt;/A&gt;.&lt;/P&gt; 
&lt;P&gt;Hope it will help.&lt;/P&gt; 
&lt;P&gt;Best regards&lt;/P&gt; 
&lt;P&gt;Sabrina&lt;/P&gt;</description>
      <pubDate>Wed, 26 Sep 2018 09:01:26 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/How-to-load-data-into-HDFS-using-Spark-streaming-or-Batch-job/m-p/2372464#M135300</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2018-09-26T09:01:26Z</dc:date>
    </item>
  </channel>
</rss>

