<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic TFileInputDelimited for Big Data dpark cannot read mutiple gz files in Talend Studio</title>
    <link>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344745#M112357</link>
    <description>Hi, I'm using the Talend Big Data studio Enterprise edition and need to read (extract) multiple gz files and then apply transformations on them.
&lt;BR /&gt;
&lt;BR /&gt;On normal DI I used tFileUnarchive for this purpose but it's not present in Spark Big Data.
&lt;BR /&gt;
&lt;BR /&gt;I know that tFileInputDelimited for big data can read gz files by default but I've yet to find a way to allow it to take multiple files as input
&lt;BR /&gt;
&lt;BR /&gt;My files are in the format
&lt;BR /&gt;
&lt;BR /&gt;File1-00001.out.gz
&lt;BR /&gt;File1-00002.out.gz
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;File1-0075.out.gz
&lt;BR /&gt;File2-00001.out.gz
&lt;BR /&gt;File2-00002.out.gz
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;File2-00075.out.gz</description>
    <pubDate>Sat, 16 Nov 2024 05:41:47 GMT</pubDate>
    <dc:creator>Anonymous</dc:creator>
    <dc:date>2024-11-16T05:41:47Z</dc:date>
    <item>
      <title>TFileInputDelimited for Big Data dpark cannot read mutiple gz files</title>
      <link>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344745#M112357</link>
      <description>Hi, I'm using the Talend Big Data studio Enterprise edition and need to read (extract) multiple gz files and then apply transformations on them.
&lt;BR /&gt;
&lt;BR /&gt;On normal DI I used tFileUnarchive for this purpose but it's not present in Spark Big Data.
&lt;BR /&gt;
&lt;BR /&gt;I know that tFileInputDelimited for big data can read gz files by default but I've yet to find a way to allow it to take multiple files as input
&lt;BR /&gt;
&lt;BR /&gt;My files are in the format
&lt;BR /&gt;
&lt;BR /&gt;File1-00001.out.gz
&lt;BR /&gt;File1-00002.out.gz
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;File1-0075.out.gz
&lt;BR /&gt;File2-00001.out.gz
&lt;BR /&gt;File2-00002.out.gz
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;.
&lt;BR /&gt;File2-00075.out.gz</description>
      <pubDate>Sat, 16 Nov 2024 05:41:47 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344745#M112357</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2024-11-16T05:41:47Z</dc:date>
    </item>
    <item>
      <title>Re: TFileInputDelimited for Big Data dpark cannot read mutiple gz files</title>
      <link>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344746#M112358</link>
      <description>&lt;P&gt;Please, what do you mean by spark big data?&lt;/P&gt; 
&lt;P&gt;&lt;SPAN class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="talend big data - tFileUnarchive.PNG" style="width: 980px;"&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="0683p000009M5MT.png"&gt;&lt;img src="https://community.qlik.com/t5/image/serverpage/image-id/130701i53E3D4CCDB8EF3D5/image-size/large?v=v2&amp;amp;px=999" role="button" title="0683p000009M5MT.png" alt="0683p000009M5MT.png" /&gt;&lt;/span&gt;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 31 May 2019 17:22:46 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344746#M112358</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2019-05-31T17:22:46Z</dc:date>
    </item>
    <item>
      <title>Re: TFileInputDelimited for Big Data dpark cannot read mutiple gz files</title>
      <link>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344747#M112359</link>
      <description>Sorry for sounding so vague, I'm new to Talend.
&lt;BR /&gt;
&lt;BR /&gt;I'm talking about the Big Data Batch jobs which run on Spark framework. The Standard Batch jobs has components like tFileList and tFileUnarchive but a Big Data Batch job doesn't.</description>
      <pubDate>Fri, 31 May 2019 17:35:17 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344747#M112359</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2019-05-31T17:35:17Z</dc:date>
    </item>
    <item>
      <title>Re: TFileInputDelimited for Big Data dpark cannot read mutiple gz files</title>
      <link>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344748#M112360</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt; 
&lt;P&gt;&amp;nbsp; &amp;nbsp; You will have to do an orchestration using DI job and BD job to solve this problem. Why don't you try to give each gz file as parameter to a BD job where it will perform the balance steps. The parent job which will send one file at time can be a normal DI job.&lt;/P&gt; 
&lt;P&gt;&amp;nbsp;&lt;/P&gt; 
&lt;P&gt;Warm Regards,&lt;BR /&gt;Nikhil Thampi&lt;/P&gt; 
&lt;P&gt;Please appreciate our Talend community members by giving Kudos for sharing their time for your query. If your query is answered, please mark the topic as resolved &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 31 May 2019 21:22:01 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/TFileInputDelimited-for-Big-Data-dpark-cannot-read-mutiple-gz/m-p/2344748#M112360</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2019-05-31T21:22:01Z</dc:date>
    </item>
  </channel>
</rss>

