<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic tFileInputParquet - How to read generic parquet files and extract schema? in Talend Studio</title>
    <link>https://community.qlik.com/t5/Talend-Studio/tFileInputParquet-How-to-read-generic-parquet-files-and-extract/m-p/2302191#M74206</link>
    <description>&lt;P&gt;Hello.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I want to make a job that reads a parquet file, in order to apply transformations to the data. This job needs to be generic, which means that I cannot know in advance the column names of my parquet files.&lt;/P&gt;&lt;P&gt;Unfortunately, tFileInputParquet is not compatible neither with dynamic fields, nor with reading the lines as a single string, which would allow me to circumvent this issue.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;How can I solve this issue without resorting to a custom Spark program? Thank you in advance.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;imho, I am very surprised that schema guessing is not already a feature for most components. Since Spark's error stacktrace actually lists the column names in the error message, this looks like a technical possibility. Even default index numbers would help tremendously. As far as I am concerned, industrializing any of our jobs through Talend is either difficult or almost impossible.&lt;/P&gt;</description>
    <pubDate>Sat, 16 Nov 2024 01:14:22 GMT</pubDate>
    <dc:creator>CPorrot1602485748</dc:creator>
    <dc:date>2024-11-16T01:14:22Z</dc:date>
    <item>
      <title>tFileInputParquet - How to read generic parquet files and extract schema?</title>
      <link>https://community.qlik.com/t5/Talend-Studio/tFileInputParquet-How-to-read-generic-parquet-files-and-extract/m-p/2302191#M74206</link>
      <description>&lt;P&gt;Hello.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I want to make a job that reads a parquet file, in order to apply transformations to the data. This job needs to be generic, which means that I cannot know in advance the column names of my parquet files.&lt;/P&gt;&lt;P&gt;Unfortunately, tFileInputParquet is not compatible neither with dynamic fields, nor with reading the lines as a single string, which would allow me to circumvent this issue.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;How can I solve this issue without resorting to a custom Spark program? Thank you in advance.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;imho, I am very surprised that schema guessing is not already a feature for most components. Since Spark's error stacktrace actually lists the column names in the error message, this looks like a technical possibility. Even default index numbers would help tremendously. As far as I am concerned, industrializing any of our jobs through Talend is either difficult or almost impossible.&lt;/P&gt;</description>
      <pubDate>Sat, 16 Nov 2024 01:14:22 GMT</pubDate>
      <guid>https://community.qlik.com/t5/Talend-Studio/tFileInputParquet-How-to-read-generic-parquet-files-and-extract/m-p/2302191#M74206</guid>
      <dc:creator>CPorrot1602485748</dc:creator>
      <dc:date>2024-11-16T01:14:22Z</dc:date>
    </item>
  </channel>
</rss>

