Do not input private or sensitive data. View Qlik Privacy & Cookie Policy.
Skip to main content

Announcements
Qlik Open Lakehouse is Now Generally Available! Discover the key highlights and partner resources here.
cancel
Showing results for 
Search instead for 
Did you mean: 
Anonymous
Not applicable

Passing Context Variables in BigData Spark Job

Hi Team

 

I have designed a Job to convert csv file to parquet format as shown in pic

0683p000009Lrdr.png

In tfileinputdelimited i am using context variable to define file path

i am getting following error stating No input path specified

cant we parameterise BigDataSpark Job.

How this can be resolved. Do we have any other approch 

please help me out.

 

ErrorMessage

par_dir_name...HR Services Action Report 02.07.2016.csv
[ERROR]: org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand - Aborting job.
java.io.IOException: No input paths specified in job
at org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:201)
at org.talend.hadoop.mapred.lib.file.TDelimitedFileInputFormat.listStatus(TDelimitedFileInputFormat.java:70)
at org.talend.hadoop.mapred.lib.file.TDelimitedFileInputFormat.getSplits(TDelimitedFileInputFormat.java:96)

 

 

Thanks

Labels (4)
1 Solution

Accepted Solutions
Anonymous
Not applicable
Author

Hello,

Can you successfully execute your spark job when use your file path directly in tfileinputdelimited component without context variables?

Best regards

Sabrina

View solution in original post

4 Replies
Anonymous
Not applicable
Author

Hello,

Can you successfully execute your spark job when use your file path directly in tfileinputdelimited component without context variables?

Best regards

Sabrina

badri-nair
Contributor
Contributor

Hi Sabrina, 

I have tried the same. Context load component does not work in a spark batch job.

the hard coded path to files and directories work.

 

Thanks

Badri Nair 

 

torradot
Contributor
Contributor

Hello

Have you some news about this topic ? I have the same problem to iterate several files with a variable name in the tfileinputdelimited .

Thanks for the help...

Tiago
How1
Contributor
Contributor

We have the exact same issue, I would welcome any feedback on this at all.

 

Hardcoded values in the tFileInputDelimited work fine, but the minute I want the filename to be a context variable, the job fails with a not found error.

 

Even if we have the tFileInputDelimited path as:

 

"C:\test\" + context.foldername + "\file.csv"

 

the Spark job fails with:

 

org.apache.hadoop.mapred.InvalidInputException: Input path does not exist: C:\test\file.csv