Unlock a world of possibilities! Login now and discover the exclusive benefits awaiting you.
Hi Team
I have designed a Job to convert csv file to parquet format as shown in pic
In tfileinputdelimited i am using context variable to define file path
i am getting following error stating No input path specified
cant we parameterise BigDataSpark Job.
How this can be resolved. Do we have any other approch
please help me out.
ErrorMessage
par_dir_name...HR Services Action Report 02.07.2016.csv
[ERROR]: org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand - Aborting job.
java.io.IOException: No input paths specified in job
at org.apache.hadoop.mapred.FileInputFormat.listStatus(FileInputFormat.java:201)
at org.talend.hadoop.mapred.lib.file.TDelimitedFileInputFormat.listStatus(TDelimitedFileInputFormat.java:70)
at org.talend.hadoop.mapred.lib.file.TDelimitedFileInputFormat.getSplits(TDelimitedFileInputFormat.java:96)
Thanks
Hello,
Can you successfully execute your spark job when use your file path directly in tfileinputdelimited component without context variables?
Best regards
Sabrina
Hello,
Can you successfully execute your spark job when use your file path directly in tfileinputdelimited component without context variables?
Best regards
Sabrina
Hi Sabrina,
I have tried the same. Context load component does not work in a spark batch job.
the hard coded path to files and directories work.
Thanks
Badri Nair
We have the exact same issue, I would welcome any feedback on this at all.
Hardcoded values in the tFileInputDelimited work fine, but the minute I want the filename to be a context variable, the job fails with a not found error.
Even if we have the tFileInputDelimited path as:
"C:\test\" + context.foldername + "\file.csv"
the Spark job fails with:
org.apache.hadoop.mapred.InvalidInputException: Input path does not exist: C:\test\file.csv