Skip to main content
Announcements
Introducing Qlik Answers: A plug-and-play, Generative AI powered RAG solution. READ ALL ABOUT IT!
cancel
Showing results for 
Search instead for 
Did you mean: 
Anonymous
Not applicable

Reading Flat Large file (2GB) in Big data job (spark)

I am trying to process 2gb flat file using spark big data job. 

it takes very long time (3+) to just read the file. 

 

i have also updated no of executors and executor memory (reluctantly). it doesn't work either. 

 

any suggestions is appreciated  

 

 0683p000009Ma0s.png

Labels (3)
5 Replies
manodwhb
Champion II
Champion II

@uganesh , how are you executing your job ? are you executing from Studio?

Anonymous
Not applicable
Author

@manodwhb thanks for quick response. 

 

I am running Job on RE (submit from studio) , hosted on my AWS env and connecting to EMR cluster.  

manodwhb
Champion II
Champion II

@uganesh, so you were building a job and copying that zip file in the remote engine and executing .sh file?

Anonymous
Not applicable
Author

@manodwhb : 

Below is screenshot of my studio. 

 

Studio ---publish-->  RE --> EMR 

 

0683p000009Ma0x.png

manodwhb
Champion II
Champion II

@uganesh, all executors are utilizing?