When you want to read data which is present in sub directories then Spark prompts below error -
Caused by: java.io.IOException: Not a file: hdfs://sdatalakedev/a/b/c=dinesh/subDir1
Solution is to set below property -
set mapreduce.input.fileinputformat.input.dir.recursive=true
Comments
Post a Comment