I have very huge amount of data stored in a text delimited table whose DDL
I created through Impala
, I have written it with data in Hive
and I want to get it to local file system
in Unix
in a single file.
I have multiple files in HDFS
table location path, this is the command I used:
hadoop fs -cat '<tablelocationinHDFS>/*' > /unixlocalpathandfilename.csv
I'm getting 2 errors:
cat: is a directory error
Because there is one impala_insert_staging directory inside it, how do I avoid this error?
cat: Unable to write to output stream
Or Should I be getting data into Local using get
command or getmerge
command into a single file?
How do I avoid the above errors and have a successful run of command.