I want to know what are the different ways through which I can bring data into HDFS.
I am a newbie to Hadoop and was a java web developer till this time. I want to know if I have a web application that is creating log files, how can i import the log files into HDFS.
You create a single Sqoop import command that imports data from diverse data sources, such as a relational database, into HDFS. You enter the Sqoop import command on the command line of your cluster to import data from a data source into HDFS.
You can import the following file types from a Hadoop Distributed File System (HDFS): . avro, . csv, . json, .
There are lot's of ways on how you can ingest data into HDFS, let me try to illustrate them here:
hdfs dfs -put
- simple way to insert files from local file system to HDFSBest solution for bringing web application logs to HDFS is through Flume.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With