First create the DB and table where you want to populate.
echo "create database staff2; use staff2; CREATE TABLE editorial (id INT(100) unsigned not null AUTO_INCREMENT, name VARCHAR(20), email VARCHAR(20), primary key (id));" | mysql -u root -p
Once done, we have the data we want to copy in HDFS.
hdfs dfs -cat /home/training/staff/editorial/part-m-* 1,Peter,peter@example.com 2,Jack,jack@example.com
Now dump into MySQL using sqoop.
sqoop export --connect jdbc:mysql://localhost/staff2 --username root -P --table editorial --export-dir /home/training/staff/editorial 17/02/27 12:51:56 INFO sqoop.Sqoop: Running Sqoop version: 1.4.5-cdh5.2.0 Enter password: 17/02/27 12:51:58 INFO manager.SqlManager: Using default fetchSize of 1000 17/02/27 12:51:58 INFO tool.CodeGenTool: Beginning code generation 17/02/27 12:51:59 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM `editorial` AS t LIMIT 1 17/02/27 12:51:59 INFO manager.SqlManager: Executing SQL statement: SELECT t.* FROM `editorial` AS t LIMIT 1 17/02/27 12:51:59 INFO orm.CompilationManager: HADOOP_MAPRED_HOME is /usr/lib/hadoop-0.20-mapreduce Note: /tmp/sqoop-training/compile/e560499b42a9738bbc5ef127712adc7b/editorial.java uses or overrides a deprecated API. Note: Recompile with -Xlint:deprecation for details. 17/02/27 12:52:03 INFO orm.CompilationManager: Writing jar file: /tmp/sqoop-training/compile/e560499b42a9738bbc5ef127712adc7b/editorial.jar 17/02/27 12:52:03 INFO mapreduce.ExportJobBase: Beginning export of editorial 17/02/27 12:52:06 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 17/02/27 12:52:08 INFO input.FileInputFormat: Total input paths to process : 2 17/02/27 12:52:08 INFO input.FileInputFormat: Total input paths to process : 2 17/02/27 12:52:09 INFO mapred.JobClient: Running job: job_201702221239_0006 17/02/27 12:52:10 INFO mapred.JobClient: map 0% reduce 0% 17/02/27 12:52:31 INFO mapred.JobClient: map 50% reduce 0% 17/02/27 12:52:45 INFO mapred.JobClient: map 100% reduce 0% 17/02/27 12:52:49 INFO mapred.JobClient: Job complete: job_201702221239_0006 17/02/27 12:52:49 INFO mapred.JobClient: Counters: 24 17/02/27 12:52:49 INFO mapred.JobClient: File System Counters 17/02/27 12:52:49 INFO mapred.JobClient: FILE: Number of bytes read=0 17/02/27 12:52:49 INFO mapred.JobClient: FILE: Number of bytes written=1176756 17/02/27 12:52:49 INFO mapred.JobClient: FILE: Number of read operations=0 17/02/27 12:52:49 INFO mapred.JobClient: FILE: Number of large read operations=0 17/02/27 12:52:49 INFO mapred.JobClient: FILE: Number of write operations=0 17/02/27 12:52:49 INFO mapred.JobClient: HDFS: Number of bytes read=759 17/02/27 12:52:49 INFO mapred.JobClient: HDFS: Number of bytes written=0 17/02/27 12:52:49 INFO mapred.JobClient: HDFS: Number of read operations=19 17/02/27 12:52:49 INFO mapred.JobClient: HDFS: Number of large read operations=0 17/02/27 12:52:49 INFO mapred.JobClient: HDFS: Number of write operations=0 17/02/27 12:52:49 INFO mapred.JobClient: Job Counters 17/02/27 12:52:49 INFO mapred.JobClient: Launched map tasks=4 17/02/27 12:52:49 INFO mapred.JobClient: Data-local map tasks=4 17/02/27 12:52:49 INFO mapred.JobClient: Total time spent by all maps in occupied slots (ms)=64216 17/02/27 12:52:49 INFO mapred.JobClient: Total time spent by all reduces in occupied slots (ms)=0 17/02/27 12:52:49 INFO mapred.JobClient: Total time spent by all maps waiting after reserving slots (ms)=0 17/02/27 12:52:49 INFO mapred.JobClient: Total time spent by all reduces waiting after reserving slots (ms)=0 17/02/27 12:52:49 INFO mapred.JobClient: Map-Reduce Framework 17/02/27 12:52:49 INFO mapred.JobClient: Map input records=2 17/02/27 12:52:49 INFO mapred.JobClient: Map output records=2 17/02/27 12:52:49 INFO mapred.JobClient: Input split bytes=661 17/02/27 12:52:49 INFO mapred.JobClient: Spilled Records=0 17/02/27 12:52:49 INFO mapred.JobClient: CPU time spent (ms)=3390 17/02/27 12:52:49 INFO mapred.JobClient: Physical memory (bytes) snapshot=422584320 17/02/27 12:52:49 INFO mapred.JobClient: Virtual memory (bytes) snapshot=2940895232 17/02/27 12:52:49 INFO mapred.JobClient: Total committed heap usage (bytes)=127401984 17/02/27 12:52:49 INFO mapreduce.ExportJobBase: Transferred 759 bytes in 42.9426 seconds (17.6748 bytes/sec) 17/02/27 12:52:49 INFO mapreduce.ExportJobBase: Exported 2 records.
Now we can see the content in MySQL DB named staff2.
echo "use staff2; SELECT * FROM editorial;" | mysql -u root -p Enter password: id name email 1 Peter peter@example.com 2 Jack jack@example.com