##Classic .. WordCount Example
hadoop jar target/hadoop-applications-1.0.jar assets.a01.WordCount /home/tony/hadoop-applications/a01/clusterFile /home/tony/hadoop-applications/a01/output{N}
- Dive into the generic type of Mappper interface
- Dive into the generic type of Reducer interface
##Classic .. WordCount2 Example
hadoop jar target/hadoop-applications-1.0.jar assets.a02.WordCount2 /home/tony/hadoop-applications/a02/clusterFile /home/tony/hadoop-applications/a02/output{N}
- Dive into the generic type of Reducer interface
##Parsing API endpont logs : NginxLogParser
hdfs dfs -copyFromLocal src/main/resources/nginxlog /home/tony/hadoop-applications/a03/input
#access logs
hadoop jar target/hadoop-applications-1.0.jar assets.a03.NginxLogParser /home/tony/hadoop-applications/a03/input/nginxlog/access.*.gz /home/tony/hadoop-applications/a03/output20
#error logs
hadoop jar target/hadoop-applications-1.0.jar assets.a03.NginxLogParser /home/tony/hadoop-applications/a03/input/nginxlog/error.*.gz /home/tony/hadoop-applications/a03/output21
output look liks
10.XXX.XXX.XXX 361
10.XXX.XXX.OOO 3
10.XXX.XXX.YYY 3
##HDFS file system manipulation
#ls
hadoop fs -ls /home/tony/hadoop-applications/hdfs01
#CheckExist
hadoop jar target/hadoop-applications-1.0.jar assets.hdfs01.CheckExist
#CheckExistCreate
hadoop jar target/hadoop-applications-1.0.jar assets.hdfs01.CheckExistCreate
#ls again
hadoop fs -ls /home/tony/hadoop-applications/hdfs01