ubuntu安装与测试hadoop1.1.0版本
最近没什么事,就使用之前测试openstack的机器,测试一下hadoop,看看他到底是干嘛的?
测试环境为ubuntu 12.04 64位系统,hadoop版本是1.1.0
下面是安装过程
- apt-get install openjdk-7-jdk
- root@openstack:~/hadoop/conf# java -version
- java version "1.6.0_24"
- OpenJDK Runtime Environment (IcedTea6 1.11.5) (6b24-1.11.5-0ubuntu1~12.04.1)
- OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode)
- wget http://mirror.bit.edu.cn/apache/hadoop/common/hadoop-1.1.0/hadoop-1.1.0.tar.gz
- tart zxvf hadoop-1.1.0.tar.gz
- mv hadoop-1.1.0 hadoop
- <configuration>
- <property>
- <name>fs.default.name</name>
- <value>hdfs://localhost:9000</value>
- </property>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>/home/hadoop/tmp</value>
- </property>
- </configuration>
- <configuration>
- <property>
- <name>dfs.replication</name>
- <value>1</value>
- </property>
- </configuration>
- <configuration>
- <property>
- <name>mapred.job.tracker</name>
- <value>localhost:9001</value>
- </property>
- </configuration>
- export JAVA_HOME=/usr/lib/jvm/java-1.6.0-openjdk-amd64
- root@openstack:~/hadoop/conf# java -version
- java version "1.6.0_24"
- OpenJDK Runtime Environment (IcedTea6 1.11.5) (6b24-1.11.5-0ubuntu1~12.04.1)
- OpenJDK 64-Bit Server VM (build 20.0-b12, mixed mode)
- bin/hadoop namenode -format
- root@openstack:~/hadoop/conf# cd ..
- root@openstack:~/hadoop# bin/hadoop namenode -format
- 12/11/27 14:10:43 INFO namenode.NameNode: STARTUP_MSG:
- /************************************************************
- STARTUP_MSG: Starting NameNode
- STARTUP_MSG: host = openstack/127.0.1.1
- STARTUP_MSG: args = [-format]
- STARTUP_MSG: version = 1.1.0
- STARTUP_MSG: build = https://svn.apache.org/repos/asf/hadoop/common/branches/branch-1.1 -r 1394289; compiled by 'hortonfo' on Thu Oct 4 22:06:49 UTC 2012
- ************************************************************/
- 12/11/27 14:10:43 INFO util.GSet: VM type = 64-bit
- 12/11/27 14:10:43 INFO util.GSet: 2% max memory = 17.77875 MB
- 12/11/27 14:10:43 INFO util.GSet: capacity = 2^21 = 2097152 entries
- 12/11/27 14:10:43 INFO util.GSet: recommended=2097152, actual=2097152
- 12/11/27 14:10:44 INFO namenode.FSNamesystem: fsOwner=root
- 12/11/27 14:10:44 INFO namenode.FSNamesystem: supergroupsupergroup=supergroup
- 12/11/27 14:10:44 INFO namenode.FSNamesystem: isPermissionEnabled=true
- 12/11/27 14:10:44 INFO namenode.FSNamesystem: dfs.block.invalidate.limit=100
- 12/11/27 14:10:44 INFO namenode.FSNamesystem: isAccessTokenEnabled=false accessKeyUpdateInterval=0 min(s), accessTokenLifetime=0 min(s)
- 12/11/27 14:10:44 INFO namenode.NameNode: Caching file names occuring more than 10 times
- 12/11/27 14:10:44 INFO common.Storage: Image file of size 110 saved in 0 seconds.
- 12/11/27 14:10:44 INFO namenode.FSEditLog: closing edit log: position=4, editlog=/home/hadoop/tmp/dfs/name/current/edits
- 12/11/27 14:10:44 INFO namenode.FSEditLog: close success: truncate to 4, editlog=/home/hadoop/tmp/dfs/name/current/edits
- 12/11/27 14:10:46 INFO common.Storage: Storage directory /home/hadoop/tmp/dfs/name has been successfully formatted.
- 12/11/27 14:10:46 INFO namenode.NameNode: SHUTDOWN_MSG:
- /************************************************************
- SHUTDOWN_MSG: Shutting down NameNode at openstack/127.0.1.1
- ************************************************************/
- bin/start-all.sh
- root@openstack:~/hadoop# bin/start-all.sh
- starting namenode, logging to /root/hadoop/libexec/../logs/hadoop-root-namenode-openstack.out
- localhost: starting datanode, logging to /root/hadoop/libexec/../logs/hadoop-root-datanode-openstack.out
- localhost: starting secondarynamenode, logging to /root/hadoop/libexec/../logs/hadoop-root-secondarynamenode-openstack.out
- starting jobtracker, logging to /root/hadoop/libexec/../logs/hadoop-root-jobtracker-openstack.out
- localhost: starting tasktracker, logging to /root/hadoop/libexec/../logs/hadoop-root-tasktracker-openstack.out
- root@openstack:~/hadoop# jps
- 9340 SecondaryNameNode
- 9665 TaskTracker
- 9426 JobTracker
- 9822 Jps
- 8853 NameNode
- 9091 DataNode
- echo "hello world,this is my hadoop test">/tmp/test1
- echo "welcome to hadoop world">/tmp/test2
- root@openstack:~/hadoop# bin/hadoop dfs -ls /tmp/test
- Found 2 items
- -rw-r--r-- 1 root supergroup 35 2012-11-27 15:01 /tmp/test/test1
- -rw-r--r-- 1 root supergroup 24 2012-11-27 15:01 /tmp/test/test2
- root@openstack:~/hadoop# bin/hadoop jar hadoop-examples-1.1.0.jar wordcount /tmp/test/test* /tmp/test/result
- 12/11/27 15:28:21 INFO input.FileInputFormat: Total input paths to process : 2
- 12/11/27 15:28:21 INFO util.NativeCodeLoader: Loaded the native-hadoop library
- 12/11/27 15:28:21 WARN snappy.LoadSnappy: Snappy native library not loaded
- 12/11/27 15:28:21 INFO mapred.JobClient: Running job: job_201211271500_0002
- 12/11/27 15:28:22 INFO mapred.JobClient: map 0% reduce 0%
- 12/11/27 15:28:31 INFO mapred.JobClient: map 100% reduce 0%
- 12/11/27 15:28:40 INFO mapred.JobClient: map 100% reduce 33%
- 12/11/27 15:28:42 INFO mapred.JobClient: map 100% reduce 100%
- 12/11/27 15:28:44 INFO mapred.JobClient: Job complete: job_201211271500_0002
- 12/11/27 15:28:44 INFO mapred.JobClient: Counters: 29
- 12/11/27 15:28:44 INFO mapred.JobClient: Job Counters
- 12/11/27 15:28:44 INFO mapred.JobClient: Launched reduce tasks=1
- 12/11/27 15:28:44 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=12789
- 12/11/27 15:28:44 INFO mapred.JobClient: Total time spent by all reduces waiting after reserving slots (ms)=0
- 12/11/27 15:28:44 INFO mapred.JobClient: Total time spent by all maps waiting after reserving slots (ms)=0
- 12/11/27 15:28:44 INFO mapred.JobClient: Launched map tasks=2
- 12/11/27 15:28:44 INFO mapred.JobClient: Data-local map tasks=2
- 12/11/27 15:28:44 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=10558
- 12/11/27 15:28:44 INFO mapred.JobClient: File Output Format Counters
- 12/11/27 15:28:44 INFO mapred.JobClient: Bytes Written=70
- 12/11/27 15:28:44 INFO mapred.JobClient: FileSystemCounters
- 12/11/27 15:28:44 INFO mapred.JobClient: FILE_BYTES_READ=125
- 12/11/27 15:28:44 INFO mapred.JobClient: HDFS_BYTES_READ=261
- 12/11/27 15:28:44 INFO mapred.JobClient: FILE_BYTES_WRITTEN=71664
- 12/11/27 15:28:44 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=70
- 12/11/27 15:28:44 INFO mapred.JobClient: File Input Format Counters
- 12/11/27 15:28:44 INFO mapred.JobClient: Bytes Read=59
- 12/11/27 15:28:44 INFO mapred.JobClient: Map-Reduce Framework
- 12/11/27 15:28:44 INFO mapred.JobClient: Map output materialized bytes=131
- 12/11/27 15:28:44 INFO mapred.JobClient: Map input records=2
- 12/11/27 15:28:44 INFO mapred.JobClient: Reduce shuffle bytes=131
- 12/11/27 15:28:44 INFO mapred.JobClient: Spilled Records=20
- 12/11/27 15:28:44 INFO mapred.JobClient: Map output bytes=99
- 12/11/27 15:28:44 INFO mapred.JobClient: CPU time spent (ms)=3620
- 12/11/27 15:28:44 INFO mapred.JobClient: Total committed heap usage (bytes)=302252032
- 12/11/27 15:28:44 INFO mapred.JobClient: Combine input records=10
- 12/11/27 15:28:44 INFO mapred.JobClient: SPLIT_RAW_BYTES=202
- 12/11/27 15:28:44 INFO mapred.JobClient: Reduce input records=10
- 12/11/27 15:28:44 INFO mapred.JobClient: Reduce input groups=9
- 12/11/27 15:28:44 INFO mapred.JobClient: Combine output records=10
- 12/11/27 15:28:44 INFO mapred.JobClient: Physical memory (bytes) snapshot=431722496
- 12/11/27 15:28:44 INFO mapred.JobClient: Reduce output records=9
- 12/11/27 15:28:44 INFO mapred.JobClient: Virtual memory (bytes) snapshot=3213344768
- 12/11/27 15:28:44 INFO mapred.JobClient: Map output records=10
- root@openstack:~/hadoop# bin/hadoop dfs -ls /tmp/test
- Found 3 items
- drwxr-xr-x - root supergroup 0 2012-11-27 15:28 /tmp/test/result
- -rw-r--r-- 1 root supergroup 35 2012-11-27 15:01 /tmp/test/test1
- -rw-r--r-- 1 root supergroup 24 2012-11-27 15:01 /tmp/test/test2
- root@openstack:~/hadoop# bin/hadoop dfs -ls /tmp/test/result
- Found 3 items
- -rw-r--r-- 1 root supergroup 0 2012-11-27 15:28 /tmp/test/result/_SUCCESS
- drwxr-xr-x - root supergroup 0 2012-11-27 15:28 /tmp/test/result/_logs
- -rw-r--r-- 1 root supergroup 70 2012-11-27 15:28 /tmp/test/result/part-r-00000
- root@openstack:~/hadoop# bin/hadoop dfs -cat /tmp/test/result/part-r-00000
- hadoop 2
- hello 1
- is 1
- my 1
- test 1
- to 1
- welcome 1
- world 1
- world,this 1
- root@openstack:~/hadoop# cat /tmp/test1
- hello world,this is my hadoop test
- root@openstack:~/hadoop# cat /tmp/test2
- welcome to hadoop world
BTW:如果大家认为我写的不错,希望能给我的博客投个票,谢谢!
http://blog.51cto.com/contest2012/2382721
本文转自 reinxu 51CTO博客,原文链接:http://blog.51cto.com/dl528888/1072963,如需转载请自行联系原作者









