在Ubuntu上装Hadoop
在装Hadoop之前首先需要:
1.java1.6.x 最好是sun的,1.5.x也可以
2.ssh
安装ssh
?
# The java implementation to use. Required. # export JAVA_HOME=/usr/lib/j2sdk1.5-sun?
?修改成java的安装目录:(我的是:/usr/lib/jvm/java-6-sun-1.6.0.15)
?
<?xml version="1.0"?><?xml-stylesheet type="text/xsl" href="configuration.xsl"?><!-- Put site-specific property overrides in this file. --><configuration> <property><name>hadoop.tmp.dir</name> <value>/home/hadoop/hadoop-datastore/hadoop-${user.name}</value> </property> <property><name>fs.default.name</name><value>hdfs://localhost:9000</value> </property></configuration>?
?hadoop.tmp.dir配置为你想要的路径,${user.name}会自动扩展为运行hadoop的用户名
?
conf/hdfs-site.xml
?
<configuration> <property> <name>dfs.replication</name> <value>1</value> </property></configuration>
?
?dfs.replication为默认block复制数量
conf/mapred-site.xml
?
<configuration> <property> <name>mapred.job.tracker</name> <value>localhost:9001</value> </property></configuration>
?
?执行
?
格式化分布式文件系统:
?
$ bin/start-all.sh?
?可以从
?
NameNode - http://localhost:50070/?
?查看NameNode和JobTracker
?
运行例子:
?
?
$ bin/hadoop fs -put conf input$ bin/hadoop jar hadoop-*-examples.jar grep input output 'dfs[a-z.]+'?
look at the run result:
$ bin/hadoop fs -get output output?
参考:1、http://hadoop.apache.org/common/docs/current/quickstart.html