Linux_hadoop_install
1、 Build Linux env
my env is VM RedHat Linux 6.5 64bit
set fixed IP
vim /etc/sysconfig/network-scripts/ifcfg-eth0
set IP to : 192.168.38.128
modify hostname: vim /etc/hosts
set hostname to : itbuilder1
2、install JDK
config JDK env variables
3、install Hadoop env
download Apache hadoop pkg
addr:http://archive.apache.org/dist/hadoop/core/stable2/hadoop-2.7.1.tar.gz
3.1 Extract the package to the specified directory
create a dir : mkdir /usr/local/hadoop
extract file to dir : /usr/local/hadoop :tar -zxvf hadoop-2.7.1.tar.gz -C /usr/local/hadoop
3.2 Modify the configuration file
hadoop2.7.1 version need to modify 5 config files :
1、hadoop-env.sh
2、core-site.xml
3、hdfs-site.xml
4、mapred-site.xml(mapred-site.xml.template)
5、yarn-site.xml
these file all under etc of hadoop, the detail dir is : /usr/local/hadoop/hadoop-2.7.1/etc/hadoop/
3.2.1 Modfiy env variable (hadoop-env.sh)
vim hadoop-env.sh
set up JDK root directory, as shown below:

export JAVA_HOME=/usr/java/jdk1.8.0_20
3.2.2 core-site.xml ,set namenode and temp file addr of HDFS.
<configuration>
<!--set HDFS addr (NameNode) -->
<property>
<name>fs.defaultFS</name>
<value>hdfs://itbuilder1:9000</value>
</property>
<!--set dir of Hadoop runtime file storage directory-->
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/hadoop/hadoop-2.7.1/tmp</value>
</property>
</configuration>
3.2.3 hdfs-site.xml (set duplicate quantity)
<configuration>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
3.2.4 mapred-site.xml ( tell hadoop that later MR runs on yarn )
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
3.2.5 yarn-site.xml
<configuration>
<!-- tell nodemanager the way to get data is shuffle -->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<!--set yarn addr (ResourceManager) -->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>itbuilder1</value>
</property>
</configuration>
4、add hadoop to env variable
vim /etc/profile
export JAVA_HOME=/usr/java/jdk1.8.0_20
export HADOOP_HOME=/usr/local/hadoop/hadoop-2.7.1
export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin
#refresh /etc/profile
source /etc/profile
5、Initialize (format) file system (HDFS)
#hadoop namenode -format
hdfs namenode -format
6、start hadoop (hdfs yarn)
./start-all.sh (need to input linux password)
./start-hdfs.sh
./start-yarn.sh
View the current process of opening by JPs command
[root@linuxidc ~]# jps
3461 ResourceManager
3142 DataNode
3751 NodeManager
3016 NameNode
5034 Jps
3307 SecondaryNameNode
Access the management interface :
http://192.168.38.128:50070 (hdfs management interface)
http://192.168.38.128:8088 (mr management interface)
Linux_hadoop_install的更多相关文章
随机推荐
- [Locked] Longest Substring with At Most Two Distinct Characters
Longest Substring with At Most Two Distinct Characters Given a string, find the length of the longes ...
- POJ 3050 穷举
题意:给定一个5*5的地图,每个格子上有一个数字.从一个格子出发(上下左右4个方向),走5步将数字连起来可以构造出一个6位数.问该地图可以构造出多少个不同的6位数. 分析:可以对每个格子做深度优先遍历 ...
- Java实现人民币大写代码解析
想要实现人民币大写,在发票等场景中使用?? 1234.56显示为:壹仟贰佰叁拾肆元伍角陆分,那就往下看看吧! 本程序可以实现 0 到 9999 9999 9999.994 以内的人民币大写转换,精确到 ...
- 删除表中多余的重复记录(多个字段),只留有rowid最小的记录
假如表Users,其中ID为自增长. ID,Name,Sex 1 张三,男 2 张三,男 3 李四,女 4 李四,女 5 王五,男 --查找出最小行号ID的重复记录 select Name,Sex,C ...
- Android 省市县 三级联动(android-wheel的使用)[转]
转载:http://blog.csdn.net/lmj623565791/article/details/23382805 今天没事跟群里面侃大山,有个哥们说道Android Wheel这个控件,以为 ...
- MYSQL 体系结构图-LRU FREELIST FLUSH LIST
- JDK5-静态导入
import static 1. 导入一个类内所有静态成员 import static java.lang.Math.*; public class StaticImport { public sta ...
- codevs 2612 最有分解方案 (贪心)
/* 数字不重复 将一个正整数分解成若干的整数的和 数字不重复 且数字不相同 保证不重复的话 贪心策略是从2开始分 然后把最后剩下的数均匀分到后面 证明嘛 这里写的可能不是很严谨 对于一个n 如果我们 ...
- HDU5308-脑补-对拍
先贴上对拍的结果: 感想:可以明显发现下标相差1的关系,所以对付这种需要大量脑补的水题,真的蛋疼,以前没用过对拍,只知道有这么回事,调程序什么的都是手算,人工模拟(经常大脑发热,严重缺氧不足),所以试 ...
- CSS从大图中抠取小图完整教程(background-position应用) (转)
自认为把background-position的应用讲得非常通俗易懂的教材.做个记号. 相信很多喜欢研究网页界面的童鞋都遇到过一个奇妙的现象:网页中很多图片素材被合成在一张图片上. 起初小菜模仿网站的 ...