hadoop(九)启动|关闭集群(完全分布式六)|11
前置章节:hadoop集群namenode启动ssh免密登录(hadoop完全分布式五)|11
集群启动
配置workers(3.x之前是slaves),
删除localhost,添加102/103/104
[shaozhiqi@hadoop102 hadoop]$ pwd
/opt/module/hadoop-3.1.2/etc/hadoop
[shaozhiqi@hadoop102 hadoop]$ vi workers
[shaozhiqi@hadoop102 hadoop]$ cat workers
hadoop102
hadoop103
hadoop104
[shaozhiqi@hadoop102 hadoop]$
注:workers文件结尾不能有空格,文件中不能有空行
同步workers到其他机器
[shaozhiqi@hadoop102 hadoop]$ testxsync workers
fname=workers
pdir=/opt/module/hadoop-3.1.2/etc/hadoop
-------hadoop103 -----------------------
sending incremental file list
workers
sent 113 bytes received 37 bytes 100.00 bytes/sec
total size is 40 speedup is 0.27
------hadoop104 -----------------------
sending incremental file list
workers
sent 113 bytes received 37 bytes 100.00 bytes/sec
total size is 40 speedup is 0.27
[shaozhiqi@hadoop102 hadoop]$
验证是否同步成功
[shaozhiqi@hadoop103 hadoop]$ cat workers
hadoop102
hadoop103
hadoop104
[shaozhiqi@hadoop103 hadoop]$
[shaozhiqi@hadoop104 hadoop]$ cat workers
hadoop102
hadoop103
hadoop104
[shaozhiqi@hadoop104 hadoop]$
停止上节启动的namenode和datanode
[shaozhiqi@hadoop103 hadoop-3.1.2]$ sbin/hadoop-daemon.sh stop datanode
shaozhiqi@hadoop104 hadoop-3.1.2]$ sbin/hadoop-daemon.sh stop datanode
shaozhiqi@hadoop102 hadoop-3.1.2]$ sbin/hadoop-daemon.sh stop datanode
shaozhiqi@hadoop102 hadoop-3.1.2]$ sbin/hadoop-daemon.sh stop namenode
版本提示用这个命令:hdfs --daemon stop
启动hdfs
sbin/start-dfs.sh
(hdfs包括了namenode和datanode两个节点。具体可查看大数据及hadoop简要概念|6
[shaozhiqi@hadoop102 hadoop-3.1.2]$ sbin/start-dfs.sh
Starting namenodes on [hadoop102]
hadoop102: WARNING: /opt/module/hadoop-3.1.2/logs does not exist. Creating.
Starting datanodes
hadoop103: WARNING: /opt/module/hadoop-3.1.2/logs does not exist. Creating.
hadoop104: WARNING: /opt/module/hadoop-3.1.2/logs does not exist. Creating.
Starting secondary namenodes [hadoop104]
[shaozhiqi@hadoop102 hadoop-3.1.2]$
Jps发现其他都正常,但是namenode就是没有启动。。。
删除data logs,格式化namenode
[shaozhiqi@hadoop102 hadoop-3.1.2]$ bin/hdfs namenode –format
再次启动hdfs
[shaozhiqi@hadoop102 hadoop-3.1.2]$ sbin/start-dfs.sh
Starting namenodes on [hadoop102]
Starting datanodes
hadoop104: WARNING: /opt/module/hadoop-3.1.2/logs does not exist. Creating.
hadoop103: WARNING: /opt/module/hadoop-3.1.2/logs does not exist. Creating.
Starting secondary namenodes [hadoop104]
[shaozhiqi@hadoop102 hadoop-3.1.2]$ jps
7619 DataNode
7877 Jps
7494 NameNode
[shaozhiqi@hadoop102 hadoop-3.1.2]$
验证
102
[shaozhiqi@hadoop102 hadoop-3.1.2]$ jps
7619 DataNode
7877 Jps
7494 NameNode
[shaozhiqi@hadoop102 hadoop-3.1.2]$
103
[shaozhiqi@hadoop103 hadoop-3.1.2]$ jps
2931 Jps
2847 DataNode
[shaozhiqi@hadoop103 hadoop-3.1.2]$
104
[shaozhiqi@hadoop104 hadoop-3.1.2]$ jps
5424 SecondaryNameNode
5467 Jps
5308 DataNode
[shaozhiqi@hadoop104 hadoop-3.1.2]$
参照我们的规划配置,具体查看hadoop集群配置同步(hadoop完全分布式四)|10

验证通过
也可以看先我们的hdfs的namenodeweb,可以看到有三台机器ok

启动yarn
根据我们在[**hadoop****集群配置同步[hadoop集群配置同步(hadoop完全分布式四)|10]
的集群分配原则,我们的yarn的recourceManger在103所以yarn的启动只能在103起,否则会报错
sbin/start-yarn.sh
[shaozhiqi@hadoop103 hadoop-3.1.2]$ sbin/start-yarn.sh
Starting resourcemanager
Starting nodemanagers
[shaozhiqi@hadoop103 hadoop-3.1.2]$ jps
3154 ResourceManager
3271 NodeManager
3611 Jps
2847 DataNode
[shaozhiqi@hadoop103 hadoop-3.1.2]$
103ok
[shaozhiqi@hadoop104 hadoop-3.1.2]$ jps
5424 SecondaryNameNode
5782 Jps
5657 NodeManager
5308 DataNode
[shaozhiqi@hadoop104 hadoop-3.1.2]$
104ok
[shaozhiqi@hadoop102 hadoop-3.1.2]$ jps
7619 DataNode
8052 NodeManager
8164 Jps
7494 NameNode
[shaozhiqi@hadoop102 hadoop-3.1.2]$
102ok
启动成功
下节测试我们的hadoop集群
由于要关闭电脑,所以停止hadoop集群
启动时先起hdfs,然后yarn
先停yarn 再停hdfs
停集群yarn
[shaozhiqi@hadoop103 hadoop-3.1.2]$ sbin/stop-yarn.sh
Stopping nodemanagers
Stopping resourcemanager
[shaozhiqi@hadoop103 hadoop-3.1.2]$ jps
4145 Jps
2847 DataNode
[shaozhiqi@hadoop103 hadoop-3.1.2]$
停集群hdsf
[shaozhiqi@hadoop102 hadoop-3.1.2]$ sbin/stop-dfs.sh
Stopping namenodes on [hadoop102]
Stopping datanodes
Stopping secondary namenodes [hadoop104]
[shaozhiqi@hadoop102 hadoop-3.1.2]$ jps
8941 Jps
[shaozhiqi@hadoop102 hadoop-3.1.2]$
记得下次启动集群前格式namenode
hadoop(九)启动|关闭集群(完全分布式六)|11的更多相关文章
- CentOS系统下Hadoop 2.4.1集群安装配置(简易版)
安装配置 1.软件下载 JDK下载:jdk-7u65-linux-i586.tar.gz http://www.oracle.com/technetwork/java/javase/downloads ...
- Hadoop 2.6.1 集群安装配置教程
集群环境: 192.168.56.10 master 192.168.56.11 slave1 192.168.56.12 slave2 下载安装包/拷贝安装包 # 存放路径: cd /usr/loc ...
- Hadoop HA高可用集群搭建(Hadoop+Zookeeper+HBase)
声明:作者原创,转载注明出处. 作者:帅气陈吃苹果 一.服务器环境 主机名 IP 用户名 密码 安装目录 master188 192.168.29.188 hadoop hadoop /home/ha ...
- 完全关闭及再次启动cdh集群
关闭集群 - 关闭集群所有组件 关闭Cloudera Management Service 关闭cdh所有客户端节点 sudo /opt/cloudera-manager/cm-5.11.1/etc/ ...
- 【Redis学习之九】Redis集群:Twemproxy和HA
环境 虚拟机:VMware 10 Linux版本:CentOS-6.5-x86_64 客户端:Xshell4 FTP:Xftp4 jdk8 redis-3.0.4 主从模式对写压力没有分担,解决思路就 ...
- shell脚本启动所有集群节点
#profile变量追加到.bashrc中 cat /etc/profile >> ~/.bashrc #start-all-cluster.sh 启动脚本 #!/bin/bash ec ...
- 使用Docker搭建Hadoop集群(伪分布式与完全分布式)
之前用虚拟机搭建Hadoop集群(包括伪分布式和完全分布式:Hadoop之伪分布式安装),但是这样太消耗资源了,自学了Docker也来操练一把,用Docker来构建Hadoop集群,这里搭建的Hado ...
- 大数据系列(3)——Hadoop集群完全分布式坏境搭建
前言 上一篇我们讲解了Hadoop单节点的安装,并且已经通过VMware安装了一台CentOS 6.8的Linux系统,咱们本篇的目标就是要配置一个真正的完全分布式的Hadoop集群,闲言少叙,进入本 ...
- 3-3 Hadoop集群完全分布式配置部署
Hadoop集群完全分布式配置部署 下面的部署步骤,除非说明是在哪个服务器上操作,否则默认为在所有服务器上都要操作.为了方便,使用root用户. 1.准备工作 1.1 centOS6服务器3台 手动指 ...
随机推荐
- vscode 的tab空格设置设置为4的方法
1.点击“文件>首选项>设置” 进入设置页面,设置如下几个选项 2.在“文件>首选项>设置” 的“用户设置”里添加 "editor.detectIndentation ...
- 关于Idea中不能使用Scanner在console
遇到了麻烦,在Idea中使用@Test运行程序时,scanner在控制台无法输入,然后来回折腾... 创建了一个新的类里面含有main方法,可以完美运行scanner: 重新回来,发现还是不行, 创建 ...
- 蓝桥杯vip 字符串对比
蓝桥杯vip 字符串对比 题目如下 给定两个仅由大写字母或小写字母组成的字符串(长度介于1到10之间),它们之间的关系是以下4中情况之一: 1:两个字符串长度不等.比如 Beijing 和 Hebei ...
- oracle去除重复数据与oracle分页
一.去除oracle中重复数据,可以使用rowid列,rowid列是一个伪列,该列在数据库中灭一个表中都有,但是我们查询数据库的时候,默认都没有给我们返回这一列,这一列用来区分数据库中的每一行时间,可 ...
- Pyinstaller通过spec文件打包py程序(多个py脚本)
Pyinstaller pyinstaller是python的一个第三方模块,使用它可以将python程序打包为可执行文件,实现打包后的程序在没有python环境的机器上也可以运行.pyinstall ...
- 在ASP.NET MVC中如何预防Cookie的窃取攻击(转载)
Cookie Cookie is a small piece of data sent by a web server to a web browser. The browser stores thi ...
- .NET Core 3 WPF MVVM框架 Prism系列之区域管理器
本文将介绍如何在.NET Core3环境下使用MVVM框架Prism的使用区域管理器对于View的管理 一.区域管理器 我们在之前的Prism系列构建了一个标准式Prism项目,这篇文章将会讲解之前项 ...
- 单线程多任务协程vip电影爬取
单线程多任务协程vip电影爬取 --仅供学习使用勿作商用如有违规后果自负!!! 这几天一直在使用python爬取电影,主要目的也是为了巩固前段时间强化学习的网络爬虫,也算是一个不错的检验吧,面对众 ...
- 为什么信息熵要定义成-Σp*log(p)?
作者:西贝链接:https://www.zhihu.com/question/30828247/answer/64816509来源:知乎著作权归作者所有.商业转载请联系作者获得授权,非商业转载请注明出 ...
- Building Applications with Force.com and VisualForce(Dev401)(十二):Implementing Business Processes:Automating Business Processes Part 1
ev401-013:Implementing Business Processes:Automating Business Processes Part 1 Module Objectives1.Li ...