原文博客链接地址:10gR2 rac怎样重跑root.sh ?

前几天遇到一客户的10205 rac,出现LMD进程IPC SEND TIMEOUT问题。

准备深入研究下Oracle RAC

的LMON,LMD以及LMS进程,发现自己的VM RAC无法启动了,最后看了下。竟然是有个节点的分区不见了。

++++Node2

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
[root@rac2 raw]# ls -ltr  /dev/sdf*
brw-r----- 1 root   disk     8, 84 Dec  4  2013 /dev/sdf4
brw-r----- 1 root   disk     8, 83 Dec  4  2013 /dev/sdf3
brw-r----- 1 root   disk     8, 82 Dec  4  2013 /dev/sdf2
brw-r----- 1 root   disk     8, 81 Dec  4  2013 /dev/sdf1
brwxrwxr-x 1 oracle oinstall 8, 80 Dec  4  2013 /dev/sdf
[root@rac2 bin]# cat /etc/rc.d/rc.local
#!/bin/sh
#
# This script will be executed *after* all the other init scripts.
# You can put your own initialization stuff in here
if you don't
# want to
do the full Sys V style init stuff.
touch /var/lock/subsys/local
chown
-R oracle:oinstall /dev/sdf
chown
-R oracle:oinstall /dev/sde
chown
-R oracle:oinstall /dev/sdb
chown
-R oracle:oinstall /dev/sdd
chown
-R oracle:oinstall /dev/sdc
chmod
-R 775 /dev/sdf
chmod
-R 775 /dev/sde
chmod
-R 775 /dev/sdb
chmod
-R 775 /dev/sdd
chmod
-R 775 /dev/sdc
raw /dev/raw/raw1    /dev/sdf1
raw /dev/raw/raw2    /dev/sdf2
raw /dev/raw/raw3    /dev/sdf3
raw /dev/raw/raw4    /dev/sdf4
chown
-R oracle:dba /dev/raw

++++Node1

1
2
3
4
[root@rac1 bin]# partprobe
[root@rac1 bin]# ls -ltr  /dev/sdf*
brwxrwxr-x 1 oracle oinstall 8, 80 Jun 29 01:37 /dev/sdf
[root@rac1 bin]#

我这里是将当中一个共享盘进行了分区,然后绑定为raw。发现当中一个节点的分区不见了,最后把节点2

reboot之后。2个节点的分区都看不到了。无语了。

印象中Linux有一些方法能够恢复partition。尝试使用gpart来试试。例如以下是我尝试:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
[root@rac1 repodata]# gpart /dev/sdf
Begin scan...
End
scan.
Checking partitions...
Ok.
Guessed primary partition table:
Primary partition(1)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Primary partition(2)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Primary partition(3)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Primary partition(4)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
[root@rac1 repodata]# gpart -W /dev/sdf  /dev/sdf
Begin scan...
End
scan.
Checking partitions...
Ok.
Guessed primary partition table:
Primary partition(1)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Primary partition(2)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Primary partition(3)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Primary partition(4)
type: 000(0x00)(unused)
size: 0mb #s(0) s(0-0)
chs:  (0/0/0)-(0/0/0)d (0/0/0)-(0/0/0)r
Edit this table (y,n) : y
Edit which partition (1..4, q to quit) : q
Activate which partition (1..4, q to quit) : 1
Write this partition table (y,n) : y
* Warning: partition table written, you should reboot now.
[root@rac1 repodata]# ls -ltr /dev/sdf*
brwxrwxr-x 1 oracle oinstall 8, 80 Jun 29 03:57 /dev/sdf

大家能够看到,我这里gpart扫出来的,尽管能看到有4个分区。可是全是空的。不知道为什么,太怪异了。

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
[root@rac1 ~]# fdisk -l
Disk /dev/sda: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot      Start        
End      Blocks   Id  System
/dev/sda1   *           1           6       48163+  83  Linux
/dev/sda2               7         515     4088542+  83  Linux
/dev/sda3             516         776     2096482+  82  Linux swap / Solaris
/dev/sda4             777        2610    14731605    5  Extended
/dev/sda5             777        2610    14731573+  83  Linux
Disk /dev/sdb: 524 MB, 524288000 bytes
64 heads, 32 sectors/track, 500 cylinders
Units = cylinders of 2048 * 512 = 1048576 bytes
Disk /dev/sdb doesn't contain a valid partition table
Disk /dev/sdc: 4294 MB, 4294967296 bytes
255 heads, 63 sectors/track, 522 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk /dev/sdc doesn't contain a valid partition table
Disk /dev/sdd: 4294 MB, 4294967296 bytes
255 heads, 63 sectors/track, 522 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk /dev/sdd doesn't contain a valid partition table
Disk /dev/sde: 4294 MB, 4294967296 bytes
255 heads, 63 sectors/track, 522 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk /dev/sde doesn't contain a valid partition table
Disk /dev/sdf: 2147 MB, 2147483648 bytes
255 heads, 63 sectors/track, 261 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot      Start        
End      Blocks   Id  System
/dev/sdf1   *           1           1           0    0 
Empty
Partition 1 has different physical/logical beginnings (non-Linux?):
phys=(0, 0, 0) logical=(0, 0, 1)
Partition 1 has different physical/logical endings:
phys=(0, 0, 0) logical=(267349, 89, 4)
Partition 1 does not
end on cylinder boundary.

因此,最后gpart写回去也一样没用。由于数据都没了。

由于我这里也没有ocr的备份。因此仅仅能重建了。

在10gR2 版本号中。我们能够不必要又一次安装,仅仅跑root.sh脚本就可以。那么怎么搞呢 ?

假设你直接执行root.sh,那么会遇到类似例如以下的问题:

[root@rac1 crs]# ./root.sh

WARNING: directory ‘/home/oracle/app/oracle/product/10.2.0′ is not owned by root

WARNING: directory ‘/home/oracle/app/oracle/product’ is not owned by root

WARNING: directory ‘/home/oracle/app/oracle’ is not owned by root

WARNING: directory ‘/home/oracle/app’ is not owned by root

WARNING: directory ‘/home/oracle’ is not owned by root

No value set for the CRS parameter CRS_OCR_LOCATIONS. Using Values in paramfile.crs

Checking to see if Oracle CRS stack is already configured

Oracle CRS stack is already configured and will be running under init(1M)

[root@rac1 crs]#

##### 清理部分文件,准备运行root.sh

1
2
3
4
5
6
7
8
9
10
11
12
13
1. 删除/etc/oracle/ocr.loc
[root@rac1 crs]# mv /etc/oracle/ocr.loc  /etc/oracle/ocr.loc.bak
mv: overwrite `/etc/oracle/ocr.loc.bak'?

y

2. 删除cssfatal文件
[root@rac1 crs]# cd /etc/oracle/
[root@rac1 oracle]# pwd
/etc/oracle/scls_scr/rac1/oracle
[root@rac1 oracle]# rm cssfatal
rm: remove regular file `cssfatal'? n
[root@rac1 oracle]# mv cssfatal cssfatal.bak
3.  删除/etc/inittab中的信息

最后我分开运行root.sh脚本就可以,例如以下是过程:

节点1:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
[root@rac1 rac1]# /home/oracle/app/oracle/product/10.2.0/crs/root.sh
WARNING: directory '/home/oracle/app/oracle/product/10.2.0' is not owned by root
WARNING: directory '/home/oracle/app/oracle/product' is not owned by root
WARNING: directory '/home/oracle/app/oracle' is not owned by root
WARNING: directory '/home/oracle/app' is not owned by root
WARNING: directory '/home/oracle' is not owned by root
No value set for the CRS parameter CRS_OCR_LOCATIONS. Using Values in paramfile.crs
Checking to see if Oracle CRS stack is already configured
Setting the permissions on OCR backup directory
Setting up NS directories
Oracle Cluster Registry configuration upgraded successfully
WARNING: directory '/home/oracle/app/oracle/product/10.2.0' is not owned by root
WARNING: directory '/home/oracle/app/oracle/product' is not owned by root
WARNING: directory '/home/oracle/app/oracle' is not owned by root
WARNING: directory '/home/oracle/app' is not owned by root
WARNING: directory '/home/oracle' is not owned by root
Successfully accumulated necessary OCR keys.
Using ports: CSS=49895 CRS=49896 EVMC=49898 and EVMR=49897.
node <nodenumber>: <nodename> <private interconnect name> <hostname>
node 1: rac1 rac1-priv rac1
node 2: rac2 rac2-priv rac2
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Now formatting voting device: /dev/raw/raw2
Format of 1 voting devices complete.
Startup will be queued to init within 30 seconds.
Adding daemons to inittab
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
rac1
CSS is inactive on these nodes.
rac2
Local node checking complete.
Run root.sh on remaining nodes to start CRS daemons.

节点2:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
[root@rac2 oracle]# /home/oracle/app/oracle/product/10.2.0/crs/root.sh
WARNING: directory '/home/oracle/app/oracle/product/10.2.0' is not owned by root
WARNING: directory '/home/oracle/app/oracle/product' is not owned by root
WARNING: directory '/home/oracle/app/oracle' is not owned by root
WARNING: directory '/home/oracle/app' is not owned by root
WARNING: directory '/home/oracle' is not owned by root
No value set for the CRS parameter CRS_OCR_LOCATIONS. Using Values in paramfile.crs
Checking to see if Oracle CRS stack is already configured
Setting the permissions on OCR backup directory
Setting up NS directories
Oracle Cluster Registry configuration upgraded successfully
WARNING: directory '/home/oracle/app/oracle/product/10.2.0' is not owned by root
WARNING: directory '/home/oracle/app/oracle/product' is not owned by root
WARNING: directory '/home/oracle/app/oracle' is not owned by root
WARNING: directory '/home/oracle/app' is not owned by root
WARNING: directory '/home/oracle' is not owned by root
clscfg: EXISTING configuration version 3 detected.
clscfg: version 3 is 10G Release 2.
Successfully accumulated necessary OCR keys.
Using ports: CSS=49895 CRS=49896 EVMC=49898 and EVMR=49897.
node <nodenumber>: <nodename> <private interconnect name> <hostname>
node 1: rac1 rac1-priv rac1
node 2: rac2 rac2-priv rac2
clscfg: Arguments check out successfully.
NO KEYS WERE WRITTEN. Supply -force parameter to override.
-force is destructive and will destroy any previous cluster
configuration.
Oracle Cluster Registry for cluster has already been initialized
Startup will be queued to init within 30 seconds.
Adding daemons to inittab
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
rac1
rac2
CSS is active on all nodes.
Waiting for the Oracle CRSD and EVMD to start
Oracle CRS stack installed and running under init(1M)
Running vipca(silent) for configuring nodeapps
Creating VIP application resource on (2) nodes...
Creating GSD application resource on (2) nodes...
Creating ONS application resource on (2) nodes...
Starting VIP application resource on (2) nodes...
Starting GSD application resource on (2) nodes...
Starting ONS application resource on (2) nodes...
Done.

最后我们看到,crs进程都正常起来了。例如以下:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
[root@rac1 oracle]# ps -ef|grep d.bin
oracle   12371 12370  0 04:34 ?        00:00:00 /home/oracle/app/oracle/product/10.2.0/crs/bin/evmd.bin
root     12446 11819  0 04:34 ?        00:00:00 /home/oracle/app/oracle/product/10.2.0/crs/bin/crsd.bin reboot
root     12688 12452  0 04:34 ?        00:00:00 /home/oracle/app/oracle/product/10.2.0/crs/bin/oprocd.bin run -t 1000 -m 500
oracle   12914 12520  0 04:34 ?        00:00:00 /home/oracle/app/oracle/product/10.2.0/crs/bin/ocssd.bin
root     15267  5027  0 04:41 pts/1    00:00:00 grep d.bin
[root@rac1 oracle]# cd /home/oracle/app/oracle/product/10.2.0/crs/bin
[root@rac1 bin]# ./ocrcheck
Status of Oracle Cluster Registry is as follows :
Version                  :          2
Total space (kbytes)     :     521836
Used space (kbytes)      :       4604
Available space (kbytes) :     517232
ID                       :  559767577
Device/File Name         : /dev/raw/raw1
Device/File integrity check succeeded
Device/File not configured
Cluster registry integrity check succeeded
[root@rac1 bin]# ./crsctl query css votedisk
0.     0    /dev/raw/raw4
located 1 votedisk(s).
[root@rac1 bin]#
[root@rac2 bin]# ./crs_stat -t -v
Name           Type           R/RA   F/FT   Target    State     Host
----------------------------------------------------------------------
ora.rac1.gsd   application    0/5    0/0    ONLINE    ONLINE    rac1
ora.rac1.ons   application    0/3    0/0    ONLINE    ONLINE    rac1
ora.rac1.vip   application    0/0    0/0    ONLINE    ONLINE    rac1
ora.rac2.gsd   application    0/5    0/0    ONLINE    ONLINE    rac2
ora.rac2.ons   application    0/3    0/0    ONLINE    ONLINE    rac2
ora.rac2.vip   application    0/0    0/0    ONLINE    ONLINE    rac2
[root@rac2 bin]#

+++注冊db和asm

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
[oracle@rac1 bdump]$ srvctl add database -d roger -o /home/oracle/app/oracle/product/10.2.0/db_1
[oracle@rac1 bdump]$ srvctl add instance -d roger -i roger1 -n rac1
[oracle@rac1 bdump]$ srvctl add instance -d roger -i roger2 -n rac2
[oracle@rac1 bdump]$ srvctl add asm -n rac1 -i +ASM1 -o /home/oracle/app/oracle/product/10.2.0/db_1
null
[PRKS-1030 : Failed to add configuration for ASM instance "+ASM1" on node "rac1" in cluster registry, [PROC-5: User does not have permission to perform a cluster
registry operation on this key. Authentication error [User does not have permission to perform this operation] [0]]
[PROC-5: User does not have permission to perform a cluster registry operation on this key. Authentication error [User does not have permission to perform this
operation] [0]]]
[oracle@rac1 bdump]$
[root@rac2 bin]# ./crs_getperm ora.rac1.vip
Name: ora.rac1.vip
owner:root:rwx,pgrp:oinstall:r-x,other::r--,user:oracle:r-x,
[root@rac2 bin]# ./crs_getperm ora.rac2.vip
Name: ora.rac2.vip
owner:root:rwx,pgrp:oinstall:r-x,other::r--,user:oracle:r-x,
[root@rac2 bin]#

能够看到上面运行报错了,開始以为是vip资源的问题,于是改动了owner,最后发现这是错位的步骤:

1
2
3
4
5
6
7
8
9
10
[root@rac2 bin]# ./crs_setperm ora.rac1.vip  -o oracle
[root@rac2 bin]# ./crs_setperm ora.rac1.vip  -g oinstall
[root@rac2 bin]# ./crs_setperm ora.rac2.vip  -o oracle
[root@rac2 bin]# ./crs_setperm ora.rac2.vip  -g oinstall
[root@rac2 bin]#
[oracle@rac1 bdump]$ srvctl add asm -n rac1 -i +ASM1 -o /home/oracle/app/oracle/product/10.2.0/db_1
null
[PRKS-1030 : Failed to add configuration for ASM instance "+ASM1" on node "rac1" in cluster registry, [PROC-5: User does not have permission to perform a cluster
registry operation on this key. Authentication error [User does not have permission to perform this operation] [0]]
[PROC-5: User does not have permission to perform a cluster registry operation on this key. Authentication error [User does not have permission to perform this
operation] [0]]]

尽管加入失败,我就改用root运行,进行加入,例如以下:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
[root@rac1 bin]# ./srvctl add asm -n rac1 -i +ASM1 -o /home/oracle/app/oracle/product/10.2.0/db_1
[root@rac1 bin]# ./srvctl add asm -n rac2 -i +ASM2 -o /home/oracle/app/oracle/product/10.2.0/db_1
[root@rac1 bin]# ./crs_stat -p|grep asm
NAME=ora.rac1.ASM1.asm
NAME=ora.rac2.ASM2.asm
[root@rac1 bin]# ./crs_setperm ora.rac1.ASM1.asm -o oracle
[root@rac1 bin]# ./crs_setperm ora.rac1.ASM2.asm -o oracle
[root@rac1 bin]# ./crs_setperm ora.rac1.ASM1.asm -g oinstall
[root@rac1 bin]# ./crs_setperm ora.rac2.ASM2.asm -g oinstall
[oracle@rac1 bdump]$ crs_start ora.rac1.ASM1.asm
Attempting to start `ora.rac1.ASM1.asm` on member `rac1`
Start of `ora.rac1.ASM1.asm` on member `rac1` succeeded.
[oracle@rac1 bdump]$ crs_start ora.rac2.ASM2.asm
Attempting to start `ora.rac2.ASM2.asm` on member `rac2`
Start of `ora.rac2.ASM2.asm` on member `rac2` succeeded.
[oracle@rac1 bdump]$ crs_stat -t
Name           Type           Target    State     Host
------------------------------------------------------------
ora....SM1.asm application    ONLINE    ONLINE    rac1
ora....C1.lsnr application    ONLINE    OFFLINE
ora.rac1.gsd   application    ONLINE    ONLINE    rac1
ora.rac1.ons   application    ONLINE    ONLINE    rac1
ora.rac1.vip   application    ONLINE    OFFLINE
ora....SM2.asm application    ONLINE    ONLINE    rac2
ora....C2.lsnr application    ONLINE    OFFLINE
ora.rac2.gsd   application    ONLINE    ONLINE    rac2
ora.rac2.ons   application    ONLINE    ONLINE    rac2
ora.rac2.vip   application    ONLINE    OFFLINE
ora.roger.db   application    ONLINE    ONLINE    rac2
ora....r1.inst application    ONLINE    ONLINE    rac1
ora....r2.inst application    ONLINE    ONLINE    rac2

最后发现vip和lsnr资源死活起不来。看了下日志。才发现開始vip资源搞错了,vip资源的owner本身就应该是root才对。

通过查看crsd.log 日志。会发现例如以下信息:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
2014-06-29 09:04:56.578: [  CRSRES][2719009680]0startRunnable: setting CLI values
2014-06-29 09:04:56.775: [  CRSRES][2708519824]0startRunnable: setting CLI values
2014-06-29 09:04:56.820: [  CRSRES][2687540112]0startRunnable: setting CLI values
2014-06-29 09:04:56.903: [  CRSRES][2719009680]0Attempting to start `ora.rac1.vip` on member `rac1`
2014-06-29 09:04:56.929: [  CRSRES][2708519824]0Attempting to start `ora.rac1.ASM1.asm` on member `rac1`
2014-06-29 09:04:56.951: [  CRSRES][2687540112]0Attempting to start `ora.roger.roger1.inst` on member `rac1`
2014-06-29 09:04:58.798: [  CRSAPP][2719009680]0StartResource error for ora.rac1.vip error code = 1
2014-06-29 09:04:59.579: [  CRSRES][2719009680]0Start of `ora.rac1.vip` on member `rac1` failed.
2014-06-29 09:05:00.007: [ COMMCRS][2644503440]clsc_send_msg: (0x98bede0) NS err (12571, 12560), transport (530, 111, 0)
2014-06-29 09:05:00.007: [ CRSCOMM][2719009680]0CLSC connect failed torac2ret = 9
2014-06-29 09:05:00.008: [  CRSEVT][2719009680]0invokepeer ret 200
2014-06-29 09:05:00.040: [  CRSRES][2719009680]0Remote start never sent to rac2: X_E2E_NotSent : Failed to connect to node: rac2
(File: caa_CmdRTI.cpp, line: 504
2014-06-29 09:05:00.040: [  CRSRES][2719009680][ALERT]0Remote start for `ora.rac1.vip` failed on member `rac2`
2014-06-29 09:05:01.047: [  CRSRES][2719009680]0startRunnable: setting CLI values
2014-06-29 09:05:01.147: [  CRSRES][2719009680]0Attempting to start `ora.rac1.vip` on member `rac1`
2014-06-29 09:05:02.400: [  CRSAPP][2719009680]0StartResource error for ora.rac1.vip error code = 1
2014-06-29 09:05:03.702: [  CRSRES][2719009680]0Start of `ora.rac1.vip` on member `rac1` failed.
2014-06-29 09:05:04.811: [  CRSRES][2613033872]0startRunnable: setting CLI values
2014-06-29 09:05:04.967: [  CRSRES][2613033872]0Attempting to start `ora.rac1.vip` on member `rac1`
2014-06-29 09:05:05.268: [  CRSAPP][2613033872]0StartResource error for ora.rac1.vip error code = 1
2014-06-29 09:05:06.769: [  CRSRES][2613033872]0Start of `ora.rac1.vip` on member `rac1` failed.
2014-06-29 09:05:11.078: [  CRSRES][2613033872]0startRunnable: setting CLI values
2014-06-29 09:05:11.342: [  CRSRES][2613033872]0Attempting to start `ora.rac1.ons` on member `rac1`
2014-06-29 09:05:13.926: [  CRSRES][2613033872]0Start of `ora.rac1.ons` on member `rac1` succeeded.
2014-06-29 09:05:13.966: [  CRSRES][2708519824]0Start of `ora.rac1.ASM1.asm` on member `rac1` succeeded.
2014-06-29 09:05:45.321: [  CRSRES][2708519824]0CRS-1002: Resource 'ora.rac1.ons' is already running on member 'rac1'
2014-06-29 09:05:46.461: [  CRSRES][2687540112]0Start of `ora.roger.roger1.inst` on member `rac1` succeeded.
2014-06-29 09:05:46.472: [  CRSRES][2698029968]0Skip online resource: ora.rac1.ons
2014-06-29 09:05:49.505: [  CRSRES][2687540112]0startRunnable: setting CLI values
2014-06-29 09:05:49.969: [  CRSRES][2613033872]0startRunnable: setting CLI values
2014-06-29 09:05:50.186: [  CRSRES][2613033872]0Attempting to start `ora.rac1.vip` on member `rac1`
2014-06-29 09:05:50.307: [  CRSRES][2687540112]0Attempting to start `ora.rac1.gsd` on member `rac1`
2014-06-29 09:05:50.788: [  CRSRES][2677050256]0Attempting to start `ora.rac2.vip` on member `rac2`
2014-06-29 09:05:50.906: [  CRSRES][2698029968]0Attempting to start `ora.rac2.gsd` on member `rac2`
2014-06-29 09:05:50.985: [  CRSRES][2719009680]0Attempting to start `ora.rac2.ons` on member `rac2`
2014-06-29 09:05:51.079: [  CRSRES][2708519824]0Attempting to start `ora.roger.db` on member `rac2`
2014-06-29 09:05:51.082: [  CRSAPP][2613033872]0StartResource error for ora.rac1.vip error code = 1
2014-06-29 09:05:51.978: [  CRSRES][2613033872]0Start of `ora.rac1.vip` on member `rac1` failed.
2014-06-29 09:05:52.059: [  CRSRES][2613033872]0rac2 : CRS-1019: Resource ora.rac1.LISTENER_RAC1.lsnr (application) cannot run on rac2
2014-06-29 09:05:53.001: [  CRSRES][2687540112]0Start of `ora.rac1.gsd` on member `rac1` succeeded.
2014-06-29 09:05:54.193: [  CRSRES][2708519824]0Start of `ora.roger.db` on member `rac2` succeeded.
2014-06-29 09:05:54.505: [  CRSRES][2698029968]0Start of `ora.rac2.gsd` on member `rac2` succeeded.
2014-06-29 09:05:54.869: [  CRSRES][2634013584]0CRS-1002: Resource 'ora.roger.db' is already running on member 'rac2'
2014-06-29 09:05:55.054: [  CRSRES][2677050256]0Start of `ora.rac2.vip` on member `rac2` failed.
2014-06-29 09:05:55.226: [  CRSRES][2677050256]0startRunnable: setting CLI values
2014-06-29 09:05:55.277: [  CRSRES][2677050256]0Attempting to start `ora.rac2.vip` on member `rac1`
2014-06-29 09:05:55.585: [  CRSAPP][2677050256]0StartResource error for ora.rac2.vip error code = 1
2014-06-29 09:05:55.714: [  CRSRES][2719009680]0Start of `ora.rac2.ons` on member `rac2` succeeded.
2014-06-29 09:05:55.910: [  CRSRES][2677050256]0Start of `ora.rac2.vip` on member `rac1` failed.
2014-06-29 09:05:56.363: [  CRSRES][2677050256]0Attempting to start `ora.rac2.vip` on member `rac2`
2014-06-29 09:05:57.180: [  CRSRES][2677050256]0Start of `ora.rac2.vip` on member `rac2` failed.
2014-06-29 09:05:57.993: [  CRSRES][2654993296]0startRunnable: setting CLI values
2014-06-29 09:05:58.611: [  CRSAPP][2654993296]0StartResource error for ora.rac1.vip error code = 1
2014-06-29 09:05:59.333: [  CRSRES][2708519824]0startRunnable: setting CLI values
2014-06-29 09:06:00.129: [  CRSAPP][2708519824]0StartResource error for ora.rac2.vip error code = 1
2014-06-29 09:06:06.328: [  CRSRES][2708519824]0startRunnable: setting CLI values
2014-06-29 09:06:06.916: [  CRSAPP][2708519824]0StartResource error for ora.rac1.vip error code = 1

最后将vip资源改回为root,一切正常:

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
[root@rac1 bin]# ./crs_setperm ora.rac1.vip -o root
[root@rac1 bin]# ./crs_setperm ora.rac1.vip -g root
[root@rac1 bin]# ./crs_start ora.rac1.vip
Attempting to start `ora.rac1.vip` on member `rac1`
Start of `ora.rac1.vip` on member `rac1` succeeded.
[root@rac1 bin]# ./crs_setperm ora.rac2.vip -o root
[root@rac1 bin]# ./crs_setperm ora.rac2.vip -g root
[root@rac1 bin]# ./crs_start ora.rac2.vip
Attempting to start `ora.rac2.vip` on member `rac2`
Start of `ora.rac2.vip` on member `rac2` succeeded.
[root@rac1 bin]#
[oracle@rac1 racg]$ crs_stat -t -v
Name           Type           R/RA   F/FT   Target    State     Host
----------------------------------------------------------------------
ora....SM1.asm application    0/5    0/0    ONLINE    ONLINE    rac1
ora....C1.lsnr application    0/5    0/0    ONLINE    ONLINE    rac1
ora.rac1.gsd   application    0/5    0/0    ONLINE    ONLINE    rac1
ora.rac1.ons   application    0/3    0/0    ONLINE    ONLINE    rac1
ora.rac1.vip   application    0/0    0/0    ONLINE    ONLINE    rac1
ora....SM2.asm application    0/5    0/0    ONLINE    ONLINE    rac2
ora....C2.lsnr application    0/5    0/0    ONLINE    ONLINE    rac2
ora.rac2.gsd   application    0/5    0/0    ONLINE    ONLINE    rac2
ora.rac2.ons   application    0/3    0/0    ONLINE    ONLINE    rac2
ora.rac2.vip   application    0/0    0/0    ONLINE    ONLINE    rac2
ora.roger.db   application    0/0    0/1    ONLINE    ONLINE    rac2
ora....r1.inst application    0/5    0/0    ONLINE    ONLINE    rac1
ora....r2.inst application    0/5    0/0    ONLINE    ONLINE    rac2
[oracle@rac1 racg]$
1
非常久没有搞10gR2 rac。有点生疏了,我檫!

10gR2 rac怎样重跑root.sh ?的更多相关文章

  1. 安装GRID时跑root.sh脚本报错(ORA-27091: unable to queue I/O)

    在安装GRID过程中,运行root.sh脚本时报如下信息: Adding Clusterware entries to upstart CRS-2672: Attempting to start 'o ...

  2. centos 7安装rac 11gR2时运行root.sh报错找不到ohas服务(ohasd failed to start)

    单独在linux 7中为ohasd设置一个服务.步骤如下1. 创建服务ohas.service的服务文件并赋予权限touch /usr/lib/systemd/system/ohas.servicec ...

  3. Oracle 11g RAC 第二节点root.sh执行失败后再次执行root.sh

    Oracle 11g RAC 第二节点root.sh执行失败后再次执行root.sh前,要先清除之前的crs配置信息 # /u01/app/11.2.0/grid/crs/install/rootcr ...

  4. RAC安装gird,第一个节点执行root.sh报"The ora.asm resource is not ONLINE"错误

    RAC版本:11.2.0.4 OS版本:linux 6.4 RAC安装gird,第一个节点执行root.sh运行失败,报"The ora.asm resource is not ONLINE ...

  5. Oracle 12.2.0.1 RAC for rhel 7.X 数据库安装(节点1执行root.sh失败)

    说明: 最开始是用的rehat7.2安装12.2.0.1,后面安装GI节点一执行root.sh脚本失败,排查原因,最开始以为是操作系统的问题,换成rehat7.6,同样的出现问题,经过一番折腾,后面通 ...

  6. RAC执行root.sh报libcap.so.1: cannot open shared object file

    Failed to create keys in the OLR, rc = 127, Message: /opt/app/11.2.0/grid/bin/clscfg.bin: error whil ...

  7. RAC安装GI时运行root.sh脚本结果

    第一节点运行root.sh脚本的结果: # /u01/app//grid/root.sh Performing root user operation for Oracle 11g The follo ...

  8. RAC安装重新运行root.sh

    rac1执行root.sh成功,rac2执行失败. 在重新执行root.sh前,在rac2上把crs等配置信息删除: # /u01/app//grid/crs/install/rootcrs.pl - ...

  9. Oracle:Redhat 7.4+Oracle Rac 11.2.0.4 执行root.sh报错处理

    一.报错信息 二.原因分析 因为RHEL 7使用systemd而不是initd运行进程和重启进程,而root.sh通过传统的initd运行ohasd进程 三.解决办法 在RHEL 7中ohasd需要被 ...

随机推荐

  1. jquery中的jsonp跨域调用

                                                    jquery jsonp跨域调用接口

  2. jvm 垃圾回收概念和算法

    1.概念 GC 中的垃圾,特指存在于内存中.不会再被使用的对象.垃圾回收有很多种算法,如引用计数法.复制算法.分代.分区的思想. 2.算法 1.引用计数法:对象被其他所引用时计数器加 1,而当引用失效 ...

  3. java整型byte,short,int,long取值范围大小

     byte 1个字节 short 2个字节 int 4个字节long 8 个字节 varchar 可变长度的非Unicode数据,最长为8000个字符nvarchar 可变长度Unicode数据,最长 ...

  4. OS - 线程和进程的差别

    进程是资源分配的基本单位,又是调度执行的基本单位.比如.用户执行自己的程序,系统就创建一个进程.并为它分配资源,包含各种表.内存空间.磁盘空间.I/O设备等. 然后.把该进程放入进程的就绪队列.进程调 ...

  5. php设计模式之责任链模式

    php设计模式之责任链模式 实际问题 你的论坛有举报功能,版主能解决粗口方面的举报,警察能解决严重一点的黄赌毒方面的举报,更严重的反政府的举报就需要由国安局来完成. 职场中每个人都有直属的上级,如果到 ...

  6. BZOJ 3339 线段树

    思路: 考虑离线处理 显然 l固定时 r越大 ans越大 那我们不妨按照l从小到大排序 l->l+1的时候 l到next[l]这段区间都跟a[l]取min就好了 搞颗线段树维护一下 //By S ...

  7. JavaScript进阶练习

    题目:编写如下页面 当用户点击”统计“按钮时,在窗口中弹出文本框中出现次数最多的字符并显示其出现的次数 点击统计按钮时效果如图所示: 实现代码: <!DOCTYPE html> <h ...

  8. Generating SSH Keys for github

    由于最近电脑重装了Windows 8.1, 想用github维护一些代码.故不得不重新生成一下ssh key. 按https://help.github.com/articles/generating ...

  9. 使用rman恢复数据小结

    恢复前提有数据备份 以 alter database open resetlogs 开机以后多要做一次全备(以前的备份失效了) 恢复参数文件: restore spfile from '/home/o ...

  10. 函数式编程-只用"表达式",不用"语句"()

    把函数当作普通的运算符使用. 2. 只用"表达式",不用"语句"() "表达式"(expression)是一个单纯的运算过程,总是有返回值: ...