[terry笔记]redhat5.5_11gR2_RAC_安装

 redhat5.5_11gR2_RAC_安装,这篇主要记录RAC安装的执行步骤,最烦琐的就是前期配置,到后面图形界面runInstaller,asmca,dbca就很容易了。

 

--hostname检查--

[root@node1 Server]# hostname

node1

 

--hostname修改,假定两个节点,node1,node2—

[root@node1 Server]#vi /etc/sysconfig/network

[root@node1 Server]#hostname xxx

 

--修改hosts文件,注意对应的hostname和命名,注意127那两列,不要出现主机名,每台机器需要两个网卡,一个绑定public-ip,一个绑定private-ip,virtual-ip和scan-ip会在安装完grid后自动开通--

[root@node1 Server]# cat /etc/hosts

# Do not remove the following line, or various programs
# that require network functionality will fail.
127.0.0.1        localhost.localdomain localhost
::1             localhost6.localdomain6 localhost6

#public ip
192.168.100.101 node1
192.168.100.102 node2

#virtual ip
192.168.100.201 node1-vip
192.168.100.202 node2-vip

#private ip
10.10.10.1 node1-priv
10.10.10.2 node2-priv

#scan ip
192.168.100.250 racscan

 

--创建grid(管理grid)和oracle(管理database)用户,以及oinstall、dba、oper、asmadmin、asmoper、asmdba组—

--如果只创建一个oracle用户我试验也可以,但是会造成一定的混乱,因为grid和database两个的软件的目录中,存在相同的命令--

[root@node1 Server]#groupadd -g 1000 oinstall

[root@node1 Server]#groupadd -g 1001 dba

[root@node1 Server]#groupadd -g 1002 oper

[root@node1 Server]#groupadd -g 1003 asmadmin

[root@node1 Server]#groupadd -g 1004 asmoper

[root@node1 Server]#groupadd -g 1001 asmdba

[root@node1 Server]#useradd –u 1000 –g oinstall –G dba,asmdba,oper oracle

[root@node1 Server]#passwd oracle

[root@node1 Server]#useradd –u 1001 –g oinstall –G dba,asmadmin,asmoper,asmdba,oper grid

[root@node1 Server]#passwd grid

[oracle@node6 ~]$ id
uid=1000(oracle) gid=1000(oinstall) groups=1000(oinstall),1001(dba),1002(oper),1005(asmdba)

[grid@node6 ~]$ id
uid=1001(grid) gid=1000(oinstall) groups=1000(oinstall),1001(dba),1002(oper),1003(asmadmin),1004(asmoper),1005(asmdba)

 

--创建相关目录--

[root@node1 Server]# mkdir -p /u01/ora11g/oracle

[root@node1 Server]# mkdir -p /u01/ora11g/grid

[root@node1 Server]# mkdir -p /opt/ora11g

[root@node1 Server]# chown -R oracle:oinstall /u01/ora11g

[root@node1 Server]# chown -R grid:oinstall /u01/ora11g/grid

[root@node1 Server]# chown -R grid:oinstall /opt/ora11g

[root@node1 Server]# chmod 777 /u01

 

--关闭系统ntp服务(同步时间),11gR2自带同步时间的服务,移除ntp服务的配置文件,不然GRID还是会找ntp--

[root@node1 /]# service ntpd stop

[root@node1 /]# chkconfig ntpd off

[root@node1 /]# chkconfig sendmail off

[root@node1 /]# mv /etc/ntp.conf /etc/ntp.conf.bk

 

--安装相关软件,注意32位和64位,现在基本上都是64位的吧--

[root@node1 Server]# pwd
/media/RHEL_5.5 i386 DVD/Server

[root@node1 Server]# rpm -ivh compat-gcc-34-3.4.6-4.i386.rpm gcc-4.1.2-48.el5.i386.rpm gcc-c++-4.1.2-48.el5.i386.rpm glibc-devel-2.5-49.i386.rpm libstdc++-devel-4.1.2-48.el5.i386.rpm libXp-1.0.0-8.1.el5.i386.rpm openmotif22-2.2.3-18.i386.rpm elfutils-libelf-devel-0.137-3.el5.i386.rpm elfutils-libelf-devel-static-0.137-3.el5.i386.rpm sysstat-7.0.2-3.el5.i386.rpm libaio-devel-0.3.106-5.i386.rpm libgomp-4.4.0-6.el5.i386.rpm glibc-headers-2.5-49.i386.rpm kernel-headers-2.6.18-194.el5.i386.rpm

[root@localhost Server]# rpm -ivh compat-gcc-34-3.4.6-4.x86_64.rpm gcc-4.1.2-48.el5.x86_64.rpm gcc-c++-4.1.2-48.el5.x86_64.rpm glibc-devel-2.5-49.x86_64.rpm libstdc++-devel-4.1.2-48.el5.x86_64.rpm libXp-1.0.0-8.1.el5.x86_64.rpm openmotif22-2.2.3-18.x86_64.rpm elfutils-libelf-devel-0.137-3.el5.x86_64.rpm elfutils-libelf-devel-static-0.137-3.el5.x86_64.rpm sysstat-7.0.2-3.el5.x86_64.rpm libaio-devel-0.3.106-5.x86_64.rpm libgomp-4.4.0-6.el5.x86_64.rpm glibc-headers-2.5-49.x86_64.rpm kernel-headers-2.6.18-194.el5.x86_64.rpm

 

--设置系统参数,添加到最下方,可以根据自身系统进行修改--

[root@node1 Server]# vi /etc/sysctl.conf
kernel.shmall = 2097152
kernel.shmmax = 2147483648
kernel.shmmni = 4096
kernel.sem = 250 32000 100 128
fs.file-max = 6815744
fs.aio-max-nr = 1048576
net.ipv4.ip_local_port_range = 9000 65000
net.core.rmem_default = 1048576
net.core.rmem_max = 4194304
net.core.wmem_default = 262144
net.core.wmem_max = 1048576

[root@node1 Server]# sysctl -p

[root@node1 Server]# vi /etc/security/limits.conf
oracle soft nproc 2047
oracle hard nproc 16384
oracle soft nofile 1024
oracle hard nofile 65536
grid soft nproc 2047
grid hard nproc 16384
grid soft nofile 1024
grid hard nofile 65536

[root@node1 /]# vi /etc/pam.d/login
session    required     pam.limits.so

[root@node1 Server]# vi /etc/selinux/config
SELINUX=disabled
(重启生效)

 

--设置tmpfs的大小,/dev/shm限制了数据库可以配置的内存--

[root@node1 Server]# vi /etc/fstab

tmpfs  /dev/shm  /tmpfs  defaults,size=30G  0 0

[root@node1 Server]# mount –o remount /dev/shm

 

--两个用户的环境变量,grid用户BASE与HOME不要设置为上下级关系,不然安装时会报错ins-32026。oracle用户添加一个ORACLE_UNQNAME,此环境变量可以打开em,emctl start dbconsole。根据需要添加NLS_LANG和NLS_DATE_FORMAT。注意修改各节点的SID—

export NLS_LANG="SIMPLIFIED CHINESE_CHINA”.ZHS16GBK & UTF8
export NLS_DATE_FORMAT='YYYY-MM-DD HH24:MI:SS'

[grid@node1 ~]$ vi .bash_profile
export TMP=/tmp
export TMPDIR=$TMP
export ORACLE_SID=+ASM1
export ORACLE_BASE=/opt/ora11g
export ORACLE_HOME=/u01/ora11g/grid/product/11.2.0
export PATH=$ORACLE_HOME/bin:/usr/sbin:$PATH
if [ $USER = "oracle" ] || [ $USER = "grid" ];then
   if [ $SHELL = "/bin/ksh" ];then
      ulimit -p 16384
      ulimit -n 65536
  else
    ulimit -u 16384 -n 65536
   fi
   umask 022
fi


[oracle@node1 ~]$ vi .bash_profile
export TMP=/tmp
export TMPDIR=$TMP
export ORACLE_SID=racdb1
export ORACLE_TREM=xterm
export ORACLE_BASE=/u01/ora11g
export ORACLE_HOME=$ORACLE_BASE/oracle/product/11.2.0
export PATH=$ORACLE_HOME/bin:/usr/sbin:$PATH
export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib
export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib
export ORACLE_UNQNAME=racdb
if [ $USER = "oracle" ] || [ $USER = "grid" ];then
   if [ $SHELL = "/bin/ksh" ];then
      ulimit -p 16384
      ulimit -n 65536
  else
    ulimit -u 16384 -n 65536
   fi
   umask 022
fi

 

--查看共享磁盘,此时还未格式化,两节点必须相同,11gRAC一般要有3个共享磁盘,一个grid用,一个data,一个recovery--

[root@node1 ~]# fdisk -l

Disk /dev/sda: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *           1          13      104391   83  Linux
/dev/sda2              14         274     2096482+  82  Linux swap / Solaris
/dev/sda3             275        2610    18763920   83  Linux


Disk /dev/sdb: 1073 MB, 1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Disk /dev/sdb doesn't contain a valid partition table


Disk /dev/sdc: 1073 MB, 1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Disk /dev/sdc doesn't contain a valid partition table

 
Disk /dev/sdd: 1073 MB, 1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Disk /dev/sdd doesn't contain a valid partition table


Disk /dev/sde: 10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Disk /dev/sde doesn't contain a valid partition table


Disk /dev/sdf: 10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

Disk /dev/sdf doesn't contain a valid partition table

 

 

----------------------------以下方式是利用asmlib进行asm创建---------------------------------------------------------------------

--安装asm相关软件,注意安装32位或64位。也可以不用安装,利用系统自带的udev进行asm创建(udev我在redhat5.5中没有试验成功,在OEL6.4中成功)--

[root@node2 redhat]# uname -a
Linux node2 2.6.18-194.el5 #1 SMP Tue Mar 16 21:52:43 EDT 2010 i686 i686 i386 GNU/Linux

[root@node2 redhat]# ls
oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm
oracleasmlib-2.0.4-1.el5.i386.rpm
oracleasm-support-2.1.8-1.el5.i386.rpm

[root@node1 x86_64]# ls
oracleasm-2.6.18-194.el5-2.0.5-1.el5.x86_64.rpm
oracleasmlib-2.0.4-1.el5.x86_64.rpm
oracleasm-support-2.1.8-1.el5.x86_64.rpm

[root@node2 redhat]# rpm –ivh *

 

 

--格式化过程,sdb~sdf,按顺序:n、p、1、enter、enter、w--

[root@node1 ~]# fdisk /dev/sdb

Device contains neither a valid DOS partition table, nor Sun, SGI or OSF disklabel
Building a new DOS disklabel. Changes will remain in memory only,
until you decide to write them. After that, of course, the previous
content won't be recoverable.

Warning: invalid flag 0x0000 of partition table 4 will be corrected by w(rite)

Command (m for help): n

Command action
   e   extended
   p   primary partition (1-4)
p

Partition number (1-4): 1
First cylinder (1-130, default 1):
Using default value 1
Last cylinder or +size or +sizeM or +sizeK (1-130, default 130):
Using default value 130


Command (m for help): w
The partition table has been altered!

Calling ioctl() to re-read partition table.
Syncing disks.

 

--再次查看磁盘,已经格式化完毕,两节点必须一样--

[root@node1 ~]# fdisk -l

Disk /dev/sda: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *           1          13      104391   83  Linux
/dev/sda2              14         274     2096482+  82  Linux swap / Solaris
/dev/sda3             275        2610    18763920   83  Linux

 
Disk /dev/sdb: 1073 MB, 1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sdb1               1         130     1044193+  83  Linux

 
Disk /dev/sdc: 1073 MB, 1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sdc1               1         130     1044193+  83  Linux

 
Disk /dev/sdd: 1073 MB, 1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sdd1               1         130     1044193+  83  Linux

 
Disk /dev/sde: 10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sde1               1        1305    10482381   83  Linux


Disk /dev/sdf: 10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
 
   Device Boot      Start         End      Blocks   Id  System
/dev/sdf1               1        1305    10482381   83  Linux

 

 

--查看节点node1的磁盘文件--

[root@node1 ~]# ll /dev/sd*

brw-r----- 1 root disk 8,  0 2013-03-28 /dev/sda

brw-r----- 1 root disk 8,  1 03-28 09:16 /dev/sda1

brw-r----- 1 root disk 8,  2 2013-03-28 /dev/sda2

brw-r----- 1 root disk 8,  3 03-28 09:16 /dev/sda3

brw-r----- 1 root disk 8, 16 03-28 09:24 /dev/sdb

brw-r----- 1 root disk 8, 17 03-28 09:24 /dev/sdb1

brw-r----- 1 root disk 8, 32 03-28 09:24 /dev/sdc

brw-r----- 1 root disk 8, 33 03-28 09:25 /dev/sdc1

brw-r----- 1 root disk 8, 48 03-28 09:26 /dev/sdd

brw-r----- 1 root disk 8, 49 03-28 09:26 /dev/sdd1

brw-r----- 1 root disk 8, 64 03-28 11:27 /dev/sde

brw-r----- 1 root disk 8, 65 03-28 11:27 /dev/sde1

brw-r----- 1 root disk 8, 80 03-28 11:29 /dev/sdf

brw-r----- 1 root disk 8, 81 03-28 11:29 /dev/sdf1

 

--发现节点node2看不到全部格式化好的磁盘文件,重启--

[root@node2 ~]# ll /dev/sd*

brw-r----- 1 root disk 8,  0 2013-03-28 /dev/sda

brw-r----- 1 root disk 8,  1 03-28 09:16 /dev/sda1

brw-r----- 1 root disk 8,  2 2013-03-28 /dev/sda2

brw-r----- 1 root disk 8,  3 03-28 09:16 /dev/sda3

brw-r----- 1 root disk 8, 16 2013-03-28 /dev/sdb

brw-r----- 1 root disk 8, 32 2013-03-28 /dev/sdc

brw-r----- 1 root disk 8, 48 2013-03-28 /dev/sdd

brw-r----- 1 root disk 8, 64 2013-03-28 /dev/sde

brw-r----- 1 root disk 8, 80 2013-03-28 /dev/sdf

[root@node2 ~]#init 6

 

[root@node2 ~]# ll /dev/sd*

brw-r----- 1 root disk 8,  0 2013-03-28 /dev/sda

brw-r----- 1 root disk 8,  1 03-28 11:40 /dev/sda1

brw-r----- 1 root disk 8,  2 2013-03-28 /dev/sda2

brw-r----- 1 root disk 8,  3 03-28 11:40 /dev/sda3

brw-r----- 1 root disk 8, 16 2013-03-28 /dev/sdb

brw-r----- 1 root disk 8, 17 2013-03-28 /dev/sdb1

brw-r----- 1 root disk 8, 32 2013-03-28 /dev/sdc

brw-r----- 1 root disk 8, 33 2013-03-28 /dev/sdc1

brw-r----- 1 root disk 8, 48 2013-03-28 /dev/sdd

brw-r----- 1 root disk 8, 49 2013-03-28 /dev/sdd1

brw-r----- 1 root disk 8, 64 2013-03-28 /dev/sde

brw-r----- 1 root disk 8, 65 2013-03-28 /dev/sde1

brw-r----- 1 root disk 8, 80 2013-03-28 /dev/sdf

brw-r----- 1 root disk 8, 81 2013-03-28 /dev/sdf1

 

--节点node1,初始化oracleasm,并创建asm--

[root@node1 ~]# /etc/init.d/oracleasm configure

Configuring the Oracle ASM library driver.

This will configure the on-boot properties of the Oracle ASM library
driver.  The following questions will determine whether the driver is
loaded on boot and what permissions it will have.  The current values
will be shown in brackets ('[]').  Hitting <ENTER> without typing an
answer will keep that current value.  Ctrl-C will abort.


Default user to own the driver interface []: grid
Default group to own the driver interface []: oinstall
Start Oracle ASM library driver on boot (y/n) [n]: y
Scan for Oracle ASM disks on boot (y/n) [y]: y
Writing Oracle ASM library driver configuration: done
Initializing the Oracle ASMLib driver:                     [  OK  ]
Scanning the system for Oracle ASMLib disks:               [  OK  ]


[root@node1 ~]# /etc/init.d/oracleasm createdisk CRS1 /dev/sdb1
Marking disk "CRS1" as an ASM disk:                        [  OK  ]
[root@node1 ~]# /etc/init.d/oracleasm createdisk CRS2 /dev/sdc1
Marking disk "CRS2" as an ASM disk:                        [  OK  ]
[root@node1 ~]# /etc/init.d/oracleasm createdisk CRS3 /dev/sdd1
Marking disk "CRS3" as an ASM disk:                        [  OK  ]
[root@node1 ~]# /etc/init.d/oracleasm createdisk DATA1 /dev/sde1
Marking disk "DATA1" as an ASM disk:                       [  OK  ]
[root@node1 ~]# /etc/init.d/oracleasm createdisk DATA2 /dev/sdf1
Marking disk "DATA2" as an ASM disk:                       [  OK  ]
[root@node1 ~]# /etc/init.d/oracleasm scandisks
Scanning the system for Oracle ASMLib disks:               [  OK  ]
[root@node1 ~]# /etc/init.d/oracleasm listdisks

CRS1
CRS2
CRS3
DATA1
DATA2

 

--节点node2,初始化oracleasm,并扫描asm--

[root@node2 ~]# /etc/init.d/oracleasm configure
Configuring the Oracle ASM library driver.

This will configure the on-boot properties of the Oracle ASM library
driver.  The following questions will determine whether the driver is
loaded on boot and what permissions it will have.  The current values
will be shown in brackets ('[]').  Hitting <ENTER> without typing an
answer will keep that current value.  Ctrl-C will abort.

Default user to own the driver interface []: grid
Default group to own the driver interface []: oinstall
Start Oracle ASM library driver on boot (y/n) [n]: y
Scan for Oracle ASM disks on boot (y/n) [y]: y
Writing Oracle ASM library driver configuration: done
Initializing the Oracle ASMLib driver:                     [  OK  ]
Scanning the system for Oracle ASMLib disks:               [  OK  ]

[root@node2 ~]# /etc/init.d/oracleasm scandisks
Scanning the system for Oracle ASMLib disks:               [  OK  ]
[root
@node2 ~]# /etc/init.d/oracleasm listdisks CRS1 CRS2 CRS3 DATA1 DATA2

 

--------------以下方式是利用udev进行asm创建,redhat5.5没有找到利用udev进行设置的方法,用OEL6.4做个例子------------------

首先确认是否安装了udev软件包

[root@node6 ~]# rpm -qa|grep udev
udev-095-14.21.el5

通过scsi_id获取设备的块设备的唯一标识名(在OEL6.4上面试验成功)
scsi_id -g -u /dev/sd*

在Oracle Linux 5下,可以使用如下命令(redhat5没成功):
# /sbin/scsi_id -g -u -s /block/sdb/sdb1
SATA_VBOX_HARDDISK_VBd306dbe0-df3367e3_

创建并配置UDEVRules 文件
[root@rac1 rules.d]# touch /etc/udev/rules.d/99-oracle-asmdevices.rules

添加如下内容:
KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VB83552343-28d5a489",NAME="asm-disk1", OWNER="oracle", GROUP="dba",MODE="0660"
KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name", RESULT=="1ATA_VBOX_HARDDISK_VBb96d5ecb-4eae1d96",NAME="asm-disk2", OWNER="oracle", GROUP="dba",MODE="0660"
KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VBfd7bba6c-b91fba70",NAME="asm-disk3", OWNER="oracle", GROUP="dba",MODE="0660"
KERNEL=="sd?1",BUS=="scsi", PROGRAM=="/sbin/scsi_id -g -u /dev/$name",RESULT=="1ATA_VBOX_HARDDISK_VB3239ed0d-db15bbec",NAME="asm-disk4", OWNER="oracle", GROUP="dba",MODE="0660"

 

--设置oracle与grid用户不需要密码ssh互联,并且要分别执行ssh xxx1/2 date,以防runcluvfy过不去--

--在10g RAC中必须手动配置成功,在11g RAC中不必,因为grid和db的图形安装中可以进行自动配置,简单粗暴。但是为了跑检查脚本,还是手动配置上吧--

[oracle@node1 ~]$ mkdir .ssh
[oracle@node1 ~]$ ssh-keygen -t rsa

Generating public/private rsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_rsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /home/oracle/.ssh/id_rsa.
Your public key has been saved in /home/oracle/.ssh/id_rsa.pub.
The key fingerprint is:
e3:71:9f:fa:da:14:a9:cb:c1:01:17:a2:6d:9d:9d:5f oracle@node1

[oracle@node1 ~]$ ssh-keygen -t dsa

Generating public/private dsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_dsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /home/oracle/.ssh/id_dsa.
Your public key has been saved in /home/oracle/.ssh/id_dsa.pub.
The key fingerprint is:
0b:bf:0e:7f:1c:c7:e6:59:9c:62:d4:42:ec:f7:3f:b0 oracle@node1

[oracle@node1 ~]$ cd .ssh
[oracle@node1 .ssh]$ ls
id_dsa  id_dsa.pub  id_rsa  id_rsa.pub


[oracle@node2 ~]$ mkdir .ssh
[oracle@node2 ~]$ ssh-keygen -t rsa

Generating public/private rsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_rsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /home/oracle/.ssh/id_rsa.
Your public key has been saved in /home/oracle/.ssh/id_rsa.pub.
The key fingerprint is:
82:6c:3f:6f:20:ec:6f:e8:cc:1a:18:73:0b:cf:f5:04 oracle@node2

[oracle@node2 ~]$ ssh-keygen -t dsa

Generating public/private dsa key pair.
Enter file in which to save the key (/home/oracle/.ssh/id_dsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /home/oracle/.ssh/id_dsa.
Your public key has been saved in /home/oracle/.ssh/id_dsa.pub.
The key fingerprint is:
8f:e9:90:9e:d3:77:4f:87:a4:c6:5d:ff:ce:50:42:4f oracle@node2
 

[oracle@node1 ~]$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
[oracle@node1 ~]$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys
[oracle@node1 ~]$ ssh node2 cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
[oracle@node1 ~]$ ssh node2 cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys
[oracle@node1 ~]$ scp ~/.ssh/authorized_keys node2:~/.ssh/authorized_keys

[oracle@node1 ~]$ ssh node2 date
[oracle@node1 ~]$ ssh node1 date
[oracle@node2 ~]$ ssh node1 date
[oracle@node2 ~]$ ssh node2 date

 

[grid@node1 ~]$ mkdir .ssh
[grid@node1 ~]$ ssh-keygen -t rsa

Generating public/private rsa key pair.
Enter file in which to save the key (/home/grid/.ssh/id_rsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /home/grid/.ssh/id_rsa.
Your public key has been saved in /home/grid/.ssh/id_rsa.pub.
The key fingerprint is:
ce:e0:24:b0:4c:f2:25:7d:b9:e5:48:19:10:23:33:97 grid@node1

[grid@node1 ~]$ ssh-keygen -t dsa
Generating public
/private dsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again. Your identification has been saved in /home/grid/.ssh/id_dsa. Your public key has been saved in /home/grid/.ssh/id_dsa.pub. The key fingerprint is: 11:78:88:1a:47:ff:c3:9f:a1:b1:d2:d9:d9:27:48:59 grid@node1 [grid@node2 ~]$ mkdir .ssh [grid@node2 ~]$ ssh-keygen -t rsa Generating public/private rsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_rsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_rsa. Your public key has been saved in /home/grid/.ssh/id_rsa.pub. The key fingerprint is: 53:d7:d5:e8:63:5a:c6:15:c4:81:11:80:9b:5f:f9:42 grid@node2 [grid@node2 ~]$ ssh-keygen -t dsa Generating public/private dsa key pair. Enter file in which to save the key (/home/grid/.ssh/id_dsa): Enter passphrase (empty for no passphrase): Enter same passphrase again: Your identification has been saved in /home/grid/.ssh/id_dsa. Your public key has been saved in /home/grid/.ssh/id_dsa.pub. The key fingerprint is: 9f:d7:3e:cf:e4:ff:29:df:c5:8c:85:46:6b:e2:07:1b grid@node2
[grid@node1 ~]$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys [grid@node1 ~]$ scp ~/.ssh/authorized_keys node2:~/.ssh/authorized_keys [grid@node1 ~]$ ssh node2 date [grid@node1 ~]$ ssh node1 date [grid@node2 ~]$ ssh node1 date [grid@node2 ~]$ ssh node2 date

 

--在两个节点安装cvuqdisk包并验证--

在两个 Oracle RAC 节点上安装操作系统程序包 cvuqdisk。

如果没有 cvuqdisk,集群验证实用程序就无法发现共享磁盘,当运行(手动运行或在 Oracle Grid Infrastructure 安装结束时自动运行)集群验证实用程序时,会收到这样的错误消息:“Package cvuqdisk not installed”。使用适用于硬件体系结构(例如,x86_64 或 i386)的 cvuqdisk RPM。

cvuqdisk RPM 包含在 Oracle Grid Infrastructure 安装介质上的 rpm 目录中。

设置环境变量 CVUQDISK_GRP,使其指向作为 cvuqdisk 的所有者所在的组(本文为 oinstall):

[root@node1 rpm]# pwd

/oracle-tools/oracle11gR2_linux/grid/rpm

[root@node1 rpm]# ll

总计 12

-rwxr-xr-x 1 root root 8233 2011-09-22 cvuqdisk-1.0.9-1.rpm

[root@node1 rpm]# rpm -ivh *

Preparing...                                                            (100%########################################### [100%]

Using default group oinstall to install package

   1:cvuqdisk                                                           ( 99%########################################### [100####3

[grid@node1 ~]$ export CVUQDISK_GRP=oinstall

 

--安装前检查,最后会因为dns过不去,可以忽略。最后就是图形界面安装,先用grid用户安装grid,再用oracle用户安装db,之后用grid用户设置asm,最后oracle用户dbca建立数据库--

[grid@node1 grid]$ ./runcluvfy.sh stage -pre crsinst -n node1,node2 -fixup -verbose >> /opt/ora11g/check

posted @ 2013-12-06 13:58  DoubleGinger  阅读(1190)  评论(0编辑  收藏  举报