rac onenodeのrelocation、convert
35987 ワード
本論文ではrac onenodeのrelocation,convert(rac onenodeからracへの双方向変換)をテストする.
参照ドキュメント:
https://docs.oracle.com/cd/E11882_01/rac.112/e41960/onenode.htm#RACAD7898
rac onenodeのrelocation
切り替える前に、データベースとサービスはノード1で実行されます.
relocationデータベースからノード2まで、どのノードでもrelocationが可能で、root権限の下で実行する必要があります.
ノード1とノード2のalert logをそれぞれ表示
クラスタの状況を見ると、dbやサービスなどのリソースがノード2上に表示されます.
relocationテストが完了しました.
rac onenodeをracに変換
-n onenode 1、ノード1でdbが実行されていないため、エラーが発生しました.-n onenode 2が必要です.
リソースなどのステータスを表示します.dbリソースはノード2にありますか.ノード1上のdbリソースはありません.
なぜなら、instanceを追加する必要があるからです.
新しく追加したinstanceノードでdbを起動
リソースステータスを表示し、racに変換し、両方のノードのdbが開きます.注意、srvctlでdbを起動し、sqlplusでdbを起動するには、表示されるresourceに問題があります(テストの時、sqlplusでdbを起動し、起動後、リソースで見ることができます.ただしoff状態です)
srvctlコマンドを使用してdbとサービスの構成を表示
サービスの再構成(略)
racをraconeに変換する.他のインスタンスを削除して、instanceが1つしかないことを確認する必要があります.
残りのノード
サービスの削除(追加サービスは削除されていないため省略)
変換、onenodeへの変換
end
参照ドキュメント:
https://docs.oracle.com/cd/E11882_01/rac.112/e41960/onenode.htm#RACAD7898
rac onenodeのrelocation
切り替える前に、データベースとサービスはノード1で実行されます.
[grid@onenode1 ~]$ crsctl status resource -t
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER STATE_DETAILS
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.DATA.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.LISTENER.lsnr
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.OCR.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.asm
ONLINE ONLINE onenode1 Started
ONLINE ONLINE onenode2 Started
ora.gsd
OFFLINE OFFLINE onenode1
OFFLINE OFFLINE onenode2
ora.net1.network
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.ons
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.registry.acfs
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE onenode1
ora.cvu
1 ONLINE ONLINE onenode1
ora.oc4j
1 ONLINE ONLINE onenode1
ora.onenode.db
1 ONLINE ONLINE onenode1 Open
ora.onenode.one.svc
1 ONLINE ONLINE onenode1
ora.onenode1.vip
1 ONLINE ONLINE onenode1
ora.onenode2.vip
1 ONLINE ONLINE onenode2
ora.scan1.vip
1 ONLINE ONLINE onenode1
[grid@onenode1 ~]$
[grid@onenode1 ~]$ srvctl config service -d onenode
Service name: one
Service is enabled
Server pool: onenode
Cardinality: 1
Disconnect: false
Service role: PRIMARY
Management policy: AUTOMATIC
DTP transaction: false
AQ HA notifications: false
Failover type: NONE
Failover method: NONE
TAF failover retries: 0
TAF failover delay: 0
Connection Load Balancing Goal: LONG
Runtime Load Balancing Goal: NONE
TAF policy specification: BASIC
Edition:
Preferred instances: onenode_1
Available instances:
[grid@onenode1 ~]$
relocationデータベースからノード2まで、どのノードでもrelocationが可能で、root権限の下で実行する必要があります.
[grid@onenode2 ~]$ srvctl relocate database -d onenode -n onenode2
PRCD-1222 : Online relocation of database "onenode" failed but database was restored to its original state
PRCR-1037 : Failed to update cardinality of the resource ora.onenode.db to 2
PRCR-1071 : Failed to register or update resource ora.onenode.db
CRS-0245: User doesn't have enough privilege to perform the operation
[grid@onenode2 ~]$ su root
Password:
[root@onenode2 grid]# srvctl relocate database -d onenode -n onenode2
[root@onenode2 grid]#
ノード1とノード2のalert logをそれぞれ表示
[oracle@onenode2 trace]$ more alert_onenode_2.log
Mon Mar 19 16:22:09 2018
Starting ORACLE instance (normal)
************************ Large Pages Information *******************
Per process system memlock (soft) limit = UNLIMITED
Total Shared Global Region in Large Pages = 0 KB (0%)
Large Pages used by this instance: 0 (0 KB)
Large Pages unused system wide = 0 (0 KB)
Large Pages configured system wide = 0 (0 KB)
Large Page size = 2048 KB
RECOMMENDATION:
Total System Global Area size is 906 MB. For optimal performance,
prior to the next instance restart:
1. Increase the number of unused large pages by
at least 453 (page size 2048 KB, total size 906 MB) system wide to
get 100% of the System Global Area allocated with large pages
********************************************************************
LICENSE_MAX_SESSION = 0
LICENSE_SESSIONS_WARNING = 0
Initial number of CPU is 1
Number of processor cores in the system is 1
Number of processor sockets in the system is 1
Private Interface 'eth2:1' configured from GPnP for use as a private interconnect.
[name='eth2:1', type=1, ip=169.254.230.126, mac=08-00-27-9a-1e-0c, net=169.254.0.0/16, mask=255.255.0.0, use=haip:cluster_interconnect/62]
Public Interface 'eth1' configured from GPnP for use as a public interface.
[name='eth1', type=1, ip=192.168.2.62, mac=08-00-27-79-b4-56, net=192.168.2.0/24, mask=255.255.255.0, use=public/1]
Public Interface 'eth1:1' configured from GPnP for use as a public interface.
[name='eth1:1', type=1, ip=192.168.2.72, mac=08-00-27-79-b4-56, net=192.168.2.0/24, mask=255.255.255.0, use=public/1]
Shared memory segment for instance monitoring created
CELL communication is configured to use 0 interface(s):
CELL IP affinity details:
NUMA status: non-NUMA system
cellaffinity.ora status: N/A
CELL communication will use 1 IP group(s):
Grp 0:
Picked latch-free SCN scheme 3
Using LOG_ARCHIVE_DEST_1 parameter default value as USE_DB_RECOVERY_FILE_DEST
WARNING: db_recovery_file_dest is same as db_create_file_dest
Autotune of undo retention is turned on.
LICENSE_MAX_USERS = 0
SYS auditing is disabled
Starting up:
Oracle Database 11g Enterprise Edition Release 11.2.0.4.0 - 64bit Production
With the Partitioning, Real Application Clusters, OLAP, Data Mining
and Real Application Testing options.
ORACLE_HOME = /u01/app/oracle/product/11.2.0/db_1
System name: Linux
Node name: onenode2
Release: 2.6.32-696.el6.x86_64
Version: #1 SMP Tue Feb 21 00:53:17 EST 2017
Machine: x86_64
Using parameter settings in server-side pfile /u01/app/oracle/product/11.2.0/db_1/dbs/initonenode_2.ora
System parameters with non-default values:
processes = 150
spfile = "+DATA/onenode/spfileonenode.ora"
sga_target = 904M
control_files = "+DATA/onenode/controlfile/current.261.971110795"
control_files = "+DATA/onenode/controlfile/current.260.971110795"
db_block_size = 8192
compatible = "11.2.0.4.0"
log_archive_format = "%t_%s_%r.dbf"
cluster_database = TRUE
db_create_file_dest = "+DATA"
db_recovery_file_dest = "+DATA"
db_recovery_file_dest_size= 20G
remote_login_passwordfile= "EXCLUSIVE"
db_domain = ""
dispatchers = "(PROTOCOL=TCP) (SERVICE=onenodeXDB)"
remote_listener = "onenode-scan:1521"
audit_file_dest = "/u01/app/oracle/admin/onenode/adump"
audit_trail = "DB"
db_name = "onenode"
open_cursors = 300
pga_aggregate_target = 301M
diagnostic_dest = "/u01/app/oracle"
Cluster communication is configured to use the following interface(s) for this instance
169.254.230.126
cluster interconnect IPC version:Oracle UDP/IP (generic)
IPC Vendor 1 proto 2
Mon Mar 19 16:22:20 2018
PMON started with pid=2, OS id=5987
Mon Mar 19 16:22:20 2018
PSP0 started with pid=3, OS id=5989
Mon Mar 19 16:22:21 2018
VKTM started with pid=4, OS id=5991 at elevated priority
VKTM running at (1)millisec precision with DBRM quantum (100)ms
Mon Mar 19 16:22:21 2018
GEN0 started with pid=5, OS id=5995
Mon Mar 19 16:22:21 2018
DIAG started with pid=6, OS id=5997
Mon Mar 19 16:22:21 2018
DBRM started with pid=7, OS id=5999
Mon Mar 19 16:22:21 2018
PING started with pid=8, OS id=6001
Mon Mar 19 16:22:21 2018
ACMS started with pid=9, OS id=6003
Mon Mar 19 16:22:21 2018
DIA0 started with pid=10, OS id=6005
Mon Mar 19 16:22:21 2018
LMON started with pid=11, OS id=6007
Mon Mar 19 16:22:21 2018
LMD0 started with pid=12, OS id=6009
* Load Monitor used for high load check
* New Low - High Load Threshold Range = [960 - 1280]
Mon Mar 19 16:22:21 2018
LMS0 started with pid=13, OS id=6011 at elevated priority
Mon Mar 19 16:22:21 2018
RMS0 started with pid=14, OS id=6015
Mon Mar 19 16:22:21 2018
LMHB started with pid=15, OS id=6017
Mon Mar 19 16:22:21 2018
MMAN started with pid=16, OS id=6019
Mon Mar 19 16:22:21 2018
DBW0 started with pid=17, OS id=6021
Mon Mar 19 16:22:21 2018
LGWR started with pid=18, OS id=6023
Mon Mar 19 16:22:21 2018
CKPT started with pid=19, OS id=6025
Mon Mar 19 16:22:22 2018
SMON started with pid=20, OS id=6027
Mon Mar 19 16:22:22 2018
RECO started with pid=21, OS id=6029
Mon Mar 19 16:22:22 2018
RBAL started with pid=22, OS id=6031
Mon Mar 19 16:22:22 2018
ASMB started with pid=23, OS id=6033
Mon Mar 19 16:22:22 2018
MMON started with pid=24, OS id=6035
NOTE: initiating MARK startup
Mon Mar 19 16:22:22 2018
starting up 1 dispatcher(s) for network address '(ADDRESS=(PARTIAL=YES)(PROTOCOL=TCP))'...
Starting background process MARK
Mon Mar 19 16:22:22 2018
MMNL started with pid=25, OS id=6039
Mon Mar 19 16:22:22 2018
MARK started with pid=26, OS id=6041
NOTE: MARK has subscribed
starting up 1 shared server(s) ...
lmon registered with NM - instance number 2 (internal mem no 1)
Reconfiguration started (old inc 0, new inc 4)
List of instances:
1 2 (myinst: 2)
Global Resource Directory frozen
* allocate domain 0, invalid = TRUE
Communication channels reestablished
* domain 0 valid = 1 according to instance 1
Master broadcasted resource hash value bitmaps
Non-local Process blocks cleaned out
LMS 0: 0 GCS shadows cancelled, 0 closed, 0 Xw survived
Set master node info
Submitted all remote-enqueue requests
Dwn-cvts replayed, VALBLKs dubious
All grantable enqueues granted
Submitted all GCS remote-cache requests
Fix write in gcs resources
Reconfiguration complete
Mon Mar 19 16:22:24 2018
LCK0 started with pid=30, OS id=6054
Starting background process RSMN
Mon Mar 19 16:22:24 2018
RSMN started with pid=31, OS id=6056
ORACLE_BASE not set in environment. It is recommended
that ORACLE_BASE be set in the environment
Mon Mar 19 16:22:25 2018
ALTER SYSTEM SET local_listener=' (ADDRESS=(PROTOCOL=TCP)(HOST=192.168.2.72)(PORT=1521))' SCOPE=MEMORY SID='onenode_2';
ALTER DATABASE MOUNT /* db agent *//* {2:37189:197} */
NOTE: Loaded library: System
SUCCESS: diskgroup DATA was mounted
NOTE: dependency between database onenode and diskgroup resource ora.DATA.dg is established
Successful mount of redo thread 2, with mount id 2692565945
Database mounted in Shared Mode (CLUSTER_DATABASE=TRUE)
Lost write protection disabled
Create Relation IPS_PACKAGE_UNPACK_HISTORY
Completed: ALTER DATABASE MOUNT /* db agent *//* {2:37189:197} */
ALTER DATABASE OPEN /* db agent *//* {2:37189:197} */
Picked broadcast on commit scheme to generate SCNs
ARCH: STARTING ARCH PROCESSES
Mon Mar 19 16:22:35 2018
ARC0 started with pid=33, OS id=6087
Mon Mar 19 16:22:36 2018
ARC0: Archival started
ARCH: STARTING ARCH PROCESSES COMPLETE
ARC0: STARTING ARCH PROCESSES
Mon Mar 19 16:22:36 2018
ARC1 started with pid=35, OS id=6089
Mon Mar 19 16:22:36 2018
ARC2 started with pid=36, OS id=6091
ARC1: Archival started
ARC2: Archival started
ARC1: Becoming the 'no FAL' ARCH
ARC1: Becoming the 'no SRL' ARCH
ARC2: Becoming the heartbeat ARCH
Mon Mar 19 16:22:36 2018
ARC3 started with pid=37, OS id=6093
Mon Mar 19 16:22:36 2018
Thread 2 opened at log sequence 2
Current log# 4 seq# 2 mem# 0: +DATA/onenode/onlinelog/group_4.273.971111363
Current log# 4 seq# 2 mem# 1: +DATA/onenode/onlinelog/group_4.274.971111371
Successful open of redo thread 2
MTTR advisory is disabled because FAST_START_MTTR_TARGET is not set
Mon Mar 19 16:22:36 2018
SMON: enabling cache recovery
ARC3: Archival started
ARC0: STARTING ARCH PROCESSES COMPLETE
[6060] Successfully onlined Undo Tablespace 5.
Undo initialization finished serial:0 start:172224 end:173164 diff:940 (9 seconds)
Verifying file header compatibility for 11g tablespace encryption..
Verifying 11g file header compatibility for tablespace encryption completed
SMON: enabling tx recovery
Database Characterset is WE8MSWIN1252
Mon Mar 19 16:22:40 2018
minact-scn: Inst 2 is a slave inc#:4 mmon proc-id:6035 status:0x2
minact-scn status: grec-scn:0x0000.00000000 gmin-scn:0x0000.00000000 gcalc-scn:0x0000.00000000
No Resource Manager plan active
Starting background process GTX0
Mon Mar 19 16:22:44 2018
GTX0 started with pid=39, OS id=6101
Starting background process RCBG
Mon Mar 19 16:22:44 2018
RCBG started with pid=40, OS id=6113
replication_dependency_tracking turned off (no async multimaster replication found)
Starting background process QMNC
Mon Mar 19 16:22:45 2018
QMNC started with pid=41, OS id=6124
Mon Mar 19 16:22:53 2018
Completed: ALTER DATABASE OPEN /* db agent *//* {2:37189:197} */
Mon Mar 19 16:22:58 2018
ALTER SYSTEM SET service_names='one' SCOPE=MEMORY SID='onenode_2';
Mon Mar 19 16:22:59 2018
Starting background process CJQ0
Mon Mar 19 16:22:59 2018
CJQ0 started with pid=47, OS id=6290
Mon Mar 19 16:23:02 2018
minact-scn: Inst 2 is now the master inc#:4 mmon proc-id:6035 status:0x7
minact-scn status: grec-scn:0x0000.00000000 gmin-scn:0x0000.00000000 gcalc-scn:0x0000.00000000
minact-scn: Master returning as live inst:1 has inc# mismatch instinc:0 cur:4 errcnt:0
Mon Mar 19 16:23:20 2018
Reconfiguration started (old inc 4, new inc 6)
List of instances:
2 (myinst: 2)
Global Resource Directory frozen
* dead instance detected - domain 0 invalid = TRUE
Communication channels reestablished
Master broadcasted resource hash value bitmaps
Non-local Process blocks cleaned out
Mon Mar 19 16:23:20 2018
LMS 0: 0 GCS shadows cancelled, 0 closed, 0 Xw survived
Set master node info
Submitted all remote-enqueue requests
Dwn-cvts replayed, VALBLKs dubious
All grantable enqueues granted
Post SMON to start 1st pass IR
Mon Mar 19 16:23:20 2018
Instance recovery: looking for dead threads
Submitted all GCS remote-cache requests
Post SMON to start 1st pass IR
Fix write in gcs resources
Reconfiguration complete
Instance recovery: lock domain invalid but no dead threads
Mon Mar 19 16:23:23 2018
Restarting dead background process DIA0
Mon Mar 19 16:23:23 2018
DIA0 started with pid=10, OS id=6336
Mon Mar 19 16:24:20 2018
Decreasing number of real time LMS from 1 to 0
[oracle@onenode2 trace]$
Mon Mar 19 16:22:23 2018
Reconfiguration started (old inc 2, new inc 4)
List of instances:
1 2 (myinst: 1)
Global Resource Directory frozen
Communication channels reestablished
Master broadcasted resource hash value bitmaps
Non-local Process blocks cleaned out
Mon Mar 19 16:22:23 2018
LMS 0: 0 GCS shadows cancelled, 0 closed, 0 Xw survived
Set master node info
Submitted all remote-enqueue requests
Dwn-cvts replayed, VALBLKs dubious
All grantable enqueues granted
Mon Mar 19 16:22:23 2018
minact-scn: master found reconf/inst-rec after rec-scn scan old-inc#:2 new-inc#:4
Submitted all GCS remote-cache requests
Fix write in gcs resources
Reconfiguration complete
minact-scn: Master returning as live inst:2 has inc# mismatch instinc:0 cur:4 errcnt:0
Mon Mar 19 16:22:57 2018
ALTER SYSTEM SET service_names='onenode' SCOPE=MEMORY SID='onenode_1';
Shutting down instance (transactional local)
Stopping background process SMCO
Shutting down instance: further logons disabled
Stopping background process QMNC
Mon Mar 19 16:23:00 2018
Stopping background process CJQ0
Stopping background process MMNL
Stopping background process MMON
Local transactions complete. Performing immediate shutdown
License high water mark = 9
All dispatchers and shared servers shutdown
ALTER SYSTEM SET _shutdown_completion_timeout_mins=30 SCOPE=MEMORY;
ALTER DATABASE CLOSE NORMAL /* db agent *//* {2:37189:246} */
Mon Mar 19 16:23:05 2018
SMON: disabling tx recovery
Stopping background process RCBG
SMON: disabling cache recovery
Mon Mar 19 16:23:07 2018
NOTE: Deferred communication with ASM instance
NOTE: deferred map free for map id 20
Mon Mar 19 16:23:07 2018
NOTE: Deferred communication with ASM instance
Redo thread 1 internally disabled at seq 1 (LGWR)
Shutting down archive processes
Archiving is disabled
Mon Mar 19 16:23:08 2018
ARCH shutting down
ARC3: Archival stopped
Mon Mar 19 16:23:08 2018
ARCH shutting down
ARC1: Archival stopped
Mon Mar 19 16:23:08 2018
ARCH shutting down
ARC0: Archival stopped
Mon Mar 19 16:23:09 2018
ARC2: Archiving disabled thread 1 sequence 1
Thread 1 closed at log sequence 1
Successful close of redo thread 1
Mon Mar 19 16:23:13 2018
NOTE: Deferred communication with ASM instance
NOTE: deferred map free for map id 5
Mon Mar 19 16:23:13 2018
Completed: ALTER DATABASE CLOSE NORMAL /* db agent *//* {2:37189:246} */
ALTER DATABASE DISMOUNT /* db agent *//* {2:37189:246} */
Shutting down archive processes
Archiving is disabled
Archived Log entry 2 added for thread 1 sequence 1 ID 0xa07c1e47 dest 1:
ARC2: Archival disabled due to shutdown: 1089
Shutting down archive processes
Archiving is disabled
Archive process shutdown avoided: 0 active
ARCH shutting down
ARC2: Archival stopped
Mon Mar 19 16:23:17 2018
NOTE: Deferred communication with ASM instance
NOTE: deferred map free for map id 3
Completed: ALTER DATABASE DISMOUNT /* db agent *//* {2:37189:246} */
ARCH: Archival disabled due to shutdown: 1089
Shutting down archive processes
Archiving is disabled
Mon Mar 19 16:23:18 2018
NOTE: force a map free for map id 3
Mon Mar 19 16:23:18 2018
Stopping background process VKTM
ARCH: Archival disabled due to shutdown: 1089
Shutting down archive processes
Archiving is disabled
Mon Mar 19 16:23:18 2018
NOTE: force a map free for map id 35
NOTE: force a map free for map id 34
Mon Mar 19 16:23:18 2018
NOTE: Shutting down MARK background process
Mon Mar 19 16:23:19 2018
freeing rdom 0
Mon Mar 19 16:23:25 2018
Instance shutdown complete
[oracle@onenode1 trace]$
クラスタの状況を見ると、dbやサービスなどのリソースがノード2上に表示されます.
[root@onenode2 grid]# crsctl status resource -t
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER STATE_DETAILS
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.DATA.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.LISTENER.lsnr
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.OCR.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.asm
ONLINE ONLINE onenode1 Started
ONLINE ONLINE onenode2 Started
ora.gsd
OFFLINE OFFLINE onenode1
OFFLINE OFFLINE onenode2
ora.net1.network
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.ons
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.registry.acfs
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE onenode1
ora.cvu
1 ONLINE ONLINE onenode1
ora.oc4j
1 ONLINE ONLINE onenode1
ora.onenode.db
2 ONLINE ONLINE onenode2 Open
ora.onenode.one.svc
1 ONLINE ONLINE onenode2
ora.onenode1.vip
1 ONLINE ONLINE onenode1
ora.onenode2.vip
1 ONLINE ONLINE onenode2
ora.scan1.vip
1 ONLINE ONLINE onenode1
[root@onenode2 grid]#
[root@onenode2 grid]# srvctl config service -d onenode
Service name: one
Service is enabled
Server pool: onenode
Cardinality: 1
Disconnect: false
Service role: PRIMARY
Management policy: AUTOMATIC
DTP transaction: false
AQ HA notifications: false
Failover type: NONE
Failover method: NONE
TAF failover retries: 0
TAF failover delay: 0
Connection Load Balancing Goal: LONG
Runtime Load Balancing Goal: NONE
TAF policy specification: BASIC
Edition:
Preferred instances: onenode_2
Available instances:
[root@onenode2 grid]#
relocationテストが完了しました.
rac onenodeをracに変換
-n onenode 1、ノード1でdbが実行されていないため、エラーが発生しました.-n onenode 2が必要です.
[root@onenode1 ~]# su - oracle
[oracle@onenode1 ~]$ srvctl convert database -d onenode -c RAC -n onenode1
PRCD-1148 : Failed to convert the configuration of RAC One Node database onenode into its equivalent configuration for RAC database
PRCD-1151 : Failed to convert the configuration of RAC One Node database onenode into its equivalent RAC database configuration because RAC One Node database is running on a different server onenode2 than the given target server onenode1
[oracle@onenode1 ~]$
[oracle@onenode1 ~]$ srvctl convert database -d onenode -c RAC -n onenode2
[oracle@onenode1 ~]$
リソースなどのステータスを表示します.dbリソースはノード2にありますか.ノード1上のdbリソースはありません.
[grid@onenode1 ~]$ crsctl status resource -t
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER STATE_DETAILS
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.DATA.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.LISTENER.lsnr
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.OCR.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.asm
ONLINE ONLINE onenode1 Started
ONLINE ONLINE onenode2 Started
ora.gsd
OFFLINE OFFLINE onenode1
OFFLINE OFFLINE onenode2
ora.net1.network
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.ons
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.registry.acfs
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE onenode1
ora.cvu
1 ONLINE ONLINE onenode1
ora.oc4j
1 ONLINE ONLINE onenode1
ora.onenode.db
2 ONLINE ONLINE onenode2 Open
ora.onenode.one.svc
1 ONLINE ONLINE onenode2
ora.onenode1.vip
1 ONLINE ONLINE onenode1
ora.onenode2.vip
1 ONLINE ONLINE onenode2
ora.scan1.vip
1 ONLINE ONLINE onenode1
[grid@onenode1 ~]$
なぜなら、instanceを追加する必要があるからです.
srvctl add instance -d onenode -i onenode_1 -n onenode1
[oracle@onenode1 ~]$ srvctl add instance -d onenode -i onenode_1 -n onenode1
[oracle@onenode1 ~]$
srvctl config database -d onenode
[oracle@onenode1 ~]$ srvctl config database -d onenode
Database unique name: onenode
Database name: onenode
Oracle home: /u01/app/oracle/product/11.2.0/db_1
Oracle user: oracle
Spfile: +DATA/onenode/spfileonenode.ora
Domain:
Start options: open
Stop options: immediate
Database role: PRIMARY
Management policy: AUTOMATIC
Server pools: onenode
Database instances: onenode_1,onenode_2
Disk Groups: DATA
Mount point paths:
Services: one
Type: RAC
Database is administrator managed
[oracle@onenode1 ~]$
新しく追加したinstanceノードでdbを起動
srvctl start database -d onenode
[oracle@onenode1 ~]$ srvctl start database -d onenode
[oracle@onenode1 ~]$
リソースステータスを表示し、racに変換し、両方のノードのdbが開きます.注意、srvctlでdbを起動し、sqlplusでdbを起動するには、表示されるresourceに問題があります(テストの時、sqlplusでdbを起動し、起動後、リソースで見ることができます.ただしoff状態です)
[grid@onenode1 ~]$ crsctl status resource -t
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER STATE_DETAILS
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.DATA.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.LISTENER.lsnr
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.OCR.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.asm
ONLINE ONLINE onenode1 Started
ONLINE ONLINE onenode2 Started
ora.gsd
OFFLINE OFFLINE onenode1
OFFLINE OFFLINE onenode2
ora.net1.network
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.ons
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.registry.acfs
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE onenode1
ora.cvu
1 ONLINE ONLINE onenode1
ora.oc4j
1 ONLINE ONLINE onenode1
ora.onenode.db
1 ONLINE ONLINE onenode1 Open
2 ONLINE ONLINE onenode2 Open
ora.onenode.one.svc
1 ONLINE ONLINE onenode2
ora.onenode1.vip
1 ONLINE ONLINE onenode1
ora.onenode2.vip
1 ONLINE ONLINE onenode2
ora.scan1.vip
1 ONLINE ONLINE onenode1
[grid@onenode1 ~]$
srvctlコマンドを使用してdbとサービスの構成を表示
[grid@onenode1 ~]$ srvctl config service -d onenode
Service name: one
Service is enabled
Server pool: onenode_one
Cardinality: 1
Disconnect: false
Service role: PRIMARY
Management policy: AUTOMATIC
DTP transaction: false
AQ HA notifications: false
Failover type: NONE
Failover method: NONE
TAF failover retries: 0
TAF failover delay: 0
Connection Load Balancing Goal: LONG
Runtime Load Balancing Goal: NONE
TAF policy specification: BASIC
Edition:
Preferred instances: onenode_2
Available instances:
[grid@onenode1 ~]$ srvctl config database -d onenode
Database unique name: onenode
Database name: onenode
Oracle home: /u01/app/oracle/product/11.2.0/db_1
Oracle user: oracle
Spfile: +DATA/onenode/spfileonenode.ora
Domain:
Start options: open
Stop options: immediate
Database role: PRIMARY
Management policy: AUTOMATIC
Server pools: onenode
Database instances: onenode_1,onenode_2
Disk Groups: DATA
Mount point paths:
Services: one
Type: RAC
Database is administrator managed
[grid@onenode1 ~]$
サービスの再構成(略)
srvctl remove service -d onenode -s one
srvctl add service -d onenode -s onenodesrv -r RACDB1,RACDB2 -P BASIC
racをraconeに変換する.他のインスタンスを削除して、instanceが1つしかないことを確認する必要があります.
[oracle@onenode1 ~]$ srvctl remove instance -d onenode -i onenode_1
Remove instance from the database onenode? (y/[n]) y
PRCD-1052 : Failed to remove instance from database onenode
PRCD-1101 : Failed to remove running instance onenode_1 for database onenode
[oracle@onenode1 ~]$
[oracle@onenode1 ~]$ srvctl remove instance -d onenode -i onenode_1
Remove instance from the database onenode? (y/[n]) y
[oracle@onenode1 ~]$
残りのノード
[grid@onenode1 ~]$ srvctl config database -d onenode
Database unique name: onenode
Database name: onenode
Oracle home: /u01/app/oracle/product/11.2.0/db_1
Oracle user: oracle
Spfile: +DATA/onenode/spfileonenode.ora
Domain:
Start options: open
Stop options: immediate
Database role: PRIMARY
Management policy: AUTOMATIC
Server pools: onenode
Database instances: onenode_2
Disk Groups: DATA
Mount point paths:
Services: one
Type: RAC
Database is administrator managed
[grid@onenode1 ~]$
サービスの削除(追加サービスは削除されていないため省略)
srvctl remove service -d onenode -s onenodesrv
srvctl add service -d onenode -s one -P BASIC
変換、onenodeへの変換
srvctl convert database -d onenode -c RACONENODE -w 30 -i onenode
[oracle@onenode2 ~]$ srvctl convert database -d onenode -c RACONENODE -w 30 -i onenode
[oracle@onenode2 ~]$
[oracle@onenode2 ~]$ srvctl config database -d onenode
Database unique name: onenode
Database name: onenode
Oracle home: /u01/app/oracle/product/11.2.0/db_1
Oracle user: oracle
Spfile: +DATA/onenode/spfileonenode.ora
Domain:
Start options: open
Stop options: immediate
Database role: PRIMARY
Management policy: AUTOMATIC
Server pools: onenode
Database instances:
Disk Groups: DATA
Mount point paths:
Services: one
Type: RACOneNode
Online relocation timeout: 30
Instance name prefix: onenode
Candidate servers: onenode2
Database is administrator managed
[oracle@onenode2 ~]$
[grid@onenode2 ~]$ crsctl status resource -t
--------------------------------------------------------------------------------
NAME TARGET STATE SERVER STATE_DETAILS
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.DATA.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.LISTENER.lsnr
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.OCR.dg
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.asm
ONLINE ONLINE onenode1 Started
ONLINE ONLINE onenode2 Started
ora.gsd
OFFLINE OFFLINE onenode1
OFFLINE OFFLINE onenode2
ora.net1.network
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.ons
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
ora.registry.acfs
ONLINE ONLINE onenode1
ONLINE ONLINE onenode2
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE onenode1
ora.cvu
1 ONLINE ONLINE onenode1
ora.oc4j
1 ONLINE ONLINE onenode1
ora.onenode.db
2 ONLINE ONLINE onenode2 Open
ora.onenode.one.svc
1 ONLINE ONLINE onenode2
ora.onenode1.vip
1 ONLINE ONLINE onenode1
ora.onenode2.vip
1 ONLINE ONLINE onenode2
ora.scan1.vip
1 ONLINE ONLINE onenode1
[grid@onenode2 ~]$
end