1 / 43

Oracle RAC 11gR2

Oracle RAC 11gR2. RAC Joins the Cloud. RAC Grows Up. RAC 11gR2 (aka 12gR1) adds a number of features that transforms RAC clusters into database clouds Many features are geared towards enterprises but there are new features that any organization can take advantage of

rivka
Download Presentation

Oracle RAC 11gR2

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Oracle RAC 11gR2 RAC Joins the Cloud

  2. RAC Grows Up • RAC 11gR2 (aka 12gR1) adds a number of features that transforms RAC clusters into database clouds • Many features are geared towards enterprises but there are new features that any organization can take advantage of • With today’s high power x86 and 11gR2 RAC it is entirely possible to replace many silo clusters with fewer 11gR2 “clouds” • Many 11gR1 and older RAC commands are depreciated

  3. 11gR2 RAC New Features • Brand New Installer • Clusterware is now Grid Infrastructure • ASM is now part of the Grid Infrastructure installation • Voting and OCR disk are ASM diskgroup • Raw devices are only supported for upgraded clusters • Future Linux release may drop support for raw devices • Plug and Play For Clusters • Cluster is now a DNS service • DHCP for server VIPs • New tools are introduced old tools retired

  4. RAC 11gR2 Install

  5. Non Linear Installer

  6. Grid Plug and Play

  7. SSH Keys

  8. OCR/Voting Disks

  9. OCR/Voting Disks • OCR/Voting disks can be ASM Diskgroup • Raw devices are only supported for upgraded clusters • Other options are NFS and certified clustered filesystems • 3 disks for normal redundancy, 5 disks for high

  10. IPMI

  11. IPMI (cont) • Intelligent Platform Management Interface • Defines a cross vendor standard to monitor and manage hardware • Dell, HP, IBM, Intel, AMD and dozens more • IPMI in relation to RAC allows other nodes to manage the server regardless of the OS state

  12. ASM Roles

  13. Fix Common Problems • Installer can create a script that can fix some problems • Won’t install missing packages

  14. Finish Install • After GUI finished is will ask to run the command oraInrootinst.sh and root.sh

  15. Cluster Startup Sequence • Init starts Oracle High Availability Service Deamon (OHASD) • OHASD starts • cssdagent – starts cssd • orarootagent – manages all root owned ohasd processes • oraagent – manages all oracle owned ohsad resources • cssdmonintor – monitors cssd and node health • OHASD rootagent starts • crsd – main deamon for managing cluster resources • ctssd – time synchronization • Diskmon • ACFS – ASM cluster file system drivers

  16. Cluster Startup Sequence • Oraagent starts – level 2 • mdnsd – used for DNS lookups • gipcd – inter-process and inter-node communication • gpnpd – Grid Plug & Play deamon • EVMD – event monitor • ASM – resource for monitoring ASM instance • CRSD – level 3 • orarootagent • oraagent • CRSD rootagent – level 4 • SCAN VIP(s) - Single Client Access Name • Node VIPs – one per node • ACFS Registry – For mounting ASM Cluster File System • GNS VIP - optional

  17. SCAN • SIMPLE CLIENT ACCESS NAME • A single DNS entry that represents the cluster has a whole • Clients no longer need to know the actual servers that are in a cluster • THIN connections or connections not using a VIP are better protected against a node failure • If using GNS SCAN must be in the domain managed by GNS • DNS needs to be able to delegate resolution to GNS • If not using GNS DNS should be configured to round robin up to 3 IPs for the SCAN name. Recommend use 11.2 client • If using GNS 3 IPs will be requested from DHCP

  18. SCAN Listener • Listener that is dependent on a SCAN VIP • If the SCAN VIP moves to a different node in the cluster the listener will move with the VIP • Depending on the number of nodes in a cluster a single node may have more than one SCAN VIP and listener • remote_listener parameter should be set to a SCAN VIP for 11gR2 databases • Simplify RAC Dataguard connections for clients • Two new sqlnet parameters • CONNECT_TIME = timeout in seconds • RETRY_COUNT = number of tries • In a single TNS entry enter both primary and standby servers SCAN address. If the connection is made to standby it will failover to primary.

  19. SCAN Load Balancing • Client receives IP from DNS or GNS that resolves to SCAN VIP • SCAN Listener bound to SCAN VIP redirects client to local listener that has the least load • Client connects to the local listener on the node VIP • Client needs to be able to connect to SCAN VIP and to the local node VIP and be able to resolve all names

  20. SCAN • SCAN Listeners Overview

  21. Grid - Plug And Play • Greatly simplifies the adding/removing and connecting to a cluster • Basically run cluster verify script and then addNode.sh • Configures cluster listener and ASM • No longer use static IPs for VIPs

  22. GNS • Grid Naming Service • Handles name resolution for the cluster • Requires DHCP for VIP addresses! • DNS needs to delegate resolution for the cluster to GNS • Use the GNS VIP not server or scan VIP • Bind Example • prodcluster.db.oracle.com NS gns.prodcluster.db.oracle.com • gns.prodcluster.db.oracle.com. 10.90.51.20 • When a client connects the cluster will provide the IP address • As RAC servers are added DHCP assigns a VIP address, GNS is updated with the new server’s VIP address and now can route connects to the new server. • No client connection strings are never changed • Remove a node is basically reversing the process

  23. Client Connections With GNS

  24. ASM Changes • ASM part of grid infrastructure – one less HOME to manage • New GUI management tool – asmca • Replaces DBCA for ASM management • Create diskgroups, volumes, ACFS mounts • Enhanced controls to manage ASM attributes • SPFILE is stored in ASM diskgroup • Cluster reads the diskgroup before ASM starts • ASMCMD is greatly enhanced and mostly replaces sqlplus • ASM roles can play a large role in managing ASM • Diskgroups can be cluster resources

  25. ASM Changes (cont) • New O/S aware cluster file system (ACFS) • ASM disks can be used to create filesystem mounted by O/S • Only supported on Redhat/Oracle Linux, other O/S support coming in the future • If a file can natively support ASM then the file is not supported on ACFS! • No data, control or redo files • DirectIO is not available • Take advantage of ASM performance over other filesystems • ACFS is able to do snapshots • Still able to access ASM diskgroups with FTP/HTTP via XMLDB • Copy command can copy out and in of ASM (11.1.0.7)

  26. ASM Changes (cont) • Intelligent placement of data • ASM is able to put hot datafiles on the outer edge of the disk • Can provide a 50% increase in I/O performance • Compatible parameters for both ASM & RDBMS 11.2 • Work best with the geometry of the disks are know • JOBD – uses disks sectors for data placment • LUN – logical sectoring number which may not have any relation to the physical layout of the lun on the disk • Diskgroup should be 50% full for full benefit • alter diskgroup data add template datafile_hot attributes (hot mirrorhot); • ascmdlstmpl –l –G data

  27. ASM – Access Control Lists • Set permissions at the ASM file level • Set permissions for users and groups • User is the oracle database software owner • Only available for Linux and Unix • Not so much for security as for separation of duties • Files created in ASM are owned by DBUSER • Create a separate OSDBA group for each database using a separate ORACLE_HOME. Need different groups for OSASM and OSDBA for ASM • Compatible 11.2 diskgroup both for ASM and RDBMS

  28. ASM – Access Control Lists (cont) • Disk attributes • ACCESS_CONTROL.ENABLED = TRUE • ACCESS_CONTROL.MASK has to be set • Mask values • 6 removes all • 2 removes write • 0 removes nothing • Mask of 026 sets values of 640 • read-write for owner, read for group and nothing for everyone else • asmcmd> setattr –d data access_control_enabled true • sql> alter diskgroup data set attribute ‘access_control.enabled’ = true;

  29. ASM Permissions • ALTER DISKGROUP ADD USERGROUP … WITH MEMBER • ALTER DISKGROUP DROP USERGROUP • ALTER DISKGROUP MODIFY USERGROUP ADD MEMBER • ALTER DISKGROUP MODIFY USERGROUP DROP MEMBER • ALTER DISKGROUP ADD USER • ALTER DISKGROUP DROP USER • ALTER DISKGROUP SET PERMISSIONALTER DISKGROUP SET OWNERSHIP • SELECT * FROM V$ASM_USER • SELECT * FROM V$ASM_USERGROUP • ALTER DISKGROUP ADD USERGROUP myfiles WITH MEMBER bill;

  30. ASM Roles • OS groups control rights in ASM • OSDBA – can access ASM files and set ACLs • ASMOPER – can start/stop ASM • OSASM – full ASM control • Allow people to administer ASM without having sysdba rights to the databases

  31. ASM Commandline Tool

  32. ASM Commandline Tool (cont) • Create diskgroups from ascmd • Same syntax as sqlplus • Create diskgroupmydata external redundancy • DISK ‘ORCL:DISK1’ NAME mydata1, ‘ORCL:DISK2’ NAME mydata2; • Create diskgroup from XML via asmcmd • asmcmd> chkdg

  33. ASM Commandline Tool (cont) • lsdg – list diskgroup information • Very similar to select * from v$asm_diskgroup • lsdsk – list disk information • Same information found in select * from v$asm_disk

  34. Cluster Registry & Voting Disks • Raw devices only supported for upgraded clusters • Can move OCR and Voting disks to ASM diskgroup after upgrade • ocrconfig add <diskgroup name> • ocrcheck • ocrconfig delete <raw device> • crsctl query cssvotedisk • crsctl add cssvotedisk <diskgroup name> • crsctl delete cssvotedisk <raw device>

  35. Cluster Management • crsctl in the grid infrastructure home manages almost every cluster command • crs_* scripts are depreicated

  36. Depreciated Commands • crs_stat • crs_register • crs_unregister • crs_start • crs_stop • crs_getperm • crs_profile • crs_relocate • crs_setperm • crsctl check crsd • crsctl check cssd • crsctl check evmd • crsctl debug log • crsctl set cssvotedisk • crsctl start resources • crsctl stop resources

  37. Server Pools • Create “mini” clusters inside of larger clusters for policy based databases • Two default pools Free/Generic • Free pool contains servers not assigned to another pool • Generic pool is for running pre 11.2 databases and non policy managed databases • Specify min/max nodes and priority • Cluster manages the members of the pool based on load, availability, priority and min/max requirements • Cluster will move servers from free and lower priority pools to meet the needs of the server pool • Can create ACL on server pools for role based management

  38. Vote/OCR Disk Management • Raw is supported for upgraded clusters but it is possible to migrate to ASM disks after 11.2 cluster upgrade • Move vote disks to ASM • crsctl query cssvotedisk • crsctl replace votedisk +asm_disk_group • Move OCR to ASM • ocrconfig -add +new_disk_group • ocrconfig -delete old_storage_location

  39. Troubleshooting Cluster Problems • Golden Rule – Ensure time is synced across cluster • Help with comparing log files • Ensures nodes aren’t evicted due to time skew • If NTP is not used Oracle will use Cluste Time Synchronization service • Recommend to have cluster down if changing the time • Logs are stored under GRID_HOME/log/hostname • Main alert log • Separate directory for each cluster process • Diagnostic Script • GRID_HOME/bin/diagcollection.pl • Collects cluster and system logs for debugging

  40. Troubleshooting (cont) • Modify time before node is evicted • Default is 30 seconds • crsctl set css miscount 45 • On busy systems logs may not flush before node reboots • Crsctl set cssdiagwait 13 – force • Not set by default • Requires cluster outage to set

  41. Adding A Node • Much easier in 11gR2 • Setup ssh keys for new host • Run cluster verify tool • cluvfy stage –pre crsint –n <new host> <-fixup> • Fixes any problems • cluvfy stage –pre nodeadd –n <new host> • $ORACLE_HOME/oui/bin/addNode.sh –silent “CLUSTER_NEW_NODES=<new host>” “CLUSTER_NEW_VIRTUAL_HOSTNAMES=<new host vip>”

  42. Instance Caging • Puts a cap on the number of CPUs a instance will use • Set the cpu_count parameter • Enable resource manager with CPU polices

  43. Oracle Joins The Cloud • With 11gR2 it is now possible to create few larger clusters while being able to maintain performance, security and reassure customers concerns • Increased utilization rates • Lower administration costs • Happy Management

More Related