Good Contents Are Everywhere, But Here, We Deliver The Best of The Best.Please Hold on!
Database Management Services, Oracle Database Management Solution, Oracle Databases, Oracle Exadata



An engineered system comprising of Compute Nodes, Storage Cells and Infiniband  – all of it packaged inside a single physical cabinet called “Exadata Rack


Exadata Hardware Generation At A Glance

 

Exadata Database Machine X8-2




0

Uncategorized
Oracle announced the next-generation Oracle Exadata X8 with significant hardware and software enhancements in overall performance, storage capacity, network bandwidth, and automation. Exadata X7 delivers extreme performance and reliability to run the largest, most business-critical database workloads.


Oracle Exadata X8-2 quick glance:


  • – Latest Intel Xeon (8260) Processors (2.4GHz) 2*24 cores per database server (384 core in a Full Rack)
  •  
  • – Exadata X8 Capacity-on-Demand enables at least 14 cores per server
  •  
  • – Delivers up to 60 percent faster throughput than previous models
  •  
  • – NO changes in Database server Physical memory (upto 12TB in a full Rack)
  •  
  • – Latest Intel Xeon (5218) Processors (2.3GHz) 2*16 cores per Storage server (448 core in a Full Rack)
  •  
  • – NO increase in the capacity of Extreme Flash storage (716.8TB in a full Rack)
  •  
  • – 40% increase in disk capacity (2352TB in a full Rack)
  •  
  • – Exadata X8 introduces new Exadata storage option – Extended (XT) Storage Server
  •  
  • – Each Exadata XT Storage Server includes twelve 14 TB SAS disk drives
  •  
  • – A full rack Exadata X8-2 system has:
    •  
    • – Raw capacity of 2.3 petabytes of disk storage & 358.4TB Flash
    •  
    • – 720 terabytes of NVMe all-Flash storage
    •  
    • – Raw capacity of 2.3 petabytes of disk storage & No flash
  •  
  • – Additional network 2x 10/25 Gb optical Ethernet (client – optional)
  •  
  • – Available in Oracle Public Cloud – Oracle Database Exadata Cloud Service





For more information please visit –

https://www.oracle.com/a/ocom/docs/engineered-systems/exadata/exadata-x8-2-ds.pdf
 

https://www.oracle.com/a/ocom/docs/engineered-systems/exadata/exadata-x8-8-ds.pdf



0

Oracle provides “Exachk” utility to conduct a comprehensive Health Check on Oracle SuperCluster to validate hardware, firmware and configuration. Exachk Utility is available for Oracle Engineered Systems such as Exadata (V2 and above), Exalogic, Exalytics, SuperCluster, MiniCluster, ZDLRA & Big Data. 

When Exachk is run from the primary LDOM as user ‘root’ it will discover and run exachk utility for each component:
  • Configuration checks for Compute nodes, Storage cells and InfiniBand Switches
  • Grid Infrastructure, Database and ASM and Operating System software checks

When Exachk is run in a Database zone or Virtualized environment it will collect data for:
  • All RAC Node
  • All Database Instance
  • Grid Infrastructure

You can also run Exachk on a specific component such as:
  • Database Servers
  • Storage Cells
  • Infiniband Switches
  • Grid Infrastructure, Database & ASM and so on

It is recommended to run Exachk as root user and have SSH equivalence setup in the SuperCluster. But you can run Exachk as ordinary user and without having root ssh setup.

It is recommended to execute the latest exachk at the following situation:
  • Monthly
  • Before any planned maintenance activity
  • Immediately after completion of planned maintenance activity
  • Immediately after an outage or incident

Exachk Binary and output file location:
  • Default Exachk Location: /opt/oracle.SupportTools/exachk
  • Defautl Exachk Output Location: /opt/oracle.SupportTools/exachk


Courtesy Oracle


Steps to Deploy and Execute Exachk utility on SuperCluster


  • Download Latest Exachk Utility
You can download the latest Exachk from MOS note 1070954.1

  • Download deploy_exachk.sh script to deploy and install Exachk in all Primary LDOM and in each Zone

  • Copy the downloaded Exachk utility and deploy_exachk.sh into /opt/oracle.SupportTools
# cd /opt/oracle.SupportTools
# mv exachk Exachk-bkp

  • Deploy Exachk as follows
# cd /opt/oracle.SupportTools/
# ./deploy_exachk.sh exachk.zip
# ls -ltr
# cd exachk
# ls -l exachk

As of writing the latest Exachk available is 18.2.0_20180518

  • Verify Exachk Version on LDOM
# cd /opt/oracle.SupportTools/exachk
# ./exachk -v

  • To verify Exachk version on all zones in a LDOM
# zoneadm list | grep -v global > zone_list
# hostname >> zone_list
# /opt/oracle.supercluster/bin/dcli -g zone_list -l root /opt/oracle.SupportTools/exachk/exachk -v

Note: root RSA keys should be set up for SSH

  • Execute Exachk on Primary LDOM or Global Zone
# cd /opt/oracle.SupportTools/exachk
# ./exachk

  • Execute Exachk in non-global zone local zone
Login to non-global zone local zone using zlogin and execute the following commands

# zlogin <hostname>
# cd /opt/oracle.SupportTools/exachk
# ./exachk

Important Note: In zones there is currently an issue with discovery, and so one must set the RAT_ORACLE_HOME and RAT_GRID_HOME environment variables in some cases.


Conclusion
In this article we have learned to perform Oracle SuperCluster Stack Health Check using Exachk utility. Exachk Utility is available for Oracle Engineered Systems such as Exadata (V2 and above), Exalogi, Exalytics, SuperCluster, MiniCluster, ZDLRA & Big Data.

1


Oracle provides “Exachk” utility to conduct a comprehensive Exadata Health Check on Exadata Database Machine to validate hardware, firmware and configuration.


Exachk Utility is available for Oracle engineered systems such as Exadata (V2 and above), Exalogic, Exalytics, SuperCluster, MiniCluster, ZDLRA & Big Data. Exachk utility performs the following checks:
  • Configuration checks for Compute nodes, Storage cells and InfiniBand Switches
  • Grid Infrastructure, Database and ASM and Operating System software checks
  • MAA Scorecard which conducts an automatic MAA Review
  • Exadata Software Planner, Software prechecks, Exadata and Database Critical Issue alerts


It is recommended to execute the latest exachk at the following situation:
  • Monthly
  • Before any planned maintenance activity
  • Immediately after completion of planned maintenance activity
  • Immediately after an outage or incident


Steps to Exadata Health Exachk Using Exachk Utility

  • Download latest Exachk utility from the MOS note. As of writing the latest Exachk verion available is “12.2.0.1.4_20171212”
Oracle Exadata Database Machine exachk or Health Check (Doc ID 1070954.1)


Note: It is recommended to use latest Exachk to perform Exadata Health Check



  • As root user, create ‘Exachk’ directory on compute node 1 as follows
[root@dm01db01 ~]# cd /root
[root@dm01db01 ~]# mkdir Exachk

  • Using Winscp Copy the Downloaded Exachk utility from your desktop/laptop to the Exadata compute node 1 under /root/Exachk





  • As root user, Login to Exadata Compute node 1 and unzip the Exachk utility
[root@dm01db01 ~]# cd /root/Exachk/


[root@dm01db01 Exachk]# ls -ltr
total 112576
-rw-r–r– 1 root root 115158363 Apr 10 05:11 exachk.zip


[root@dm01db01 Exachk]# unzip exachk.zip





  • Ensure that the SSH is setup across all Compute nodes, Storage cells and Ibswitches
[root@dm01db01 Exachk]# dcli -g ~/all_group -l root ‘uptime’




To Setup SSH across the cluster, use the following command:


[root@dm01db01 ~]# cd /opt/oracle.SupportTools/
[root@dm01db01 oracle.SupportTools]# ./setup_ssh_eq.sh ~/all_group root welcome1

  • As root user, Execute the Exachk utility
[root@dm01db01 ~]# cd /root/Exachk/


[root@dm01db01 Exachk]# ls -ltr


[root@dm01db01 Exachk]# ./exachk





Depending on the Exadata Cluster Size and number of databases it may take several minutes to complete Exachk execution.
  • Using Winscp, copy the Exachk zip file and/or html file to your desktop/laptop to review
  • Open the html file, review it and take necessary action if necessary
  • Under the table of contents there are different component listed. Look out for the CRITICAL and FAIL components.
Click on the ‘view’ hyperlink for more details and the recommendation to fix the problem.




MAA Scorecard




Conclusion

In this article we have learned to perform complete Exadata Stack Health Check using Exachk utility. Exachk Utility is available for Oracle engineered systems such as Exadata (V2 and above), Exalogi, Exalytics, SuperCluster, MiniCluster, ZDLRA & Big Data.
4


Introduction

By default Oracle ACS configure root file system with 30GB space on Exadata Computed nodes X2 and above. In most of the cases this space is sufficient to store operating system, Exadata software, log file and diagnostic files. Over time if you store patches, software and log files are not purged this space will be filled faster. Exadata X2 and above uses volume group and it is easy to extend the logical volume space on which the root file system is mounted.

Root file system is created on two system partitions LVDb Sys1 and LVDb Sys2 and both system partitions must be size equally at the same time. Only one system partition is active at any time and other is inactive.

In this article, I will demonstrate how you can extend root file system size on Exadata Compute nodes online without any downtime.

Environemt

Exadata X5-2 Half Rack
Exadata storage software version 12.1.2.3.4

Current Root File System Allocation

[root@exa01db01 ~]# df -h /
Filesystem            Size  Used Avail Use% Mounted on
/dev/mapper/VGExaDb-LVDbSys1
                       30G   25G  3.5G  88% /
 


List Logical Volume and It’s Details

lvm> lvs

lvm> lvs -o lv_name,lv_path,vg_name,lv_size
  LV        Path                   VG      LSize
  LVDbOra1  /dev/VGExaDb/LVDbOra1  VGExaDb 200.00g
  LVDbSwap1 /dev/VGExaDb/LVDbSwap1 VGExaDb  24.00g
  LVDbSys1  /dev/VGExaDb/LVDbSys1  VGExaDb  30.00g
  LVDbSys2  /dev/VGExaDb/LVDbSys2  VGExaDb  30.00g
  perflv    /dev/VGExaDb/perflv    VGExaDb   5.00g



Get the Current Active System Partition

[root@exa01db01 ~]# imageinfo

Kernel version: 2.6.39-400.294.1.el6uek.x86_64 #1 SMP Wed Jan 11 08:46:38 PST 2017 x86_64
Image kernel version: 2.6.39-400.294.1.el6uek
Image version: 12.1.2.3.4.170111
Image activated: 2017-04-08 12:14:23 -0500
Image status: success
System partition on device: /dev/mapper/VGExaDb-LVDbSys1




Steps to Increase Root File System on Compute Nodes:

  • Get the Current Root File System Utilization

[root@exa01db01 ~]# df -h /
Filesystem            Size  Used Avail Use% Mounted on
/dev/mapper/VGExaDb-LVDbSys1
                       30G   25G  3.5G  88% /


  • Get Current Logical Volume Configuration

lvm> lvs -o lv_name,lv_path,vg_name,lv_size
  LV        Path                   VG      LSize
  LVDbOra1  /dev/VGExaDb/LVDbOra1  VGExaDb 200.00g
  LVDbSwap1 /dev/VGExaDb/LVDbSwap1 VGExaDb  24.00g
  LVDbSys1  /dev/VGExaDb/LVDbSys1  VGExaDb  30.00g
  LVDbSys2  /dev/VGExaDb/LVDbSys2  VGExaDb  30.00g
  perflv    /dev/VGExaDb/perflv    VGExaDb   5.00g


  • Ensure Root File System Can be Resized Online

[root@exa01db01 ~]# tune2fs -l /dev/mapper/VGExaDb-LVDbSys1 | grep resize_inode
Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isize

All Nodes:



[root@exa01db01 ~]# dcli -g dbs_group -l root ‘tune2fs -l /dev/mapper/VGExaDb-LVDbSys1 | grep resize_inode’exa01db01: Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isizeexa01db02: Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isizeexa01db03: Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isizeexa01db04: Filesystem features:      has_journal ext_attr resize_inode dir_index filetype needs_recovery extent flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isize


  • Get the free space available in the Volume Group

lvm> vgdisplay -s
  “VGExaDb” 1.63 TiB  [295.00 GiB used / 1.34 TiB free]


  • Extend both logical volumes using lvextend command. Here we are extending the root file system by 50GB, so the file system become 80GB in total.

[root@exa01db01 ~]# lvextend -L +50G /dev/VGExaDb/LVDbSys1
  Size of logical volume VGExaDb/LVDbSys1 changed from 30.00 GiB (7680 extents) to 80.00 GiB (20480 extents).
  Logical volume LVDbSys1 successfully resized

[root@exa01db01 ~]# lvextend -L +50G /dev/VGExaDb/LVDbSys2
  Size of logical volume VGExaDb/LVDbSys2 changed from 30.00 GiB (7680 extents) to 80.00 GiB (20480 extents).
  Logical volume LVDbSys2 successfully resized


  • Now resize the file system using resize2fs command.

[root@exa01db01 ~]# resize2fs /dev/VGExaDb/LVDbSys1
resize2fs 1.43-WIP (20-Jun-2013)
Filesystem at /dev/VGExaDb/LVDbSys1 is mounted on /; on-line resizing required
old_desc_blocks = 2, new_desc_blocks = 4
Performing an on-line resize of /dev/VGExaDb/LVDbSys1 to 15728640 (4k) blocks.
The filesystem on /dev/VGExaDb/LVDbSys1 is now 15728640 blocks long.

[root@exa01db02 ~]# e2fsck -f /dev/VGExaDb/LVDbSys1
e2fsck 1.43-WIP (20-Jun-2013)
/dev/VGExaDb/LVDbSys1 is mounted.
e2fsck: Cannot continue, aborting.




The resize command ro LVDbSys2 is failed as it is inactive. So we must execute the fsck first before resizing.

[root@exa01db01 ~]# resize2fs /dev/VGExaDb/LVDbSys2
resize2fs 1.43-WIP (20-Jun-2013)
Please run ‘e2fsck -f /dev/VGExaDb/LVDbSys2’ first.

[root@exa01db01 ~]# e2fsck -f /dev/VGExaDb/LVDbSys2
e2fsck 1.43-WIP (20-Jun-2013)
Pass 1: Checking inodes, blocks, and sizes
Pass 2: Checking directory structure
Pass 3: Checking directory connectivity
Pass 4: Checking reference counts
Pass 5: Checking group summary information
/dev/VGExaDb/LVDbSys2: 122199/3932160 files (0.3% non-contiguous), 5496667/7864320 blocks

Now execute the resize file system again

[root@exa01db01 ~]# resize2fs /dev/VGExaDb/LVDbSys2
resize2fs 1.43-WIP (20-Jun-2013)
Resizing the filesystem on /dev/VGExaDb/LVDbSys2 to 15728640 (4k) blocks.
The filesystem on /dev/VGExaDb/LVDbSys2 is now 15728640 blocks long.


  • Validate the root file system

[root@exa01db01 ~]# df -h /
Filesystem            Size  Used Avail Use% Mounted on
/dev/mapper/VGExaDb-LVDbSys1
                       80G   25G   55G  31% /
 


Conclusion

In this article we have demonstrated how to resize root file system on Exadata Compute node online without any outage. It is important to note that root file system is create on two system partitions for high availability. 
0