![]() | Oracle System Handbook - ISO 7.0 May 2018 Internal/Partner Edition | ||
|
|
![]() |
||||||||||||||||
Solution Type Technical Instruction Sure Solution 1472716.1 : Sun ZFS Storage Appliance: How to set up and use Analytics for benchmarking applications
This document describes which ZFS Storage Appliance Analytics to enable in performance-related benchmarks or proofs of concept (POC) and how to interpret the results of those benchmarks. This document also includes references to the client-side accounting tools that should also be used to monitor the client system during benchmarking or performance testing. This document contains details relevant to the Oracle University Systems Tech Talk, "Direct NFS and the ZFS Storage Appliance; An Integrated Architecture for High Performance Databases." The Oracle University eSeminar may be accessed from this link: http://ilearning.oracle.com/ilearn/en/learner/jsp/rco_details.jsp?classid=1244206820&rcoid=1198683411. In this Document
Applies to:Sun ZFS Backup Appliance - Version All Versions to All Versions [Release All Releases]Sun ZFS Storage 7420 - Version All Versions to All Versions [Release All Releases] Sun ZFS Storage 7320 - Version All Versions to All Versions [Release All Releases] Sun ZFS Storage 7120 - Version All Versions to All Versions [Release All Releases] 7000 Appliance OS (Fishworks) GoalThis document provides the reader with the minimum level of instrumentation to execute performance test and benchmarks with the ZFS Storage Appliance or ZFS Backup Appliance. SolutionThe content in this document was presented at an IOUG webinar. The webinar is accessible from this link: http://www.ioug.org/p/cm/ld/fid=153. A similar presentation was executed as a System Tech Talk, which was published at http://ilearning.oracle.com/ilearn/en/learner/jsp/rco_details.jsp?classid=1244206820&rcoid=1198683411. The content delivered in these presentations add significant detail to the text presented in this MOS note. Detailed performance monitoring of the ZFS Storage Appliance for general use cases can be accomplished by first enabling Advanced Analytics (Configuration->Preferences, check the Enable Advanced Analytics box) and then enabling the following Analytics:
Details for interpreting each accounting statistics are available from the ZFS Storage Appliance and ZFS Backup Appliance browser interface (BUI) help menu: Help->Analytis->Statistics. The client operating system accessing the ZFS Storage Appliance or ZFS Backup Appliance instrumentation should also be enabled and recorded during performance evaluations or benchmarking exercises. For Solaris begin with the following instrumentation (please note the 5 second timing is a good first suggestion, but specific circumstances may dictate a different value):
For the Linux operating system begin with the following instrumentation:
For the Windows operating system begin the Windows PerfMon tool and enable the following instrumentation:
In the specific use case of Oracle Database Access to the ZFS Storage Appliance or ZFS Backup Appliance the Automatic Workload Repository (AWR) report should be enabled and used during any benchmarking or performance testing effort. AWR snapshots should be triggered at the beginning and end of the test workload and the AWR report should be generated from these snapshots. Review the ZFS Storage Appliance Analytics and and operating system accounting statistics in the context of the workload shown in the AWR load profile and tablespace I/O statistics and the storage related wait events. Pay specific attention to the following details:
The Oracle Database documenation specific to your release, available at docs.oracle.com, contains descriptions of all of the wait events shown in the AWR report. Details regarding setup and usage of AWR are show in Document id 1363422. In the case of Oracle 11g, you can find the descriptions at this link: http://docs.oracle.com/cd/B28359_01/server.111/b28320/waitevents003.htm#BGGIBDJI. By comparing workload and response times reported by the test application, the database (if used), the operating system, and the storage system bottlenecks can be quickly and accurately identified. In practical systems, bottlenecks can be created by software bugs, misconfigured software, hardware faults, and misconfigured hardware. By carefully instrumenting the system and analyzing the results performance problems can be classified and solved. In general, system throughput may be limited by several factors
Hardware saturation is an important goal of any benchmarking activity. Network, CPU, and storage media all have well-defined limits as to how much throughput they can process, and hitting these limits usually indicates a well-tuned system. If the hardware is not saturated, then the problem limiting throughput may be related to software processing on the hardware. Troubleshooting this type of issue usually requires drawing a software block diagram showing the software components executed during the test, instrumenting those components, and identifying which software component is holding up the work, and finally reconfiguring or fixing that software component. This exercise is most easily undertaken by starting with the software/hardware block diagram in the system. Supply side bottlenecks show up when we double the workload at the application, throughput does not change, application response time increases, and storage response time stays the same. If there were no supply side bottleneck, doubling the number of I/O the application queues to the storage should double the time it takes for the storage to respond to the I/O. If the storage response time does not increase, then something between the application and the storage bottlenecked the I/O. This is a common problem when benchmarking systems that requires expert-level attention to the details of the system. Hardware faults, while rare, are a possibility in any real system. Always confirm there are no errors in the operating system log, the network driver status, and there are no faults reported in the storage system's status pages. Software bugs that are not present in lightly loaded conditions can show up under extreme load conditions. This includes race conditions that cause I/O to hang and processing problems that cause I/O to appear to stall. In cases where there are no obvious hardware faults, saturation, or bottlenecks pay careful attention to application code. In the context of the ZFSSA, if you hit a case where doubling the application workload does not cause a doubling of the front-end protocol response time and the operating system reports the same response time as the ZFSSA then you may have an application-level software bug.
Please note: Collecting 'excessive' amounts of Analytics data can cause performance, kernel memory and/or zpool capacity issues.
Please configure the Analytics retention settings to retain sufficient data for review over a reasonable timeframe. See Document ID 1461959.1 - How to configure Analytics for dataset retention policies
References<NOTE:1567691.1> - How to use ZFS Storage Appliance Analytics to diagnose application, client OS, network, and storage tuning problems.<NOTE:1363422.1> - Automatic Workload Repository (AWR) Reports - Main Information Sources <NOTE:1509652.1> - ZFS Storage Appliance Sizing: Frequently Asked Questions and Answers Attachments This solution has no attachment |
||||||||||||||||
|