& c – Fields: D F
For changing to the different views type:
4& '# CPU Core cycles used by a VM. High values are an indicator CPU load average for the last one, five and
m Memory i Interrupts v Disk VM 4 - - # Counter showing how long a VM has to wait for swapped
for VMs causing performance problems on ESXi Hosts. 15 minutes
c CPU d Disk Adapter p Power states pages read from disk. A reason for this could be memory
n Network u Disk Device overcommitment. Pay attention if %SWPWT is >5!
≥5
f for add/remove fields
V show only virtual machine instances
2 highlight a row scrolling down
8 highlight a row scrolling up
4!,! '# Counter showing percentage of time a ready to run vCPU
spacebar: refresh screen was not scheduled because of a CPU limit setting. Remove the limit for
s 2: refresh screen every two seconds better performance.
≥1
+ n – Fields: A B C D E F K L
4 5 # Percentage of time spent by system to process interrupts 4 # This value is interesting if you are using vSMP virtual
and to perform other system activities on behalf of the world. machines. It shows the percentage of time a ready to run VM has spent
4'. / 4'. . # Dropped Packages transmitted/Dropped Packages received. Possible cause: maybe caused by high I/O VM in coFdeschedule state.
Values larger 0 are a sign for high network utilization >20 ≥3
≥1 If value is >3 decrease the number of vCPUs from the VM concerned.
40!-$) # percentage of time a VM was waiting for some VMkernel activity to complete (such as I/O) before it
can continue. Includes %SWPWT and "blocked", but not IDLE Time (as %WAIT does). 4.'5# Percentage of time a VM was waiting to be scheduled.If you note values between
100 five and ten percent take care.
Possible cause: Storage performance issue | latency to a device in the VM configuration eg. USB device, serial >10
& 6 "* 6 +) # provide information what
passFthrough device or parallel passFthrough device Possible reasons: too many vCPUs, too many vSMP VMs or a CPU limit setting (check
%MLMTD)
physical NIC a VM is actually using.
! " m – Fields: B D J K Q ' d – Fields: A B G J
! " # '$01# Latency at the device driver level $2. * #
! , (# Amount of guest physical memory (MB) the ESXi Host is
reclaiming by balloon driver. A reason for this is memory high enough free memory available If the storage system has not responded within 60 seconds VMs with an
soft < 4% free memory: Host reclaim memory by balloon driver Indicator for storage performance
overcommitment. troubles >25 Windows Operating System will issue an abort. ≥1
≥1 hard < 2% free memory: Host starts to swap, you will see performance troubles
low < 1% free memory: ESX stop the VMs to allocate more RAM
average memory overcommitment for the last
one, five and 15 minutes
() * # Values larger 0 indicate that the host is
actively compressing memory.
&+() * # Values larger 0 indicate that the 3$01# Latency caused by VMKernel ≥3 1$01# GAVG = DAVG + KAVG >25
host is accessing compressed memory. Possible cause: Queuing (wrong queue depth
parameter or wrong failover policy) ≥1
. * # number of commands reset per second
Reason for this behaviour is memory
overcommitment.
≥1
+&!$ m (change to memory view) – Fields: D G
- &.# Memory (in MB) that has been swapped by VMKernel. $ % & '# Memory (in MB) compressed by ESXi Host
Possible cause: memory overcommitment. ≥1
≥1 +!+# Numa Node where the VM is located +4,# Percentage of VM Memory located at the local NUMA Node. If this
value is less than 80 Percent the VM will experience performance issues. <80
-.* / --* # Rate at which the ESXi Host is writing to or reading from swapped memory. Possible
cause: memory overcommitment. ≥1
6 " +.! !# VM Memory (in MB)
+,! !# VM Memory (in MB) located at local Node
located at remote Node
7 889 Copyright © 2012F2013 runningFsystem.com | Designed By: Andi Lesslhumer | Version 1.1