<div dir="ltr">So I have a fairly basic setup using glusterfs between 2 nodes. The nodes have 10 gig connections and the bricks reside on SSD LVM LUNs:<div><br></div><div><div>Brick1: media1-be:/gluster/brick1/gluster_volume_0</div><div>Brick2: media2-be:/gluster/brick1/gluster_volume_0</div></div><div><br></div><div><br></div><div>On this volume I have a LIO iscsi target with 1 fileio backstore that's being shared out to vmware ESXi hosts. The volume is around 900 gig and the fileio store is around 850g:</div><div><br></div><div>-rw-r--r-- 1 root root 912680550400 Oct  5 20:47 iscsi.disk.3<br></div><div><br></div><div>I set the WWN to be the same so the ESXi hosts see the nodes as 2 paths to the same target. I believe this is what I want. The issues I'm seeing is that while the IO wait is low I'm seeing high CPU usage with only 3 VMs running on only 1 of the ESX servers:</div><div><br></div><div>this is media2-be:</div><div><div> PID USER    PR  NI   VIRT   RES   SHR S  %CPU %MEM   TIME+ COMMAND                         </div><div> 1474 root    20  0 1396620  37912  5980 S 135.0  0.1 157:01.84 glusterfsd                        </div><div> 1469 root    20  0  747996  13724  5424 S  2.0  0.0  1:10.59 glusterfs                        </div></div><div><br></div><div>And this morning it seemed like I had to restart the LIO service on media1-be as the VMware was seeing time-out issues. I'm seeing issues like this on the VMware ESX servers:</div><div><br></div><div><div>2016-10-06T00:51:41.100Z cpu0:32785)WARNING: ScsiDeviceIO: 1223: Device naa.600140501ce79002e724ebdb66a6756d performance has deteriorated. I/O latency increased from average value of 33420 microseconds to 732696 microseconds.</div></div><div><br></div><div>Are there any special settings I need to have gluster+LIO+vmware to work? Has anyone gotten this to work fairly well that it is stable? What am I missing?</div><div><br></div><div>thanks,<br>Mike</div><div><br></div><div><br></div></div>