Bug 1394654

Summary: [Perf] : Large file reads are off target by 35 percent on Ganesha v3/v4 mounts
Product: [Red Hat Storage] Red Hat Gluster Storage Reporter: Ambarish <asoman>
Component: nfs-ganeshaAssignee: Kaleb KEITHLEY <kkeithle>
Status: CLOSED DUPLICATE QA Contact: Ambarish <asoman>
Severity: high Docs Contact:
Priority: unspecified    
Version: rhgs-3.2CC: amukherj, asoman, bturner, dang, ffilz, jthottan, kkeithle, mbenjamin, rhinduja, rhs-bugs, skoduri, storage-qa-internal
Target Milestone: ---Keywords: Performance
Target Release: RHGS 3.2.0   
Hardware: x86_64   
OS: Linux   
Whiteboard:
Fixed In Version: Doc Type: If docs needed, set a value
Doc Text:
Story Points: ---
Clone Of: Environment:
Last Closed: 2017-02-09 14:04:36 UTC Type: Bug
Regression: --- Mount Type: ---
Documentation: --- CRM:
Verified Versions: Category: ---
oVirt Team: --- RHEL 7.3 requirements from Atomic Host:
Cloudforms Team: --- Target Upstream Version:
Embargoed:

Description Ambarish 2016-11-14 07:19:49 UTC
Description of problem:


Version-Release number of selected component (if applicable):


How reproducible:


Steps to Reproduce:
1.
2.
3.

Actual results:


Expected results:


Additional info:

Comment 3 Ambarish 2016-11-14 07:32:36 UTC
*The bug  got submitted to the system before I could write anything,writing description here,Apologies* :

Description of problem:
-----------------------

A regression was introduced in the latest Ganesha bits in large file sequential reads.

Gluster 3.8.4-2,Ganesha 2.4.0-2 : 2156113.75 kB/sec

Gluster 3.8.4-4,Ganesha 2.4.1-1 : 1421350.29 kB/sec

Regression : 35% 



Version-Release number of selected component (if applicable):
------------------------------------------------------------


nfs-ganesha-gluster-2.4.1-1.el7rhgs.x86_64
glusterfs-ganesha-3.8.4-4.el7rhgs.x86_64


How reproducible:
----------------

100%

Steps to Reproduce:
-------------------

1. Create 2*2 volume.Mount via NFS Ganesha vers=4 and vers=3. 

2. Run Iozone Seq read workload(iozone -+m <conf file>  -C -w -c -e -i 1 -+n -r 64k -s 8g -t 16)

3. Check for regression with older builds.

Actual results:
---------------

35% regression on large file reads on latest builds.

Expected results:
-----------------

Regression within +-10% is acceptable between releases/dev builds.

Additional info:
----------------

OS : RHEL 7.3

*Vol Config*  :

Volume Name: testvol
Type: Distributed-Replicate
Volume ID: e84889ee-7bed-426f-b187-2b15fb244175
Status: Started
Snapshot Count: 0
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: gqas013.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick0
Brick2: gqas005.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick1
Brick3: gqas006.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick2
Brick4: gqas011.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick3
Options Reconfigured:
ganesha.enable: on
features.cache-invalidation: on
server.allow-insecure: on
performance.stat-prefetch: off
transport.address-family: inet
performance.readdir-ahead: on
nfs.disable: on
nfs-ganesha: enable
cluster.enable-shared-storage: enable

Comment 4 Ambarish 2016-11-14 07:35:37 UTC
I tried my perf regression tests on latest gluster build on old Ganesha(2.4.0-2),and I could see sequential reads upto 2000 MBps.

Looks like the regression was introduced with the Ganesha rebase.

Comment 5 Soumya Koduri 2016-11-14 10:49:06 UTC
@Frank,

Do you have any clue on what patch (which went into 2.4.1) could have caused this regression when compared to 2.4.0 build.

Comment 6 Daniel Gryniewicz 2016-11-14 15:44:57 UTC
Nothing in the history jumps out at me.  Bisecting the releases might help narrow it down.

Comment 12 Kaleb KEITHLEY 2016-11-22 13:52:27 UTC
when possible please collect wireshark dumps of gnfs and nfs-ganesha during the the v3 large file writes.

Thanks.

Comment 13 Kaleb KEITHLEY 2016-11-22 13:53:52 UTC
cancel needinfo, wrong bug