| Summary: | [Perf] : Large file random writes regressed on FUSE mounts by 11-24% on RHGS 3.1.3 | ||||||
|---|---|---|---|---|---|---|---|
| Product: | Red Hat Gluster Storage | Reporter: | Ambarish <asoman> | ||||
| Component: | write-behind | Assignee: | Csaba Henk <csaba> | ||||
| Status: | CLOSED WONTFIX | QA Contact: | Rahul Hinduja <rhinduja> | ||||
| Severity: | high | Docs Contact: | |||||
| Priority: | high | ||||||
| Version: | rhgs-3.1 | CC: | amukherj, chrisw, csaba, ndevos, nlevinki, pkarampu, pprakash, ravishankar, rcyriac, rgowdapp, rhinduja, rhs-bugs, sasundar, sheggodu, smohan | ||||
| Target Milestone: | --- | Keywords: | Performance, Regression, ZStream | ||||
| Target Release: | --- | ||||||
| Hardware: | x86_64 | ||||||
| OS: | Linux | ||||||
| Whiteboard: | |||||||
| Fixed In Version: | Doc Type: | Bug Fix | |||||
| Doc Text: | Story Points: | --- | |||||
| Clone Of: | Environment: | ||||||
| Last Closed: | 2018-04-16 18:17:40 UTC | Type: | Bug | ||||
| Regression: | --- | Mount Type: | --- | ||||
| Documentation: | --- | CRM: | |||||
| Verified Versions: | Category: | --- | |||||
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |||||
| Cloudforms Team: | --- | Target Upstream Version: | |||||
| Attachments: |
|
||||||
Ugggh! I meant Version number = glusterfs-3.7.9-1.el6rhs.x86_64 Hitting the reported issue with 3.7.9-2 build as well. Reproduced on 3.7.9-2.The issue is intermittent ,though |
Created attachment 1144313 [details] Console logs showing Iozone's Random R/W throughput Description of problem: I see a slight regression on large file random writes with FUSE mounted volumes. This is from one of the automated runs : With 3.1.2 (baseline): mean rand write throughput = 389985.975000 KB/s With 3.1.3 : mean rand write throughput= 293772.400000 KB/s Regression : -24.67 percent Version-Release number of selected component (if applicable): glusterfs-3.7.5-19.el6rhs.x86_64 How reproducible: 2/2 Steps to Reproduce: 1. Run iozone random R/W test(I=2) on fuse mounts with 3.1.2 thrice 2. Run same test thrice after upgrading to RHGS 3.1.3 3. The mean throughputs should not vary by more than 10% Actual results: 24% off target random write performance Expected results: Regression Threshold is 10%. Additional info: OS : RHEL 6.7 Iozone was used in a distributed multithreaded manner with a 2G file size ,record size of 64K and a total of 16 threads. Setup consisted of 4 servers,4 clients (1X mount per server) on 10GbE network. Volume Settings : [root@gqas001 ~]# gluster v info Volume Name: testvol Type: Distributed-Replicate Volume ID: 2a668beb-7f26-48f9-8550-157108fe1a55 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: gqas001.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick0 Brick2: gqas014.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick1 Brick3: gqas015.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick2 Brick4: gqas016.sbu.lab.eng.bos.redhat.com:/bricks/testvol_brick3 Options Reconfigured: performance.readdir-ahead: on performance.stat-prefetch: off server.allow-insecure: on [root@gqas001 ~]# [root@gqas001 ~]# Console logs attached for two tests(machines were reimaged in between)