Bug 1156637
Summary: | Gluster small-file creates do not scale with brick count | |||
---|---|---|---|---|
Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Ben England <bengland> | |
Component: | distribute | Assignee: | Shyamsundar <srangana> | |
Status: | CLOSED ERRATA | QA Contact: | Ben Turner <bturner> | |
Severity: | high | Docs Contact: | ||
Priority: | urgent | |||
Version: | rhgs-3.0 | CC: | achauras, annair, ira, jharriga, mpillai, nsathyan, perfbz, srangana, vagarwal | |
Target Milestone: | --- | |||
Target Release: | RHGS 3.1.0 | |||
Hardware: | Unspecified | |||
OS: | Unspecified | |||
Whiteboard: | ||||
Fixed In Version: | glusterfs-3.7.1-1 | Doc Type: | Bug Fix | |
Doc Text: | Story Points: | --- | ||
Clone Of: | ||||
: | 1219637 (view as bug list) | Environment: | ||
Last Closed: | 2015-07-29 04:36:40 UTC | Type: | Bug | |
Regression: | --- | Mount Type: | --- | |
Documentation: | --- | CRM: | ||
Verified Versions: | Category: | --- | ||
oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | ||
Cloudforms Team: | --- | Target Upstream Version: | ||
Embargoed: | ||||
Bug Depends On: | 1220064 | |||
Bug Blocks: | 1202842 |
Description
Ben England
2014-10-24 21:22:49 UTC
sadly there is another piece to the scalability puzzle for SMB/RHS. When a small-file create is done there, we see GETXATTR user.glusterfs.get_real_filename:my-filename.type is sent to all servers. Protocol trace and annotation are at this .tcpdump.gz and .odp respectively: http://perf1.perf.lab.eng.bos.redhat.com/bengland/public/rhs/insignia/ 1-GbE.tcpdump.gz - the tcpdump, viewable with wireshark day-in-the-life-smb-gluster.odp - the annotation I am guessing that this is how we avoid doing READDIRPLUS per file to handle windows case-insensitivity - ask each brick to check for us. There is no way to know which brick the file is on if we are case-insensitive, since consistent hashing does not map the filename to upper/lower case before hashing. Could this be a solution - have a volume option to map filename to all upper-case before hashing? IMHO A lot of users would not care if they were prevented from creating files FOO.BAR and foo.bar in the same directory, even Linux users, and would consider this a small price to pay for being able to use shared storage. Current work is to get the patch posted by Jeff ( https://review.gluster.org/#/c/7702/ ). This is for the problem as posted in the #description. I think we may need to fork comment #2 into its own bug, as that does not fall under the same lookup DHT issue. 3.7 upstream has this feature merged, and so 3.1 RHGS should get it when upstream is pulled downstream for the release. Any backwards compatibility issues here? For example, if you upgrade server first, but not all clients get upgraded, what will happen? Or if you upgrade clients first, will this be ok? @Ben, bug #1221747 captures the backward compat. issues that you detail in comment #10, work in progress to address the same upstream. While rebalance is happening, the stat of files show error and the files are not accessible. They come back to normal when rebalance is finished.Also, I see one file was corrupted. This seems to be an issue. Need to be looked into unless documented. Marking this bug as failed. This BZ is being superceded by 1222053 as the lookup-unhashed option is being depracated. Will execute the same testcases with the new option of lookup-optimize. The feature/fix provided for the bug is under the lookup-optimize switch and not the older lookup-unhashed switch. As a result, request verification of the same (as posted in comment #13). Marking this on QE again. Feature also requires documentation, refer to this commit for the same: http://review.gluster.org/#/c/11109/ Verified on glusterfs-3.7.1-9.el6rhs.x86_64. The numbers are: 3.0.4 - 1480 files / sec 3.1 default - 1600 files / sec 3.1 + cluster.lookup-optimize - 2003 files / sec Just to note average with client and server event threads set to 4 and lookup optimize enabled: 2146 files / second Since the problem described in this bug report should be resolved in a recent advisory, it has been closed with a resolution of ERRATA. For information on the advisory, and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://rhn.redhat.com/errata/RHSA-2015-1495.html |