Description of problem: This is likely a job tracker memory leak. Filing this problem with the glusterfs-hadoop plug-in, for tracking and documentation purposes. Hadoop benchmark WordCount (100GB input) fails with Job initialization failed: java.lang.OutOfMemoryError: Java heap space at java.util.Arrays.copyOfRange(Arrays.java:3209) at java.lang.String.(String.java:215) at java.lang.StringBuffer.toString(StringBuffer.java:585) at java.net.URI.toString(URI.java:1907) at java.net.URI.(URI.java:732) at org.apache.hadoop.fs.Path.(Path.java:65) at org.apache.hadoop.fs.Path.(Path.java:50) at org.apache.hadoop.mapreduce.JobSubmissionFiles.getJobSplitFile(JobSubmissionFiles.java:51) at org.apache.hadoop.mapreduce.split.SplitMetaInfoReader.readSplitMetaInfo(SplitMetaInfoReader.java:75) at org.apache.hadoop.mapred.JobInProgress.createSplits(JobInProgress.java:835) at org.apache.hadoop.mapred.JobInProgress.initTasks(JobInProgress.java:725) at org.apache.hadoop.mapred.JobTracker.initJob(JobTracker.java:3890) at org.apache.hadoop.mapred.EagerTaskInitializationListener$InitJob.run(EagerTaskInitializationListener.java:79) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) after running successfully 16 times in a row WordCount with 100GB text input file fails due to apparent Jobtracker memory leak. Version-Release number of selected component (if applicable): Hadoop Version: hadoop-1.2.0.1.3.2.0-110 (part of Hortonworks 1.3) RHS 2.0.5 glusterfs-hadoop 2.1 How reproducible: takes 10+ hours to reproduce and NOT 100% reproducible Steps to Reproduce: 1. 2. 3. Actual results: Expected results: Additional info:
I vote to close this: ITs probably not a gluster related bug, but rather ,related to the way the JT stores job metadata. Is this possibly a duplicate of an existing hadoop JIRA, maybe: https://issues.apache.org/jira/browse/HADOOP-4018 or https://issues.apache.org/jira/browse/HADOOP-3670 ?
Per 11/13 bug triage meeting, re-assigning to bchilds and mark low.
gluster-hadoop project is not being actively maintained, and the team is not planning to contribute further on to it.