Activity

I ran this on my box and it looks like it's super slow due to MAPREDUCE-2392. It timed out after 15 minutes but spent a lot of its time waiting on TTs to shut down. I'm re-running now with Tom's patch from that JIRA applied for comparison.

Todd Lipcon
added a comment - 17/Mar/11 18:18 I ran this on my box and it looks like it's super slow due to MAPREDUCE-2392 . It timed out after 15 minutes but spent a lot of its time waiting on TTs to shut down. I'm re-running now with Tom's patch from that JIRA applied for comparison.

Todd Lipcon
added a comment - 17/Mar/11 18:29 With Tom's patch, the test runtime was 10 minutes on my box (with SSD).
We should consider speeding this up or splitting it into multiple test classes, just in case build machines run a bit slower.

Ramkumar Vadali
added a comment - 17/Mar/11 19:54 Breaks TestBlockFixer into several tests. The file TestBlockFixer.java now has tests that do not use a MiniMRCluster. The other TestBlockFixer*.java files have a few tests each that use MiniMRCluster.

org.apache.hadoop.fs.ChecksumException: Checksum error: /blk_-5688804776476378478:of:/user/dhruba/raidtest/file2 at 3584 exp: 924592111 got: -781589875
at org.apache.hadoop.fs.FSInputChecker.verifySums(FSInputChecker.java:320)
at org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:276)
at org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:225)
at org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:193)
at org.apache.hadoop.hdfs.BlockReader.read(BlockReader.java:117)
at org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:477)
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:528)
at java.io.DataInputStream.read(DataInputStream.java:83)
at org.apache.hadoop.hdfs.TestRaidDfs.validateFile(TestRaidDfs.java:402)
at org.apache.hadoop.raid.TestBlockFixerDistConcurrency.testConcurrentJobs(TestBlockFixerDistConcurrency.java:161)

Todd Lipcon
added a comment - 21/Mar/11 20:44 Hi Ramkumar. It looks like one of the tests is failing:
org.apache.hadoop.fs.ChecksumException: Checksum error: /blk_-5688804776476378478:of:/user/dhruba/raidtest/file2 at 3584 exp: 924592111 got: -781589875
at org.apache.hadoop.fs.FSInputChecker.verifySums(FSInputChecker.java:320)
at org.apache.hadoop.fs.FSInputChecker.readChecksumChunk(FSInputChecker.java:276)
at org.apache.hadoop.fs.FSInputChecker.read1(FSInputChecker.java:225)
at org.apache.hadoop.fs.FSInputChecker.read(FSInputChecker.java:193)
at org.apache.hadoop.hdfs.BlockReader.read(BlockReader.java:117)
at org.apache.hadoop.hdfs.DFSInputStream.readBuffer(DFSInputStream.java:477)
at org.apache.hadoop.hdfs.DFSInputStream.read(DFSInputStream.java:528)
at java.io.DataInputStream.read(DataInputStream.java:83)
at org.apache.hadoop.hdfs.TestRaidDfs.validateFile(TestRaidDfs.java:402)
at org.apache.hadoop.raid.TestBlockFixerDistConcurrency.testConcurrentJobs(TestBlockFixerDistConcurrency.java:161)

Todd Lipcon
added a comment - 06/Apr/11 23:55 Yes, I ran this test a few times and it looks like that failure only happens intermittently (though I was able to reproduce it once). I will commit this to trunk momentarily.