Nutch org.apache.hadoop.util.DiskChecker$DiskErrorException

今天在用nutch抓數據是總是報錯:

> org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local
directory for taskTracker/jobcache/job_local_0001/attempt_local_0001_m_000000_0/output/spill0.out
>         at org.apache.hadoop.fs.LocalDirAllocator$AllocatorPerContext.getLocalPathForWrite(LocalDirAllocator.java:335)
>         at org.apache.hadoop.fs.LocalDirAllocator.getLocalPathForWrite(LocalDirAllocator.java:124)
>         at org.apache.hadoop.mapred.MapOutputFile.getSpillFileForWrite(MapOutputFile.java:107)
>         at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:930)
>         at org.apache.hadoop.mapred.MapTask$MapOutputBuffer.flush(MapTask.java:842)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
>         at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:138)
> 2011-07-10 19:02:25,778 FATAL crawl.Generator - Generator: java.io.IOException: Job failed!
>         at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1232)
>         at org.apache.nutch.crawl.Generator.generate(Generator.java:472)
>         at org.apache.nutch.crawl.Generator.run(Generator.java:618)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>         at org.apache.nutch.crawl.Generator.main(Generator.java:581)


最後看了下網上的信息,最後發現是hadoop把/tmp文件夾給塞滿了。只要把/tmp/hadoop-root/mapred 文件夾刪除即可。再重新運行就不會出錯了。



發表評論
所有評論
還沒有人評論,想成為第一個評論的人麼? 請在上方評論欄輸入並且點擊發布.
相關文章