English 中文(简体)
shluffle 错误: 超过最大失败_ unique_matche: 正在救助
原标题:shuffle error:exceeded max_failed_unique_matche : bailing out

我是一个新人,正在尝试执行字数示例。 我电脑上有由虚拟机器制作的4个节点组。 每次任务完成地图任务时, 都会有大约16%的减少任务显示这个错误 :

打乱错误: 超过 MAX_ FAILED_ UNIQUE_ FETCHES; 援助退出 。

12/05/24 04:43:12 WARN 地图red.JobClient:阅读任务输出错误

看来奴隶们无法从其他奴隶那里检索到数据。 有些链接上我发现它可能来自/ etc/ 主持人文件中的不一致。 但我已经交叉检查过它们, 它们都是一致的。 谁能帮我?

最佳回答

是否有防火墙来防止在普通的 Hadoop 端口的集束节点之间的通信( 在本案中任务跟踪器为 50060 ) 。 在端口 50060 上从一个节点到另一个节点进行测试, 并检查您是否得到了 http 响应代码 :

curl -I http://node1:50060/

一定要将上述节点1替换为 $HADOOP_HOME/conf/slaves 文件中的每一个值

EDIT 因此,结果发现这很可能是一个 DNS 问题, 这里您应该尝试的是 :

  • Examine the ${HADOOP_HOME}/conf/slaves file - each entry in here needs to be in the /etc/hosts file for each node in your cluster, or you must have them in your networks DNS server
  • Once you ve asserted the hosts file ON EVERY NODE in your cluster (or configured your DNS server), log into each node and check that you can ping the other cluster nodes by the names in the slaves file. Finally assert you can curl the tasktracker (port 50060) from each node to the other nodes (again using the machine names in the slaves file)
  • Restart your mapreduce services, just to be safe
问题回答

在每个节点中通过在终端中输入 $hostname 来检查主机名 。 请确保您的机器名相同( 主节点中的主机名和奴隶节点中的奴隶主机名) 。 如果没有, 请用您的节点名( master/ slave) 更改 / etc/ hostname 。 然后重新启动系统。 这将有效 。

< a href=> "http://suprogec.blogspot.in" rel="no follow" >SIMPLE Group





相关问题
Hadoop - namenode is not starting up

I am trying to run hadoop as a root user, i executed namenode format command hadoop namenode -format when the Hadoop file system is running. After this, when i try to start the name node server, it ...

What default reducers are available in Elastic MapReduce?

I hope I m asking this in the right way. I m learning my way around Elastic MapReduce and I ve seen numerous references to the "Aggregate" reducer that can be used with "Streaming" job flows. In ...

Establishing Eclipse project environment for HadoopDB

I have checked-out a project from SourceForge named HadoopDB. It uses some class in another project named Hive. I have used Eclipse Java build path setting to link source to the Hive project root ...

Hadoop: intervals and JOIN

I m very new to Hadoop and I m currently trying to join two sources of data where the key is an interval (say [date-begin/date-end]). For example: input1: 20091001-20091002 A 20091011-20091104 ...

hadoop- determine if a file is being written to

Is there a way to determine if a file in hadoop is being written to? eg- I have a process that puts logs into hdfs. I have another process that monitors for the existence of new logs in hdfs, but I ...

Building Apache Hive - impossible to resolve dependencies

I am trying out the Apache Hive as per http://wiki.apache.org/hadoop/Hive/GettingStarted and am getting this error from Ivy: Downloaded file size doesn t match expected Content Length for http://...

热门标签