如何指定一个mapreduce应该运行的单个任务路径(节点)? [英] How to specify individual tasktrackers(nodes) on which a mapreduce should run?
本文介绍了如何指定一个mapreduce应该运行的单个任务路径(节点)?的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!
问题描述
在mapreduce作业中,单个任务将在随机节点上运行,是否有任何方法可以限制任务应运行的节点?
解决方案
Hadoop不会选择随机运行任务的节点。数据局部性被考虑,否则会有很多网络开销。
任务和它们运行的节点之间没有亲缘关系。 Hadoop不提供任何此类功能。
During the mapreduce job, the individual tasks will be running on random nodes, Is there any way to restrict the nodes in which the tasks should run ?
解决方案
Hadoop doesn't pick the nodes to run the tasks on random. Data locality is considered or else there would be a lot of network overhead.
There is no affinity between the tasks and the nodes on which they run. Hadoop doesn't offer any such functionality.
这篇关于如何指定一个mapreduce应该运行的单个任务路径(节点)?的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!
查看全文