hadoop学习(一)yarn jobs pending 和 jobs map % reduce 0%

yarn运行mapreduce程序wordcount

yarn jar ./share/hadoop/mapreduce/hadoop-mapreduce-examples-2.9.2.jar wordcount /input/text1 /output/text2_out

坑一:
yarn jobs 一直是 pending状态 而不是 running 状态
是因为yarn-site.xml中对yarn.nodemanager.resource.memory-mb,yarn.scheduler.maximum-allocation-mb大小设置出现了 问题,经过踩坑得出结论: 至少2048+ ,否则真的会卡的动不了…

搜索回答的时候看到有人讲是Hosts的问题 修改/etc/hosts文件 删掉无关节点 虽然我不是这个问题但还是记录一下吧~
下面是我的配置:

 <property>
        <name>yarn.resourcemanager.hostname</name>
        <value>master</value>
   </property>

   <property>
        <description>The address of the applications manager interface in the RM.</description>
        <name>yarn.resourcemanager.address</name>
        <value>${yarn.resourcemanager.hostname}:8032</value>
   </property>
   <property>
        <description>The address of the scheduler interface.</description>
        <name>yarn.resourcemanager.scheduler.address</name>
        <value>${yarn.resourcemanager.hostname}:8030</value>
   </property>
<property>
        <description>The http address of the RM web application.</description>
        <name>yarn.resourcemanager.webapp.address</name>
        <value>${yarn.resourcemanager.hostname}:8088</value>
   </property>
   <property>
        <description>The https adddress of the RM web application.</description>
        <name>yarn.resourcemanager.webapp.https.address</name>
        <value>${yarn.resourcemanager.hostname}:8090</value>
   </property>
<property>
        <name>yarn.resourcemanager.resource-tracker.address</name>
        <value>${yarn.resourcemanager.hostname}:8031</value>
   </property>
   <property>
        <description>The address of the RM admin interface.</description>
        <name>yarn.resourcemanager.admin.address</name>
        <value>${yarn.resourcemanager.hostname}:8033</value>
   </property>
   <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
   </property>
   <property>
        <name>yarn.log-aggregation.retain-seconds</name>
        <value>604800</value>
   </property>
   <property>
        <name>yarn.nodemanager.resource.memory-mb</name>
        <value>2048</value>
   </property>
   <property>
        <name>yarn.scheduler.maximum-allocation-mb</name>
        <value>2048</value>
   </property>

坑二:
jobs 是running状态的情况下 但是map reduce job的进度一直是0%
搜了好久,试了好多方法也想了想得出的结论如下:

  1. 可能是内存分配太少的问题,我的total memory显示只有4GB,4GB的电脑我已经尽力了…
    似乎正常都是二十几GB左右的样子,按理说我的主节点设置了2GB,子节点为1GB,可能虚拟内存的时候只计算了主节点乘节点数嘛… 大神求解,嘤嘤嘤~~~
  2. 我最后是因为mapreduce的配置出现了问题,修改了maprd.site.xml配置,之前配置也太粗心了,mapred.job.tracker属性的value应该是基于hdfs:// 而不是https://… 对原理还是理解不清晰,课程讲的太快了老师也不讲清楚 就要考试了还是一知半解 绝望了…
<configuration>
      <property>  
        <name>mapreduce.framework.name</name>  
        <value>yarn</value>  
      </property>
      <property>
        <name>mapred.job.tracker</name>
        <value>hdfs://192.168.11.5:9001</value>
      </property>
      <property>
      <name>mapred.local.dir</name>
       <value>/root/hadoop/var</value>
      </property>
      <property>
	<name>mapreduced.jobhistory.webapp.address</name>
	<value>master:19888</value>
      </property>
</configuration>

啊啊啊啊虽然编译器那块还是个坑,但是也算迈出一小步了,
明天还要加油鸭!!!
老大的新剧就要上了 5.7 破冰行动 耶斯!

猜你喜欢

转载自blog.csdn.net/qq_42910523/article/details/89648010