Hadoop2.6.0.CDH 源码编译

  • 需要软件和下载地址
软件 地址
hadoop-2.6.0-cdh5.6.0-src.tar.gz http://archive-primary.cloudera.com/cdh5/cdh/5/hadoop-2.6.0-cdh5.6.0-src.tar.gz
JDK1.7 https://www.oracle.com/technetwork/java/javase/downloads/java-archive-downloads-javase7-521261.html
Maven 3.0+ http://mirrors.tuna.tsinghua.edu.cn/apache/maven/maven-3/3.6.0/binaries/apache-maven-3.6.0-bin.tar.gz
protobuf(最好下载2.5.0版本) https://pypi.org/project/protobuf/2.5.0/
findbugs https://sourceforge.net/projects/findbugs/files/findbugs/1.3.9/findbugs-1.3.9.tar.gz/download
Tomcat-6.+ https://archive.apache.org/dist/tomcat/tomcat-6/v6.0.53/bin/apache-tomcat-6.0.53.tar.gz
snappy http://pkgs.fedoraproject.org/repo/pkgs/snappy/snappy-1.1.1.tar.gz/8887e3b7253b22a31f5486bca3cbc1c2/snappy-1.1.1.tar.gz
或者直接下载:(这里是我下载好的) 链接:https://pan.baidu.com/s/1N3GHgWv-sOzT5v1gzu_gww

提取码:0o35
复制这段内容后打开百度网盘手机App,操作更方便哦

开始安装…

一. 安装配置

我的环境变量配置在 root用户的当前环境变量,
因为我是使用root用户操作的.hadoop用户也可以,但是我这里hadoop用户会报以前创建文件夹或执行文件的权限
所以我使用root用户操作,vim ~/.bash_profile

  • 安装JDK
  1. 解压
  2. 配置环境变量
  • 安装Maven
  1. 解压Maven
  2. 配置环境变量
  3. 配置setting.xml
<mirror>
        <id>alimaven</id>
         <name>aliyun maven</name>
        <url>http://maven.aliyun.com/nexus/content/groups/public/</url>
        <mirrorOf>central</mirrorOf>
        </mirror>
  • 安装protobuf
  1. 解压/安装
    tar -zxvf protobuf-2.5.0.tar.gz -C …/servers/
    cd /export/servers/protobuf-2.5.0
    ./configure
    make && make install

  2. 配置环境变量

  3. 测试是否安装成功
    protoc --version

  • 安装以下依赖包
    yum install autoconf automake libtool cmake
    yum install ncurses-devel
    yum install openssl-devel
    yum install lzo-devel zlib-devel gcc gcc-c++

  • bzip2压缩需要的依赖包

  1. 下载
    yum install -y bzip2-devel
  • 安装protobuf
  1. 解压/安装

tar -zxvf protobuf-2.5.0.tar.gz -C …/servers/
cd /export/servers/protobuf-2.5.0
./configure
make && make install

  • 配置环境变量

二. 最终环境变量

export JAVA_HOME=/home/hadoop/app/jdk1.7.0_79
export PROTOC_HOME=/root/protobuf-2.5.0
export MAVEN_HOME=/home/hadoop/app/apache-maven-3.3.9
export FINDBUGS_HOME=/home/hadoop/software/compile/findbugs-1.3.9
export PATH=$JAVA_HOME/bin:$PROTOC_HOME/bin:$MAVEN_HOME/bin:$FINDBUGS_HOME/bin:$PATH


三.hadoop源码编译

  • hadoop源码安装
  1. 解压
  2. 编译安装
    mvn package -Pdist,native -DskipTests –Dtar
  • 支持snappy需要下载安装snappy
  1. 安装/编译snappy
    tar -zxf snappy-1.1.1.tar.gz -C …/servers/
    cd …/servers/snappy-1.1.1/
    ./configure
    make && make install
  2. 支持snappy的源码编译
    mvn package -DskipTests -Pdist,native -Dtar -Drequire.snappy -e -X

-X查看详细日志

四.结果

  • 编译成功

[INFO] --- maven-javadoc-plugin:2.8.1:jar (module-javadocs) @ hadoop-dist ---
[INFO] Building jar: /home/hadoop/software/compile/hadoop-2.6.0-cdh5.7.0/hadoop-dist/target/hadoop-dist-2.6.0-cdh5.7.0-javadoc.jar
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Hadoop Main ................................. SUCCESS [  4.355 s]
[INFO] Apache Hadoop Project POM .......................... SUCCESS [  2.377 s]
[INFO] Apache Hadoop Annotations .......................... SUCCESS [  6.614 s]
[INFO] Apache Hadoop Assemblies ........................... SUCCESS [  1.079 s]
[INFO] Apache Hadoop Project Dist POM ..................... SUCCESS [  2.813 s]
[INFO] Apache Hadoop Maven Plugins ........................ SUCCESS [  6.289 s]
[INFO] Apache Hadoop MiniKDC .............................. SUCCESS [  7.335 s]
[INFO] Apache Hadoop Auth ................................. SUCCESS [  7.251 s]
[INFO] Apache Hadoop Auth Examples ........................ SUCCESS [  5.306 s]
[INFO] Apache Hadoop Common ............................... SUCCESS [04:02 min]
[INFO] Apache Hadoop NFS .................................. SUCCESS [ 10.869 s]
[INFO] Apache Hadoop KMS .................................. SUCCESS [01:13 min]
[INFO] Apache Hadoop Common Project ....................... SUCCESS [  0.094 s]
[INFO] Apache Hadoop HDFS ................................. SUCCESS [05:32 min]
[INFO] Apache Hadoop HttpFS ............................... SUCCESS [03:57 min]
[INFO] Apache Hadoop HDFS BookKeeper Journal .............. SUCCESS [ 14.586 s]
[INFO] Apache Hadoop HDFS-NFS ............................. SUCCESS [ 10.348 s]
[INFO] Apache Hadoop HDFS Project ......................... SUCCESS [  0.365 s]
[INFO] hadoop-yarn ........................................ SUCCESS [  0.144 s]
[INFO] hadoop-yarn-api .................................... SUCCESS [02:29 min]
[INFO] hadoop-yarn-common ................................. SUCCESS [ 42.984 s]
[INFO] hadoop-yarn-server ................................. SUCCESS [  0.211 s]
[INFO] hadoop-yarn-server-common .......................... SUCCESS [ 17.051 s]
[INFO] hadoop-yarn-server-nodemanager ..................... SUCCESS [ 28.840 s]
[INFO] hadoop-yarn-server-web-proxy ....................... SUCCESS [  4.770 s]
[INFO] hadoop-yarn-server-applicationhistoryservice ....... SUCCESS [ 10.825 s]
[INFO] hadoop-yarn-server-resourcemanager ................. SUCCESS [ 33.446 s]
[INFO] hadoop-yarn-server-tests ........................... SUCCESS [  2.175 s]
[INFO] hadoop-yarn-client ................................. SUCCESS [  9.191 s]
[INFO] hadoop-yarn-applications ........................... SUCCESS [  0.096 s]
[INFO] hadoop-yarn-applications-distributedshell .......... SUCCESS [  4.704 s]
[INFO] hadoop-yarn-applications-unmanaged-am-launcher ..... SUCCESS [  3.566 s]
[INFO] hadoop-yarn-site ................................... SUCCESS [  0.139 s]
[INFO] hadoop-yarn-registry ............................... SUCCESS [  9.109 s]
[INFO] hadoop-yarn-project ................................ SUCCESS [ 10.264 s]
[INFO] hadoop-mapreduce-client ............................ SUCCESS [  0.259 s]
[INFO] hadoop-mapreduce-client-core ....................... SUCCESS [ 34.068 s]
[INFO] hadoop-mapreduce-client-common ..................... SUCCESS [ 27.019 s]
[INFO] hadoop-mapreduce-client-shuffle .................... SUCCESS [  7.110 s]
[INFO] hadoop-mapreduce-client-app ........................ SUCCESS [ 17.977 s]
[INFO] hadoop-mapreduce-client-hs ......................... SUCCESS [ 13.277 s]
[INFO] hadoop-mapreduce-client-jobclient .................. SUCCESS [ 11.089 s]
[INFO] hadoop-mapreduce-client-hs-plugins ................. SUCCESS [  4.022 s]
[INFO] hadoop-mapreduce-client-nativetask ................. SUCCESS [02:03 min]
[INFO] Apache Hadoop MapReduce Examples ................... SUCCESS [  8.514 s]
[INFO] hadoop-mapreduce ................................... SUCCESS [  8.381 s]
[INFO] Apache Hadoop MapReduce Streaming .................. SUCCESS [  7.468 s]
[INFO] Apache Hadoop Distributed Copy ..................... SUCCESS [ 14.055 s]
[INFO] Apache Hadoop Archives ............................. SUCCESS [  4.625 s]
[INFO] Apache Hadoop Archive Logs ......................... SUCCESS [  4.193 s]
[INFO] Apache Hadoop Rumen ................................ SUCCESS [  9.228 s]
[INFO] Apache Hadoop Gridmix .............................. SUCCESS [  7.490 s]
[INFO] Apache Hadoop Data Join ............................ SUCCESS [  4.210 s]
[INFO] Apache Hadoop Ant Tasks ............................ SUCCESS [  3.682 s]
[INFO] Apache Hadoop Extras ............................... SUCCESS [  4.795 s]
[INFO] Apache Hadoop Pipes ................................ SUCCESS [ 11.485 s]
[INFO] Apache Hadoop OpenStack support .................... SUCCESS [  9.499 s]
[INFO] Apache Hadoop Amazon Web Services support .......... SUCCESS [  9.641 s]
[INFO] Apache Hadoop Azure support ........................ SUCCESS [  7.469 s]
[INFO] Apache Hadoop Client ............................... SUCCESS [  9.368 s]
[INFO] Apache Hadoop Mini-Cluster ......................... SUCCESS [  4.012 s]
[INFO] Apache Hadoop Scheduler Load Simulator ............. SUCCESS [  6.967 s]
[INFO] Apache Hadoop Tools Dist ........................... SUCCESS [ 13.934 s]
[INFO] Apache Hadoop Tools ................................ SUCCESS [  0.069 s]
[INFO] Apache Hadoop Distribution ......................... SUCCESS [ 56.287 s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD SUCCESS
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 29:06 min
[INFO] Finished at: 2018-12-05T22:51:11+08:00
[INFO] Final Memory: 146M/412M


五. 编译Lzo压缩配置

  1. 参考博客

    https://blog.csdn.net/wenyusuran/article/details/25904839

    https://blog.csdn.net/u011586820/article/details/49763003

  2. 验证lzop

[hadoop@hadoop001 native]$ lzop --version
lzop 1.03
LZO library 2.06
Copyright (C) 1996-2010 Markus Franz Xaver Johannes Oberhumer

  1. 操作上面完成之后
  • 查看/home/hadoop/app/hadoop-lzo-master/build/native/Linux-amd64-64/lib是否生成lzo库的链接

[hadoop@hadoop001 lib]$ ll
总用量 196
-rw-r--r-- 1 hadoop hadoop 116430 12月  3 23:09 libgplcompression.a
-rw-rw-r-- 1 hadoop hadoop   1132 12月  3 23:09 libgplcompression.la
lrwxrwxrwx 1 hadoop hadoop     26 12月  3 23:09 libgplcompression.so -> libgplcompression.so.0.0.0
lrwxrwxrwx 1 hadoop hadoop     26 12月  3 23:09 libgplcompression.so.0 -> libgplcompression.so.0.0.0
-rwxrwxr-x 1 hadoop hadoop  77424 12月  3 23:09 libgplcompression.so.0.0.0

  1. 复制home/hadoop/app/hadoop-lzo-master/build/native/Linux-amd64-64/lib下的所有包到你自己安装好的$HADOOP_HOME/lib/nativ

    (网上是一顿配置,我这里直接把生成的文件复制过去)

猜你喜欢

转载自blog.csdn.net/huonan_123/article/details/84764503