hadoop (cdh version) that is configured to download and install

  1. Download hadoop archive, here is the download version 5.14.2 (address: http://archive.cloudera.com/cdh5/cdh/5/ )
    hadoop (cdh version) that is configured to download and install

hadoop (cdh version) that is configured to download and install
2. Xftp tool bag onto the compressed virtual machine / opt directory (Personal optional)
hadoop (cdh version) that is configured to download and install
3. Extract hadoop archive (command: the tar-2.6.0-cdh5.14.2.tar.gz -zxvf hadoop) hadoop (cdh version) that is configured to download and install
4. where for clarity, a new file folder with bigData stored separately after decompression, and rename
the new: mkdir bigdata
mobile: Music Videos Hadoop-2.6.0-cdh5.14.2 ./bigdata/hadoop260 hadoop (cdh version) that is configured to download and install
5. the head weight: modify the configuration file, move to / etc Up / down hadoop directory, ls command, circled in the figure below in this document is that we need to configure hadoop (cdh version) that is configured to download and install
6. configuration 1: vi hadoop-env.sh amended as follows position, to modify JAVA_HOME own path, you can echo $ JAVA_HOME View , revised and stored exit
hadoop (cdh version) that is configured to download and install
7. configuration 2: vi core-site.xml modify a positionhadoop (cdh version) that is configured to download and install
hadoop (cdh version) that is configured to download and install

<property>
                <name>fs.defaultFS</name>
                <value>hdfs://192.168.56.109:9000</value>
        </property>
        <property>
                <name>hadoop.tmp.dir</name>
                <value>/opt/hadoopdata</value>
        </property>
        <property>
                <name>hadoop.proxyuser.root.users</name>
                <value>*</value>
        </property>
        <property>
                <name>hadoop.proxyuser.root.groups</name>
                <value>*</value>
        </property>

8. Configuration 3: vi hdfs-site.xml, as in the previous step and adding the following code Tagshadoop (cdh version) that is configured to download and install

       <property>
                <name>dfs.replication</name>
                <value>1</value>
        </property>

9. Configuration 4, needs its own Copy the file, and the previous step was added as the following code Tags
command: cp mapred-site.xml.template mapred-site.xml
configuration: vi mapred-site.xmlhadoop (cdh version) that is configured to download and install

        <property>
                <name>mapreduce.framework.name</name>
                <value>yarn</value>
        </property>

10. Configuration 5: vi yarn-site.xml, middle-line comments can be removed directlyhadoop (cdh version) that is configured to download and install

<property>
                <name>yarn.resourcemanager.localhost</name>
                <value>localhost</value>
        </property>
        <property>
                <name>yarn.nodemanager.aux-services</name>
                <value>mapreduce_shuffle</value>
        </property>

11. Six-: vi / etc / profile, moved to the last, the following code is addedhadoop (cdh version) that is configured to download and install

export HADOOP_HOME=/opt/bigdata/hadoop260
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_COMMON_HOME=$HADOOP_HOME
export HADOOP_HDFS_HOME=$HADOOP_HOME
export YARN_HOME=$HADOOP_HOME
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin
export HADOOP_INSTALL=$HADOOP_HOME

12. The activation profile: Source / etc / Profile
13. Format NameNode: HDFS NameNode -format
14. A run: start-all.sh, then each step input Yes
15. A check on the operation: jps, to see if the following 5 process, represents configured to run a successful, if less which one went to check the corresponding configuration filehadoop (cdh version) that is configured to download and install

Guess you like

Origin blog.51cto.com/14497235/2433079