One of the big data: Hadoop installation

First, install Hadoop

1. The first extract hadoop to the relevant folder

1566901830564

2. Go to the folder after extracting good, relevant documents to be modified

1566902099154

1566893588624

3. Configure hadoop-env.sh

vi hadoop-env.sh:
export= JAVA_HOME=/opt/inst/jdk181

1566893705161

4. placed core-site.xml

vi core-site.xml
        <property>
                <name>fs.defaultFS</name>
                <value>hdfs://bigdata:9000</value>
        </property>
        <property>
                <name>hadoop.tmp.dir</name>
                <value>/opt/hadoopdata</value>
        </property>
        <property>
                <name>hadoop.proxyuser.root.users</name>
                <value>*</value>
        </property>
        <property>
                <name>hadoop.proxyuser.root.groups</name>
                <value>*</value>
        </property>

1566893927457

1566894986655

5. Configure hdfs-site.xml

vi hdfs-site.xml
        <property>
                <name>dfs.replication</name>
                <value>1</value>
        </property>

1566895205777

6.配置cp mapred-site.xml.template mapred-site.xml

cp mapred-site.xml.template mapred-site.xml

1566895334652

7.vi mapred-site.xml

        <property>
                <name>mapreduce.framework.name</name>
                <value>yarn</value>
        </property>

1566896513659

8.vi yarn-site.xml

vi 

1566896452284

9.vi /etc/profile

export HADOOP_HOME=/opt/bigdata/hadoop260
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_COMMON_HOME=$HADOOP_HOME
export HADOOP_HDFS_HOME=$HADOOP_HOME
export YARN_HOME=$HADOOP_HOME
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin
export HADOOP_INSTALL=$HADOOP_HOME

1566896579450

10. Update and format

source  /etc/profile
hdfs namenode -format

1566897164905

11. Start

start-all.sh

1566897200256

jps
#查看进程

1566897333891

1566897884766

12. The linux into the hadoop

hdfs dfs -put /opt/a.txt  /cm/
hdfs dfs -ls /cm

1566898227114

Hadoop understand

Three core

1.HDFS: Hadoop Distribute File System Distributed File System

NameNode: master node directory

DataNode: slave node data

Backup master node: SecondaryNameNode

2.YARN: Yet Another Resource Negotiator Resource Management Scheduler

Scheduling is considered the force of memory and CPU resources

(Only one) to schedule by ResourceManager

ResourceManager major role:

1. The client requests

2. Monitoring NodeManager

3. Start or monitoring ApplicationMaster ()

4. The scheduling or resource allocation

NodeManager (more)

NodeManager major role:

1. management resources on a single node

2. The processing of commands from the ResourceManager

3. The processing of commands from ApplicationMaster

3.MapReduce: distributed computing framework

Operations

1566913676798

Guess you like

Origin blog.51cto.com/14522074/2434368