First, install Hadoop
1. The first extract hadoop to the relevant folder
2. Go to the folder after extracting good, relevant documents to be modified
3. Configure hadoop-env.sh
vi hadoop-env.sh:
export= JAVA_HOME=/opt/inst/jdk181
4. placed core-site.xml
vi core-site.xml
<property>
<name>fs.defaultFS</name>
<value>hdfs://bigdata:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/opt/hadoopdata</value>
</property>
<property>
<name>hadoop.proxyuser.root.users</name>
<value>*</value>
</property>
<property>
<name>hadoop.proxyuser.root.groups</name>
<value>*</value>
</property>
5. Configure hdfs-site.xml
vi hdfs-site.xml
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
6.配置cp mapred-site.xml.template mapred-site.xml
cp mapred-site.xml.template mapred-site.xml
7.vi mapred-site.xml
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
8.vi yarn-site.xml
vi
9.vi /etc/profile
export HADOOP_HOME=/opt/bigdata/hadoop260
export HADOOP_MAPRED_HOME=$HADOOP_HOME
export HADOOP_COMMON_HOME=$HADOOP_HOME
export HADOOP_HDFS_HOME=$HADOOP_HOME
export YARN_HOME=$HADOOP_HOME
export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
export PATH=$PATH:$HADOOP_HOME/sbin:$HADOOP_HOME/bin
export HADOOP_INSTALL=$HADOOP_HOME
10. Update and format
source /etc/profile
hdfs namenode -format
11. Start
start-all.sh
jps
#查看进程
12. The linux into the hadoop
hdfs dfs -put /opt/a.txt /cm/
hdfs dfs -ls /cm
Hadoop understand
Three core
1.HDFS: Hadoop Distribute File System Distributed File System
NameNode: master node directory
DataNode: slave node data
Backup master node: SecondaryNameNode
2.YARN: Yet Another Resource Negotiator Resource Management Scheduler
Scheduling is considered the force of memory and CPU resources
(Only one) to schedule by ResourceManager
ResourceManager major role:
1. The client requests
2. Monitoring NodeManager
3. Start or monitoring ApplicationMaster ()
4. The scheduling or resource allocation
NodeManager (more)
NodeManager major role:
1. management resources on a single node
2. The processing of commands from the ResourceManager
3. The processing of commands from ApplicationMaster
3.MapReduce: distributed computing framework
Operations