HDFS伪分布式环境搭建记录

一、环境准备
IP地址查看
ifconfig
192.168.2.46

修改主机名
hostnamectl set-hostname master
hostname

关闭防火墙
systemctl disable firewalld

配置hosts文件
vi /etc/hosts
192.168.2.46 master

重启网络
systemctl restart network

创建hadoop用户
useradd hadoop
passwd hadoop
密码hadoop7890
su - hadoop
ssh-keygen -t rsa -P ''
ls .ssh/
cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys

设置SSH配置
su - root
vi /etc/ssh/sshd_config

PubkeyAuthentication yes # 启用公钥私钥配对认证方式
AuthorizedKeysFile .ssh/authorized_keys # 公钥文件路径(默认不需要修改)

systemctl restart sshd

验证ssh
su - hadoop
ssh master
exit
su - root

二、安装jdk
tar -zxvf jdk-8u162-linux-x64.tar.gz
mv /root/res/jdk1.8.0_162/ /usr/local/src/
cd /usr/local/src/

设置环境变量
vi ~/.bash_profile

export JAVA_HOME=/usr/local/src/jdk1.8.0_162
export JRE_HOME=/usr/local/src/jdk1.8.0_162/jre
export CLASSPATH=.:$JAVA_HOME/lib:$JRE_HOME/lib:$CLASSPATH
export PATH=$JAVA_HOME/bin:$JRE_HOME/bin:$JAVA_HOME:$PATH

source ~/.bash_profile
echo $JAVA_HOME
java -version

三、安装hadoop
tar -zxvf hadoop-2.7.7.tar.gz
mv hadoop-2.7.7 /usr/local/src/
cd /usr/local/src/
mv hadoop-2.7.7 hadoop

设置环境变量
vi ~/.bash_profile

export HADOOP_HOME=/usr/local/src/hadoop
export PATH=$HADOOP_HOME:$JAVA_HOME/bin:$JRE_HOME/bin:$JAVA_HOME:$PATH

echo $HADOOP_HOME

修改hadoop-env.sh
cd /usr/local/src/hadoop/etc/hadoop/
vi hadoop-env.sh

export JAVA_HOME=/usr/local/src/jdk1.8.0_162

vi core-site.xml

<property>
<name>fs.defaultFS</name>
<value>hdfs://master:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/usr/local/src/hadoop/tmp</value>
</property>

mkdir /usr/local/src/hadoop/tmp

vi hdfs-site.xml

<property>
<name>dfs.datanode.ipc.address</name>
<value>0.0.0.0:50020</value>
</property>
<property>
<name>dfs.datanode.http.address</name>
<value>0.0.0.0:50075</value>
</property>
<property>
<name>dfs.replication</name>
<value>1</value>
</property>

vi yarn-site.xml

<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>

cp mapred-site.xml.template mapred-site.xml
vi mapred-site.xml

<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>master:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>master:19888</value>
</property>

vi masters
master
vi slaves
master
注:因为在伪分布式模式下,作为master的namenode与作为slave的datanode是同一台服务器,所以配置文件中的ip是一样的

cp ~/res/fastjson-1.2.41.jar /usr/local/src/hadoop/share/hadoop/common/

四、测试hadoop
su - hadoop
cd /usr/local/src/hadoop/
bin/hdfs namenode -format
sbin/start-all.sh
bin/hdfs dfsadmin -report
su - root
jps

五、常用命令
ls 命令
bin/hdfs dfs -ls /
df 命令 查看根目录磁盘空间
bin/hdfs dfs -df /
put 命令
bin/hdfs dfs -put NOTICE.txt /
cat 命令
bin/hdfs dfs -cat /NOTICE.txt
cp 命令
bin/hdfs dfs -cp /NOTICE.txt /test.txt
du 命令
bin/hdfs dfs -du /
rm命令
bin/hdfs dfs -rm /NOTICE.txt
mkdir命令
bin/hdfs dfs -mkdir /user
bin/hdfs dfs -mkdir /user/zhang
rmdir命令
bin/hdfs dfs -rmdir /user/zhang

猜你喜欢

转载自www.cnblogs.com/zhanglinfeng715/p/12168316.html