1. Zookeeper Dump
访问HBase的web页面:http://192.168.26.140:16030/zk.jsp
HBase is rooted at /hbase Active master address: hadoop.master,16020,1428562242762 Backup master addresses: Region server holding hbase:meta: hadoop.master,16201,1428562244095 Region servers: hadoop.master,16201,1428562244095 /hbase/replication: /hbase/replication/peers: /hbase/replication/rs: /hbase/replication/rs/hadoop.master,16201,1428562244095: Quorum Server Statistics: localhost:2181 Zookeeper version: 3.4.6-1569965, built on 02/20/2014 09:09 GMT Clients: /127.0.0.1:47168[1](queued=0,recved=2916,sent=2920) /127.0.0.1:47251[1](queued=0,recved=2752,sent=2752) /0:0:0:0:0:0:0:1:46915[1](queued=0,recved=2753,sent=2753) /0:0:0:0:0:0:0:1:46907[1](queued=0,recved=2763,sent=2763) /0:0:0:0:0:0:0:1:47060[1](queued=0,recved=2717,sent=2717) /127.0.0.1:47173[1](queued=0,recved=2754,sent=2754) /127.0.0.1:47174[1](queued=0,recved=3310,sent=3310) /127.0.0.1:47167[1](queued=0,recved=7394,sent=7465) /127.0.0.1:54521[1](queued=0,recved=33,sent=33) /127.0.0.1:47172[1](queued=0,recved=2753,sent=2753) /127.0.0.1:54583[0](queued=0,recved=1,sent=0) Latency min/avg/max: 0/0/790 Received: 30227 Sent: 30301 Connections: 11 Outstanding: 0 Zxid: 0xe5 Mode: standalone Node count: 45
1.包含的信息:HBase在Zookeeper中的根目录:/hbase
2.Master地址:hadoop.master,16020,1428562242762(域名,端口和启动的时间戳)
3. meta表所在的Region Server:hadoop.master,16201,1428562244095(域名,端口和启动的时间戳)
4. HBase集群包含的Region Servers:这里是一个Standalone集群,因此只有一个:hadoop.master,16201,1428562244095(同时也是meta表所在的Region Server)
2. meta表包含的信息
hbase命令行上执行如下命令,打印meta表的所有的数据
hbase(main):021:0> scan 'hbase:meta' ROW COLUMN+CELL hbase:namespace,,1426129053122.2e6 column=info:regioninfo, timestamp=1426129054266, value={ENCODED => 2e6738351085a34e0b83933c9dbe98b6, NA 738351085a34e0b83933c9dbe98b6. ME => 'hbase:namespace,,1426129053122.2e6738351085a34e0b83933c9dbe98b6.', STARTKEY => '', ENDKEY => ''} hbase:namespace,,1426129053122.2e6 column=info:seqnumDuringOpen, timestamp=1428562261802, value=\x00\x00\x00\x00\x00\x00\x00\x0C 738351085a34e0b83933c9dbe98b6. hbase:namespace,,1426129053122.2e6 column=info:server, timestamp=1428562261802, value=hadoop.master:16201 738351085a34e0b83933c9dbe98b6. hbase:namespace,,1426129053122.2e6 column=info:serverstartcode, timestamp=1428562261802, value=1428562244095 738351085a34e0b83933c9dbe98b6. //手工添加注释:行table3,,1428632308630.bb738e209c3b4d5833bb123549b324c3. 列族info 列:regioninfo, seqnumDuringOpen,server,serverstartcode //关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围 //table3的第一个Region(0-300) table3,,1428632308630.bb738e209c3b column=info:regioninfo, timestamp=1428632310195, value={ENCODED => bb738e209c3b4d5833bb123549b324c3, NA 4d5833bb123549b324c3. ME => 'table3,,1428632308630.bb738e209c3b4d5833bb123549b324c3.', STARTKEY => '', ENDKEY => 'row-300'} table3,,1428632308630.bb738e209c3b column=info:seqnumDuringOpen, timestamp=1428632313497, value=\x00\x00\x00\x00\x00\x00\x00\x02 4d5833bb123549b324c3. table3,,1428632308630.bb738e209c3b column=info:server, timestamp=1428632313497, value=hadoop.master:16201 4d5833bb123549b324c3. table3,,1428632308630.bb738e209c3b column=info:serverstartcode, timestamp=1428632313497, value=1428562244095 4d5833bb123549b324c3. //关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围 //table3的第二个Region(300-500) table3,row-300,1428632308630.2cf83 column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 2cf837baf289fb122c82d77214562c0a, NA 7baf289fb122c82d77214562c0a. ME => 'table3,row-300,1428632308630.2cf837baf289fb122c82d77214562c0a.', STARTKEY => 'row-300', ENDKEY = > 'row-500'} table3,row-300,1428632308630.2cf83 column=info:seqnumDuringOpen, timestamp=1428632313425, value=\x00\x00\x00\x00\x00\x00\x00\x02 7baf289fb122c82d77214562c0a. table3,row-300,1428632308630.2cf83 column=info:server, timestamp=1428632313425, value=hadoop.master:16201 7baf289fb122c82d77214562c0a. table3,row-300,1428632308630.2cf83 column=info:serverstartcode, timestamp=1428632313425, value=1428562244095 7baf289fb122c82d77214562c0a. //关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围 //table3的第三个Region(500-700) table3,row-500,1428632308630.1fa2e column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 1fa2e49c7404d3cd39afc39a99cc1c26, NA 49c7404d3cd39afc39a99cc1c26. ME => 'table3,row-500,1428632308630.1fa2e49c7404d3cd39afc39a99cc1c26.', STARTKEY => 'row-500', ENDKEY = > 'row-700'} table3,row-500,1428632308630.1fa2e column=info:seqnumDuringOpen, timestamp=1428632312856, value=\x00\x00\x00\x00\x00\x00\x00\x02 49c7404d3cd39afc39a99cc1c26. table3,row-500,1428632308630.1fa2e column=info:server, timestamp=1428632312856, value=hadoop.master:16201 49c7404d3cd39afc39a99cc1c26. table3,row-500,1428632308630.1fa2e column=info:serverstartcode, timestamp=1428632312856, value=1428562244095 49c7404d3cd39afc39a99cc1c26. //关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围 //table3的第四个Region(700-900) table3,row-700,1428632308630.1d3d6 column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 1d3d6e64741683e551e5b8d0722d7ff9, NA e64741683e551e5b8d0722d7ff9. ME => 'table3,row-700,1428632308630.1d3d6e64741683e551e5b8d0722d7ff9.', STARTKEY => 'row-700', ENDKEY = > 'row-900'} table3,row-700,1428632308630.1d3d6 column=info:seqnumDuringOpen, timestamp=1428632312817, value=\x00\x00\x00\x00\x00\x00\x00\x02 e64741683e551e5b8d0722d7ff9. table3,row-700,1428632308630.1d3d6 column=info:server, timestamp=1428632312817, value=hadoop.master:16201 e64741683e551e5b8d0722d7ff9. table3,row-700,1428632308630.1d3d6 column=info:serverstartcode, timestamp=1428632312817, value=1428562244095 e64741683e551e5b8d0722d7ff9. //关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围 //table3的第五个Region(900-正无穷) table3,row-900,1428632308630.4e048 column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 4e048f1c6a7a9a58994629ae957d4ddf, NA f1c6a7a9a58994629ae957d4ddf. ME => 'table3,row-900,1428632308630.4e048f1c6a7a9a58994629ae957d4ddf.', STARTKEY => 'row-900', ENDKEY = > ''} table3,row-900,1428632308630.4e048 column=info:seqnumDuringOpen, timestamp=1428632312795, value=\x00\x00\x00\x00\x00\x00\x00\x02 f1c6a7a9a58994629ae957d4ddf. table3,row-900,1428632308630.4e048 column=info:server, timestamp=1428632312795, value=hadoop.master:16201 f1c6a7a9a58994629ae957d4ddf. table3,row-900,1428632308630.4e048 column=info:serverstartcode, timestamp=1428632312795, value=1428562244095 f1c6a7a9a58994629ae957d4ddf.
3. -ROOT-表、meta表以及Region Server、Region的关系
4. Hadoop生态系统与HBase的集成
- MapReduce为HBase提供了高性能的计算能力
- Zookeeper为HBase提供了稳定服务和failover机制。
- Pig和Hive还为HBase提供了高层语言支持,使得在HBase上进行数据统计处理变的非常简单;
- Sqoop则为HBase提供了方便的RDBMS数据导入功能,使得传统数据库数据向HBase中迁移变的非常方便。