3.4 HDFS集群模式

第3章 HDFS:分布式文件系统

3.4 HDFS集群模式

节点 IP 角色
node1 192.168.80.131 NameNode,DataNode
node2 192.168.80.132 SecondaryNameNode,DataNode
node3 192.168.80.133 DataNode

3.4.1 Hadoop环境变量

[root@node1 ~]# vi /etc/profile.d/custom.sh
      
      
  • 1
  • 1

      
      
  1. #Hadoop path
  2. export HADOOP_HOME=/opt/hadoop- 2.7. 3
  3. export PATH= $PATH: $HADOOP_HOME/bin: $HADOOP_HOME/sbin
  • 1
  • 2
  • 3
  • 这里写图片描述
[root@node1 ~]# source /etc/profile.d/custom.sh
      
      
  • 1

【2018-01-27补充】 
感谢吴家行hang的提醒,这里node2和node3也要进行类似的环境变量配置。

3.4.2 准备工作

由于前面在node1上部署了Hadoop单机模式,需要停止Hadoop所有服务并清除数据目录。顺便检验一下设置的Hadoop环境变量。

这里写图片描述

清除Hadoop数据目录

[root@node1 ~]# rm -rf /tmp/hadoop-root/
      
      
  • 1

3.4.2 core-site.xml


      
      
  1. [root @node1 ~] # cd /opt/hadoop-2.7.3/etc/hadoop/
  2. [root @node1 hadoop] # vi core-site.xml
  • 1
  • 2

core-site.xml文件内容如下:


      
      
  1. <?xml version="1.0" encoding="UTF-8"?>
  2. <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3. <configuration>
  4. <property>
  5. <name>fs.defaultFS </name>
  6. <value>hdfs://node1:9000 </value>
  7. </property>
  8. <property>
  9. <name>hadoop.tmp.dir </name>
  10. <value>/var/data/hadoop </value>
  11. </property>
  12. <property>
  13. <name>io.file.buffer.size </name>
  14. <value>65536 </value>
  15. </property>
  16. </configuration>
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16

3.4.3 hdfs-site.xml

[root@node1 hadoop]# vi hdfs-site.xml
      
      
  • 1

hdfs-site.xml文件内容如下:


      
      
  1. <?xml version="1.0" encoding="UTF-8"?>
  2. <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3. <configuration>
  4. <property>
  5. <name>dfs.replication </name>
  6. <value>3 </value>
  7. </property>
  8. <property>
  9. <name>dfs.namenode.secondary.http-address </name>
  10. <value>node2:50090 </value>
  11. </property>
  12. <property>
  13. <name>dfs.namenode.secondary.https-address </name>
  14. <value>node2:50091 </value>
  15. </property>
  16. </configuration>
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16

3.4.4 slaves

编辑slaves文件

[root@node1 hadoop]# vi slaves
      
      
  • 1

slaves文件内容设置为:


      
      
  1. node1
  2. node2
  3. node3
  • 1
  • 2
  • 3

这里写图片描述

3.4.5 分发文件

将Hadoop软件包复制到node2和node3节点上

[root@node1 ~]# scp -r /opt/hadoop-2.7.3/ node2:/opt
      
      
  • 1
[root@node1 ~]# scp -r /opt/hadoop-2.7.3/ node3:/opt
      
      
  • 1

将环境变量文件复制到node2和node3节点上

[root@node1 ~]# scp /etc/profile.d/custom.sh node2:/etc/profile.d
      
      
  • 1
[root@node1 ~]# scp /etc/profile.d/custom.sh node3:/etc/profile.d
      
      
  • 1

最后source一下

[root@node2 ~]# source /etc/profile.d/custom.sh
      
      
  • 1
[root@node3 ~]# source /etc/profile.d/custom.sh
      
      
  • 1

3.4.6 NameNode格式化

[root@node1 ~]# hdfs namenode -format
      
      
  • 1

      
      
  1. ************************************************************/
  2. 17/ 05/ 14 09: 17: 28 INFO namenode .NameNode: registered UNIX signal handlers for [ TERM, HUP, INT]
  3. 17/ 05/ 14 09: 17: 28 INFO namenode .NameNode: createNameNode [-format]
  4. Formatting using clusterid: CID- 29bae3d3- 1786- 4428- 8359- 077976fe15e5
  5. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: No KeyProvider found.
  6. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: fsLock is fair: true
  7. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .DatanodeManager: dfs .block .invalidate .limit= 1000
  8. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .DatanodeManager: dfs .namenode .datanode .registration .ip-hostname-check= true
  9. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: dfs .namenode .startup .delay .block .deletion .sec is set to 000: 00: 00: 00.000
  10. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: The block deletion will start around 2017 May 14 09: 17: 30
  11. 17/ 05/ 14 09: 17: 30 INFO util .GSet: Computing capacity for map BlocksMap
  12. 17/ 05/ 14 09: 17: 30 INFO util .GSet: VM type = 64-bit
  13. 17/ 05/ 14 09: 17: 30 INFO util .GSet: 2.0% max memory 966.7 MB = 19.3 MB
  14. 17/ 05/ 14 09: 17: 30 INFO util .GSet: capacity = 2^ 21 = 2097152 entries
  15. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: dfs .block .access .token .enable= false
  16. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: defaultReplication = 3
  17. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: maxReplication = 512
  18. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: minReplication = 1
  19. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: maxReplicationStreams = 2
  20. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: replicationRecheckInterval = 3000
  21. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: encryptDataTransfer = false
  22. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: maxNumBlocksToLog = 1000
  23. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: fsOwner = root (auth: SIMPLE)
  24. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: supergroup = supergroup
  25. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: isPermissionEnabled = true
  26. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: HA Enabled: false
  27. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: Append Enabled: true
  28. 17/ 05/ 14 09: 17: 31 INFO util .GSet: Computing capacity for map INodeMap
  29. 17/ 05/ 14 09: 17: 31 INFO util .GSet: VM type = 64-bit
  30. 17/ 05/ 14 09: 17: 31 INFO util .GSet: 1.0% max memory 966.7 MB = 9.7 MB
  31. 17/ 05/ 14 09: 17: 31 INFO util .GSet: capacity = 2^ 20 = 1048576 entries
  32. 17/ 05/ 14 09: 17: 31 INFO namenode .FSDirectory: ACLs enabled? false
  33. 17/ 05/ 14 09: 17: 31 INFO namenode .FSDirectory: XAttrs enabled? true
  34. 17/ 05/ 14 09: 17: 31 INFO namenode .FSDirectory: Maximum size of an xattr: 16384
  35. 17/ 05/ 14 09: 17: 31 INFO namenode .NameNode: Caching file names occuring more than 10 times
  36. 17/ 05/ 14 09: 17: 31 INFO util .GSet: Computing capacity for map cachedBlocks
  37. 17/ 05/ 14 09: 17: 31 INFO util .GSet: VM type = 64-bit
  38. 17/ 05/ 14 09: 17: 31 INFO util .GSet: 0.25% max memory 966.7 MB = 2.4 MB
  39. 17/ 05/ 14 09: 17: 31 INFO util .GSet: capacity = 2^ 18 = 262144 entries
  40. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: dfs .namenode .safemode .threshold-pct = 0.9990000128746033
  41. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: dfs .namenode .safemode .min .datanodes = 0
  42. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: dfs .namenode .safemode .extension = 30000
  43. 17/ 05/ 14 09: 17: 31 INFO metrics .TopMetrics: NNTop conf: dfs .namenode .top .window .num .buckets = 10
  44. 17/ 05/ 14 09: 17: 31 INFO metrics .TopMetrics: NNTop conf: dfs .namenode .top .num .users = 10
  45. 17/ 05/ 14 09: 17: 31 INFO metrics .TopMetrics: NNTop conf: dfs .namenode .top .windows .minutes = 1, 5, 25
  46. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: Retry cache on namenode is enabled
  47. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
  48. 17/ 05/ 14 09: 17: 31 INFO util .GSet: Computing capacity for map NameNodeRetryCache
  49. 17/ 05/ 14 09: 17: 31 INFO util .GSet: VM type = 64-bit
  50. 17/ 05/ 14 09: 17: 31 INFO util .GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
  51. 17/ 05/ 14 09: 17: 31 INFO util .GSet: capacity = 2^ 15 = 32768 entries
  52. 17/ 05/ 14 09: 17: 31 INFO namenode .FSImage: Allocated new BlockPoolId: BP- 698786385- 192.168 .80 .131- 1494767851416
  53. 17/ 05/ 14 09: 17: 31 INFO common .Storage: Storage directory / var/data/hadoop/dfs/name has been successfully formatted.
  54. 17/ 05/ 14 09: 17: 31 INFO namenode .FSImageFormatProtobuf: Saving image file / var/data/hadoop/dfs/name/current/fsimage .ckpt_0000000000000000000 using no compression
  55. 17/ 05/ 14 09: 17: 31 INFO namenode .FSImageFormatProtobuf: Image file / var/data/hadoop/dfs/name/current/fsimage .ckpt_0000000000000000000 of size 351 bytes saved in 0 seconds.
  56. 17/ 05/ 14 09: 17: 31 INFO namenode .NNStorageRetentionManager: Going to retain 1 images with txid >= 0
  57. 17/ 05/ 14 09: 17: 31 INFO util .ExitUtil: Exiting with status 0
  58. 17/ 05/ 14 09: 17: 31 INFO namenode .NameNode: SHUTDOWN_MSG:
  59. /************************************************************
  60. SHUTDOWN_MSG: Shutting down NameNode at node1/192.168.80.131
  61. ************************************************************/
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • 31
  • 32
  • 33
  • 34
  • 35
  • 36
  • 37
  • 38
  • 39
  • 40
  • 41
  • 42
  • 43
  • 44
  • 45
  • 46
  • 47
  • 48
  • 49
  • 50
  • 51
  • 52
  • 53
  • 54
  • 55
  • 56
  • 57
  • 58
  • 59
  • 60
  • 61

      
      
  1. [ root@node1 ~ ] # ll /var/data/hadoop/dfs/name/current/
  2. total 16
  3. - rw - r - - r - - 1 root root 351 May 14 09:17 fsimage_0000000000000000000
  4. - rw - r - - r - - 1 root root 62 May 14 09:17 fsimage_0000000000000000000 . md5
  5. - rw - r - - r - - 1 root root 2 May 14 09:17 seen_txid
  6. - rw - r - - r - - 1 root root 206 May 14 09:17 VERSION
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6

3.4.7 启动HDFS


      
      
  1. [root@node1 ~] # start-dfs.sh
  2. Starting namenodes on [node1]
  3. node1: starting namenode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-namenode-node1 .out
  4. node2: starting datanode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-datanode-node2 .out
  5. node3: starting datanode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-datanode-node3 .out
  6. node1: starting datanode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-datanode-node1 .out
  7. Starting secondary namenodes [node2]
  8. node2: starting secondarynamenode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-secondaryn
  9. [root@node1 ~] #
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9

这里写图片描述

查看三个节点上的Java进程:

这里写图片描述

这里写图片描述

这里写图片描述

3.4.8 HDFS Web界面

打开http://192.168.80.131:50070

这里写图片描述

在”Datanodes”可以看到三个DataNode节点的信息:

这里写图片描述

灰常灰常感谢原博主的辛苦工作,为防止删博,所以转载,只供学习使用,不做其他任何商业用途。 https://blog.csdn.net/chengyuqiang/article/details/72058013

第3章 HDFS:分布式文件系统

3.4 HDFS集群模式

节点 IP 角色
node1 192.168.80.131 NameNode,DataNode
node2 192.168.80.132 SecondaryNameNode,DataNode
node3 192.168.80.133 DataNode

3.4.1 Hadoop环境变量

[root@node1 ~]# vi /etc/profile.d/custom.sh
    
    
  • 1
  • 1

    
    
  1. #Hadoop path
  2. export HADOOP_HOME=/opt/hadoop- 2.7. 3
  3. export PATH= $PATH: $HADOOP_HOME/bin: $HADOOP_HOME/sbin
  • 1
  • 2
  • 3
  • 这里写图片描述
[root@node1 ~]# source /etc/profile.d/custom.sh
    
    
  • 1

【2018-01-27补充】 
感谢吴家行hang的提醒,这里node2和node3也要进行类似的环境变量配置。

3.4.2 准备工作

由于前面在node1上部署了Hadoop单机模式,需要停止Hadoop所有服务并清除数据目录。顺便检验一下设置的Hadoop环境变量。

这里写图片描述

清除Hadoop数据目录

[root@node1 ~]# rm -rf /tmp/hadoop-root/
    
    
  • 1

3.4.2 core-site.xml


    
    
  1. [root @node1 ~] # cd /opt/hadoop-2.7.3/etc/hadoop/
  2. [root @node1 hadoop] # vi core-site.xml
  • 1
  • 2

core-site.xml文件内容如下:


    
    
  1. <?xml version="1.0" encoding="UTF-8"?>
  2. <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3. <configuration>
  4. <property>
  5. <name>fs.defaultFS </name>
  6. <value>hdfs://node1:9000 </value>
  7. </property>
  8. <property>
  9. <name>hadoop.tmp.dir </name>
  10. <value>/var/data/hadoop </value>
  11. </property>
  12. <property>
  13. <name>io.file.buffer.size </name>
  14. <value>65536 </value>
  15. </property>
  16. </configuration>
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16

3.4.3 hdfs-site.xml

[root@node1 hadoop]# vi hdfs-site.xml
    
    
  • 1

hdfs-site.xml文件内容如下:


    
    
  1. <?xml version="1.0" encoding="UTF-8"?>
  2. <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
  3. <configuration>
  4. <property>
  5. <name>dfs.replication </name>
  6. <value>3 </value>
  7. </property>
  8. <property>
  9. <name>dfs.namenode.secondary.http-address </name>
  10. <value>node2:50090 </value>
  11. </property>
  12. <property>
  13. <name>dfs.namenode.secondary.https-address </name>
  14. <value>node2:50091 </value>
  15. </property>
  16. </configuration>
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16

3.4.4 slaves

编辑slaves文件

[root@node1 hadoop]# vi slaves
    
    
  • 1

slaves文件内容设置为:


    
    
  1. node1
  2. node2
  3. node3
  • 1
  • 2
  • 3

这里写图片描述

3.4.5 分发文件

将Hadoop软件包复制到node2和node3节点上

[root@node1 ~]# scp -r /opt/hadoop-2.7.3/ node2:/opt
    
    
  • 1
[root@node1 ~]# scp -r /opt/hadoop-2.7.3/ node3:/opt
    
    
  • 1

将环境变量文件复制到node2和node3节点上

[root@node1 ~]# scp /etc/profile.d/custom.sh node2:/etc/profile.d
    
    
  • 1
[root@node1 ~]# scp /etc/profile.d/custom.sh node3:/etc/profile.d
    
    
  • 1

最后source一下

[root@node2 ~]# source /etc/profile.d/custom.sh
    
    
  • 1
[root@node3 ~]# source /etc/profile.d/custom.sh
    
    
  • 1

3.4.6 NameNode格式化

[root@node1 ~]# hdfs namenode -format
    
    
  • 1

    
    
  1. ************************************************************/
  2. 17/ 05/ 14 09: 17: 28 INFO namenode .NameNode: registered UNIX signal handlers for [ TERM, HUP, INT]
  3. 17/ 05/ 14 09: 17: 28 INFO namenode .NameNode: createNameNode [-format]
  4. Formatting using clusterid: CID- 29bae3d3- 1786- 4428- 8359- 077976fe15e5
  5. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: No KeyProvider found.
  6. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: fsLock is fair: true
  7. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .DatanodeManager: dfs .block .invalidate .limit= 1000
  8. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .DatanodeManager: dfs .namenode .datanode .registration .ip-hostname-check= true
  9. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: dfs .namenode .startup .delay .block .deletion .sec is set to 000: 00: 00: 00.000
  10. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: The block deletion will start around 2017 May 14 09: 17: 30
  11. 17/ 05/ 14 09: 17: 30 INFO util .GSet: Computing capacity for map BlocksMap
  12. 17/ 05/ 14 09: 17: 30 INFO util .GSet: VM type = 64-bit
  13. 17/ 05/ 14 09: 17: 30 INFO util .GSet: 2.0% max memory 966.7 MB = 19.3 MB
  14. 17/ 05/ 14 09: 17: 30 INFO util .GSet: capacity = 2^ 21 = 2097152 entries
  15. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: dfs .block .access .token .enable= false
  16. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: defaultReplication = 3
  17. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: maxReplication = 512
  18. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: minReplication = 1
  19. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: maxReplicationStreams = 2
  20. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: replicationRecheckInterval = 3000
  21. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: encryptDataTransfer = false
  22. 17/ 05/ 14 09: 17: 30 INFO blockmanagement .BlockManager: maxNumBlocksToLog = 1000
  23. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: fsOwner = root (auth: SIMPLE)
  24. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: supergroup = supergroup
  25. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: isPermissionEnabled = true
  26. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: HA Enabled: false
  27. 17/ 05/ 14 09: 17: 30 INFO namenode .FSNamesystem: Append Enabled: true
  28. 17/ 05/ 14 09: 17: 31 INFO util .GSet: Computing capacity for map INodeMap
  29. 17/ 05/ 14 09: 17: 31 INFO util .GSet: VM type = 64-bit
  30. 17/ 05/ 14 09: 17: 31 INFO util .GSet: 1.0% max memory 966.7 MB = 9.7 MB
  31. 17/ 05/ 14 09: 17: 31 INFO util .GSet: capacity = 2^ 20 = 1048576 entries
  32. 17/ 05/ 14 09: 17: 31 INFO namenode .FSDirectory: ACLs enabled? false
  33. 17/ 05/ 14 09: 17: 31 INFO namenode .FSDirectory: XAttrs enabled? true
  34. 17/ 05/ 14 09: 17: 31 INFO namenode .FSDirectory: Maximum size of an xattr: 16384
  35. 17/ 05/ 14 09: 17: 31 INFO namenode .NameNode: Caching file names occuring more than 10 times
  36. 17/ 05/ 14 09: 17: 31 INFO util .GSet: Computing capacity for map cachedBlocks
  37. 17/ 05/ 14 09: 17: 31 INFO util .GSet: VM type = 64-bit
  38. 17/ 05/ 14 09: 17: 31 INFO util .GSet: 0.25% max memory 966.7 MB = 2.4 MB
  39. 17/ 05/ 14 09: 17: 31 INFO util .GSet: capacity = 2^ 18 = 262144 entries
  40. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: dfs .namenode .safemode .threshold-pct = 0.9990000128746033
  41. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: dfs .namenode .safemode .min .datanodes = 0
  42. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: dfs .namenode .safemode .extension = 30000
  43. 17/ 05/ 14 09: 17: 31 INFO metrics .TopMetrics: NNTop conf: dfs .namenode .top .window .num .buckets = 10
  44. 17/ 05/ 14 09: 17: 31 INFO metrics .TopMetrics: NNTop conf: dfs .namenode .top .num .users = 10
  45. 17/ 05/ 14 09: 17: 31 INFO metrics .TopMetrics: NNTop conf: dfs .namenode .top .windows .minutes = 1, 5, 25
  46. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: Retry cache on namenode is enabled
  47. 17/ 05/ 14 09: 17: 31 INFO namenode .FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
  48. 17/ 05/ 14 09: 17: 31 INFO util .GSet: Computing capacity for map NameNodeRetryCache
  49. 17/ 05/ 14 09: 17: 31 INFO util .GSet: VM type = 64-bit
  50. 17/ 05/ 14 09: 17: 31 INFO util .GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
  51. 17/ 05/ 14 09: 17: 31 INFO util .GSet: capacity = 2^ 15 = 32768 entries
  52. 17/ 05/ 14 09: 17: 31 INFO namenode .FSImage: Allocated new BlockPoolId: BP- 698786385- 192.168 .80 .131- 1494767851416
  53. 17/ 05/ 14 09: 17: 31 INFO common .Storage: Storage directory / var/data/hadoop/dfs/name has been successfully formatted.
  54. 17/ 05/ 14 09: 17: 31 INFO namenode .FSImageFormatProtobuf: Saving image file / var/data/hadoop/dfs/name/current/fsimage .ckpt_0000000000000000000 using no compression
  55. 17/ 05/ 14 09: 17: 31 INFO namenode .FSImageFormatProtobuf: Image file / var/data/hadoop/dfs/name/current/fsimage .ckpt_0000000000000000000 of size 351 bytes saved in 0 seconds.
  56. 17/ 05/ 14 09: 17: 31 INFO namenode .NNStorageRetentionManager: Going to retain 1 images with txid >= 0
  57. 17/ 05/ 14 09: 17: 31 INFO util .ExitUtil: Exiting with status 0
  58. 17/ 05/ 14 09: 17: 31 INFO namenode .NameNode: SHUTDOWN_MSG:
  59. /************************************************************
  60. SHUTDOWN_MSG: Shutting down NameNode at node1/192.168.80.131
  61. ************************************************************/
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • 11
  • 12
  • 13
  • 14
  • 15
  • 16
  • 17
  • 18
  • 19
  • 20
  • 21
  • 22
  • 23
  • 24
  • 25
  • 26
  • 27
  • 28
  • 29
  • 30
  • 31
  • 32
  • 33
  • 34
  • 35
  • 36
  • 37
  • 38
  • 39
  • 40
  • 41
  • 42
  • 43
  • 44
  • 45
  • 46
  • 47
  • 48
  • 49
  • 50
  • 51
  • 52
  • 53
  • 54
  • 55
  • 56
  • 57
  • 58
  • 59
  • 60
  • 61

    
    
  1. [ root@node1 ~ ] # ll /var/data/hadoop/dfs/name/current/
  2. total 16
  3. - rw - r - - r - - 1 root root 351 May 14 09:17 fsimage_0000000000000000000
  4. - rw - r - - r - - 1 root root 62 May 14 09:17 fsimage_0000000000000000000 . md5
  5. - rw - r - - r - - 1 root root 2 May 14 09:17 seen_txid
  6. - rw - r - - r - - 1 root root 206 May 14 09:17 VERSION
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6

3.4.7 启动HDFS


    
    
  1. [root@node1 ~] # start-dfs.sh
  2. Starting namenodes on [node1]
  3. node1: starting namenode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-namenode-node1 .out
  4. node2: starting datanode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-datanode-node2 .out
  5. node3: starting datanode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-datanode-node3 .out
  6. node1: starting datanode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-datanode-node1 .out
  7. Starting secondary namenodes [node2]
  8. node2: starting secondarynamenode, logging to /opt/hadoop- 2.7 .3/logs/hadoop-root-secondaryn
  9. [root@node1 ~] #
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9

这里写图片描述

查看三个节点上的Java进程:

这里写图片描述

这里写图片描述

这里写图片描述

3.4.8 HDFS Web界面

打开http://192.168.80.131:50070

这里写图片描述

在”Datanodes”可以看到三个DataNode节点的信息:

这里写图片描述

灰常灰常感谢原博主的辛苦工作,为防止删博,所以转载,只供学习使用,不做其他任何商业用途。 https://blog.csdn.net/chengyuqiang/article/details/72058013

猜你喜欢

转载自blog.csdn.net/zqwzlanbao/article/details/84069837
3.4
今日推荐