3 title: HDFS on Kubernetes
6 Repository holding helm charts for running Hadoop Distributed File System (HDFS)
9 See [charts/README.md](charts/README.md) for how to run the charts.
11 See [tests/README.md](tests/README.md) for how to run integration tests for
17 In case some pods are in pending state, check by using kubectl describe command.
20 Type Reason Age From Message
21 ---- ------ ---- ---- -------
22 Warning FailedScheduling 7s (x20 over 66s) default-scheduler pod has unbound immediate PersistentVolumeClaims (repeated 3 times)
25 Then make sure you have the storage provisioner up and running.
26 In our case, its rook that we support.
27 So, rook should be up and be the default storage proviosner.
31 rook-ceph-block (default) ceph.rook.io/block 132m
34 Delete all the previous unbound PVCs like below :
36 NAME STATUS VOLUME CAPACITY ACCESS MODES STORAGECLASS AGE
37 data-hdfs1-zookeeper-0 Pending 108m
38 editdir-hdfs1-journalnode-0 Pending 108m
39 metadatadir-hdfs1-namenode-0 Pending 108m
43 kubectl delete pvc/data-hdfs1-zookeeper-0
44 kubectl delete pvc/editdir-hdfs1-journalnode-0
45 kubectl delete pvc/metadatadir-hdfs1-namenode-0
48 #### If the dataNode restarts with the error:
50 19/07/19 21:22:55 FATAL datanode.DataNode: Initialization failed for Block pool <registering> (Datanode Uuid unassigned) service to hdfs1-namenode-1.hdfs1-namenode.hdfs1.svc.cluster.local/XXX.YY.ZZ.KK:8020. Exiting.
51 java.io.IOException: All specified directories are failed to load.
52 at org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:478)
53 at org.apache.hadoop.hdfs.server.datanode.DataNode.initStorage(DataNode.java:1358)
54 at org.apache.hadoop.hdfs.server.datanode.DataNode.initBlockPool(DataNode.java:1323)
55 at org.apache.hadoop.hdfs.server.datanode.BPOfferService.verifyAndSetNamespaceInfo(BPOfferService.java:317)
56 at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.connectToNNAndHandshake(BPServiceActor.java:223)
57 at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:802)
60 * SOLUTION: Make sure that whatever host path you set for the dataNode is deleted and doesnt exist before you run the hdfs helm chart.
66 In case you are reinstalling the HDFS, delete the host path : /hdfs-data
67 before you proceed or else the above error shall come.