尝试在Ubuntu 18上启动和运行Hadoop 3.2.2。
遵循本教程。https://phoenixnap.com/kb/install-hadoop-ubuntu
我试着排除故障,我读到了这个帖子:Datanode process not running in Hadoop
已检查Datanode日志,已查看权限问题。
因为上面链接的帖子中12票答案的第五部分似乎最相关,所以我这样做了:
cd dfsdata
sudo chmod -R 755 datanode
cd ..
cd hadoop-3.2.2
cd sbin
./stop-all.sh
hadoop namenode -format
start-all.sh
jps但列表中仍然没有DataNode。
(这与建议的过程有点不同;我没有停止-所有这些都是在更改权限之前进行的,这有关系吗?)
Datanode日志文件:
************************************************************/
2021-06-02 15:14:24,546 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: registered UNIX signal handlers for [TERM, HUP, INT]
2021-06-02 15:14:25,607 INFO org.apache.hadoop.hdfs.server.datanode.checker.ThrottledAsyncChecker: Scheduling a check for [DISK]file:/home/hdoop/dfsdata/datanode
2021-06-02 15:14:25,737 WARN org.apache.hadoop.hdfs.server.datanode.checker.StorageLocationChecker: Exception checking StorageLocation [DISK]file:/home/hdoop/dfsdata/datanode
EPERM: Operation not permitted
at org.apache.hadoop.io.nativeio.NativeIO$POSIX.chmodImpl(Native Method)
at org.apache.hadoop.io.nativeio.NativeIO$POSIX.chmod(NativeIO.java:382)
at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:863)
at org.apache.hadoop.fs.ChecksumFileSystem$1.apply(ChecksumFileSystem.java:510)
at org.apache.hadoop.fs.ChecksumFileSystem$FsOperation.run(ChecksumFileSystem.java:491)
at org.apache.hadoop.fs.ChecksumFileSystem.setPermission(ChecksumFileSystem.java:513)
at org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:234)
at org.apache.hadoop.util.DiskChecker.checkDirInternal(DiskChecker.java:141)
at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:116)
at org.apache.hadoop.hdfs.server.datanode.StorageLocation.check(StorageLocation.java:239)
at org.apache.hadoop.hdfs.server.datanode.StorageLocation.check(StorageLocation.java:52)
at org.apache.hadoop.hdfs.server.datanode.checker.ThrottledAsyncChecker$1.call(ThrottledAsyncChecker.java:142)
at com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
at com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
at com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2021-06-02 15:14:25,740 ERROR org.apache.hadoop.hdfs.server.datanode.DataNode: Exception in secureMain
org.apache.hadoop.util.DiskChecker$DiskErrorException: Too many failed volumes - current valid volumes: 0, volumes configured: 1, volumes failed: 1, volume failures tolerated: 0
at org.apache.hadoop.hdfs.server.datanode.checker.StorageLocationChecker.check(StorageLocationChecker.java:231)
at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2806)
at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2721)
at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:2763)
at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2907)
at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2931)
2021-06-02 15:14:25,742 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 1: org.apache.hadoop.util.DiskChecker$DiskErrorException: Too many failed volumes - current valid volumes: 0, volumes configured: 1, volumes failed: 1, volume failures tolerated: 0
2021-06-02 15:14:25,760 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at ip-172-31-94-192.ec2.internal/172.31.94.192
************************************************************/以前的线程对我来说有点混乱,因为它们指示不同的文件夹名称和类似的东西,可能是由于使用以前版本的hadoop和/或不同的设置过程。
编辑:还尝试了stop-all,更改datanode和namenode上的权限,以及start-all。似乎什么都没改变。
编辑2:这个线程实际上看起来非常相似:Hadoop datanode not running
发布于 2021-06-03 00:42:13
这个帖子提供了答案:Hadoop datanode not running
我已经创建了dfsdata文件夹作为根帐户,而不是hdoop帐户,所以只需要修改-R并“重新启动”hadoop进程。
https://stackoverflow.com/questions/67808417
复制相似问题