10 Temmuz 2015 Cuma

Hadoop on Solaris - Namenode Problem


For a project we need to install Hadoop on Solaris servers.

After installation we have faced with the following error in the namenode log;

2015-07-06 14:11:14,834 INFO org.apache.hadoop.hdfs.server.namenode.FSEditLog: Number of transactions: 2 Total time for transactions(ms): 2 Number of transactions batched in Syncs: 0 Number of syncs: 3 SyncTimes(ms): 576
2015-07-06 14:11:14,843 INFO org.apache.hadoop.hdfs.server.namenode.FileJournalManager: Finalizing edits file /data/hdfs_name/current/edits_inprogress_0000000000000000005 -> /data/hdfs_name/current/edits_0000000000000000005-0000000000000000006
2015-07-06 14:11:14,851 INFO org.apache.hadoop.ipc.Server: Stopping server on 19000
2015-07-06 14:11:14,937 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Stopping services started for active state
2015-07-06 14:11:14,937 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Stopping services started for standby state
2015-07-06 14:11:14,953 INFO org.mortbay.log: Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@0.0.0.0:50070
2015-07-06 14:11:14,965 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Stopping NameNode metrics system...
2015-07-06 14:11:14,968 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system stopped.
2015-07-06 14:11:14,969 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system shutdown complete.
2015-07-06 14:11:15,005 FATAL org.apache.hadoop.hdfs.server.namenode.NameNode: Failed to start namenode.
ExitCodeException exitCode=1:
at org.apache.hadoop.util.Shell.runCommand(Shell.java:538)
at org.apache.hadoop.util.Shell.run(Shell.java:455)
at org.apache.hadoop.fs.DF.getFilesystem(DF.java:76)
at org.apache.hadoop.hdfs.server.namenode.NameNodeResourceChecker$CheckedVolume.<init>(NameNodeResourceChecker.java:69)
at org.apache.hadoop.hdfs.server.namenode.NameNodeResourceChecker.addDirToCheck(NameNodeResourceChecker.java:165)
at org.apache.hadoop.hdfs.server.namenode.NameNodeResourceChecker.<init>(NameNodeResourceChecker.java:134)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startCommonServices(FSNamesystem.java:1103)
at org.apache.hadoop.hdfs.server.namenode.NameNode.startCommonServices(NameNode.java:629)
at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:615)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:762)
at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:746)
at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1438)
at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1504)
2015-07-06 14:11:15,018 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 1
As you can see, it is not a clear error message. When we investigate the root cause , we have found out that it can't execute the following command ;
df -k -P XXXXX

In solaris -P parameter is not supported so I have done a small patch for it in hadoop-common library.It is attached.

My colleagues solution is better he has the linked the default df command to "/usr/xpg4/bin/df" which supports -P  command. Thanks to Can Sevilmis.



Hiç yorum yok:

Yorum Gönder