What is the biggest problem of extremely large hadoop cluster ?
Hi , I am curious to know what is the biggest problem of extremely large hadoop cluster. What I can imagine now is the memory cost of meta data of hdfs in name node. One solution I can think about now is to use other storage implementation such as database to store the metadata, although it has performance cost. Is there any other solutions or any problems of extremely large hadoop cluster ? -- Best Regards Jeff Zhang
[jira] Created: (HDFS-994) Provide methods for obtaining delegation token from Namenode for hftp and other uses
Provide methods for obtaining delegation token from Namenode for hftp and other uses Key: HDFS-994 URL: https://issues.apache.org/jira/browse/HDFS-994 Project: Hadoop HDFS Issue Type: Improvement Reporter: Jakob Homan Assignee: Jakob Homan In hftp, destination clusters will require an RPC-version-agnostic means of obtaining delegation tokens from the source cluster. The easiest method is provide a webservice to retrieve a token over http. This can be encrypted via SSL (backed by Kerberos, done in another JIRA), providing security for cross-cluster hftp operations. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HDFS-995) Replace usage of FileStatus#isDir()
Replace usage of FileStatus#isDir() --- Key: HDFS-995 URL: https://issues.apache.org/jira/browse/HDFS-995 Project: Hadoop HDFS Issue Type: Improvement Components: name-node Affects Versions: 0.22.0 Reporter: Eli Collins Assignee: Eli Collins Fix For: 0.22.0 HADOOP-6585 is going to deprecate FileStatus#isDir(). This jira is for replacing all uses of isDir() in HDFS with checks of isDirectory(), isFile(), or isSymlink() as needed. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.