[ https://issues.apache.org/jira/browse/HDFS-13538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=18016955#comment-18016955 ]
ASF GitHub Bot commented on HDFS-13538: --------------------------------------- violetnspct commented on PR #7915: URL: https://github.com/apache/hadoop/pull/7915#issuecomment-3235946882 @aswinmprabhu ## Suggested Unit Test Scenarios ### File: `hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/DiskChecker.java` **Method:** `doDiskIo` **Recommended Test Scenarios:** 1. Test handling of actual disk failure error message - should be classified as disk failure 2. Test handling of mixed error case where disk is both full and has other issues 3. Test with null/empty error message - should handle gracefully 4. Test with different variations of 'No space left' error messages (case sensitivity, different formats) ### File: `hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java` **Method:** `checkDirs` **Recommended Test Scenarios:** 1. Test that when checkDirWithDiskIo flag is disabled, the method calls checkDir() 2. Test that the checkDirWithDiskIo configuration flag properly reads from the configuration settings 3. Test that the method handles null configuration gracefully 4. Test that changing the configuration flag value at runtime properly switches the checking method **Edge Cases to Cover:** 1. Configuration flag is false and performs regular checks without disk I/O 2. Same directory path checked with both methods produces consistent results > HDFS DiskChecker should handle disk full situation > -------------------------------------------------- > > Key: HDFS-13538 > URL: https://issues.apache.org/jira/browse/HDFS-13538 > Project: Hadoop HDFS > Issue Type: Bug > Reporter: Kihwal Lee > Assignee: Arpit Agarwal > Priority: Critical > Labels: pull-request-available > Attachments: HDFS-13538.01.patch > > > Fix disk checker issues reported by [~kihwal] in HADOOP-13738: > When space is low, the os returns ENOSPC. Instead simply stop writing, the > drive is marked bad and replication happens. This make cluster-wide space > problem worse. If the number of "failed" drives exceeds the DFIP limit, the > datanode shuts down. -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org