diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeHdfsFileSystemContract.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeHdfsFileSystemContract.java index dcfa051c3902d..3d0e705e6a98c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeHdfsFileSystemContract.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeHdfsFileSystemContract.java @@ -103,7 +103,7 @@ protected String getDefaultWorkingDirectory() { } @Override - @Test + @Test(timeout = 60000) public void testAppend() throws IOException { AppendTestUtil.testAppend(fs, new Path("/append/f")); } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeWithHdfsScheme.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeWithHdfsScheme.java index 650a4722798e8..76eee6d727c06 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeWithHdfsScheme.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/fs/viewfs/TestViewFileSystemOverloadSchemeWithHdfsScheme.java @@ -496,7 +496,7 @@ private boolean isFallBackExist(Configuration config) { * When InnerCache disabled, all matching ViewFileSystemOverloadScheme * initialized scheme file systems would not use FileSystem cache. */ - @Test(timeout = 3000) + @Test(timeout = 30000) public void testViewFsOverloadSchemeWithNoInnerCacheAndHdfsTargets() throws Exception { final Path hdfsTargetPath = new Path(defaultFSURI + HDFS_USER_FOLDER); @@ -523,7 +523,7 @@ public void testViewFsOverloadSchemeWithNoInnerCacheAndHdfsTargets() * initialized scheme file systems should continue to take advantage of * FileSystem cache. */ - @Test(timeout = 3000) + @Test(timeout = 30000) public void testViewFsOverloadSchemeWithNoInnerCacheAndLocalSchemeTargets() throws Exception { final Path localTragetPath = new Path(localTargetDir.toURI()); @@ -545,7 +545,7 @@ public void testViewFsOverloadSchemeWithNoInnerCacheAndLocalSchemeTargets() /** * Tests the rename with nfly mount link. */ - @Test(timeout = 3000) + @Test(timeout = 30000) public void testNflyRename() throws Exception { final Path hdfsTargetPath1 = new Path(defaultFSURI + HDFS_USER_FOLDER); final Path hdfsTargetPath2 = new Path(defaultFSURI + HDFS_USER_FOLDER + 1); @@ -577,7 +577,7 @@ public void testNflyRename() throws Exception { /** * Tests the write and read contents with nfly mount link. */ - @Test(timeout = 3000) + @Test(timeout = 30000) public void testNflyWriteRead() throws Exception { final Path hdfsTargetPath1 = new Path(defaultFSURI + HDFS_USER_FOLDER); final Path hdfsTargetPath2 = new Path(defaultFSURI + HDFS_USER_FOLDER + 1); @@ -604,7 +604,7 @@ public void testNflyWriteRead() throws Exception { * target file. 3. Tests the read works with repairOnRead flag. 4. Tests that * previously deleted file fully recovered and exists. */ - @Test(timeout = 3000) + @Test(timeout = 30000) public void testNflyRepair() throws Exception { final NflyFSystem.NflyKey repairKey = NflyFSystem.NflyKey.repairOnRead; final Path hdfsTargetPath1 = new Path(defaultFSURI + HDFS_USER_FOLDER); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java index 6da46dec378d2..3a8528968dc6c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java @@ -63,7 +63,7 @@ protected String getDefaultWorkingDirectory() { return defaultWorkingDirectory; } - @Test + @Test(timeout = 60000) public void testAppend() throws IOException { AppendTestUtil.testAppend(fs, new Path("/testAppend/f")); } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSnapshotCommands.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSnapshotCommands.java index 32ac2980cd590..52bc12f4d8098 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSnapshotCommands.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestSnapshotCommands.java @@ -224,7 +224,7 @@ public void testSnapshotCommandsWithURI()throws Exception { fs.delete(new Path("/Fully/QPath"), true); } - @Test (timeout=60000) + @Test (timeout=120000) public void testSnapshotDiff()throws Exception { Configuration config = new HdfsConfiguration(); Path snapDirPath = new Path(fs.getUri().toString() + "/snap_dir"); diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/command/TestDiskBalancerCommand.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/command/TestDiskBalancerCommand.java index 1d6331d83c558..d0107b507ade9 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/command/TestDiskBalancerCommand.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/diskbalancer/command/TestDiskBalancerCommand.java @@ -158,7 +158,7 @@ public void testSubmitPlanInNonRegularStatus() throws Exception { * Tests running multiple commands under on setup. This mainly covers * {@link org.apache.hadoop.hdfs.server.diskbalancer.command.Command#close} */ - @Test(timeout = 60000) + @Test(timeout = 120000) public void testRunMultipleCommandsUnderOneSetup() throws Exception { final int numDatanodes = 1; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNamenodeStorageDirectives.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNamenodeStorageDirectives.java index 5bcb3a571a75c..0469b91988451 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNamenodeStorageDirectives.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNamenodeStorageDirectives.java @@ -192,7 +192,7 @@ private void testStorageTypes(StorageType[][] storageTypes, * Types. * @throws IOException */ - @Test(timeout=60000) + @Test(timeout=120000) public void testTargetStorageTypes() throws ReconfigurationException, InterruptedException, TimeoutException, IOException { // DISK and not anything else. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHdfsFileSystemContract.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHdfsFileSystemContract.java index 722ebad72d239..6821d9b431f87 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHdfsFileSystemContract.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/web/TestWebHdfsFileSystemContract.java @@ -396,7 +396,7 @@ public void testOffsetPlusLengthParamsLongerThanFile() throws IOException { } } - @Test + @Test(timeout = 60000) public void testResponseCode() throws IOException { final WebHdfsFileSystem webhdfs = (WebHdfsFileSystem)fs; final Path root = new Path("/");