Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
31 changes: 1 addition & 30 deletions hadoop-ozone/recon/dev-support/findbugsExcludeFile.xml
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
limitations under the License.
-->
<FindBugsFilter>
<!-- Generated Code -->
<Match>
<Package name="org.hadoop.ozone.recon.schema"/>
</Match>
Expand All @@ -25,34 +26,4 @@
<Match>
<Package name="org.hadoop.ozone.recon.schema.tables.pojos"/>
</Match>

<!-- Test -->
<Match>
<Class name="org.apache.hadoop.ozone.recon.TestReconUtils"/>
<Bug pattern="RV_RETURN_VALUE_IGNORED_BAD_PRACTICE" />
</Match>
<Match>
<Class name="org.apache.hadoop.ozone.recon.recovery.TestReconOmMetadataManagerImpl"/>
<Bug pattern="RV_RETURN_VALUE_IGNORED_BAD_PRACTICE" />
</Match>
<Match>
<Class name="org.apache.hadoop.ozone.recon.spi.impl.TestOzoneManagerServiceProviderImpl"/>
<Bug pattern="NP_NULL_ON_SOME_PATH_FROM_RETURN_VALUE" />
</Match>
<Match>
<Class name="org.apache.hadoop.ozone.recon.spi.impl.TestOzoneManagerServiceProviderImpl"/>
<Bug pattern="OBL_UNSATISFIED_OBLIGATION" />
</Match>
<Match>
<Class name="org.apache.hadoop.ozone.recon.spi.impl.TestOzoneManagerServiceProviderImpl"/>
<Bug pattern="RV_RETURN_VALUE_IGNORED_BAD_PRACTICE" />
</Match>
<Match>
<Class name="~org.apache.hadoop.ozone.recon.tasks.TestNSSummaryTask.*"/>
<Bug pattern="SIC_INNER_SHOULD_BE_STATIC" /> <!-- Only non-static inner classes can be @Nested -->
</Match>
<Match>
<Class name="org.apache.hadoop.ozone.recon.TestReconUtils"/>
<Bug pattern="OBL_UNSATISFIED_OBLIGATION_EXCEPTION_EDGE" />
</Match>
</FindBugsFilter>
Original file line number Diff line number Diff line change
Expand Up @@ -27,13 +27,9 @@
import static org.mockito.Mockito.mock;
import static org.mockito.Mockito.when;

import java.io.BufferedWriter;
import java.io.File;
import java.io.FileInputStream;
import java.io.FileOutputStream;
import java.io.IOException;
import java.io.InputStream;
import java.io.OutputStreamWriter;
import java.net.HttpURLConnection;
import java.net.URL;
import java.nio.charset.Charset;
Expand All @@ -56,7 +52,7 @@
* Test Recon Utility methods.
*/
public class TestReconUtils {
private static PipelineID randomPipelineID = PipelineID.randomId();
private static final PipelineID RANDOM_PIPELINE_ID = PipelineID.randomId();

@TempDir
private Path temporaryFolder;
Expand All @@ -76,31 +72,18 @@ public void testGetReconDbDir() throws Exception {
public void testCreateTarFile(@TempDir File tempSnapshotDir)
throws Exception {

FileInputStream fis = null;
FileOutputStream fos = null;
File tarFile = null;

try {
String testDirName = tempSnapshotDir.getPath();

File file = new File(testDirName + "/temp1.txt");
OutputStreamWriter writer = new OutputStreamWriter(
new FileOutputStream(file), UTF_8);
writer.write("Test data 1");
writer.close();

file = new File(testDirName + "/temp2.txt");
writer = new OutputStreamWriter(
new FileOutputStream(file), UTF_8);
writer.write("Test data 2");
writer.close();
FileUtils.write(new File(testDirName + "/temp1.txt"), "Test data 1", UTF_8);
FileUtils.write(new File(testDirName + "/temp2.txt"), "Test data 2", UTF_8);

tarFile = createTarFile(Paths.get(testDirName));
assertNotNull(tarFile);

} finally {
org.apache.hadoop.io.IOUtils.closeStream(fis);
org.apache.hadoop.io.IOUtils.closeStream(fos);
FileUtils.deleteDirectory(tempSnapshotDir);
FileUtils.deleteQuietly(tarFile);
}
Expand All @@ -113,19 +96,11 @@ public void testUntarCheckpointFile() throws Exception {
temporaryFolder.resolve("NewDir")).toFile();
File file1 = Paths.get(newDir.getAbsolutePath(), "file1")
.toFile();
String str = "File1 Contents";
BufferedWriter writer = new BufferedWriter(new OutputStreamWriter(
new FileOutputStream(file1.getAbsoluteFile()), UTF_8));
writer.write(str);
writer.close();
FileUtils.write(file1, "File1 Contents", UTF_8);

File file2 = Paths.get(newDir.getAbsolutePath(), "file2")
.toFile();
str = "File2 Contents";
writer = new BufferedWriter(new OutputStreamWriter(
new FileOutputStream(file2.getAbsoluteFile()), UTF_8));
writer.write(str);
writer.close();
FileUtils.write(file2, "File2 Contents", UTF_8);

//Create test tar file.
File tarFile = createTarFile(newDir.toPath());
Expand All @@ -142,54 +117,39 @@ public void testMakeHttpCall() throws Exception {
String url = "http://localhost:9874/dbCheckpoint";
File file1 = Paths.get(temporaryFolder.toString(), "file1")
.toFile();
BufferedWriter writer = new BufferedWriter(new OutputStreamWriter(
new FileOutputStream(file1.getAbsoluteFile()), UTF_8));
writer.write("File 1 Contents");
writer.close();
InputStream fileInputStream = new FileInputStream(file1);

String contents;
URLConnectionFactory connectionFactoryMock =
mock(URLConnectionFactory.class);
HttpURLConnection urlConnectionMock = mock(HttpURLConnection.class);
when(urlConnectionMock.getInputStream()).thenReturn(fileInputStream);
when(connectionFactoryMock.openConnection(any(URL.class), anyBoolean()))
.thenReturn(urlConnectionMock);
try (InputStream inputStream = new ReconUtils()
.makeHttpCall(connectionFactoryMock, url, false).getInputStream()) {
contents = IOUtils.toString(inputStream, Charset.defaultCharset());
FileUtils.write(file1, "File 1 Contents", UTF_8);
try (InputStream fileInputStream = Files.newInputStream(file1.toPath())) {

String contents;
URLConnectionFactory connectionFactoryMock =
mock(URLConnectionFactory.class);
HttpURLConnection urlConnectionMock = mock(HttpURLConnection.class);
when(urlConnectionMock.getInputStream()).thenReturn(fileInputStream);
when(connectionFactoryMock.openConnection(any(URL.class), anyBoolean()))
.thenReturn(urlConnectionMock);
try (InputStream inputStream = new ReconUtils()
.makeHttpCall(connectionFactoryMock, url, false).getInputStream()) {
contents = IOUtils.toString(inputStream, Charset.defaultCharset());
}

assertEquals("File 1 Contents", contents);
}

assertEquals("File 1 Contents", contents);
}

@Test
public void testGetLastKnownDB(@TempDir File newDir) throws IOException {

File file1 = Paths.get(newDir.getAbsolutePath(), "valid_1")
.toFile();
String str = "File1 Contents";
BufferedWriter writer = new BufferedWriter(new OutputStreamWriter(
new FileOutputStream(file1.getAbsoluteFile()), UTF_8));
writer.write(str);
writer.close();
FileUtils.write(file1, "File1 Contents", UTF_8);

File file2 = Paths.get(newDir.getAbsolutePath(), "valid_2")
.toFile();
str = "File2 Contents";
writer = new BufferedWriter(new OutputStreamWriter(
new FileOutputStream(file2.getAbsoluteFile()), UTF_8));
writer.write(str);
writer.close();

FileUtils.write(file2, "File2 Contents", UTF_8);

File file3 = Paths.get(newDir.getAbsolutePath(), "invalid_3")
.toFile();
str = "File3 Contents";
writer = new BufferedWriter(new OutputStreamWriter(
new FileOutputStream(file3.getAbsoluteFile()), UTF_8));
writer.write(str);
writer.close();
FileUtils.write(file3, "File3 Contents", UTF_8);

ReconUtils reconUtils = new ReconUtils();
File latestValidFile = reconUtils.getLastKnownDB(newDir, "valid");
Expand Down Expand Up @@ -255,7 +215,7 @@ private static ContainerInfo.Builder getDefaultContainerInfoBuilder(
public static ContainerInfo getContainer(
final HddsProtos.LifeCycleState state) {
return getDefaultContainerInfoBuilder(state)
.setPipelineID(randomPipelineID)
.setPipelineID(RANDOM_PIPELINE_ID)
.build();
}
}
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ public void testStart() throws Exception {
File snapshotFile = new File(
checkpoint.getCheckpointLocation().getParent() + "/" +
"om.snapshot.db_" + System.currentTimeMillis());
checkpoint.getCheckpointLocation().toFile().renameTo(snapshotFile);
assertTrue(checkpoint.getCheckpointLocation().toFile().renameTo(snapshotFile));

//Create new Recon OM Metadata manager instance.
File reconOmDbDir = Files.createDirectory(
Expand Down
Loading