Skip to content

Conversation

@unknowntpo
Copy link
Contributor

What changes were proposed in this pull request?

HDDS-13168. Fix content type error response format in CheckUploadContentTypeFilter.java

The CheckUploadContentTypeFilter was modified to return properly formatted JSON error responses instead of using the generic sendError() method. When a data upload request lacks the required application/octet-stream content-type, the filter now returns a consistent JSON error response with the correct application/json content-type header.

This change ensures API consistency across all WebHDFS endpoints and makes error responses easier for clients to parse programmatically. Previously, error responses from this filter were not properly formatted as JSON, creating inconsistency with other API responses.

The specific changes include:

  • Replaced the sendError() call with a proper JSON response construction
  • Set appropriate HTTP status code (400 Bad Request)
  • Ensured the response content-type is application/json;charset=utf-8
  • Maintained the same error message content for backward compatibility.

What is the link to the Apache JIRA

https://issues.apache.org/jira/browse/HDDS-13168

How was this patch tested?

  1. Build distribution
mvn clean package -DskipTests=true
  1. Run docker compose
cd ./hadoop-ozone/dist/target/ozone-2.1.0-SNAPSHOT/compose/ozone
OZONE_DATANODES=3 ./run.sh -d
  1. Create a file called test.txt in the local directory
echo 'hello world' > export.txt
  1. Send request without content-type
curl -v -X PUT -T ./export.txt "http://localhost:14000/webhdfs/v1/exports/text/export.txt?op=CREATE&data=true&user.name=hadoop"
  1. Verify that response code is Bad Request (400), and the error message is:

and the response will be:

curl -v -X PUT -T ./test.txt  "http://localhost:14000/webhdfs/v1/exports/text/export.txt?op=CREATE&data=true&user.name=hadoop"
Note: Unnecessary use of -X or --request, PUT is already inferred.
* Host localhost:14000 was resolved.
* IPv6: ::1
* IPv4: 127.0.0.1
*   Trying [::1]:14000...
* Connected to localhost (::1) port 14000
> PUT /webhdfs/v1/exports/text/export.txt?op=CREATE&data=true&user.name=hadoop HTTP/1.1
> Host: localhost:14000
> User-Agent: curl/8.7.1
> Accept: */*
> Content-Length: 12
> 
* upload completely sent off: 12 bytes
< HTTP/1.1 400 Bad Request
< Date: Thu, 12 Jun 2025 15:33:13 GMT
< Cache-Control: no-cache
< Expires: Thu, 12 Jun 2025 15:33:13 GMT
< Pragma: no-cache
< X-Content-Type-Options: nosniff
< X-XSS-Protection: 1; mode=block
< Set-Cookie: hadoop.auth="u=hadoop&p=hadoop&t=simple-dt&e=1749778393435&s=qrnNFoyNNWT+lE4tzPx6ke2XMpGUUH4Eb7W13zdodYQ="; Path=/; HttpOnly
< Content-Type: application/json;charset=utf-8
< Transfer-Encoding: chunked
< 
* Connection #0 to host localhost left intact
{"error":"Data upload requests must have content-type set to 'application/octet-stream'"}

Which has status code Bad Request (400) and JSON response will be:

{"error":"Data upload requests must have content-type set to 'application/octet-stream'"}
  1. Now, send request with content-type set to application/octet-stream
curl -v -X PUT -T ./export.txt -H"Content-Type: application/octet-stream" "http://localhost:14000/webhdfs/v1/exports/text/export.txt?op=CREATE&data=true&user.name=hadoop"

and the response code will be 201

Note: Unnecessary use of -X or --request, PUT is already inferred.
* Host localhost:14000 was resolved.
* IPv6: ::1
* IPv4: 127.0.0.1
*   Trying [::1]:14000...
* Connected to localhost (::1) port 14000
> PUT /webhdfs/v1/exports/text/export.txt?op=CREATE&data=true&user.name=hadoop HTTP/1.1
> Host: localhost:14000
> User-Agent: curl/8.7.1
> Accept: */*
> Content-Type: application/octet-stream
> Content-Length: 12
> 
* upload completely sent off: 12 bytes
< HTTP/1.1 201 Created
< Date: Thu, 12 Jun 2025 15:30:48 GMT
< Cache-Control: no-cache
< Expires: Thu, 12 Jun 2025 15:30:48 GMT
< Pragma: no-cache
< X-Content-Type-Options: nosniff
< X-XSS-Protection: 1; mode=block
< Set-Cookie: hadoop.auth="u=hadoop&p=hadoop&t=simple-dt&e=1749778248717&s=m66Bktyw7AXq6xuDt64EdpF9llvWFGAt7tx0K64Fmd8="; Path=/; HttpOnly
< Location: http://localhost:14000/webhdfs/v1/exports/text/export.txt
< Content-Type: application/json
< Content-Length: 72
< 
* Connection #0 to host localhost left intact
{"Location":"http://localhost:14000/webhdfs/v1/exports/text/export.txt"}

Which has status code Created (201) and JSON response will be:

{"Location":"http://localhost:14000/webhdfs/v1/exports/text/export.txt"}

…entTypeFilter

Changed from sendError() to proper JSON response format to ensure correct content-type
is sent to client when upload content-type validation fails.
@unknowntpo unknowntpo marked this pull request as ready for review June 12, 2025 15:53
@unknowntpo
Copy link
Contributor Author

@peterxcli Would you like to review this PR?

Copy link
Member

@peterxcli peterxcli left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @unknowntpo for this quick patch!

And since current way is quite verbose, we can have add a utility to do this for other sendError calls in follow-up. (I guess the others get the same issue which the message we set for them doesnt show in the response.

image

@unknowntpo @adoroszlai WDYT?

@adoroszlai adoroszlai changed the title HDDS-13168. Fix content type error response format in CheckUploadContentTypeFilter.java HDDS-13168. Fix error response format in CheckUploadContentTypeFilter Jun 12, 2025
@unknowntpo
Copy link
Contributor Author

Thanks @unknowntpo for this quick patch!

And since current way is quite verbose, we can have add a utility to do this for other sendError calls in follow-up. (I guess the others get the same issue which the message we set for them doesnt show in the response.

image @unknowntpo @adoroszlai WDYT?

Agree, and in order to give user consistent style of response, we may also need to define a response model, specify the schema of the response JSON object.

Comment on lines 104 to 110
// Create JSON response
String jsonResponse = "{\"error\":\"" + errorMessage + "\"}";

PrintWriter writer = httpRes.getWriter();
writer.write(jsonResponse);
writer.flush();
writer.close();
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please consider using a utility library to craft JSON output, for example, Jackson:

import com.fasterxml.jackson.databind.ObjectMapper;
import java.util.HashMap;
import java.util.Map;

Map<String, String> errorMap = new HashMap<>();
errorMap.put("error", errorMessage);

ObjectMapper mapper = new ObjectMapper();
String jsonResponse = mapper.writeValueAsString(errorMap);

PrintWriter writer = httpRes.getWriter();
writer.write(jsonResponse);

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I use JsonUtil to do this, which use Jackson under the hood.

@adoroszlai adoroszlai merged commit 1e47217 into apache:master Jun 13, 2025
42 checks passed
@adoroszlai
Copy link
Contributor

Thanks @unknowntpo for the patch, @jojochuang, @peterxcli for the review.

And since current way is quite verbose, we can have add a utility to do this for other sendError calls in follow-up. (I guess the others get the same issue which the message we set for them doesnt show in the response.

@peterxcli can you please file a task for that?

@peterxcli
Copy link
Member

peterxcli commented Jun 13, 2025

@unknowntpo unknowntpo deleted the HDDS-13168-wrong-content-type-not-send-to-client branch June 13, 2025 07:49
sadanand48 pushed a commit to sadanand48/hadoop-ozone that referenced this pull request Jun 16, 2025
aswinshakil added a commit to aswinshakil/ozone that referenced this pull request Jun 20, 2025
…239-container-reconciliation

Commits: 62

da53b5b HDDS-13299. Fix failures related to delete (apache#8665)
8c1b439 HDDS-13296. Integration check always passes due to missing output (apache#8662)
7329859 HDDS-13023. Container checksum is missing after container import (apache#8459)
a0af93e HDDS-13292. Change `<? extends KeyValue>` to `<KeyValue>` in test (apache#8657)
f3050cf HDDS-13276. Use KEY_ONLY/VALUE_ONLY iterator in SCM/Datanode. (apache#8638)
e9c0a45 HDDS-13262. Simplify key name validation (apache#8619)
f713e57 HDDS-12482. Avoid using CommonConfigurationKeys (apache#8647)
b574709 HDDS-12924. datanode used space calculation optimization (apache#8365)
de683aa HDDS-13263. Refactor DB Checkpoint Utilities. (apache#8620)
97262aa HDDS-13256. Updated OM Snapshot Grafana Dashboard to reflect metric updates from HDDS-13181. (apache#8639)
9d2b415 HDDS-13234. Expired secret key can abort leader OM startup. (apache#8601)
d9049a2 HDDS-13220. Change Recon 'Negative usedBytes' message loglevel to DEBUG (apache#8648)
6df3077 HDDS-9223. Use protobuf for SnapshotDiffJobCodec (apache#8503)
a7fc290 HDDS-13236. Change Table methods not to throw IOException. (apache#8645)
9958f5b HDDS-13287. Upgrade commons-beanutils to 1.11.0 due to CVE-2025-48734 (apache#8646)
48aefea HDDS-13277. [Docs] Native C/C++ Ozone clients (apache#8630)
052d912 HDDS-13037. Let container create command support STANDALONE , RATIS and EC containers (apache#8559)
90ed60b HDDS-13279. Skip verifying Apache Ranger binaries in CI (apache#8633)
9bc53b2 HDDS-11513. All deletion configurations should be configurable without restart (apache#8003)
ac511ac HDDS-13259. Deletion Progress - Grafana Dashboard (apache#8617)
3370f42 HDDS-13246. Change `<? extend KeyValue>` to `<KeyValue>` in hadoop-hdds (apache#8631)
7af8c44 HDDS-11454. Ranger integration for Docker Compose environment (apache#8575)
5a3e4e7 HDDS-13273. Bump awssdk to 2.31.63 (apache#8626)
77138b8 HDDS-13254. Change table iterator to optionally read key or value. (apache#8621)
ce288b6 HDDS-13265. Simplify the page Access Ozone using HTTPFS REST API (apache#8629)
36fe888 HDDS-13275. Improve CheckNative implementation (apache#8628)
d38484e HDDS-13274. Bump sqlite-jdbc to 3.50.1.0 (apache#8627)
3f3ec43 HDDS-13266. `ozone debug checknative` to show OpenSSL lib (apache#8623)
8983a63 HDDS-13272. Bump junit to 5.13.1 (apache#8625)
a927113 HDDS-13271. [Docs] Minor text updates, reference links. (apache#8624)
7e77058 HDDS-13112. [Docs] OM Bootstrap can also happen when follower falls behind too much. (apache#8600)
fd13300 HDDS-10775. Support bucket ownership verification (apache#8558)
3ecf345 HDDS-13207. [Docs] Third party systems compatible with Ozone S3. (apache#8584)
ad5a507 HDDS-13035. SnapshotDeletingService should hold write locks while purging deleted snapshots (apache#8554)
38a9186 HDDS-12637. Increase max buffer size for tar entry read/write (apache#8618)
f31c264 HDDS-13045. Implement Immediate Triggering of Heartbeat when Volume Full (apache#8590)
0701d6a HDDS-13248. Remove `ozone debug replicas verify` option --output-dir (apache#8612)
ca1afe8 HDDS-13257. Remove separate split for shell integration tests (apache#8616)
5d6fe94 HDDS-13216. Standardize Container[Replica]NotFoundException messages (apache#8599)
1e47217 HDDS-13168. Fix error response format in CheckUploadContentTypeFilter (apache#8614)
6d4d423 HDDS-13181. Added metrics for internal Snapshot Operations. (apache#8606)
4a461b2 HDDS-10490. Intermittent NPE in TestSnapshotDiffManager#testLoadJobsOnStartUp (apache#8596)
bf29f7f HDDS-13235. The equals/hashCode methods in anonymous KeyValue classes may not work. (apache#8607)
6ff3ad6 HDDS-12873. Improve ContainerData statistics synchronization. (apache#8305)
09d3b27 HDDS-13244. TestSnapshotDeletingServiceIntegrationTest should close snapshots after deleting them (apache#8611)
931bc2d HDDS-13243. copy-rename-maven-plugin version is missing (apache#8605)
3b5985c HDDS-13244. Disable TestSnapshotDeletingServiceIntegrationTest
6bf009c HDDS-12927. metrics and log to indicate datanode crossing disk limits (apache#8573)
752da2b HDDS-12760. Intermittent Timeout in testImportedContainerIsClosed (apache#8349)
8c32363 HDDS-13050. Update StartFromDockerHub.md. (apache#8586)
ba1887c HDDS-13241. Fix some potential resource leaks (apache#8602)
bbaf71e HDDS-13130. Rename all instances of Disk Usage to Namespace usage (apache#8571)
0628386 HDDS-13142. Correct SCMPerformanceMetrics for delete operation. (apache#8592)
516bc96 HDDS-13148. [Docs] Update Transparent Data Encryption doc. (apache#8530)
5787135 HDDS-13229. [Doc] Fix incorrect CLI argument order in OM upgrade docs (apache#8598)
ba95074 HDDS-13107. Support limiting output of `ozone admin datanode list` (apache#8595)
e7f5544 HDDS-13171. Replace pipelineID if nodes are changed (apache#8562)
3c9d4d8 HDDS-13103. Correct transaction metrics in SCMBlockDeletingService. (apache#8516)
f62eb8a HDDS-13160. Remove SnapshotDirectoryCleaningService and refactor AbstractDeletingService (apache#8547)
b46e6b2 HDDS-13150. Fixed SnapshotLimitCheck when failures occur. (apache#8532)
203c1d3 HDDS-13206. Update documentation for Apache Ranger (apache#8583)
2072ef0 HDDS-13214. populate-cache fails due to unused dependency (apache#8594)

Conflicts:
	hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/impl/ContainerData.java
	hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/KeyValueContainer.java
	hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/helpers/KeyValueContainerUtil.java
	hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/statemachine/background/BlockDeletingTask.java
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants