Skip to content

Conversation

@vitaliili-db
Copy link
Contributor

@vitaliili-db vitaliili-db commented Oct 13, 2022

What changes were proposed in this pull request?

Add a helper object to get SQL config keys for errors.

Why are the changes needed?

Allows injection of vendor specific SQL config keys, e.g. Databricks ansi_mode instead of spark.sql.ansi.enabled

Does this PR introduce any user-facing change?

No

How was this patch tested?

Careful review and existing unit tests.

@MaxGekk
Copy link
Member

MaxGekk commented Oct 13, 2022

Allows injection of vendor specific SQL config keys, e.g. Databricks ansi_mode instead of spark.sql.ansi.enabled

@vitaliili-db Can't you just add an alternative for the config, see for example

.withAlternative("spark.sql.legacy.parquet.int96RebaseModeInWrite")

@github-actions github-actions bot added the SQL label Oct 13, 2022
@vitaliili-db
Copy link
Contributor Author

Allows injection of vendor specific SQL config keys, e.g. Databricks ansi_mode instead of spark.sql.ansi.enabled

@vitaliili-db Can't you just add an alternative for the config, see for example

.withAlternative("spark.sql.legacy.parquet.int96RebaseModeInWrite")

@MaxGekk For defining config withAlternative will work, but this logic is for retrieval. I.e. we want to show correct configuration key in error message depending in which context an error occurred.

@AmplabJenkins
Copy link

Can one of the admins verify this patch?

@MaxGekk
Copy link
Member

MaxGekk commented Oct 19, 2022

I.e. we want to show correct configuration key in error message depending in which context an error occurred.

Can you do the replacement in more centralised way? For example, override the getMessageParameters() method and replace the config there? I just worry that we will forget to wrap the config in the future in some places, and your changes bring additional overhead during review.

dongjoon-hyun pushed a commit that referenced this pull request Jan 23, 2024
### What changes were proposed in this pull request?
This pr aims to upgrade Arrow from 14.0.2 to 15.0.0, this version fixes the compatibility issue with Netty 4.1.104.Final(GH-39265).

Additionally, since the `arrow-vector` module uses `eclipse-collections` to replace `netty-common` as a compile-level dependency, Apache Spark has added a dependency on `eclipse-collections` after upgrading to use Arrow 15.0.0.

### Why are the changes needed?
The new version brings the following major changes:

Bug Fixes
GH-34610 - [Java] Fix valueCount and field name when loading/transferring NullVector
GH-38242 - [Java] Fix incorrect internal struct accounting for DenseUnionVector#getBufferSizeFor
GH-38254 - [Java] Add reusable buffer getters to char/binary vectors
GH-38366 - [Java] Fix Murmur hash on buffers less than 4 bytes
GH-38387 - [Java] Fix JDK8 compilation issue with TestAllTypes
GH-38614 - [Java] Add VarBinary and VarCharWriter helper methods to more writers
GH-38725 - [Java] decompression in Lz4CompressionCodec.java does not set writer index

New Features and Improvements
GH-38511 - [Java] Add getTransferPair(Field, BufferAllocator, CallBack) for StructVector and MapVector
GH-14936 - [Java] Remove netty dependency from arrow-vector
GH-38990 - [Java] Upgrade to flatc version 23.5.26
GH-39265 - [Java] Make it run well with the netty newest version 4.1.104

The full release notes as follows:

- https://arrow.apache.org/release/15.0.0.html

### Does this PR introduce _any_ user-facing change?
No

### How was this patch tested?
Pass GitHub Actions

### Was this patch authored or co-authored using generative AI tooling?
No

Closes #44797 from LuciferYang/SPARK-46718.

Authored-by: yangjie01 <[email protected]>
Signed-off-by: Dongjoon Hyun <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants