-
Notifications
You must be signed in to change notification settings - Fork 87
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add "what" property for migration to scope down table migrations. #856
Conversation
Codecov ReportAttention:
Additional details and impacted files@@ Coverage Diff @@
## main #856 +/- ##
==========================================
+ Coverage 85.75% 85.85% +0.10%
==========================================
Files 42 42
Lines 5369 5394 +25
Branches 971 978 +7
==========================================
+ Hits 4604 4631 +27
+ Misses 547 546 -1
+ Partials 218 217 -1 ☔ View full report in Codecov by Sentry. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
keep PR small, let's focus on Scope
.
fe5d51b
to
17be05d
Compare
17be05d
to
ebb0e87
Compare
table_a = Table("a", "b", "c", "MANAGED", "DELTA", location="dbfs:/somelocation/tablename") | ||
assert table_a.is_dbfs_root | ||
assert table_a.what == What.DBFS_ROOT_DELTA | ||
table_b = Table("a", "b", "c", "MANAGED", "PARQUET", location="dbfs:/somelocation/tablename") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
what about
- PARQUET & external?
- JSON & external?
- CSV & external?
- TEXT & external?
- avro (whatever other formats) & external?
- what about the same formats & DBFS?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
External tables are divided to supported and unsupported.
All the unsupported are marked "unknown"
It may be a good idea to introduce EXTERNAL_SYNC and EXTERNAL_NO_SYNC or something like that.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't want to replace the table_format I want the field to highlight what kind of table it is when it comes to migration.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I added EXTERNAL_SYNC and EXTERNAL_NO_SYNC to correspond to the supported/unsupported formats.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm
* Added "what" property for migration to scope down table migrations ([#856](#856)). * Added job count in the assessment dashboard ([#858](#858)). * Adopted `installation` package from `databricks-labs-blueprint` ([#860](#860)). * Debug logs to print only the first 96 bytes of SQL query by default, tunable by `debug_truncate_bytes` SDK configuration property ([#859](#859)). * Extract command codes and unify the checks for spark_conf, cluster_policy, init_scripts ([#855](#855)). * Improved installation failure with actionable message ([#840](#840)). * Improved validating groups membership cli command ([#816](#816)). Dependency updates: * Updated databricks-labs-blueprint requirement from ~=0.1.0 to ~=0.2.4 ([#867](#867)).
* Added "what" property for migration to scope down table migrations ([#856](#856)). * Added job count in the assessment dashboard ([#858](#858)). * Adopted `installation` package from `databricks-labs-blueprint` ([#860](#860)). * Debug logs to print only the first 96 bytes of SQL query by default, tunable by `debug_truncate_bytes` SDK configuration property ([#859](#859)). * Extract command codes and unify the checks for spark_conf, cluster_policy, init_scripts ([#855](#855)). * Improved installation failure with actionable message ([#840](#840)). * Improved validating groups membership cli command ([#816](#816)). Dependency updates: * Updated databricks-labs-blueprint requirement from ~=0.1.0 to ~=0.2.4 ([#867](#867)).
) ## Changes <!-- Summary of your changes that are easy to understand. Add screenshots when necessary --> ### Linked issues related to #333 Resolves #.. ### Functionality - [ ] added relevant user documentation - [ ] added new CLI command - [ ] modified existing command: `databricks labs ucx ...` - [ ] added a new workflow - [ ] modified existing workflow: `...` - [ ] added a new table - [ ] modified existing table: `...` ### Tests <!-- How is this tested? Please see the checklist below and also describe any other relevant tests --> - [ ] manually tested - [ ] added unit tests - [ ] added integration tests - [ ] verified on staging environment (screenshot attached)
* Added "what" property for migration to scope down table migrations ([#856](#856)). * Added job count in the assessment dashboard ([#858](#858)). * Adopted `installation` package from `databricks-labs-blueprint` ([#860](#860)). * Debug logs to print only the first 96 bytes of SQL query by default, tunable by `debug_truncate_bytes` SDK configuration property ([#859](#859)). * Extract command codes and unify the checks for spark_conf, cluster_policy, init_scripts ([#855](#855)). * Improved installation failure with actionable message ([#840](#840)). * Improved validating groups membership cli command ([#816](#816)). Dependency updates: * Updated databricks-labs-blueprint requirement from ~=0.1.0 to ~=0.2.4 ([#867](#867)).
Changes
Linked issues
related to #333
Resolves #..
Functionality
databricks labs ucx ...
...
...
Tests