Skip to content

Only remove duplicated statistics on error#65653

Merged
balloob merged 2 commits intodevfrom
recorder_statistics_optimize_duplicate_deletion
Feb 4, 2022
Merged

Only remove duplicated statistics on error#65653
balloob merged 2 commits intodevfrom
recorder_statistics_optimize_duplicate_deletion

Conversation

@emontnemery
Copy link
Copy Markdown
Contributor

Proposed change

Only remove duplicated statistics on error

We first try to recreate the statistics indices, which will fail if there are duplicates.
If creating the index fails, remove duplicates then try again

This is done to speed up the migration step

Type of change

  • Dependency upgrade
  • Bugfix (non-breaking change which fixes an issue)
  • New integration (thank you!)
  • New feature (which adds functionality to an existing integration)
  • Breaking change (fix/feature causing existing functionality to break)
  • Code quality improvements to existing code or addition of tests

Additional information

  • This PR fixes or closes issue: fixes #
  • This PR is related to issue:
  • Link to documentation pull request:

Checklist

  • The code change is tested and works locally.
  • Local tests pass. Your PR cannot be merged unless tests pass
  • There is no commented out code in this PR.
  • I have followed the development checklist
  • The code has been formatted using Black (black --fast homeassistant tests)
  • Tests have been added to verify that the new code works.

If user exposed functionality or configuration variables are added/changed:

If the code communicates with devices, web services, or third-party tools:

  • The manifest file has all fields filled out correctly.
    Updated and included derived files by running: python3 -m script.hassfest.
  • New or updated dependencies have been added to requirements_all.txt.
    Updated by running python3 -m script.gen_requirements_all.
  • For the updated dependencies - a link to the changelog, or at minimum a diff between library versions is added to the PR description.
  • Untested files have been added to .coveragerc.

The integration reached or maintains the following Integration Quality Scale:

  • No score or internal
  • 🥈 Silver
  • 🥇 Gold
  • 🏆 Platinum

To help with the load of incoming pull requests:

@emontnemery emontnemery requested a review from a team as a code owner February 4, 2022 11:48
@probot-home-assistant
Copy link
Copy Markdown

Hey there @home-assistant/core, mind taking a look at this pull request as it has been labeled with an integration (recorder) you are listed as a code owner for? Thanks!
(message by CodeOwnersMention)

pvizeli
pvizeli previously approved these changes Feb 4, 2022
Comment on lines +599 to +640
try:
_create_index(connection, "statistics", "ix_statistics_statistic_id_start")
_create_index(
connection,
"statistics_short_term",
"ix_statistics_short_term_statistic_id_start",
)
except DatabaseError:
# There may be duplicated statistics entries, delete duplicated statistics
# and try again
delete_duplicates(instance, session)
_create_index(connection, "statistics", "ix_statistics_statistic_id_start")
_create_index(
connection,
"statistics_short_term",
"ix_statistics_short_term_statistic_id_start",
)
Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
try:
_create_index(connection, "statistics", "ix_statistics_statistic_id_start")
_create_index(
connection,
"statistics_short_term",
"ix_statistics_short_term_statistic_id_start",
)
except DatabaseError:
# There may be duplicated statistics entries, delete duplicated statistics
# and try again
delete_duplicates(instance, session)
_create_index(connection, "statistics", "ix_statistics_statistic_id_start")
_create_index(
connection,
"statistics_short_term",
"ix_statistics_short_term_statistic_id_start",
)
for _ in range(2):
try:
_create_index(connection, "statistics", "ix_statistics_statistic_id_start")
_create_index(
connection,
"statistics_short_term",
"ix_statistics_short_term_statistic_id_start",
)
except DatabaseError:
# There may be duplicated statistics entries, delete duplicated statistics
# and try again
delete_duplicates(instance, session)
else:
break

Copy link
Copy Markdown
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

with that loop, he need less doublicate code

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It's pretty nice.
Maybe we should only call delete_duplicates the first iteration though?

@emontnemery emontnemery force-pushed the recorder_statistics_optimize_duplicate_deletion branch from bbea1cc to 0b49cd6 Compare February 4, 2022 18:11
@balloob balloob merged commit 1f8e892 into dev Feb 4, 2022
@balloob balloob deleted the recorder_statistics_optimize_duplicate_deletion branch February 4, 2022 19:31
@balloob balloob mentioned this pull request Feb 4, 2022
@github-actions github-actions bot locked and limited conversation to collaborators Feb 5, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants