-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Updated README.md with instructions how to clean conflicting IAM profiles #1124
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: ekuric If they are not already assigned, you can assign the PR to them by writing The full list of commands accepted by this bot can be found here. The pull request process is described here DetailsNeeds approval from an approver in each of these files:Approvers can indicate their approval by writing |
README.md
Outdated
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Cleanup step should tear down everything created during the previous installation, however sometimes the installer may report error messages like the ones below:
README.md
Outdated
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This happens when the installer tries to create an IAM instance profile with a name that matches an existing instance profile. This usually means that IAM instance profiles from a previous installation by the user didn't get deleted during an earlier cluster teardown.
…iles in case they are present during cluster installation and causing cluster install to fail Signed-off-by: Elvir Kuric <[email protected]>
| And if you want to reinstall from scratch, `rm -rf` the asset directory beforehand. | ||
|
|
||
|
|
||
| ### Troubleshouting installation |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This doesn't belong in the project README. We have a platform-agnostic troubleshooting doc in docs/user/troubleshooting.md (already linked from the project README), and AWS-specific troubleshooting docs should go in a new docs/user/aws/troubleshooting.md (like our existing docs/user/openstack/troubleshooting.md).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I can send it there if this is not best place ... ?
|
|
||
| ### Troubleshouting installation | ||
|
|
||
| ```Cleanup``` step *should* tear down everything created during the previous installation, however sometimes the installer may report error messages like the ones below: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can you provide steps to reproduce? I'd much rather fix this bug (if it still exists) than document workarounds.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@wking either destroy step was interrupted, or stopped due to network issue, but I do not have clear reproducer. When I wanted to install new cluster it was not possible due to error described here, but possible after applying fix described here ... thanks to @dobbymoodge for suggestions
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
... either
destroystep was interrupted, or stopped due to network issue...
There are lots of ways this could go wrong. The most reliable approach to cleaning it up should be to re-launch destroy. Can you post logs showing that a fresh destroy call failed to delete the profile or other resource? I'm not even clear on how you were able to run a new create cluster without having previously had a successful destroy cluster on the previous cluster. What version of the installer were you running? 0.10.1 has #1039, with some destroy cluster fixes.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
q: Can you post logs showing that a fresh destroy call failed to delete the profile or other resource?
a: unfortunately I do not have it.
prior to every install I do git pull, ./hack/build.sh - so I was on latest.
|
Installer creates unique IAM roles and instance profiles. Also the installer now cleans up any dangling of these. /close |
|
@abhinavdahiya: Closed this PR. DetailsIn response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
|
@ekuric: The following tests failed, say
Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. |
Updated README.md with instructions how to clean conflicting IAM profiles in case
they are present during cluster installation and causing cluster install to fail
Signed-off-by: Elvir Kuric [email protected]