-
Notifications
You must be signed in to change notification settings - Fork 33
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add thanos to hub quickstart cluster & remote write from all clusters #634
Conversation
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## main #634 +/- ##
==========================================
+ Coverage 80.20% 82.50% +2.29%
==========================================
Files 64 72 +8
Lines 4492 5360 +868
==========================================
+ Hits 3603 4422 +819
- Misses 600 634 +34
- Partials 289 304 +15
Flags with carried forward coverage won't be shown. Click here to find out more.
|
hack/quickstart-setup.sh
Outdated
kubectl apply -k ${KUADARNT_THANOS_KUSTOMIZATION} | ||
success "thanos installed successfully." | ||
fi | ||
|
||
# Install observability stack | ||
info "Installing observability stack in ${KUADRANT_CLUSTER_NAME}..." | ||
kubectl kustomize ${KUADARNT_OBSERVABILITY_KUSTOMIZATION} | docker run --rm -i ryane/kfilt -i kind=CustomResourceDefinition | kubectl apply --server-side -f - |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The script fails at this point for me saying docker command not found. I do have a alias but maybe the pipe is causing weird things with it. Its not a change from this pr that's causing the issue but Im thinking we should use the CONTAINER_RUNTIME_BIN env from the top of the script instead
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Agreed.
Pushed up a fix for that
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I know this PR has been approved and merged, but I wanted to document an issue I had when trying to review this PR and have since found the solution.
My system runs on FedoraOS, and without changing any settings, works fine for using just single cluster quickstart. However, I had issues earlier when trying to run multicluster quickstart and have discovered why.
My issue was that I could not get past the INFO: Waiting for cert-manager deployments to be ready
part of the script.
The reason for this, was that inotify did not have enough resources to create more than the pods in a single cluster.
Kind has this page which explains why this issue occurs. After I ran:
sudo sysctl fs.inotify.max_user_watches=524288
sudo sysctl fs.inotify.max_user_instances=512
After deleting the faulty cluster, and re-running the quickstart as seen in the description, I was able to verify the changes working as expected.
Commenting this for future reference. :)
To verify, the PR branch
thanos
will need to be referenced until changes are back on main.KUADRANT_REF=thanos ./hack/quickstart-setup.sh
KUADRANT_REF=thanos ./hack/quickstart-setup.sh
again.In the first/hub cluster, you should see metrics from both clusters in the Grafana explore view. e.g. the
up{app="metallb",component="controller"}
metric should have 1 entry for each cluster.To access grafana in the hub, you'll need to port-forward to the service with
kubectl --context kind-kuadrant-local -n monitoring port-forward svc/grafana 3000:3000
then access it at http://127.0.0.1:3000/explore