Skip to content

Commit

Permalink
Review feedback
Browse files Browse the repository at this point in the history
  • Loading branch information
ciarams87 committed Dec 19, 2023
1 parent 55f547d commit f0b91a7
Showing 1 changed file with 11 additions and 10 deletions.
21 changes: 11 additions & 10 deletions tests/longevity/results/1.1.0/1.1.0.md
Original file line number Diff line number Diff line change
Expand Up @@ -86,22 +86,23 @@ Further investigation is out of scope of this test.

```text
resource.type="k8s_container"
resource.labels.cluster_name="ciara-1"
resource.labels.cluster_name="<CLUSTER_NAME>"
resource.labels.namespace_name="nginx-gateway"
resource.labels.container_name="nginx-gateway"
severity=ERROR
SEARCH("error")
```

There were 53 error logs across 3 pod instances. They came in 3 almost identical batches, starting just over 24 hours
after the initial deployment, and then each subsequent error batch just over 24 hours after the last. They were all
relating to leader election loss, and subsequent restart (see https://github.com/nginxinc/nginx-gateway-fabric/issues/1100).
There were 53 error logs, and 6 restarts, across 3 pod instances. The error logs came in 3 almost identical batches,
starting just over 24 hours after the initial deployment, and then each subsequent error batch just over 24 hours after
the last. They were all relating to leader election loss, and subsequent restart (see https://github.com/nginxinc/nginx-gateway-fabric/issues/1100). There were also 2 termination events, both of these occurred approximately 5 minutes
after a leader election loss and successful restart.

Each error batches caused the pod to restart, but not terminate. However, the first pod was terminated about 10 minutes
Each error batches caused the pod to restart, but not terminate. The first pod was terminated about 5 minutes
after the first error batch and subsequent restart occurred. A similar occurance happened after the third error batch.
Exactly why these pods were terminated is not currently clear, but it looks to be a cluster event (perhaps an upgrade)
as the coffee and tea pods were terminated at that time also. All the restarts happened roughly at the same time each
day.
There was no termination event after the second error batch. Exactly why these pods were terminated is not currently
clear, but it looks to be a cluster event (perhaps an upgrade) as the coffee and tea pods were terminated at that time
also. All the restarts happened roughly at the same time each day.

```text
{"level":"info", "msg":"Starting manager", "ts":"2023-12-13T17:45:10Z"} -> Start-up
Expand All @@ -120,7 +121,7 @@ Errors:

```text
resource.type=k8s_container AND
resource.labels.cluster_name="ciara-1" AND
resource.labels.cluster_name="<CLUSTER_NAME>" AND
resource.labels.container_name="nginx" AND
severity=ERROR AND
SEARCH("`[warn]`") OR SEARCH("`[error]`")
Expand All @@ -134,7 +135,7 @@ Non-200 response codes in NGINX access logs:

```text
resource.type=k8s_container AND
resource.labels.cluster_name="ciara-1" AND
resource.labels.cluster_name="<CLUSTER_NAME>" AND
resource.labels.container_name="nginx"
"GET" "HTTP/1.1" -"200"
```
Expand Down

0 comments on commit f0b91a7

Please sign in to comment.