-
Notifications
You must be signed in to change notification settings - Fork 377
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
First helm release always succeeds and doesn't wait for all pods running #672
Comments
Thanks for opening @FischlerA. Did you try using the |
Per documentation the wait attributed defaults to true. But even after explicitly setting it to true the behavior didn't change and it was still seen as success with a crashing pod. |
Ah yep, you're right – I will try and reproduce this. The provider itself doesn't do the waiting, it just passes along the wait flag to the install action in the helm package. Do you get the same issue if you do a |
@jrhouston |
I had same experiences when I use helm_release in terraform, if something goes wrong, pod status is stay at "pending" or "Error", "CreateContainer" or some other unusual status for a little longer time, Helm terraform provider will not wait until pods are running, it will exit and reported completed, However terraform state was update as failed. |
Saw the same behavior today when I deployed ingress-nginx and the very first job failed because it was rejected by another webhook. The terraform apply run waited for 5 minutes but reported a success, even though not a single resource was created successful. In fact only 1 job was there it was rejected. |
@jrhouston were you able to take a look at this? |
I'm running into this too. I pretty regularly have a successful |
I think we are hitting this as well but not entirely sure.. we are seeing helm_release pass on first run with (wait = true) where not all the pods come online because of a Gatekeeper/PSP we have in the cluster, we are not sure how get our helm_release to fail in that case |
Hi all. I'm new to Terraform. I've had to split up my Terraform deployments and include a time_sleep because of this issue. Looking forward to an update here. |
Same thing with helm job and wait_for_jobs = true.
|
I faced this issue, helm-release 'timeout' options seem not working, helm-relese stated as "successfully completed" with in 5 seconds , even though PODs are init stage. |
me too . pod status is stay at "pending" when I use helm_release in terraform, but it worked well with Helm cli. |
I don't know what happened,but it back to normal work。In the past 6 hours, I upgraded kubernetes to 1.23.1.
Versions :
|
Is anyone still encountering this issue on the latest version of the provider? I think we fixed this in #727. Just tried to reproduce this and see the error in provider version v2.0.2 but now I see the appropriate failure diagnostic in v2.6.0. |
I can't speak for everyone, but we haven't seen this issue in a while. |
This happens to me as well. |
Haven't tried it with the v2.6.0 version but will do so and report back, might take me a few days |
Reproduced on version 2.6.0 for me |
Hello @enterdv! Are you able to include the config that you used that will help us reproduce this issue? We'll want to look into it again if we're still seeing this bug |
Hello, I tried with simple helm release
|
@enterdv Hello! Thank you for providing the TF config, could you provide the output after running |
Have you fixed this problem? |
Marking this issue as stale due to inactivity. If this issue receives no comments in the next 30 days it will automatically be closed. If this issue was automatically closed and you feel this issue should be reopened, we encourage creating a new issue linking back to this one for added context. This helps our maintainers find and focus on the active issues. Maintainers may also remove the stale label at their discretion. Thank you! |
up |
up |
I am still seeing this issue in For my use case, I have a helm-release that generates self-signed certificates with cert-manager and the release gets marked successful ahead of the certificates actually being signed (causing downstream failures in other terraform modules). My workaround is to create an addition resource "time_sleep" "wait_for_signing" {
depends_on = [helm_release.cluster-issuer-self-signed]
create_duration = "60s"
}
# Export Self-Signed TLS
data "kubernetes_secret" "self-signed-tls-certs" {
...
depends_on = [time_sleep.wait_for_signing] # Can't read the secrets before they are created
} |
Terraform, Provider, Kubernetes and Helm Versions
Terraform version: 0.14.4
Provider version: 2.0.2
Kubernetes version: AWS EKS 1.18
Helm version: 3
Affected Resource(s)
Debug Output
https://gist.github.com/FischlerA/7930aff18d68a7b133ff22aadc021517
Steps to Reproduce
terraform apply
Expected Behavior
The helm deployment should fail since the pod that is being deployed is running an image that will always fail. (private image which i can't share)
Actual Behavior
The first time the helm release is deployed it always succeeds after reaching the timeout (5 min), any further deployments fail as they are supposed to after reaching the timeout (5min).
Community Note
The text was updated successfully, but these errors were encountered: