-
Notifications
You must be signed in to change notification settings - Fork 4.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[helm] worker pod is crashing after upgrading to 0.49.6+ (latest update: missing env variables. see reply for detail) #31988
Comments
We are seeing this error as well. Downgrading to 0.49.6 fixed the issue for us. |
We're also experiencing this within GKE on the latest version |
Experiencing this same issue on latest V0.49.18. Downgraded to helm chart V0.49.5, works fine now. |
We're also running into this if we upgrade the chart past |
Same issue with all charts after 0.49.6 |
After providing some environment variables in
Check here what environment variables you might be missing |
@szemek thank you so much for the info! I followed your approach and and it worked! |
for maintaining state with S3. I was able to resolve it by simply adding these two environment variable in woker section of values file : I could find this here |
Any idea on how to fix it on a EC2 deployment? |
@HatemLar helm charts is supposed to be used in k8s. I assume you are deploying airbyte with docker/etc on EC2? trying to setup same env variables above following this guide. |
@sc-yan thank you for your help! |
@HatemLar it really depends on how you want to manage your infra/deployment. generally, docker is acting like VM so the app is not supposed to read values from host machine(which is EC2 in your case), unless you want to mount some volume into docker. it's common to setup these env variables in docker-compose file, but if you do have special cases, feel free to adjust so. |
Using the below works (for GCS) since the values should likely already be in your configMap if you specified them in global.gcs:
That fixes the worker pod issue, however then I ran into the following with |
|
What I wrote is specific to the worker key in values:
|
Is there a way to make it work with IRSA authentication (as + iam role)? |
Hello all 👋 sorry the missing update here. I shared this with the engineering team and any update return here. |
Just to note that this appears to be the same solution to remediate #18016. |
this work minio:
enabled: false
worker:
extraEnv:
- name: STATE_STORAGE_S3_BUCKET_NAME
value: "XXYYZZ"
- name: STATE_STORAGE_S3_REGION
value: "eu-west-3"
- name: S3_MINIO_ENDPOINT
value: ""
global:
log4jConfig: "log4j2-no-minio.xml"
state:
storage:
type: "S3"
logs:
storage:
type: "S3"
minio:
enabled: false
s3:
enabled: true
bucket: "XXYYZZ"
bucketRegion: "eu-west-3"
accessKey:
existingSecret: "airbyte-aws-creds"
existingSecretKey: "AWS_ACCESS_KEY_ID"
secretKey:
existingSecret: "airbyte-aws-creds"
existingSecretKey: "AWS_SECRET_ACCESS_KEY"
|
I can confirm, the settings from @raphaelauv helped me to start the I'm using helm chart The Adding the following in the in my yml helped to mitigate this: server:
extraEnv:
- name: LOG4J_CONFIGURATION_FILE
valueFrom:
configMapKeyRef:
name: airbyte-env
key: LOG4J_CONFIGURATION_FILE |
(Duplicate comment as previous issue is closed) Trying the fix suggested by @marcosmarxm doesn't fix for me. After attempting upgrading from 0.49.6 -> latest since mid Jan (so 0.50.22+) it has never fixed the issue. Running the minio config in bash returns:
Not an expert in any of this at all, but it looks like the creation of the bucket isn't entirely the issue. Just wanted to provide additional info as this has been a long-open issue! Edited to add:
Edit again: |
@davinchia worker pod was deprecated right? Should we close this? |
What method are you using to run Airbyte?
Kubernetes
Platform Version or Helm Chart Version
helm 0.49.9
What step the error happened?
Upgrading the Platform or Helm Chart
Revelant information
when upgrading helm from 0.49.6 to 0.49.8/0.49.9, the worker pod keeps crashing. but if I reverted it back to 0.49.6, it's fine.
Relevant log output
The text was updated successfully, but these errors were encountered: