[bitnami/airflow] Release airflow-2.10.2-debian-12-r5 (#74089)

Signed-off-by: Bitnami Bot <bitnami-bot@vmware.com>
This commit is contained in:
Bitnami Bot
2024-10-31 17:55:24 +01:00
committed by GitHub
parent 5f829df69e
commit b0b861ff5e
5 changed files with 67 additions and 57 deletions

View File

@@ -8,11 +8,11 @@ ARG TARGETARCH
LABEL com.vmware.cp.artifact.flavor="sha256:c50c90cfd9d12b445b011e6ad529f1ad3daea45c26d20b00732fae3cd71f6a83" \
org.opencontainers.image.base.name="docker.io/bitnami/minideb:bookworm" \
org.opencontainers.image.created="2024-10-30T14:53:31Z" \
org.opencontainers.image.created="2024-10-31T15:31:42Z" \
org.opencontainers.image.description="Application packaged by Broadcom, Inc." \
org.opencontainers.image.documentation="https://github.com/bitnami/containers/tree/main/bitnami/airflow/README.md" \
org.opencontainers.image.licenses="Apache-2.0" \
org.opencontainers.image.ref.name="2.10.2-debian-12-r4" \
org.opencontainers.image.ref.name="2.10.2-debian-12-r5" \
org.opencontainers.image.source="https://github.com/bitnami/containers/tree/main/bitnami/airflow" \
org.opencontainers.image.title="airflow" \
org.opencontainers.image.vendor="Broadcom, Inc." \
@@ -30,9 +30,9 @@ RUN install_packages ca-certificates curl git krb5-user libbz2-1.0 libcom-err2 l
RUN mkdir -p /tmp/bitnami/pkg/cache/ ; cd /tmp/bitnami/pkg/cache/ ; \
COMPONENTS=( \
"wait-for-port-1.0.8-6-linux-${OS_ARCH}-debian-12" \
"python-3.12.7-3-linux-${OS_ARCH}-debian-12" \
"python-3.12.7-4-linux-${OS_ARCH}-debian-12" \
"ini-file-1.4.7-6-linux-${OS_ARCH}-debian-12" \
"airflow-2.10.2-1-linux-${OS_ARCH}-debian-12" \
"airflow-2.10.2-2-linux-${OS_ARCH}-debian-12" \
) ; \
for COMPONENT in "${COMPONENTS[@]}"; do \
if [ ! -f "${COMPONENT}.tar.gz" ]; then \

View File

@@ -3,7 +3,7 @@
"arch": "amd64",
"distro": "debian-12",
"type": "NAMI",
"version": "2.10.2-1"
"version": "2.10.2-2"
},
"ini-file": {
"arch": "amd64",
@@ -15,7 +15,7 @@
"arch": "amd64",
"distro": "debian-12",
"type": "NAMI",
"version": "3.12.7-3"
"version": "3.12.7-4"
},
"wait-for-port": {
"arch": "amd64",

View File

@@ -43,6 +43,7 @@ airflow_env_vars=(
AIRFLOW_POOL_NAME
AIRFLOW_POOL_SIZE
AIRFLOW_POOL_DESC
AIRFLOW_STANDALONE_DAG_PROCESSOR
AIRFLOW_WORKER_QUEUE
AIRFLOW_DATABASE_HOST
AIRFLOW_DATABASE_PORT_NUMBER
@@ -124,6 +125,7 @@ export AIRFLOW_HOSTNAME_CALLABLE="${AIRFLOW_HOSTNAME_CALLABLE:-}"
export AIRFLOW_POOL_NAME="${AIRFLOW_POOL_NAME:-}"
export AIRFLOW_POOL_SIZE="${AIRFLOW_POOL_SIZE:-}"
export AIRFLOW_POOL_DESC="${AIRFLOW_POOL_DESC:-}"
export AIRFLOW_STANDALONE_DAG_PROCESSOR="${AIRFLOW_STANDALONE_DAG_PROCESSOR:-no}"
AIRFLOW_WORKER_QUEUE="${AIRFLOW_WORKER_QUEUE:-"${AIRFLOW_QUEUE:-}"}"
export AIRFLOW_WORKER_QUEUE="${AIRFLOW_WORKER_QUEUE:-}"

View File

@@ -76,8 +76,9 @@ airflow_validate() {
# Check component type & executor
check_empty_value "AIRFLOW_COMPONENT_TYPE"
check_multi_value "AIRFLOW_COMPONENT_TYPE" "webserver scheduler worker"
check_multi_value "AIRFLOW_COMPONENT_TYPE" "webserver scheduler worker dag-processor"
check_empty_value "AIRFLOW_EXECUTOR"
check_yes_no_value "AIRFLOW_STANDALONE_DAG_PROCESSOR"
# Check cryptography parameters
if [[ -n "$AIRFLOW_RAW_FERNET_KEY" && -z "$AIRFLOW_FERNET_KEY" ]]; then
@@ -127,7 +128,7 @@ airflow_validate() {
done
fi
;;
scheduler|worker)
*)
# Check webserver host and port number
check_empty_value "AIRFLOW_WEBSERVER_HOST"
check_resolved_hostname "$AIRFLOW_WEBSERVER_HOST"
@@ -189,7 +190,7 @@ airflow_initialize() {
true # Avoid return false when I am not root
fi
;;
scheduler|worker)
*)
info "Waiting for Airflow Webserver to be up"
airflow_wait_for_webserver "$AIRFLOW_WEBSERVER_HOST" "$AIRFLOW_WEBSERVER_PORT_NUMBER"
if [[ "$AIRFLOW_EXECUTOR" == "CeleryExecutor" || "$AIRFLOW_EXECUTOR" == "CeleryKubernetesExecutor" ]]; then
@@ -250,7 +251,7 @@ airflow_generate_config() {
# Configure Airflow webserver authentication
airflow_configure_webserver_authentication
;;
scheduler|worker)
*)
# Generate Airflow default files
debug_execute airflow version
;;
@@ -276,6 +277,12 @@ airflow_generate_config() {
else
airflow_conf_set "core" "load_examples" "False"
fi
# Configure Dag Processor mode
if is_boolean_yes "$AIRFLOW_STANDALONE_DAG_PROCESSOR"; then
airflow_conf_set "scheduler" "standalone_dag_processor" "True"
else
airflow_conf_set "scheduler" "standalone_dag_processor" "False"
fi
fi
# Configure Airflow executor