109109
110110LOGGING_LEVEL = logging .INFO
111111
112- # the prefix to append to gunicorn worker processes after init
113- GUNICORN_WORKER_READY_PREFIX = "[ready] "
114-
115112LOG_FORMAT = conf .get ("logging" , "log_format" )
116113SIMPLE_LOG_FORMAT = conf .get ("logging" , "simple_log_format" )
117114
118115SQL_ALCHEMY_CONN : str | None = None
119116SQL_ALCHEMY_CONN_ASYNC : str | None = None
120117PLUGINS_FOLDER : str | None = None
121- DONOT_MODIFY_HANDLERS : bool | None = None
122118DAGS_FOLDER : str = os .path .expanduser (conf .get_mandatory_value ("core" , "DAGS_FOLDER" ))
123119
124120AIO_LIBS_MAPPING = {"sqlite" : "aiosqlite" , "postgresql" : "asyncpg" , "mysql" : "aiomysql" }
@@ -264,7 +260,6 @@ def configure_vars():
264260 global SQL_ALCHEMY_CONN_ASYNC
265261 global DAGS_FOLDER
266262 global PLUGINS_FOLDER
267- global DONOT_MODIFY_HANDLERS
268263
269264 SQL_ALCHEMY_CONN = conf .get ("database" , "sql_alchemy_conn" )
270265 if conf .has_option ("database" , "sql_alchemy_conn_async" ):
@@ -276,13 +271,6 @@ def configure_vars():
276271
277272 PLUGINS_FOLDER = conf .get ("core" , "plugins_folder" , fallback = os .path .join (AIRFLOW_HOME , "plugins" ))
278273
279- # If donot_modify_handlers=True, we do not modify logging handlers in task_run command
280- # If the flag is set to False, we remove all handlers from the root logger
281- # and add all handlers from 'airflow.task' logger to the root Logger. This is done
282- # to get all the logs from the print & log statements in the DAG files before a task is run
283- # The handlers are restored after the task completes execution.
284- DONOT_MODIFY_HANDLERS = conf .getboolean ("logging" , "donot_modify_handlers" , fallback = False )
285-
286274
287275def _run_openlineage_runtime_check ():
288276 """
@@ -340,10 +328,6 @@ def get_bind(
340328
341329
342330AIRFLOW_PATH = os .path .dirname (os .path .dirname (__file__ ))
343- AIRFLOW_TESTS_PATH = os .path .join (AIRFLOW_PATH , "tests" )
344- AIRFLOW_SETTINGS_PATH = os .path .join (AIRFLOW_PATH , "airflow" , "settings.py" )
345- AIRFLOW_UTILS_SESSION_PATH = os .path .join (AIRFLOW_PATH , "airflow" , "utils" , "session.py" )
346- AIRFLOW_MODELS_BASEOPERATOR_PATH = os .path .join (AIRFLOW_PATH , "airflow" , "models" , "baseoperator.py" )
347331
348332
349333def _is_sqlite_db_path_relative (sqla_conn_str : str ) -> bool :
@@ -753,9 +737,6 @@ def initialize():
753737
754738
755739# Const stuff
756-
757- KILOBYTE = 1024
758- MEGABYTE = KILOBYTE * KILOBYTE
759740WEB_COLORS = {"LIGHTBLUE" : "#4d9de0" , "LIGHTORANGE" : "#FF9933" }
760741
761742# Updating serialized DAG can not be faster than a minimum interval to reduce database
@@ -765,10 +746,6 @@ def initialize():
765746# If set to True, serialized DAGs is compressed before writing to DB,
766747COMPRESS_SERIALIZED_DAGS = conf .getboolean ("core" , "compress_serialized_dags" , fallback = False )
767748
768- # Fetching serialized DAG can not be faster than a minimum interval to reduce database
769- # read rate. This config controls when your DAGs are updated in the Webserver
770- MIN_SERIALIZED_DAG_FETCH_INTERVAL = conf .getint ("core" , "min_serialized_dag_fetch_interval" , fallback = 10 )
771-
772749CAN_FORK = hasattr (os , "fork" )
773750
774751EXECUTE_TASKS_NEW_PYTHON_INTERPRETER = not CAN_FORK or conf .getboolean (
0 commit comments