# -*- coding: utf-8 -*-
from datetime import timedelta
from time import time
from utils.operators.rich_sql_sensor import RichSqlSensor
from jms_data_back_up.dwd.dataplat.dwd_bgdm_fsimage_info_split_rk import jms_dwd__dwd_bgdm_fsimage_info_split_rk

label = f"dwd_bgdm_fsimage_info_split_rk_{int(time())}"
timeout = timedelta(minutes=10).seconds
doris_jms_dwd__dwd_bgdm_fsimage_info_split_rk = RichSqlSensor(
    task_id='doris_jms_dwd__dwd_bgdm_fsimage_info_split_rk',
    pool='broker_load_pool',
    depends_on_past=True,
    email= ['payne.jiang@jtexpress.com','yl_bigdata@yl-scm.com'],
    task_concurrency=1,
    conn_id='doris',
    pre_sql=f"""
                 TRUNCATE TABLE jms_dwd.dwd_bgdm_fsimage_info_split PARTITION (p{{{{ execution_date | cst_ds_nodash }}}});
                 LOAD LABEL jms_dwd.{label} (
                     DATA INFILE("hdfs://{{{{ var.value.hadoop_namespace }}}}/dw/hive/jms_dwd.db/external/dwd_bgdm_fsimage_info_split_all/dt={{{{  execution_date | cst_ds  }}}}/*")
                     INTO TABLE dwd_bgdm_fsimage_info_split
                     FORMAT AS 'PARQUET'
                 )
                 WITH BROKER '{{{{ var.json.doris_brokers | random_choice }}}}'
                 PROPERTIES ('timeout'='{timeout}', 'max_filter_ratio'='0.0')""",
    poke_sql=f"SHOW LOAD FROM jms_dwd WHERE label = '{label}' ORDER BY CreateTime DESC LIMIT 1",
    sql_on_kill=f"CANCEL LOAD FROM jms_dwd WHERE LABEL = '{label}'",
    success=lambda r: r[2] == 'FINISHED',
    failure=lambda r: (r[2] is not None and r[2] == 'CANCELLED', str(r[7])),
    poke_interval=60,
    execution_timeout=timedelta(seconds=timeout + 120), )

doris_jms_dwd__dwd_bgdm_fsimage_info_split_rk << jms_dwd__dwd_bgdm_fsimage_info_split_rk
