/**
 * SPDX-FileCopyrightText: 2023-2025 Sangfor Technologies Inc.
 * SPDX-License-Identifier: Mulan PSL v2
 */
exports.check = function () {
    // return true;
    return false;
};

exports.post = function () {
    return {
        "code": "0000000",
        "msg": "suceess",
        "success": true,
        "data": null
    }
};


exports.get = function () {
    return {
        "code": "0000000",
        "msg": "suceess",
        "success": true,
        "data": {
            "id": "123",
            "jobName": "XXX",
            "sqlText": "select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n ",
            "inputDataset": ["123", "321"], //输入数据集
            "outputDataset": "xxx", //输出数据集
            "outputDataspace": "124124124",
            "writeMode": "overwrite", //输出模式 overwrite 覆盖 append 追加
            "scheduleMode": "once", //执行模式 cron 周期 once 一次性
            "scheduleTime": "0", // 0表示立即执行
            "jobStatus": "running", //任务状态
            "createTime": "2022-10-21 12: 12: 12", // 创建时间
            "startTime": "2022-10-21 12: 12: 12", // 开始时间
            "endTime": "2022-10-21 12: 12: 12", // 结束时间
            "process": "50%", //执行进度
            "executeTime": "120", // 单位为秒
            "executeEnvironment": "spark 3.2.1",// 执行环境
            "resourceConfig": "cpu 10 核, 内存20G", //资源配置
            "jobType": "etl",  //  etl 代表 数据工程,     ml  代表分布式模型训练
            "executeHistory": [
                {
                    "id": "123",
                    "retryNo":'123',
                    "jobName": "XXX",
                    "sqlText": "select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n ",
                    "inputDataset": [
                        "123",
                        "321"
                    ], //输入数据集
                    "outputDataset": "xxx", //输出数据集
                    "outputDataspace": "124124124",
                    "writeMode": "overwrite", //输出模式 overwrite 覆盖 append 追加
                    "scheduleMode": "once", //执行模式 cron 周期 once 一次性
                    "scheduleTime": "0", // 0表示立即执行
                    "jobStatus": "queue", //任务状态
                    "createTime": "2022-10-21 12: 12: 12", // 创建时间
                    "startTime": "2022-10-21 12: 12: 12", // 开始时间
                    "endTime": "2022-10-21 12: 12: 12", // 结束时间
                    "process": "50%", //执行进度
                    "executeTime": "120", // 单位为秒
                    "executeEnvironment": "spark 3.2.1", // 执行环境
                    "resourceConfig": "cpu 10 核, 内存20G", //资源配置
                    "jobType": "etl", //  etl 代表 数据工程,     ml  代表分布式模型训练
                },
                {
                    "id": "123123",
                    "jobName": "XXX",
                    "retryNo":'121233',
                    "sqlText": "select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n ",
                    "inputDataset": [
                        "123",
                        "321"
                    ], //输入数据集
                    "outputDataset": "xxx", //输出数据集
                    "outputDataspace": "124124124",
                    "writeMode": "overwrite", //输出模式 overwrite 覆盖 append 追加
                    "scheduleMode": "once", //执行模式 cron 周期 once 一次性
                    "scheduleTime": "0", // 0表示立即执行
                    "jobStatus": "success", //任务状态
                    "createTime": "2022-10-21 12: 12: 12", // 创建时间
                    "startTime": "2022-10-21 12: 12: 12", // 开始时间
                    "endTime": "2022-10-21 12: 12: 12", // 结束时间
                    "process": "50%", //执行进度
                    "executeTime": "120", // 单位为秒
                    "executeEnvironment": "spark 3.2.1", // 执行环境
                    "resourceConfig": "cpu 10 核, 内存20G", //资源配置
                    "jobType": "etl", //  etl 代表 数据工程,     ml  代表分布式模型训练
                },
                {
                    "id": "123213123",
                    "retryNo":'13',
                    "jobName": "XXX",
                    "sqlText": "select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n select * from xxx \n ",
                    "inputDataset": [
                        "123",
                        "321"
                    ], //输入数据集
                    "outputDataset": "xxx", //输出数据集
                    "outputDataspace": "124124124",
                    "writeMode": "overwrite", //输出模式 overwrite 覆盖 append 追加
                    "scheduleMode": "once", //执行模式 cron 周期 once 一次性
                    "scheduleTime": "0", // 0表示立即执行
                    "jobStatus": "running", //任务状态
                    "createTime": "2022-10-21 12: 12: 12", // 创建时间
                    "startTime": "2022-10-21 12: 12: 12", // 开始时间
                    "endTime": "2022-10-21 12: 12: 12", // 结束时间
                    "process": "50%", //执行进度
                    "executeTime": "120", // 单位为秒
                    "executeEnvironment": "spark 3.2.1", // 执行环境
                    "resourceConfig": "cpu 10 核, 内存20G", //资源配置
                    "jobType": "etl", //  etl 代表 数据工程,     ml  代表分布式模型训练
                },
            ]
        }
    }
};
