// Copyright (c) 2024 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.

#include <stdio.h>
#include <string.h>
#include <sys/ipc.h>
#include <sys/msg.h>
#include <sys/types.h>
#include "paddle/extension.h"

#ifndef PD_BUILD_STATIC_OP
#define PD_BUILD_STATIC_OP(name) PD_BUILD_OP(static_op_##name)
#endif

#include "speculate_msg.h"

void SpeculateSaveWithOutputMsg(const paddle::Tensor& accept_tokens,
                                const paddle::Tensor& accept_num,
                                const paddle::Tensor& not_need_stop,
                                const paddle::Tensor& seq_lens_decoder,
                                const paddle::Tensor& prompt_lens,
                                int64_t rank_id,
                                int msg_queue_id,
                                int save_each_rank,
                                bool skip_prefill) {
    // printf("enter save output");
    if (!save_each_rank && rank_id > 0) {
        return;
    }

    int max_draft_tokens = accept_tokens.shape()[1];

    auto accept_tokens_cpu = accept_tokens.copy_to(paddle::CPUPlace(), true);
    auto accept_num_cpu = accept_num.copy_to(paddle::CPUPlace(), true);
    int64_t* accept_tokens_data = accept_tokens_cpu.data<int64_t>();
    int* accept_num_data = accept_num_cpu.data<int>();

    auto seq_lens_decoder_cpu = seq_lens_decoder.copy_to(paddle::CPUPlace(), true);
    auto prompt_lens_cpu = prompt_lens.copy_to(paddle::CPUPlace(), true);
    int* seq_lens_decoder_data = seq_lens_decoder_cpu.data<int>();
    int64_t* prompt_lens_data = prompt_lens_cpu.data<int64_t>();

    if (const char* inference_msg_queue_id_env_p =
            std::getenv("INFERENCE_MSG_QUEUE_ID")) {
        std::string inference_msg_queue_id_env_str(
            inference_msg_queue_id_env_p);
        int inference_msg_queue_id_from_env =
            std::stoi(inference_msg_queue_id_env_str);
#ifdef GET_OUTPUT_DEBUG
        std::cout << "Your INFERENCE_MSG_QUEUE_ID is: "
                  << inference_msg_queue_id_from_env << std::endl;
#endif
        msg_queue_id = inference_msg_queue_id_from_env;
    }
    static struct speculate_msgdata msg_sed;
    static key_t key = ftok("./", msg_queue_id);
    static int msgid = msgget(key, IPC_CREAT | 0666);

    msg_sed.mtype = 1;
    bool not_need_stop_data = not_need_stop.data<bool>()[0];

    int inference_msg_id_from_env = 1;
    if (const char* inference_msg_id_env_p = std::getenv("INFERENCE_MSG_ID")) {
        std::string inference_msg_id_env_str(inference_msg_id_env_p);
        inference_msg_id_from_env = std::stoi(inference_msg_id_env_str);
        if (inference_msg_id_from_env == 2) {
            // 2 and -2 is preserve for no-output indication.
            throw std::runtime_error(
                " INFERENCE_MSG_ID cannot be 2, please use other number.");
        }
        if (inference_msg_id_from_env < 0) {
            throw std::runtime_error(
                " INFERENCE_MSG_ID cannot be negative, please use other "
                "number.");
        }

#ifdef SAVE_WITH_OUTPUT_DEBUG
        std::cout << "Your INFERENCE_MSG_ID is: " << inference_msg_id_from_env
                  << std::endl;
#endif
    } else {
#ifdef SAVE_WITH_OUTPUT_DEBUG
        std::cout
            << "Failed to got INFERENCE_MSG_ID at env, use (int)1 as default."
            << std::endl;
#endif
    }

    msg_sed.mtext[0] = not_need_stop_data ? inference_msg_id_from_env
                                          : -inference_msg_id_from_env;
    int bsz = accept_tokens.shape()[0];
    msg_sed.mtext[1] = bsz;

    for (int i = 2; i < MAX_BSZ + 2; i++) {
        if (i - 2 >= bsz || (skip_prefill && seq_lens_decoder_data[i - 2] < prompt_lens_data[i - 2])) {
            msg_sed.mtext[i] = 0;
        } else {
            msg_sed.mtext[i] = (int)accept_num_data[i - 2];
        }
    }
    for (int i = MAX_BSZ + 2; i < MAX_BSZ * MAX_DRAFT_TOKENS + MAX_BSZ + 2;
         i++) {
        int token_id = i - MAX_BSZ - 2;
        int bid = token_id / MAX_DRAFT_TOKENS;
        int local_token_id = token_id % MAX_DRAFT_TOKENS;
        if (token_id / MAX_DRAFT_TOKENS >= bsz) {
            msg_sed.mtext[i] = 0;
        } else {
            msg_sed.mtext[i] =
                accept_tokens_data[bid * max_draft_tokens + local_token_id];
        }
    }
    if ((msgsnd(msgid,
                &msg_sed,
                (MAX_BSZ * MAX_DRAFT_TOKENS + MAX_BSZ + 2) * 4,
                0)) == -1) {
        printf("full msg buffer\n");
    }
    return;
}

void SpeculateSaveWithOutputMsgStatic(const paddle::Tensor& accept_tokens,
                                      const paddle::Tensor& accept_num,
                                      const paddle::Tensor& not_need_stop,
                                      const paddle::Tensor& seq_lens_decoder,
                                      const paddle::Tensor& prompt_lens,
                                      int64_t rank_id,
                                      bool save_each_rank,
                                      bool skip_prefill) {
    SpeculateSaveWithOutputMsg(
        accept_tokens, accept_num, not_need_stop, seq_lens_decoder, prompt_lens, rank_id, 1, save_each_rank, skip_prefill);
}

void SpeculateSaveWithOutputMsgDynamic(const paddle::Tensor& accept_tokens,
                                       const paddle::Tensor& accept_num,
                                       const paddle::Tensor& not_need_stop,
                                       const paddle::Tensor& seq_lens_decoder,
                                       const paddle::Tensor& prompt_lens,
                                       int64_t rank_id,
                                       int msg_queue_id,
                                       bool save_each_rank,
                                       bool skip_prefill) {
    SpeculateSaveWithOutputMsg(
        accept_tokens, accept_num, not_need_stop, seq_lens_decoder, prompt_lens, rank_id, msg_queue_id, save_each_rank, skip_prefill);
}

PD_BUILD_STATIC_OP(speculate_save_output)
    .Inputs({"accept_tokens", "accept_num", "not_need_stop", "seq_lens_decoder", "prompt_lens"})
    .Attrs({"rank_id: int64_t", "save_each_rank: bool", "skip_prefill: bool"})
    .Outputs({"x_out"})
    .SetInplaceMap({{"accept_tokens", "x_out"}})
    .SetKernelFn(PD_KERNEL(SpeculateSaveWithOutputMsgStatic));

PD_BUILD_STATIC_OP(speculate_save_output_dynamic)
    .Inputs({"accept_tokens", "accept_num", "not_need_stop", "seq_lens_decoder", "prompt_lens"})
    .Attrs({"rank_id: int64_t", "msg_queue_id: int", "save_each_rank: bool", "skip_prefill: bool"})
    .Outputs({"x_out"})
    .SetInplaceMap({{"accept_tokens", "x_out"}})
    .SetKernelFn(PD_KERNEL(SpeculateSaveWithOutputMsgDynamic));
