// Copyright 2019 yuzu Emulator Project
// Licensed under GPLv2 or any later version
// Refer to the license.txt file included.

#include <iterator>

#include "common/assert.h"
#include "common/common_types.h"
#include "common/logging/log.h"
#include "video_core/engines/maxwell_3d.h"
#include "video_core/renderer_vulkan/maxwell_to_vk.h"
#include "video_core/renderer_vulkan/vk_device.h"
#include "video_core/renderer_vulkan/wrapper.h"
#include "video_core/surface.h"

namespace Vulkan::MaxwellToVK {

using Maxwell = Tegra::Engines::Maxwell3D::Regs;

namespace Sampler {

VkFilter Filter(Tegra::Texture::TextureFilter filter) {
    switch (filter) {
    case Tegra::Texture::TextureFilter::Linear:
        return VK_FILTER_LINEAR;
    case Tegra::Texture::TextureFilter::Nearest:
        return VK_FILTER_NEAREST;
    }
    UNIMPLEMENTED_MSG("Unimplemented sampler filter={}", static_cast<u32>(filter));
    return {};
}

VkSamplerMipmapMode MipmapMode(Tegra::Texture::TextureMipmapFilter mipmap_filter) {
    switch (mipmap_filter) {
    case Tegra::Texture::TextureMipmapFilter::None:
        // TODO(Rodrigo): None seems to be mapped to OpenGL's mag and min filters without mipmapping
        // (e.g. GL_NEAREST and GL_LINEAR). Vulkan doesn't have such a thing, find out if we have to
        // use an image view with a single mipmap level to emulate this.
        return VK_SAMPLER_MIPMAP_MODE_LINEAR;
        ;
    case Tegra::Texture::TextureMipmapFilter::Linear:
        return VK_SAMPLER_MIPMAP_MODE_LINEAR;
    case Tegra::Texture::TextureMipmapFilter::Nearest:
        return VK_SAMPLER_MIPMAP_MODE_NEAREST;
    }
    UNIMPLEMENTED_MSG("Unimplemented sampler mipmap mode={}", static_cast<u32>(mipmap_filter));
    return {};
}

VkSamplerAddressMode WrapMode(const VKDevice& device, Tegra::Texture::WrapMode wrap_mode,
                              Tegra::Texture::TextureFilter filter) {
    switch (wrap_mode) {
    case Tegra::Texture::WrapMode::Wrap:
        return VK_SAMPLER_ADDRESS_MODE_REPEAT;
    case Tegra::Texture::WrapMode::Mirror:
        return VK_SAMPLER_ADDRESS_MODE_MIRRORED_REPEAT;
    case Tegra::Texture::WrapMode::ClampToEdge:
        return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE;
    case Tegra::Texture::WrapMode::Border:
        return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER;
    case Tegra::Texture::WrapMode::Clamp:
        if (device.GetDriverID() == VK_DRIVER_ID_NVIDIA_PROPRIETARY_KHR) {
            // Nvidia's Vulkan driver defaults to GL_CLAMP on invalid enumerations, we can hack this
            // by sending an invalid enumeration.
            return static_cast<VkSamplerAddressMode>(0xcafe);
        }
        // TODO(Rodrigo): Emulate GL_CLAMP properly on other vendors
        switch (filter) {
        case Tegra::Texture::TextureFilter::Nearest:
            return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE;
        case Tegra::Texture::TextureFilter::Linear:
            return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_BORDER;
        }
        UNREACHABLE();
        return VK_SAMPLER_ADDRESS_MODE_CLAMP_TO_EDGE;
    case Tegra::Texture::WrapMode::MirrorOnceClampToEdge:
        return VK_SAMPLER_ADDRESS_MODE_MIRROR_CLAMP_TO_EDGE;
    case Tegra::Texture::WrapMode::MirrorOnceBorder:
        UNIMPLEMENTED();
        return VK_SAMPLER_ADDRESS_MODE_MIRROR_CLAMP_TO_EDGE;
    default:
        UNIMPLEMENTED_MSG("Unimplemented wrap mode={}", static_cast<u32>(wrap_mode));
        return {};
    }
}

VkCompareOp DepthCompareFunction(Tegra::Texture::DepthCompareFunc depth_compare_func) {
    switch (depth_compare_func) {
    case Tegra::Texture::DepthCompareFunc::Never:
        return VK_COMPARE_OP_NEVER;
    case Tegra::Texture::DepthCompareFunc::Less:
        return VK_COMPARE_OP_LESS;
    case Tegra::Texture::DepthCompareFunc::LessEqual:
        return VK_COMPARE_OP_LESS_OR_EQUAL;
    case Tegra::Texture::DepthCompareFunc::Equal:
        return VK_COMPARE_OP_EQUAL;
    case Tegra::Texture::DepthCompareFunc::NotEqual:
        return VK_COMPARE_OP_NOT_EQUAL;
    case Tegra::Texture::DepthCompareFunc::Greater:
        return VK_COMPARE_OP_GREATER;
    case Tegra::Texture::DepthCompareFunc::GreaterEqual:
        return VK_COMPARE_OP_GREATER_OR_EQUAL;
    case Tegra::Texture::DepthCompareFunc::Always:
        return VK_COMPARE_OP_ALWAYS;
    }
    UNIMPLEMENTED_MSG("Unimplemented sampler depth compare function={}",
                      static_cast<u32>(depth_compare_func));
    return {};
}

} // namespace Sampler

namespace {

enum : u32 { Attachable = 1, Storage = 2 };

struct FormatTuple {
    VkFormat format; ///< Vulkan format
    int usage = 0;   ///< Describes image format usage
} constexpr tex_format_tuples[] = {
    {VK_FORMAT_A8B8G8R8_UNORM_PACK32, Attachable | Storage},    // ABGR8U
    {VK_FORMAT_A8B8G8R8_SNORM_PACK32, Attachable | Storage},    // ABGR8S
    {VK_FORMAT_A8B8G8R8_UINT_PACK32, Attachable | Storage},     // ABGR8UI
    {VK_FORMAT_B5G6R5_UNORM_PACK16},                            // B5G6R5U
    {VK_FORMAT_A2B10G10R10_UNORM_PACK32, Attachable | Storage}, // A2B10G10R10U
    {VK_FORMAT_A1R5G5B5_UNORM_PACK16, Attachable},              // A1B5G5R5U (flipped with swizzle)
    {VK_FORMAT_R8_UNORM, Attachable | Storage},                 // R8U
    {VK_FORMAT_R8_UINT, Attachable | Storage},                  // R8UI
    {VK_FORMAT_R16G16B16A16_SFLOAT, Attachable | Storage},      // RGBA16F
    {VK_FORMAT_R16G16B16A16_UNORM, Attachable | Storage},       // RGBA16U
    {VK_FORMAT_R16G16B16A16_SNORM, Attachable | Storage},       // RGBA16S
    {VK_FORMAT_R16G16B16A16_UINT, Attachable | Storage},        // RGBA16UI
    {VK_FORMAT_B10G11R11_UFLOAT_PACK32, Attachable | Storage},  // R11FG11FB10F
    {VK_FORMAT_R32G32B32A32_UINT, Attachable | Storage},        // RGBA32UI
    {VK_FORMAT_BC1_RGBA_UNORM_BLOCK},                           // DXT1
    {VK_FORMAT_BC2_UNORM_BLOCK},                                // DXT23
    {VK_FORMAT_BC3_UNORM_BLOCK},                                // DXT45
    {VK_FORMAT_BC4_UNORM_BLOCK},                                // DXN1
    {VK_FORMAT_BC5_UNORM_BLOCK},                                // DXN2UNORM
    {VK_FORMAT_BC5_SNORM_BLOCK},                                // DXN2SNORM
    {VK_FORMAT_BC7_UNORM_BLOCK},                                // BC7U
    {VK_FORMAT_BC6H_UFLOAT_BLOCK},                              // BC6H_UF16
    {VK_FORMAT_BC6H_SFLOAT_BLOCK},                              // BC6H_SF16
    {VK_FORMAT_ASTC_4x4_UNORM_BLOCK},                           // ASTC_2D_4X4
    {VK_FORMAT_B8G8R8A8_UNORM},                                 // BGRA8
    {VK_FORMAT_R32G32B32A32_SFLOAT, Attachable | Storage},      // RGBA32F
    {VK_FORMAT_R32G32_SFLOAT, Attachable | Storage},            // RG32F
    {VK_FORMAT_R32_SFLOAT, Attachable | Storage},               // R32F
    {VK_FORMAT_R16_SFLOAT, Attachable | Storage},               // R16F
    {VK_FORMAT_R16_UNORM, Attachable | Storage},                // R16U
    {VK_FORMAT_UNDEFINED},                                      // R16S
    {VK_FORMAT_UNDEFINED},                                      // R16UI
    {VK_FORMAT_UNDEFINED},                                      // R16I
    {VK_FORMAT_R16G16_UNORM, Attachable | Storage},             // RG16
    {VK_FORMAT_R16G16_SFLOAT, Attachable | Storage},            // RG16F
    {VK_FORMAT_UNDEFINED},                                      // RG16UI
    {VK_FORMAT_UNDEFINED},                                      // RG16I
    {VK_FORMAT_R16G16_SNORM, Attachable | Storage},             // RG16S
    {VK_FORMAT_UNDEFINED},                                      // RGB32F
    {VK_FORMAT_R8G8B8A8_SRGB, Attachable},                      // RGBA8_SRGB
    {VK_FORMAT_R8G8_UNORM, Attachable | Storage},               // RG8U
    {VK_FORMAT_R8G8_SNORM, Attachable | Storage},               // RG8S
    {VK_FORMAT_R8G8_UINT, Attachable | Storage},                // RG8UI
    {VK_FORMAT_R32G32_UINT, Attachable | Storage},              // RG32UI
    {VK_FORMAT_UNDEFINED},                                      // RGBX16F
    {VK_FORMAT_R32_UINT, Attachable | Storage},                 // R32UI
    {VK_FORMAT_R32_SINT, Attachable | Storage},                 // R32I
    {VK_FORMAT_ASTC_8x8_UNORM_BLOCK},                           // ASTC_2D_8X8
    {VK_FORMAT_UNDEFINED},                                      // ASTC_2D_8X5
    {VK_FORMAT_UNDEFINED},                                      // ASTC_2D_5X4
    {VK_FORMAT_UNDEFINED},                                      // BGRA8_SRGB
    {VK_FORMAT_BC1_RGBA_SRGB_BLOCK},                            // DXT1_SRGB
    {VK_FORMAT_BC2_SRGB_BLOCK},                                 // DXT23_SRGB
    {VK_FORMAT_BC3_SRGB_BLOCK},                                 // DXT45_SRGB
    {VK_FORMAT_BC7_SRGB_BLOCK},                                 // BC7U_SRGB
    {VK_FORMAT_R4G4B4A4_UNORM_PACK16, Attachable},              // R4G4B4A4U
    {VK_FORMAT_ASTC_4x4_SRGB_BLOCK},                            // ASTC_2D_4X4_SRGB
    {VK_FORMAT_ASTC_8x8_SRGB_BLOCK},                            // ASTC_2D_8X8_SRGB
    {VK_FORMAT_ASTC_8x5_SRGB_BLOCK},                            // ASTC_2D_8X5_SRGB
    {VK_FORMAT_ASTC_5x4_SRGB_BLOCK},                            // ASTC_2D_5X4_SRGB
    {VK_FORMAT_ASTC_5x5_UNORM_BLOCK},                           // ASTC_2D_5X5
    {VK_FORMAT_ASTC_5x5_SRGB_BLOCK},                            // ASTC_2D_5X5_SRGB
    {VK_FORMAT_ASTC_10x8_UNORM_BLOCK},                          // ASTC_2D_10X8
    {VK_FORMAT_ASTC_10x8_SRGB_BLOCK},                           // ASTC_2D_10X8_SRGB
    {VK_FORMAT_ASTC_6x6_UNORM_BLOCK},                           // ASTC_2D_6X6
    {VK_FORMAT_ASTC_6x6_SRGB_BLOCK},                            // ASTC_2D_6X6_SRGB
    {VK_FORMAT_ASTC_10x10_UNORM_BLOCK},                         // ASTC_2D_10X10
    {VK_FORMAT_ASTC_10x10_SRGB_BLOCK},                          // ASTC_2D_10X10_SRGB
    {VK_FORMAT_ASTC_12x12_UNORM_BLOCK},                         // ASTC_2D_12X12
    {VK_FORMAT_ASTC_12x12_SRGB_BLOCK},                          // ASTC_2D_12X12_SRGB
    {VK_FORMAT_ASTC_8x6_UNORM_BLOCK},                           // ASTC_2D_8X6
    {VK_FORMAT_ASTC_8x6_SRGB_BLOCK},                            // ASTC_2D_8X6_SRGB
    {VK_FORMAT_ASTC_6x5_UNORM_BLOCK},                           // ASTC_2D_6X5
    {VK_FORMAT_ASTC_6x5_SRGB_BLOCK},                            // ASTC_2D_6X5_SRGB
    {VK_FORMAT_E5B9G9R9_UFLOAT_PACK32},                         // E5B9G9R9F

    // Depth formats
    {VK_FORMAT_D32_SFLOAT, Attachable}, // Z32F
    {VK_FORMAT_D16_UNORM, Attachable},  // Z16

    // DepthStencil formats
    {VK_FORMAT_D24_UNORM_S8_UINT, Attachable},  // Z24S8
    {VK_FORMAT_D24_UNORM_S8_UINT, Attachable},  // S8Z24 (emulated)
    {VK_FORMAT_D32_SFLOAT_S8_UINT, Attachable}, // Z32FS8
};
static_assert(std::size(tex_format_tuples) == VideoCore::Surface::MaxPixelFormat);

constexpr bool IsZetaFormat(PixelFormat pixel_format) {
    return pixel_format >= PixelFormat::MaxColorFormat &&
           pixel_format < PixelFormat::MaxDepthStencilFormat;
}

} // Anonymous namespace

FormatInfo SurfaceFormat(const VKDevice& device, FormatType format_type, PixelFormat pixel_format) {
    ASSERT(static_cast<std::size_t>(pixel_format) < std::size(tex_format_tuples));

    auto tuple = tex_format_tuples[static_cast<std::size_t>(pixel_format)];
    if (tuple.format == VK_FORMAT_UNDEFINED) {
        UNIMPLEMENTED_MSG("Unimplemented texture format with pixel format={}",
                          static_cast<u32>(pixel_format));
        return {VK_FORMAT_A8B8G8R8_UNORM_PACK32, true, true};
    }

    // Use ABGR8 on hardware that doesn't support ASTC natively
    if (!device.IsOptimalAstcSupported() && VideoCore::Surface::IsPixelFormatASTC(pixel_format)) {
        tuple.format = VideoCore::Surface::IsPixelFormatSRGB(pixel_format)
                           ? VK_FORMAT_A8B8G8R8_SRGB_PACK32
                           : VK_FORMAT_A8B8G8R8_UNORM_PACK32;
    }
    const bool attachable = tuple.usage & Attachable;
    const bool storage = tuple.usage & Storage;

    VkFormatFeatureFlags usage;
    if (format_type == FormatType::Buffer) {
        usage =
            VK_FORMAT_FEATURE_STORAGE_TEXEL_BUFFER_BIT | VK_FORMAT_FEATURE_UNIFORM_TEXEL_BUFFER_BIT;
    } else {
        usage = VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT | VK_FORMAT_FEATURE_TRANSFER_DST_BIT |
                VK_FORMAT_FEATURE_TRANSFER_SRC_BIT;
        if (attachable) {
            usage |= IsZetaFormat(pixel_format) ? VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT
                                                : VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT;
        }
        if (storage) {
            usage |= VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT;
        }
    }
    return {device.GetSupportedFormat(tuple.format, usage, format_type), attachable, storage};
}

VkShaderStageFlagBits ShaderStage(Tegra::Engines::ShaderType stage) {
    switch (stage) {
    case Tegra::Engines::ShaderType::Vertex:
        return VK_SHADER_STAGE_VERTEX_BIT;
    case Tegra::Engines::ShaderType::TesselationControl:
        return VK_SHADER_STAGE_TESSELLATION_CONTROL_BIT;
    case Tegra::Engines::ShaderType::TesselationEval:
        return VK_SHADER_STAGE_TESSELLATION_EVALUATION_BIT;
    case Tegra::Engines::ShaderType::Geometry:
        return VK_SHADER_STAGE_GEOMETRY_BIT;
    case Tegra::Engines::ShaderType::Fragment:
        return VK_SHADER_STAGE_FRAGMENT_BIT;
    case Tegra::Engines::ShaderType::Compute:
        return VK_SHADER_STAGE_COMPUTE_BIT;
    }
    UNIMPLEMENTED_MSG("Unimplemented shader stage={}", static_cast<u32>(stage));
    return {};
}

VkPrimitiveTopology PrimitiveTopology([[maybe_unused]] const VKDevice& device,
                                      Maxwell::PrimitiveTopology topology) {
    switch (topology) {
    case Maxwell::PrimitiveTopology::Points:
        return VK_PRIMITIVE_TOPOLOGY_POINT_LIST;
    case Maxwell::PrimitiveTopology::Lines:
        return VK_PRIMITIVE_TOPOLOGY_LINE_LIST;
    case Maxwell::PrimitiveTopology::LineStrip:
        return VK_PRIMITIVE_TOPOLOGY_LINE_STRIP;
    case Maxwell::PrimitiveTopology::Triangles:
        return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST;
    case Maxwell::PrimitiveTopology::TriangleStrip:
        return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_STRIP;
    case Maxwell::PrimitiveTopology::TriangleFan:
        return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_FAN;
    case Maxwell::PrimitiveTopology::Quads:
        // TODO(Rodrigo): Use VK_PRIMITIVE_TOPOLOGY_QUAD_LIST_EXT whenever it releases
        return VK_PRIMITIVE_TOPOLOGY_TRIANGLE_LIST;
    case Maxwell::PrimitiveTopology::Patches:
        return VK_PRIMITIVE_TOPOLOGY_PATCH_LIST;
    default:
        UNIMPLEMENTED_MSG("Unimplemented topology={}", static_cast<u32>(topology));
        return {};
    }
}

VkFormat VertexFormat(Maxwell::VertexAttribute::Type type, Maxwell::VertexAttribute::Size size) {
    switch (type) {
    case Maxwell::VertexAttribute::Type::SignedNorm:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_8:
            return VK_FORMAT_R8_SNORM;
        case Maxwell::VertexAttribute::Size::Size_8_8:
            return VK_FORMAT_R8G8_SNORM;
        case Maxwell::VertexAttribute::Size::Size_8_8_8:
            return VK_FORMAT_R8G8B8_SNORM;
        case Maxwell::VertexAttribute::Size::Size_8_8_8_8:
            return VK_FORMAT_R8G8B8A8_SNORM;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_SNORM;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_SNORM;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_SNORM;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_SNORM;
        case Maxwell::VertexAttribute::Size::Size_10_10_10_2:
            return VK_FORMAT_A2B10G10R10_SNORM_PACK32;
        default:
            break;
        }
        break;
    case Maxwell::VertexAttribute::Type::UnsignedNorm:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_8:
            return VK_FORMAT_R8_UNORM;
        case Maxwell::VertexAttribute::Size::Size_8_8:
            return VK_FORMAT_R8G8_UNORM;
        case Maxwell::VertexAttribute::Size::Size_8_8_8:
            return VK_FORMAT_R8G8B8_UNORM;
        case Maxwell::VertexAttribute::Size::Size_8_8_8_8:
            return VK_FORMAT_R8G8B8A8_UNORM;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_UNORM;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_UNORM;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_UNORM;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_UNORM;
        case Maxwell::VertexAttribute::Size::Size_10_10_10_2:
            return VK_FORMAT_A2B10G10R10_UNORM_PACK32;
        default:
            break;
        }
        break;
    case Maxwell::VertexAttribute::Type::SignedInt:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_8:
            return VK_FORMAT_R8_SINT;
        case Maxwell::VertexAttribute::Size::Size_8_8:
            return VK_FORMAT_R8G8_SINT;
        case Maxwell::VertexAttribute::Size::Size_8_8_8:
            return VK_FORMAT_R8G8B8_SINT;
        case Maxwell::VertexAttribute::Size::Size_8_8_8_8:
            return VK_FORMAT_R8G8B8A8_SINT;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_SINT;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_SINT;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_SINT;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_SINT;
        case Maxwell::VertexAttribute::Size::Size_32:
            return VK_FORMAT_R32_SINT;
        case Maxwell::VertexAttribute::Size::Size_32_32:
            return VK_FORMAT_R32G32_SINT;
        case Maxwell::VertexAttribute::Size::Size_32_32_32:
            return VK_FORMAT_R32G32B32_SINT;
        case Maxwell::VertexAttribute::Size::Size_32_32_32_32:
            return VK_FORMAT_R32G32B32A32_SINT;
        default:
            break;
        }
        break;
    case Maxwell::VertexAttribute::Type::UnsignedInt:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_8:
            return VK_FORMAT_R8_UINT;
        case Maxwell::VertexAttribute::Size::Size_8_8:
            return VK_FORMAT_R8G8_UINT;
        case Maxwell::VertexAttribute::Size::Size_8_8_8:
            return VK_FORMAT_R8G8B8_UINT;
        case Maxwell::VertexAttribute::Size::Size_8_8_8_8:
            return VK_FORMAT_R8G8B8A8_UINT;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_UINT;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_UINT;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_UINT;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_UINT;
        case Maxwell::VertexAttribute::Size::Size_32:
            return VK_FORMAT_R32_UINT;
        case Maxwell::VertexAttribute::Size::Size_32_32:
            return VK_FORMAT_R32G32_UINT;
        case Maxwell::VertexAttribute::Size::Size_32_32_32:
            return VK_FORMAT_R32G32B32_UINT;
        case Maxwell::VertexAttribute::Size::Size_32_32_32_32:
            return VK_FORMAT_R32G32B32A32_UINT;
        default:
            break;
        }
        break;
    case Maxwell::VertexAttribute::Type::UnsignedScaled:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_8:
            return VK_FORMAT_R8_USCALED;
        case Maxwell::VertexAttribute::Size::Size_8_8:
            return VK_FORMAT_R8G8_USCALED;
        case Maxwell::VertexAttribute::Size::Size_8_8_8:
            return VK_FORMAT_R8G8B8_USCALED;
        case Maxwell::VertexAttribute::Size::Size_8_8_8_8:
            return VK_FORMAT_R8G8B8A8_USCALED;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_USCALED;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_USCALED;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_USCALED;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_USCALED;
        default:
            break;
        }
        break;
    case Maxwell::VertexAttribute::Type::SignedScaled:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_8:
            return VK_FORMAT_R8_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_8_8:
            return VK_FORMAT_R8G8_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_8_8_8:
            return VK_FORMAT_R8G8B8_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_8_8_8_8:
            return VK_FORMAT_R8G8B8A8_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_SSCALED;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_SSCALED;
        default:
            break;
        }
        break;
    case Maxwell::VertexAttribute::Type::Float:
        switch (size) {
        case Maxwell::VertexAttribute::Size::Size_32:
            return VK_FORMAT_R32_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_32_32:
            return VK_FORMAT_R32G32_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_32_32_32:
            return VK_FORMAT_R32G32B32_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_32_32_32_32:
            return VK_FORMAT_R32G32B32A32_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_16:
            return VK_FORMAT_R16_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_16_16:
            return VK_FORMAT_R16G16_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_16_16_16:
            return VK_FORMAT_R16G16B16_SFLOAT;
        case Maxwell::VertexAttribute::Size::Size_16_16_16_16:
            return VK_FORMAT_R16G16B16A16_SFLOAT;
        default:
            break;
        }
        break;
    }
    UNIMPLEMENTED_MSG("Unimplemented vertex format of type={} and size={}", static_cast<u32>(type),
                      static_cast<u32>(size));
    return {};
}

VkCompareOp ComparisonOp(Maxwell::ComparisonOp comparison) {
    switch (comparison) {
    case Maxwell::ComparisonOp::Never:
    case Maxwell::ComparisonOp::NeverOld:
        return VK_COMPARE_OP_NEVER;
    case Maxwell::ComparisonOp::Less:
    case Maxwell::ComparisonOp::LessOld:
        return VK_COMPARE_OP_LESS;
    case Maxwell::ComparisonOp::Equal:
    case Maxwell::ComparisonOp::EqualOld:
        return VK_COMPARE_OP_EQUAL;
    case Maxwell::ComparisonOp::LessEqual:
    case Maxwell::ComparisonOp::LessEqualOld:
        return VK_COMPARE_OP_LESS_OR_EQUAL;
    case Maxwell::ComparisonOp::Greater:
    case Maxwell::ComparisonOp::GreaterOld:
        return VK_COMPARE_OP_GREATER;
    case Maxwell::ComparisonOp::NotEqual:
    case Maxwell::ComparisonOp::NotEqualOld:
        return VK_COMPARE_OP_NOT_EQUAL;
    case Maxwell::ComparisonOp::GreaterEqual:
    case Maxwell::ComparisonOp::GreaterEqualOld:
        return VK_COMPARE_OP_GREATER_OR_EQUAL;
    case Maxwell::ComparisonOp::Always:
    case Maxwell::ComparisonOp::AlwaysOld:
        return VK_COMPARE_OP_ALWAYS;
    }
    UNIMPLEMENTED_MSG("Unimplemented comparison op={}", static_cast<u32>(comparison));
    return {};
}

VkIndexType IndexFormat(const VKDevice& device, Maxwell::IndexFormat index_format) {
    switch (index_format) {
    case Maxwell::IndexFormat::UnsignedByte:
        if (!device.IsExtIndexTypeUint8Supported()) {
            UNIMPLEMENTED_MSG("Native uint8 indices are not supported on this device");
            return VK_INDEX_TYPE_UINT16;
        }
        return VK_INDEX_TYPE_UINT8_EXT;
    case Maxwell::IndexFormat::UnsignedShort:
        return VK_INDEX_TYPE_UINT16;
    case Maxwell::IndexFormat::UnsignedInt:
        return VK_INDEX_TYPE_UINT32;
    }
    UNIMPLEMENTED_MSG("Unimplemented index_format={}", static_cast<u32>(index_format));
    return {};
}

VkStencilOp StencilOp(Maxwell::StencilOp stencil_op) {
    switch (stencil_op) {
    case Maxwell::StencilOp::Keep:
    case Maxwell::StencilOp::KeepOGL:
        return VK_STENCIL_OP_KEEP;
    case Maxwell::StencilOp::Zero:
    case Maxwell::StencilOp::ZeroOGL:
        return VK_STENCIL_OP_ZERO;
    case Maxwell::StencilOp::Replace:
    case Maxwell::StencilOp::ReplaceOGL:
        return VK_STENCIL_OP_REPLACE;
    case Maxwell::StencilOp::Incr:
    case Maxwell::StencilOp::IncrOGL:
        return VK_STENCIL_OP_INCREMENT_AND_CLAMP;
    case Maxwell::StencilOp::Decr:
    case Maxwell::StencilOp::DecrOGL:
        return VK_STENCIL_OP_DECREMENT_AND_CLAMP;
    case Maxwell::StencilOp::Invert:
    case Maxwell::StencilOp::InvertOGL:
        return VK_STENCIL_OP_INVERT;
    case Maxwell::StencilOp::IncrWrap:
    case Maxwell::StencilOp::IncrWrapOGL:
        return VK_STENCIL_OP_INCREMENT_AND_WRAP;
    case Maxwell::StencilOp::DecrWrap:
    case Maxwell::StencilOp::DecrWrapOGL:
        return VK_STENCIL_OP_DECREMENT_AND_WRAP;
    }
    UNIMPLEMENTED_MSG("Unimplemented stencil op={}", static_cast<u32>(stencil_op));
    return {};
}

VkBlendOp BlendEquation(Maxwell::Blend::Equation equation) {
    switch (equation) {
    case Maxwell::Blend::Equation::Add:
    case Maxwell::Blend::Equation::AddGL:
        return VK_BLEND_OP_ADD;
    case Maxwell::Blend::Equation::Subtract:
    case Maxwell::Blend::Equation::SubtractGL:
        return VK_BLEND_OP_SUBTRACT;
    case Maxwell::Blend::Equation::ReverseSubtract:
    case Maxwell::Blend::Equation::ReverseSubtractGL:
        return VK_BLEND_OP_REVERSE_SUBTRACT;
    case Maxwell::Blend::Equation::Min:
    case Maxwell::Blend::Equation::MinGL:
        return VK_BLEND_OP_MIN;
    case Maxwell::Blend::Equation::Max:
    case Maxwell::Blend::Equation::MaxGL:
        return VK_BLEND_OP_MAX;
    }
    UNIMPLEMENTED_MSG("Unimplemented blend equation={}", static_cast<u32>(equation));
    return {};
}

VkBlendFactor BlendFactor(Maxwell::Blend::Factor factor) {
    switch (factor) {
    case Maxwell::Blend::Factor::Zero:
    case Maxwell::Blend::Factor::ZeroGL:
        return VK_BLEND_FACTOR_ZERO;
    case Maxwell::Blend::Factor::One:
    case Maxwell::Blend::Factor::OneGL:
        return VK_BLEND_FACTOR_ONE;
    case Maxwell::Blend::Factor::SourceColor:
    case Maxwell::Blend::Factor::SourceColorGL:
        return VK_BLEND_FACTOR_SRC_COLOR;
    case Maxwell::Blend::Factor::OneMinusSourceColor:
    case Maxwell::Blend::Factor::OneMinusSourceColorGL:
        return VK_BLEND_FACTOR_ONE_MINUS_SRC_COLOR;
    case Maxwell::Blend::Factor::SourceAlpha:
    case Maxwell::Blend::Factor::SourceAlphaGL:
        return VK_BLEND_FACTOR_SRC_ALPHA;
    case Maxwell::Blend::Factor::OneMinusSourceAlpha:
    case Maxwell::Blend::Factor::OneMinusSourceAlphaGL:
        return VK_BLEND_FACTOR_ONE_MINUS_SRC_ALPHA;
    case Maxwell::Blend::Factor::DestAlpha:
    case Maxwell::Blend::Factor::DestAlphaGL:
        return VK_BLEND_FACTOR_DST_ALPHA;
    case Maxwell::Blend::Factor::OneMinusDestAlpha:
    case Maxwell::Blend::Factor::OneMinusDestAlphaGL:
        return VK_BLEND_FACTOR_ONE_MINUS_DST_ALPHA;
    case Maxwell::Blend::Factor::DestColor:
    case Maxwell::Blend::Factor::DestColorGL:
        return VK_BLEND_FACTOR_DST_COLOR;
    case Maxwell::Blend::Factor::OneMinusDestColor:
    case Maxwell::Blend::Factor::OneMinusDestColorGL:
        return VK_BLEND_FACTOR_ONE_MINUS_DST_COLOR;
    case Maxwell::Blend::Factor::SourceAlphaSaturate:
    case Maxwell::Blend::Factor::SourceAlphaSaturateGL:
        return VK_BLEND_FACTOR_SRC_ALPHA_SATURATE;
    case Maxwell::Blend::Factor::Source1Color:
    case Maxwell::Blend::Factor::Source1ColorGL:
        return VK_BLEND_FACTOR_SRC1_COLOR;
    case Maxwell::Blend::Factor::OneMinusSource1Color:
    case Maxwell::Blend::Factor::OneMinusSource1ColorGL:
        return VK_BLEND_FACTOR_ONE_MINUS_SRC1_COLOR;
    case Maxwell::Blend::Factor::Source1Alpha:
    case Maxwell::Blend::Factor::Source1AlphaGL:
        return VK_BLEND_FACTOR_SRC1_ALPHA;
    case Maxwell::Blend::Factor::OneMinusSource1Alpha:
    case Maxwell::Blend::Factor::OneMinusSource1AlphaGL:
        return VK_BLEND_FACTOR_ONE_MINUS_SRC1_ALPHA;
    case Maxwell::Blend::Factor::ConstantColor:
    case Maxwell::Blend::Factor::ConstantColorGL:
        return VK_BLEND_FACTOR_CONSTANT_COLOR;
    case Maxwell::Blend::Factor::OneMinusConstantColor:
    case Maxwell::Blend::Factor::OneMinusConstantColorGL:
        return VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_COLOR;
    case Maxwell::Blend::Factor::ConstantAlpha:
    case Maxwell::Blend::Factor::ConstantAlphaGL:
        return VK_BLEND_FACTOR_CONSTANT_ALPHA;
    case Maxwell::Blend::Factor::OneMinusConstantAlpha:
    case Maxwell::Blend::Factor::OneMinusConstantAlphaGL:
        return VK_BLEND_FACTOR_ONE_MINUS_CONSTANT_ALPHA;
    }
    UNIMPLEMENTED_MSG("Unimplemented blend factor={}", static_cast<u32>(factor));
    return {};
}

VkFrontFace FrontFace(Maxwell::FrontFace front_face) {
    switch (front_face) {
    case Maxwell::FrontFace::ClockWise:
        return VK_FRONT_FACE_CLOCKWISE;
    case Maxwell::FrontFace::CounterClockWise:
        return VK_FRONT_FACE_COUNTER_CLOCKWISE;
    }
    UNIMPLEMENTED_MSG("Unimplemented front face={}", static_cast<u32>(front_face));
    return {};
}

VkCullModeFlags CullFace(Maxwell::CullFace cull_face) {
    switch (cull_face) {
    case Maxwell::CullFace::Front:
        return VK_CULL_MODE_FRONT_BIT;
    case Maxwell::CullFace::Back:
        return VK_CULL_MODE_BACK_BIT;
    case Maxwell::CullFace::FrontAndBack:
        return VK_CULL_MODE_FRONT_AND_BACK;
    }
    UNIMPLEMENTED_MSG("Unimplemented cull face={}", static_cast<u32>(cull_face));
    return {};
}

VkComponentSwizzle SwizzleSource(Tegra::Texture::SwizzleSource swizzle) {
    switch (swizzle) {
    case Tegra::Texture::SwizzleSource::Zero:
        return VK_COMPONENT_SWIZZLE_ZERO;
    case Tegra::Texture::SwizzleSource::R:
        return VK_COMPONENT_SWIZZLE_R;
    case Tegra::Texture::SwizzleSource::G:
        return VK_COMPONENT_SWIZZLE_G;
    case Tegra::Texture::SwizzleSource::B:
        return VK_COMPONENT_SWIZZLE_B;
    case Tegra::Texture::SwizzleSource::A:
        return VK_COMPONENT_SWIZZLE_A;
    case Tegra::Texture::SwizzleSource::OneInt:
    case Tegra::Texture::SwizzleSource::OneFloat:
        return VK_COMPONENT_SWIZZLE_ONE;
    }
    UNIMPLEMENTED_MSG("Unimplemented swizzle source={}", static_cast<u32>(swizzle));
    return {};
}

VkViewportCoordinateSwizzleNV ViewportSwizzle(Maxwell::ViewportSwizzle swizzle) {
    switch (swizzle) {
    case Maxwell::ViewportSwizzle::PositiveX:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_POSITIVE_X_NV;
    case Maxwell::ViewportSwizzle::NegativeX:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_NEGATIVE_X_NV;
    case Maxwell::ViewportSwizzle::PositiveY:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_POSITIVE_Y_NV;
    case Maxwell::ViewportSwizzle::NegativeY:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_NEGATIVE_Y_NV;
    case Maxwell::ViewportSwizzle::PositiveZ:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_POSITIVE_Z_NV;
    case Maxwell::ViewportSwizzle::NegativeZ:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_NEGATIVE_Z_NV;
    case Maxwell::ViewportSwizzle::PositiveW:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_POSITIVE_W_NV;
    case Maxwell::ViewportSwizzle::NegativeW:
        return VK_VIEWPORT_COORDINATE_SWIZZLE_NEGATIVE_W_NV;
    }
    UNREACHABLE_MSG("Invalid swizzle={}", static_cast<int>(swizzle));
    return {};
}

} // namespace Vulkan::MaxwellToVK
