QuartDepth: Post-Training Quantization for Real-Time Depth Estimation on the Edge
Abstract
Monocular Depth Estimation (MDE) has emerged as a pivotal task in computer vision, supporting numerous real-world applications. However, deploying accurate depth estimation models on resource-limited edge devices, especially Application-Specific Integrated Circuits (ASICs), is challenging due to the high computational and memory demands. Recent advancements in foundational depth estimation deliver impressive results but further amplify the difficulty of deployment on ASICs. To address this, we propose QuartDepth which adopts post-training quantization to quantize MDE models with hardware accelerations for ASICs. Our approach involves quantizing both weights and activations to 4-bit precision, reducing the model size and computation cost. To mitigate the performance degradation, we introduce activation polishing and compensation algorithm applied before and after activation quantization, as well as a weight reconstruction method for minimizing errors in weight quantization. Furthermore, we design a flexible and programmable hardware accelerator by supporting kernel fusion and customized instruction programmability, enhancing throughput and efficiency. Experimental results demonstrate that our framework achieves competitive accuracy while enabling fast inference and higher energy efficiency on ASICs, bridging the gap between high-performance depth estimation and practical edge-device applicability. Code: https://github.com/shawnricecake/quart-depth
Community
Accepted by CVPR 2025
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- AHCPTQ: Accurate and Hardware-Compatible Post-Training Quantization for Segment Anything Model (2025)
- FP4DiT: Towards Effective Floating Point Quantization for Diffusion Transformers (2025)
- Accurate INT8 Training Through Dynamic Block-Level Fallback (2025)
- HALO: Hardware-aware quantization with low critical-path-delay weights for LLM acceleration (2025)
- Q-PETR: Quant-aware Position Embedding Transformation for Multi-View 3D Object Detection (2025)
- Post-Training Quantization for Vision Mamba with k-Scaled Quantization and Reparameterization (2025)
- Nearly Lossless Adaptive Bit Switching (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper