
CLC number:
On-line Access: 2026-02-02
Received: 2025-10-13
Revision Accepted: 2026-01-15
Crosschecked: 0000-00-00
Cited: 0
Clicked: 19
Xichuan Zhou1, Sihuan Zhao1, Rui Ding1, Jiayu Shi1, Jing Nie1, Lihui Chen1, Haijun Liu1. TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization[J]. Journal of Zhejiang University Science C, 1998, -1(-1): .
@article{title="TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization",
author="Xichuan Zhou1, Sihuan Zhao1, Rui Ding1, Jiayu Shi1, Jing Nie1, Lihui Chen1, Haijun Liu1",
journal="Journal of Zhejiang University Science C",
volume="-1",
number="-1",
pages="",
year="1998",
publisher="Zhejiang University Press & Springer",
doi="10.1631/ENG.ITEE.2025.0081"
}
%0 Journal Article
%T TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization
%A Xichuan Zhou1
%A Sihuan Zhao1
%A Rui Ding1
%A Jiayu Shi1
%A Jing Nie1
%A Lihui Chen1
%A Haijun Liu1
%J Journal of Zhejiang University SCIENCE C
%V -1
%N -1
%P
%@ 1869-1951
%D 1998
%I Zhejiang University Press & Springer
%DOI 10.1631/ENG.ITEE.2025.0081
TY - JOUR
T1 - TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization
A1 - Xichuan Zhou1
A1 - Sihuan Zhao1
A1 - Rui Ding1
A1 - Jiayu Shi1
A1 - Jing Nie1
A1 - Lihui Chen1
A1 - Haijun Liu1
J0 - Journal of Zhejiang University Science C
VL - -1
IS - -1
SP -
EP -
%@ 1869-1951
Y1 - 1998
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/ENG.ITEE.2025.0081
Abstract: vision transformers (ViTs) have achieved remarkable success across various artificial intelligence-based computer
vision applications. However, their demanding computational and memory requirements pose significant challenges for deployment on resource-constrained edge devices. Although post-training quantization (PTQ) provides a promising solution by
reducing model precision with minimal calibration data, aggressive low-bit quantization typically leads to substantial performance degradation. To address this challenge, we present the truncated uniform-log2 quantizer and progressive bit-decline
reconstruction method for vision transformers quantization (TP-ViT). It is an innovative PTQ framework specifically designed
for ViTs, featuring two key technical contributions. 1) Truncated uniform-log2 quantizer: This novel quantization approach
effectively handles outlier values in post-softmax activations, significantly reducing quantization errors. 2) Bit-decline optimization strategy: Our progressive quantization method employs transition weights to gradually reduce bit precision while
maintaining model performance under extreme quantization conditions. Comprehensive experiments on image classification,
object detection, and instance segmentation tasks demonstrate TP-ViTa??s superior performance compared to state-of-the-art
PTQ methods, particularly in challenging 3-bit quantization scenarios. Our framework achieves a notable 6.18% improvement in
Top-1 accuracy for ViT-small under 3-bit quantization. These results validate TP-ViTa??s robustness and general applicability,
paving the way for more efficient deployment of ViTs models in computer vision applications on edge hardware.
Open peer comments: Debate/Discuss/Question/Opinion
<1>