Full Text:   <20>

CLC number: 

On-line Access: 2026-02-02

Received: 2025-10-13

Revision Accepted: 2026-01-15

Crosschecked: 0000-00-00

Cited: 0

Clicked: 17

Citations:  Bibtex RefMan EndNote GB/T7714

-   Go to

Article info.
Open peer comments

Journal of Zhejiang University SCIENCE C 1998 Vol.-1 No.-1 P.

http://doi.org/10.1631/ENG.ITEE.2025.0081


TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization


Author(s):  Xichuan Zhou1, Sihuan Zhao1, Rui Ding1, Jiayu Shi1, Jing Nie1, Lihui Chen1, Haijun Liu1

Affiliation(s):  1School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China

Corresponding email(s):   Xichuan Zhou, zxc@cqu.edu.cn Sihuan Zhao, 202312131061t@stu.cqu.edu.cn Rui Ding, dingrui961210@126.com Haijun Liu, haijun_liu@126.com

Key Words:  Vision transformers, Post-training quantization, Block reconstruction, Image classification, Object detection, Instance segmentation


Share this article to: More

Xichuan Zhou1, Sihuan Zhao1, Rui Ding1, Jiayu Shi1, Jing Nie1, Lihui Chen1, Haijun Liu1. TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization[J]. Journal of Zhejiang University Science C, 1998, -1(-1): .

@article{title="TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization",
author="Xichuan Zhou1, Sihuan Zhao1, Rui Ding1, Jiayu Shi1, Jing Nie1, Lihui Chen1, Haijun Liu1",
journal="Journal of Zhejiang University Science C",
volume="-1",
number="-1",
pages="",
year="1998",
publisher="Zhejiang University Press & Springer",
doi="10.1631/ENG.ITEE.2025.0081"
}

%0 Journal Article
%T TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization
%A Xichuan Zhou1
%A Sihuan Zhao1
%A Rui Ding1
%A Jiayu Shi1
%A Jing Nie1
%A Lihui Chen1
%A Haijun Liu1
%J Journal of Zhejiang University SCIENCE C
%V -1
%N -1
%P
%@ 1869-1951
%D 1998
%I Zhejiang University Press & Springer
%DOI 10.1631/ENG.ITEE.2025.0081

TY - JOUR
T1 - TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision transformers quantization
A1 - Xichuan Zhou1
A1 - Sihuan Zhao1
A1 - Rui Ding1
A1 - Jiayu Shi1
A1 - Jing Nie1
A1 - Lihui Chen1
A1 - Haijun Liu1
J0 - Journal of Zhejiang University Science C
VL - -1
IS - -1
SP -
EP -
%@ 1869-1951
Y1 - 1998
PB - Zhejiang University Press & Springer
ER -
DOI - 10.1631/ENG.ITEE.2025.0081


Abstract: 
vision transformers (ViTs) have achieved remarkable success across various artificial intelligence-based computer vision applications. However, their demanding computational and memory requirements pose significant challenges for deployment on resource-constrained edge devices. Although post-training quantization (PTQ) provides a promising solution by reducing model precision with minimal calibration data, aggressive low-bit quantization typically leads to substantial performance degradation. To address this challenge, we present the truncated uniform-log2 quantizer and progressive bit-decline reconstruction method for vision transformers quantization (TP-ViT). It is an innovative PTQ framework specifically designed for ViTs, featuring two key technical contributions. 1) Truncated uniform-log2 quantizer: This novel quantization approach effectively handles outlier values in post-softmax activations, significantly reducing quantization errors. 2) Bit-decline optimization strategy: Our progressive quantization method employs transition weights to gradually reduce bit precision while maintaining model performance under extreme quantization conditions. Comprehensive experiments on image classification, object detection, and instance segmentation tasks demonstrate TP-ViTa??s superior performance compared to state-of-the-art PTQ methods, particularly in challenging 3-bit quantization scenarios. Our framework achieves a notable 6.18% improvement in Top-1 accuracy for ViT-small under 3-bit quantization. These results validate TP-ViTa??s robustness and general applicability, paving the way for more efficient deployment of ViTs models in computer vision applications on edge hardware.

Darkslateblue:Affiliate; Royal Blue:Author; Turquoise:Article

Open peer comments: Debate/Discuss/Question/Opinion

<1>

Please provide your name, email address and a comment





Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn
Copyright © 2000 - 2026 Journal of Zhejiang University-SCIENCE