Blog

Nov 17, 2025

Instance-Aware Group Quantization for Vision Transformers

Post-training quantization works well for CNNs but breaks down with Vision Transformers due to highly variable activation distributions. IGQ-ViT solves this by dynamically grouping channels per input instance so each group shares similar statistics, then quantizing them with shared parameters. The method also extends to softmax attention and includes a group-allocation strategy under BOP constraints. Across classification, detection, and segmentation tasks, IGQ-ViT delivers state-of-the-art quantization results for ViTs at low bit-widths without costly retraining.

Source: HackerNoon →


Share

BTCBTC
$90,285.00
2.3%
ETHETH
$3,103.08
4.48%
USDTUSDT
$1.00
0.01%
BNBBNB
$891.98
0.47%
XRPXRP
$2.03
0.47%
USDCUSDC
$1.000
0%
SOLSOL
$133.12
4.46%
STETHSTETH
$3,104.14
4.31%
TRXTRX
$0.272
1.81%
DOGEDOGE
$0.139
1.41%
ADAADA
$0.412
3.12%
FIGR_HELOCFIGR_HELOC
$1.03
0.05%
WBTWBT
$60.60
1.23%
WSTETHWSTETH
$3,796.11
4.22%
BCHBCH
$576.38
1.35%
WBTCWBTC
$90,124.00
2.47%
WBETHWBETH
$3,367.17
4.36%
USDSUSDS
$1.000
0.09%
LINKLINK
$13.73
3.7%
WEETHWEETH
$3,363.01
4.34%