Blog

Oct 18, 2025

The Illusion of Scale: Why LLMs Are Vulnerable to Data Poisoning, Regardless of Size

The research challenges the conventional wisdom that an attacker needs to control a specific percentage of the training data (e.g., 0.1% or 0.27%) to succeed. For the largest model tested (13B parameters), those 250 poisoned samples represented a minuscule 0.00016% of the total training tokens. Attack success rate remained nearly identical across all tested model scales for a fixed number of poisoned documents.

Source: HackerNoon →


Share

BTCBTC
$65,604.00
2.93%
ETHETH
$1,924.94
5.14%
USDTUSDT
$1.00
0%
BNBBNB
$611.10
2.1%
XRPXRP
$1.35
3.63%
USDCUSDC
$1.000
0.01%
SOLSOL
$81.77
5.41%
TRXTRX
$0.282
1.19%
FIGR_HELOCFIGR_HELOC
$1.05
1.42%
DOGEDOGE
$0.0933
3.78%
WBTWBT
$48.90
3.31%
ADAADA
$0.276
4%
USDSUSDS
$1.000
0.01%
BCHBCH
$463.32
4.01%
LEOLEO
$8.80
0.3%
HYPEHYPE
$27.20
4.79%
CCCC
$0.170
1.93%
LINKLINK
$8.70
4.41%
XMRXMR
$332.95
2.71%
USDEUSDE
$0.999
0.01%