Blog

Oct 18, 2025

The Illusion of Scale: Why LLMs Are Vulnerable to Data Poisoning, Regardless of Size

The research challenges the conventional wisdom that an attacker needs to control a specific percentage of the training data (e.g., 0.1% or 0.27%) to succeed. For the largest model tested (13B parameters), those 250 poisoned samples represented a minuscule 0.00016% of the total training tokens. Attack success rate remained nearly identical across all tested model scales for a fixed number of poisoned documents.

Source: HackerNoon →


Share

BTCBTC
$68,961.00
2.53%
ETHETH
$2,030.82
4.89%
USDTUSDT
$0.999
0%
XRPXRP
$1.40
3.07%
BNBBNB
$624.91
2.88%
USDCUSDC
$1.000
0%
SOLSOL
$83.72
4.71%
TRXTRX
$0.277
0.53%
DOGEDOGE
$0.0938
4.16%
FIGR_HELOCFIGR_HELOC
$1.03
0%
WBTWBT
$52.60
2.43%
BCHBCH
$515.91
1.81%
ADAADA
$0.263
3.88%
USDSUSDS
$1.000
0.01%
LEOLEO
$8.34
2.37%
HYPEHYPE
$31.49
0.99%
USDEUSDE
$0.998
0.01%
CCCC
$0.163
4.08%
LINKLINK
$8.59
4.63%
XMRXMR
$325.13
0.84%