Blog

Oct 18, 2025

The Illusion of Scale: Why LLMs Are Vulnerable to Data Poisoning, Regardless of Size

The research challenges the conventional wisdom that an attacker needs to control a specific percentage of the training data (e.g., 0.1% or 0.27%) to succeed. For the largest model tested (13B parameters), those 250 poisoned samples represented a minuscule 0.00016% of the total training tokens. Attack success rate remained nearly identical across all tested model scales for a fixed number of poisoned documents.

Source: HackerNoon →


Share

BTCBTC
$89,770.00
0.92%
ETHETH
$3,056.56
1.05%
USDTUSDT
$1.00
0.01%
BNBBNB
$893.81
1.79%
XRPXRP
$2.03
0.34%
USDCUSDC
$1.000
0.01%
SOLSOL
$133.12
0.93%
TRXTRX
$0.287
0.77%
STETHSTETH
$3,053.96
1.28%
DOGEDOGE
$0.140
0.93%
ADAADA
$0.417
0.83%
FIGR_HELOCFIGR_HELOC
$1.02
1.88%
WBTWBT
$60.59
0.09%
WSTETHWSTETH
$3,729.78
1.3%
BCHBCH
$585.62
4.26%
WBTCWBTC
$89,626.00
1.13%
WBETHWBETH
$3,313.06
1.25%
LINKLINK
$13.94
3.72%
USDSUSDS
$1.000
0.04%
BSC-USDBSC-USD
$1.00
0%