Blog

Feb 28, 2026

How Microsoft Trained a 270M-Pair AI to Power Smarter Search

Researchers at Microsoft Corporation introduce E5, a powerful text embedding model trained on 270M curated web text pairs (CCPairs). Using contrastive learning with in-batch negatives, E5 becomes the first unsupervised model to outperform BM25 on the BEIR benchmark. After fine-tuning, it tops the MTEB leaderboard—beating models 40× larger in retrieval, clustering, classification, and semantic similarity tasks.

Source: HackerNoon →


Share

BTCBTC
$71,714.00
1.83%
ETHETH
$2,219.98
1.29%
USDTUSDT
$1.00
0.01%
XRPXRP
$1.33
1.79%
BNBBNB
$596.19
1.85%
USDCUSDC
$1.00
0.01%
SOLSOL
$82.54
2.74%
TRXTRX
$0.320
0.53%
FIGR_HELOCFIGR_HELOC
$1.04
1.81%
DOGEDOGE
$0.0913
2.78%
USDSUSDS
$1.000
0%
WBTWBT
$52.52
1.37%
HYPEHYPE
$41.11
1.71%
LEOLEO
$10.12
0.14%
ADAADA
$0.244
4.12%
BCHBCH
$427.89
3.73%
LINKLINK
$8.81
3.26%
XMRXMR
$338.52
1.2%
ZECZEC
$361.93
2.35%
USDEUSDE
$1.000
0%