Blog

Feb 28, 2026

How Microsoft Trained a 270M-Pair AI to Power Smarter Search

Researchers at Microsoft Corporation introduce E5, a powerful text embedding model trained on 270M curated web text pairs (CCPairs). Using contrastive learning with in-batch negatives, E5 becomes the first unsupervised model to outperform BM25 on the BEIR benchmark. After fine-tuning, it tops the MTEB leaderboard—beating models 40× larger in retrieval, clustering, classification, and semantic similarity tasks.

Source: HackerNoon →


Share

BTCBTC
$70,787.00
0.16%
ETHETH
$2,156.72
0.12%
USDTUSDT
$1.000
0.01%
XRPXRP
$1.45
0.61%
BNBBNB
$643.15
0.28%
USDCUSDC
$1.000
0%
SOLSOL
$90.55
1.12%
TRXTRX
$0.310
1.84%
FIGR_HELOCFIGR_HELOC
$1.03
2.72%
DOGEDOGE
$0.0948
0.58%
WBTWBT
$55.40
0.55%
USDSUSDS
$1.000
0.11%
ADAADA
$0.266
1.5%
HYPEHYPE
$39.69
0.62%
BCHBCH
$471.79
1.58%
LEOLEO
$9.22
0.12%
LINKLINK
$9.14
0.05%
XMRXMR
$348.50
3.15%
USDEUSDE
$1.000
0.03%
CCCC
$0.147
2.6%