Blog

Nov 03, 2025

How Inductive Scratchpads Help Transformers Learn Beyond Their Training Data

This section examines how Transformers get beyond the local reasoning barrier with the aid of scratchpads, which are explicit sequences of reasoning stages. By teaching models intermediary steps like cumulative operations or depth-first search (DFS) traces, traditional "educated scratchpads" effectively reduce task localization. Nevertheless, these methods frequently restrict out-of-distribution generalization by overfitting to particular sequence lengths. In order to overcome this, the study presents the inductive scratchpad, a novel formulation that simulates algorithmic induction by having the model iteratively update a state variable using masked attention and reindexed word positions.

Source: HackerNoon →


Share

BTCBTC
$87,178.00
2.54%
ETHETH
$2,921.02
3.78%
USDTUSDT
$1.000
0.01%
BNBBNB
$842.03
2.64%
XRPXRP
$1.88
2.52%
USDCUSDC
$1.000
0%
SOLSOL
$123.02
3.3%
TRXTRX
$0.283
0.29%
STETHSTETH
$2,916.13
4.01%
DOGEDOGE
$0.129
3.19%
FIGR_HELOCFIGR_HELOC
$1.04
0.38%
ADAADA
$0.361
4.09%
WBTWBT
$56.84
2.5%
BCHBCH
$576.42
2.31%
WSTETHWSTETH
$3,571.45
3.84%
WBTCWBTC
$86,989.00
2.6%
WBETHWBETH
$3,174.37
3.71%
USDSUSDS
$1.000
0.01%
WEETHWEETH
$3,167.32
3.73%
BSC-USDBSC-USD
$0.999
0.08%