DiscoverBest AI papers explainedWhen can in-context learning generalize out of task distribution?
When can in-context learning generalize out of task distribution?

When can in-context learning generalize out of task distribution?

Update: 2025-10-16
Share

Description

The research empirically investigates the role of pretraining distribution and a new concept of task diversity in the emergence of ICL, particularly using models trained on linear functions. Findings indicate that increasing task diversity causes transformers to shift from a specialized solution to one that can generalize across the entire task space, a transition also observed in nonlinear regression problems. The authors constructed a phase diagram to characterize how task diversity and the number of pretraining tasks interact, while also examining the influence of factors like model depth and problem dimensionality.

Comments 
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

When can in-context learning generalize out of task distribution?

When can in-context learning generalize out of task distribution?

Enoch H. Kang