Skip to yearly menu bar Skip to main content


Poster

Initialization is Critical to Whether Transformers Fit Composite Functions by Inference or Memorizing

Zhongwang Zhang · Pengxiao Lin · Zhiwei Wang · Yaoyu Zhang · Zhi-Qin Xu

East Exhibit Hall A-C #3201
[ ]
Wed 11 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

Transformers have shown impressive capabilities across various tasks, but their performance on compositional problems remains a topic of debate. In this work, we investigate the mechanisms of how transformers behave on unseen compositional tasks. We discover that the parameter initialization scale plays a critical role in determining whether the model learns inferential solutions, which capture the underlying compositional primitives, or symmetric solutions, which simply memorize mappings without understanding the compositional structure. By analyzing the information flow and vector representations within the model, we reveal the distinct mechanisms underlying these solution types. We further find that inferential solutions exhibit low complexity bias, which we hypothesize is a key factor enabling them to learn individual mappings for single anchors. Building upon the understanding of these mechanisms, we can predict the learning behavior of models with different initialization scales when faced with data of varying complexity. Our findings provide valuable insights into the role of initialization scale in shaping the type of solution learned by transformers and their ability to learn and generalize compositional tasks.

Live content is unavailable. Log in and register to view live content