Impact of Pretraining Word Co-Occurrence on Compositional Generalization In

2 badmonster 1 7/11/2025, 6:02:31 PM arxiv.org ↗

Comments (1)

badmonster · 12h ago
a subtle but powerful insight: large multimodal models like CLIP don’t just learn individual concepts. they also depend heavily on how often those concepts appear together during training.