Are transformers really all we need? I doubt it. We tested alternative backbones for language models in low-resource scenarios — #Mamba, #xLSTM, and #HGRN2 — and they work surprisingly well!
📄 Paper: aclanthology.org/2024.conll-b...
Thanks for being part of the #BabyLM Challenge! 👶