TIL: Masked Language Models Are Surprisingly Capable Zero-Shot Learners

https://www.answer.ai/posts/2025-02-10-modernbert-instruct.html

Previous

This blog post introduces ModernBERT, a family of state-of-the-art encoder-only models representing improvements over older generation encoders across the board, with a 8,192 sequence length,...