New AI Models Challenge LLM Dominance in Key Tasks

The Ettin suite introduces paired encoder-decoder models, shifting focus from decoder-only LLMs.

Researchers have unveiled the Ettin suite, a collection of paired encoder-only and decoder-only AI models. This suite offers a fresh perspective on large language models (LLMs), moving beyond the current decoder-only trend. It aims to provide more specialized and efficient solutions for various AI tasks.

Katie Rowan

By Katie Rowan

March 14, 2026

4 min read

New AI Models Challenge LLM Dominance in Key Tasks

Key Facts

  • The Ettin suite introduces paired encoder-only and decoder-only AI models.
  • Models range from 17 million to 1 billion parameters, trained on up to 2 trillion tokens.
  • Ettin encoders beat ModernBERT, and Ettin decoders beat Llama 3.2 and SmolLM2.
  • Encoder-only models excel at classification and retrieval tasks.
  • Decoder-only models excel at generative tasks.

Why You Care

Ever wonder if your AI tools are truly the best fit for every job? What if the models everyone uses aren’t always the most efficient? A new research paper, “Seq vs Seq: An Open collection of Paired Encoders and Decoders,” suggests exactly that. It challenges the common belief that decoder-only large language models (LLMs) are the universal approach. This creation could change how you approach AI model selection and performance.

What Actually Happened

Researchers have introduced the Ettin collection of models, according to the announcement. This collection features paired encoder-only and decoder-only models. These models range in size from 17 million parameters to 1 billion. They were trained on an impressive up to 2 trillion tokens of data. The goal was to create a fair comparison between these different AI architectures. This collection aims to provide specialized tools for specific natural language processing (NLP) tasks. The team revealed that their training approach produced (SOTA) results. This applies to both encoder-only and decoder-only categories for their respective sizes. For example, the Ettin encoders outperformed ModernBERT. Meanwhile, the Ettin decoders surpassed Llama 3.2 and SmolLM2, as mentioned in the release.

Why This Matters to You

This research highlights a crucial point for anyone using or developing AI. The community often focuses on decoder-only LLMs for text generation. However, a significant portion still relies on encoder-only models. These are used for tasks like classification or retrieval, as detailed in the blog post. The Ettin collection provides a direct, apples-to-apples comparison. It uses the same training methods for both types of models. This allows for clearer insights into their strengths. Think of it as choosing the right tool from your toolbox. You wouldn’t use a hammer for a screw, would you? This study helps you identify the best AI ‘tool’ for your specific needs.

Key Findings from the Ettin collection:

  • Encoder-only models: Excel at classification and retrieval tasks.
  • Decoder-only models: Excel at generative tasks.
  • Adaptation Challenge: Adapting a decoder for encoder tasks (or vice versa) through continued training is less effective than using a purpose-built model.

For example, imagine you run a customer support chatbot. If your main goal is to understand customer intent (classification), an Ettin encoder might be more efficient. If you need to generate detailed responses, an Ettin decoder would be superior. How might this specialized approach impact your current AI workflows?

“We show that adapting a decoder model to encoder tasks (and vice versa) through continued training is subpar compared to using only the reverse objective,” the paper states. This means a dedicated model often performs better than trying to force another model into a different role. This insight can save you considerable resources and improve performance.

The Surprising Finding

Here’s the twist: while previous research often struggled with fair comparisons, the Ettin collection revealed something unexpected. Even with similar training and parameters, a specialized model is usually better than an adapted one. Specifically, the study finds that a 400M encoder outperforms a 1B decoder on MNLI. MNLI is a natural language inference task, which is a classification problem. Conversely, the decoder excels at generative tasks. This challenges the common assumption that bigger, more versatile models are always superior. It suggests that specialized, smaller models can be more effective for certain functions. This finding could lead to more efficient AI deployments. It also means you might not always need the largest LLM for every single task.

What Happens Next

The researchers have open-sourced all artifacts from this study. This includes training data and over 200 checkpoints. This move will allow future work to analyze and extend their findings, as mentioned in the release. We can expect to see new research building on the Ettin collection by late 2026. This will likely involve further exploration into specialized AI architectures. For example, imagine a content creation system using a highly efficient Ettin decoder for article generation. Meanwhile, it uses a precise Ettin encoder for content moderation. This approach could lead to faster, more accurate results with less computational overhead. My advice to you is to keep an eye on upcoming model releases. Look for those designed for specific tasks rather than general-purpose LLMs. This could significantly impact your project’s efficiency and cost. The industry implications are clear: a potential shift towards more specialized, purpose-built AI models.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice