Google brings Multi-Token Prediction drafters to Gemma 4: 3x speedup without quality loss

Google brings Multi-Token Prediction drafters to Gemma 4: 3x speedup without quality loss

One month after the introduction of Gemma 4, Google is now releasing Multi-Token Prediction (MTP) drafters for the Gemma 4 model family. These drafters use a specialized speculative decoding architecture, which enables large models such as Gemma 4 26B Mixture-of-Experts (MoE) and 31B Dense to achieve up to a threefold speedup in inference. Importantly, this performance boost comes without any drop in output quality or reasoning accuracy.

The MTP method operates by decoupling the token generation process from verification. While the primary, heavy model completes the final verification of each predicted token, a lighter drafter model predicts multiple future tokens in parallel. This approach takes advantage of idle compute resources, allowing systems to process several tokens using the drafter while the primary model would otherwise be busy with only one.

Following this architectural update, developers can reduce latency dramatically for near real-time chat, voice communication applications, and agentic workflows. Local development also benefits, as the new inference process allows large Gemma 4 models to run efficiently on personal computers and consumer GPUs for seamless, offline coding and planning. For edge devices, the optimized output speed conserves battery life, benefiting the use of E2B and E4B models. These MTP drafters are available now under the open-source Apache 2.0 license.

by Paul

  • ...

Google Gemma is an AI chatbot that belongs to a family of lightweight, state-of-the-art open models, developed using the same research and technology behind the Gemini models. Rated 5, it leverages AI-powered capabilities to provide advanced interactions. Users looking for alternatives may explore other AI chatbots offering similar functionalities.

No comments so far, maybe you want to be first?
Gu