On-Device OCR: RedNote's dots.ocr Model on Apple's Neural Engine

A new method allows sophisticated OCR models to run efficiently on your Apple devices, bypassing traditional hurdles.

RedNote's 3B parameter OCR model, dots.ocr, now runs on Apple's Neural Engine. This advancement, detailed by Hugging Face, offers superior performance and power efficiency for on-device optical character recognition. It simplifies a complex conversion process for developers.

Sarah Kline

By Sarah Kline

October 5, 2025

4 min read

On-Device OCR: RedNote's dots.ocr Model on Apple's Neural Engine

Key Facts

  • RedNote's 3B parameter OCR model, dots.ocr, now runs on Apple's Neural Engine.
  • The Neural Engine is 12x more power efficient than a CPU and 4x more power efficient than a GPU.
  • Conversion from PyTorch to Core ML is a two-step process involving graph capturing and compilation.
  • The dots.ocr model consists of a 1.2B parameter vision encoder and an LM backbone.
  • The conversion process allows for simplified model versions, focusing on single image processing for on-device apps.

Why You Care

Ever wish your iPhone or iPad could instantly read and process text from any image, even offline, with accuracy? Imagine scanning a document or a handwritten note and having its content immediately editable on your device. This isn’t just a futuristic dream anymore. A recent announcement details how the dots.ocr model now runs directly on Apple devices. This means faster, more private, and energy-efficient optical character recognition (OCR) for your everyday tasks.

What Actually Happened

RedNote’s dots.ocr, a 3B parameter OCR model, has been successfully converted to run on Apple’s Neural Engine, according to the announcement. This engine is a custom AI accelerator found in all Apple devices since 2017. The conversion process, outlined by Christopher Fleetwood and Pedro Cuenca, involved adapting the model for Core ML and MLX frameworks. This allows the OCR capabilities to operate directly on-device. The team revealed this process in a three-part series, aiming to help other developers. They hope this will highlight the necessary ideas and tools for running models on-device.

Why This Matters to You

Running AI models directly on your device offers significant advantages. Privacy is enhanced because your data never leaves your phone or tablet. What’s more, performance is often much faster than cloud-based solutions. The research shows that Apple’s Neural Engine is incredibly efficient. It’s 12 times more power efficient than a CPU and 4 times more power efficient than a GPU. This means your device can handle complex tasks without draining its battery quickly.

Consider this real-world scenario: You’re traveling internationally without reliable internet access. With on-device OCR, you could photograph a foreign menu and instantly translate it. Or imagine you’re a student, and you can quickly digitize notes from a whiteboard, even if you’re in a remote location. How might this change the way you interact with physical documents and information?

As Christopher Fleetwood states, “This process should be applicable to many other models, and we hope that this will help highlight the ideas and tools needed for developers looking to run their own models on-device.” This indicates a broader impact beyond just OCR. It paves the way for a new era of , local AI applications.

Neural Engine Efficiency Comparison

Compute UnitRelative Power Efficiency (vs. GPU)
CPU3x less efficient
GPU1x (baseline)
Neural Engine4x more efficient

The Surprising Finding

Perhaps the most surprising finding is the sheer power efficiency of the Neural Engine. While custom hardware for AI is expected to be efficient, the study finds its performance significantly outstrips general-purpose processors. Specifically, the Neural Engine is 12 times more power efficient than a CPU. This challenges the common assumption that AI always requires massive cloud servers or dedicated power-hungry GPUs. It highlights Apple’s foresight in integrating specialized AI silicon years ago. This efficiency is crucial for mobile devices where battery life is paramount. It means complex AI tasks can run without compromising your device’s usability. This opens up possibilities for always-on AI features that were previously impractical.

What Happens Next

This conversion method for dots.ocr suggests a promising future for on-device AI. Developers can now look to adapt more models for Apple’s Neural Engine. We might see more complex applications emerging within the next 6 to 12 months. For example, imagine real-time language translation apps that work entirely offline. This could also include augmented reality experiences with object recognition. The team revealed their process in a series, providing a “reasoning trace” for others. This indicates a collaborative effort to democratize on-device AI creation.

For readers, the actionable takeaway is to anticipate more and private AI features on your Apple devices. Keep an eye out for updates to your favorite apps that use these capabilities. The industry implications are vast, pushing AI processing closer to the user. This reduces reliance on cloud infrastructure and enhances data privacy. As mentioned in the release, the goal is to make this process applicable to “many other models.” This suggests a wave of new, efficient on-device AI applications is on the horizon.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice