E

Epub Master

Beta
Mar 10Plus package discount code is now available: SUPER20.
6 min read

Introducing Gemini 2.5 Flash Lite: Better Translation with Batch Processing

EpubMaster now supports Gemini 2.5 Flash Lite, delivering significantly improved translation quality and stability through batch processing.

TL;DR

We've upgraded EpubMaster to use Gemini 2.5 Flash Lite! Here's what this means for you:

  • Batch Translation: Process 10-20 paragraphs at once instead of one-by-one
  • Better Quality: AI now understands context across multiple paragraphs, producing more natural translations
  • Improved Stability: Fewer translation interruptions and errors
  • ⚠️ Trade-off: Translation speed is slightly slower due to batch processing, but still within acceptable range

What You Can Do Now

Translate More at Once

Previously, EpubMaster translated your EPUB one paragraph at a time. While functional, this approach limited the AI's ability to understand context—each paragraph was treated as an isolated piece of text.

Now, you can translate 10-20 paragraphs in a single batch. This means the AI can see the bigger picture, understand how ideas flow across paragraphs, and deliver translations that read more naturally.

Higher Translation Quality

With batch processing, Gemini 2.5 Flash Lite can understand the relationships between sentences and paragraphs. This results in:

  • Better consistency in terminology throughout your book
  • More natural sentence structures
  • Improved handling of idioms and cultural references

Greater Stability

The previous paragraph-by-paragraph approach sometimes led to translation errors or interruptions, especially with longer documents. The new batch processing system is more robust and reliable, giving you a smoother translation experience.


Technical Details

Why Gemini 2.5 Flash Lite?

We chose Gemini 2.5 Flash Lite for its excellent balance of quality and cost-efficiency. This model offers strong contextual understanding while remaining lightweight enough for practical use.

Batch Processing Architecture

Instead of sending individual paragraphs to the AI, we now group them into batches of 10-20 paragraphs. This allows the model to:

  1. See more context: Understanding how ideas connect across paragraphs
  2. Maintain consistency: Using consistent terminology throughout a section
  3. Reduce API overhead: Fewer requests mean less overhead and more efficient processing

Speed vs. Quality Trade-off

Batch processing requires more computational resources, which means translation takes slightly longer than before. However, we believe the significant improvement in translation quality and stability is worth this trade-off. Most users find the extra wait time minimal compared to the quality gains.


We're continuously working to improve EpubMaster. If you have feedback or encounter any issues, please let us know!