Translator Copilot is Unbabel’s new AI assistant constructed straight into our CAT device. It leverages massive language fashions (LLMs) and Unbabel’s proprietary High quality Estimation (QE) know-how to behave as a wise second pair of eyes for each translation. From checking whether or not buyer directions are adopted to flagging potential errors in actual time, Translator Copilot strengthens the connection between clients and translators, making certain translations should not solely correct however absolutely aligned with expectations.
Why We Constructed Translator Copilot
Translators at Unbabel obtain directions in two methods:
- Common directions outlined on the workflow degree (e.g., formality or formatting preferences)
- Venture-specific directions that apply to explicit information or content material (e.g., “Don’t translate model names”)


These seem within the CAT device and are important for sustaining accuracy and model consistency. However below tight deadlines or with complicated steerage, it’s attainable for these directions to be missed.
That’s the place Translator Copilot is available in. It was created to shut that hole by offering computerized, real-time help. It checks compliance with directions and flags any points because the translator works. Along with instruction checks, it additionally highlights grammar points, omissions, or incorrect terminology, all as a part of a seamless workflow.
How Translator Copilot Helps
The characteristic is designed to ship worth in three core areas:
- Improved compliance: Reduces danger of missed directions
- Larger translation high quality: Flags potential points early
- Diminished value and rework: Minimizes the necessity for guide revisions
Collectively, these advantages make Translator Copilot an important device for quality-conscious translation groups.
From Thought to Integration: How We Constructed It
We started in a managed playground surroundings, testing whether or not LLMs might reliably assess instruction compliance utilizing diversified prompts and fashions. As soon as we recognized the best-performing setup, we built-in it into Polyglot, our inside translator platform.
However figuring out a working setup was simply the beginning. We ran additional evaluations to know how the answer carried out throughout the precise translator expertise, gathering suggestions and refining the characteristic earlier than full rollout.
From there, we introduced every thing collectively: LLM-based instruction checks and QE-powered error detection have been merged right into a single, unified expertise in our CAT device.
What Translators See
Translator Copilot analyzes every section and makes use of visible cues (small coloured dots) to point points. Clicking on a flagged section reveals two forms of suggestions:
- AI Ideas: LLM-powered compliance checks that spotlight deviations from buyer directions
- Doable Errors: Flagged by QE fashions, together with grammar points, mistranslations, or omissions


To help translator workflows and guarantee clean adoption, we added a number of usability options:
- One-click acceptance of solutions
- Capability to report false positives or incorrect solutions
- Fast navigation between flagged segments
- Finish-of-task suggestions assortment to assemble consumer insights
The Technical Challenges We Solved
Bringing Translator Copilot to life concerned fixing a number of powerful challenges:
Low preliminary success price: In early exams, the LLM accurately recognized instruction compliance solely 30% of the time. By way of in depth immediate engineering and supplier experimentation, we raised that to 78% earlier than full rollout.
HTML formatting: Translator directions are written in HTML for readability. However this launched a brand new subject, HTML degraded LLM efficiency. We resolved this by stripping HTML earlier than sending directions to the mannequin, which required cautious immediate design to protect that means and construction.
Glossary alignment: One other early problem was that some mannequin solutions contradicted buyer glossaries. To repair this, we refined prompts to include glossary context, lowering conflicts and boosting belief in AI solutions.
How We Measure Success
To guage Translator Copilot’s impression, we applied a number of metrics:
- Error delta: Evaluating the variety of points flagged in the beginning vs. the tip of every process. A constructive error discount price signifies that the translators are utilizing Copilot to enhance high quality.


- AI solutions versus Doable Errors: AI Ideas led to a 66% error discount price, versus 57% for Doable Errors alone.


- Person habits: In 60% of duties, the variety of flagged points decreased. In 15%, there was no change, seemingly circumstances the place solutions have been ignored. We additionally observe suggestion experiences to enhance mannequin habits.
An fascinating perception emerged from our information: LLM efficiency varies by language pair. For instance, error reporting is greater in German-English, Portuguese-Italian and Portuguese-German, and decrease in english supply language pairs similar to English-Spanish or English-Norwegian, an space we’re persevering with to research.


Wanting Forward
Translator Copilot is an enormous step ahead in combining GenAI and linguist workflows. It brings instruction compliance, error detection, and consumer suggestions into one cohesive expertise. Most significantly, it helps translators ship higher outcomes, sooner.
We’re excited by the early outcomes, and much more enthusiastic about what’s subsequent! That is just the start.