Tag: raw MT

  • 7 thoughts: learnings from an offline data room of raw MT

    7 thoughts: learnings from an offline data room of raw MT

    It has been a busy few weeks. My project and process management experience have been put to the test in a different manner. I had to process large volumes of machine translations for subsequent access in an offline data room as part of an inspection mission. Extensive use of raw machine translation was the only possible approach: the quantities of translations required for gisting purposes vastly exceeded available internal and external human translation capacity. In some cases required documents could not be machine translated for various reasons. I would have suggested that “pigs might fly” if anyone had suggested anything to the contrary.

    As I have mentioned elsewhere, in some instances, there was also the risk that raw MT output could result in creating unnecessary downstream workload. Fortunately, such unnecessary work was widely avoided by means of effective but sparing use of human translation. Human translation often provided clarity where raw MT output could not. The exercise also proved useful in providing insights about non-translators’ perception of translation. It reinforced my understanding of the reality for consumers of translation.

    This post addresses several points left unanswered in my first “7 thoughts” piece in January 2025.

    1. Translation remains a blackbox for many non-linguists: their sole interest is a passable output their input. They are not interested in the procedural steps, particularly the obscure ones that assist in ensuring quality. However, issues do surface about the rendering of key terms. This opens up discussions on the need for terminology. The issue has surfaced in my M365 Copilot Champions training. In our icebreaker “speed-dating” sessions with other champions many said they use Copilot to translate. I’d usually follow-up and take them up on issues encountered: the issue of inconsistent terminology leached out.
    2. Having machine translation on tap, seemingly at no charge, can make people wasteful of resources: “seemingly limitless” amounts of capacity available for machine translation meant inflated amounts of translation, than had only human translation capacity been available. Similarly, as meetings expand to fill time slots, no limits on submissions mean that people are less disciplined in the amount of files they request for translation. We lose sight of what is truly essential. In some cases submissions were entire electronic files (with 100s of files and 1000s of pages). I added notes about existing human translations that I supplied as an oversight mechanism. This sometimes elicited surprise: people were unaware of translations’ existence. I’ve taken this away as feedback for the future regarding marketing my output internally.
    3. “Good enough” is really all some people are after: in some cases, a raw machine translation exceeds the requesting party’s own capacity for a “translation”. Logically, such users will embrace machine translation in this case as the default option. Seductive fluency of Modern NMT and LLM/GenAI-based systems also reinforces their satisfaction with the output. Similarly, there is greater tolerance in terms of expectations for MT output over human translation. This issue demonstrates what the translation profession is up against: it is difficult for some to charge premium rates for a time-consuming process, where near instanteous raw MT output apparently “satisfies” their need.
    4. MTPE has a different cognitive load to human translation: I took care when setting out my stall. I chose to centralised rather than delegate handling the documents. By centralising the task, I had a full set of documents in case I needed them. I chose not to look at the raw MT output: it was available on a strictly “as is” basis. This was the only possible option: checking them would have resulted in the need to knock them into shape. That was not my task. Indeed, I would probably have struggled to have found a happy medium in post-editing. Inconsistent terminology would quickly have become an annoyance. And I needed a strategy to do this extra task while not affecting my own working processes, to cement that I work as the expert in the lead and remain firmly in the driving seat. I would not want to suddenly be the passenger, a “human in the loop” in the translation meaning. I became ruthless: discovering limits on daily limits on numbers of documents, size of files, numbers of files, Translating 1,000s of files, while not losing my own translation time was essential. I’d fed batches of files to translate while I was eating lunch, in a call, or waiting for seminars to begin.
    5. The advantage of centralising the process with a translator: given that machine translation is such a black box, a translator needs to maintain an oversight or even override role. In the worst case, I knew how the machine translation had been conducted (i.e. engine/model used, source format etc.) In the case of a poor translation, I also wanted the option to run the source text through the MT engine and have a bilingual XLIFF or TMX output as a starting point for a re-translation, without needing to align the text myself. This was also essential for ensuring the strict segregation of human and machine translation. Also this meant that I had to receive source texts, to avoid being inundated with with target texts with no source text, e.g. texts of GenAI produced output to correct.
    6. The grey zone about internal reliance on machine translation is greater than previously realised: NMT and GenAI have increased staff members’ general awareness of solutions like DeepL (put on a pedestal over Google Translate). The same can be said of the putative ability of GenAI/LLMs to translate. This awareness leads to far more “grey” machine translation than thought. Staff members will use any available solution unless it is physically blocked. Some institutions will actively block tools like Grammarly due to security risks entailed by using it. The same is true for LLMs/GenAI. Until a company blocks access to ChatGPT or DeepSeek etc. staff members will still use them. Merely having rules prohibiting such solutions is not a sufficient deterent.
    7. Effective outcome-based human translation and poor raw MT output can reinforce people’s confidence in human translation: despite seductive fluency, and “passable” quality and speed of delivery and availability, the value of human translation can shine through when the output really matters. People still value knowing they can rely on a translation. Some jobs remain too sensitive for MT despite mitigation measures taken. The human touch also shines through subtly. In one case I had to translate two iterations of a presentation (i.e. to show the progress made on a project).
      I translated the files together (using a virtual merge). This helped ensure the consistency of terminology, which helped guide discussions. MT struggles to be as consistent (e.g. repeated segments inconsistently translated across documents). Then I added a personal touch – I added a “virtual sticky note” to the cover slides to provide a reference to the abbreviations used. By eliminating the guesswork for the reader, it also impressed that this was a human translation from the outset.

    How did I go about it:

    I took a calculated approach to avoid unnecessary cognitive downstreaming to me.

    • I did quick checks for all received folders for standardised documents with only differing content (e.g. a reused template). In certain cases I made template legend available as a human translation in addition to raw MT.
    • On occasions I translated lists e.g. of process names. In preparing the translation, I chose an approach to prioritise maximising meaning over concision – and to ensure that related documents also used these process names – to help improve coherence across many documents.
    • I prioritised document requests arising from meetings for human translation, and highlighted that they were provided as human translations rather than MT.
    • I explained internal jargon (e.g. names of certain committees, working groups, names of soft law instruments) – e.g. using the virtual sticky note.
    • I prioritised some brief softlaw instruments for translation and publication, thereby increasing the permanence of the translation, in contrast to Raw MT for “here and now”.
    • I set up workflows to allow quick alignment of MT and separate read only TMs for MT output, to avoid MT translation contaminating working TMs, while also providing a starting point for human translation.
    • I used feedback requests from specialist divisions to add terminology, and locate documents prioritised for future translation, revision and checking by subject matter experts and publication.

    Turning the process into a pipeline for ongoing translation:

    The successful conclusion to the mission should not be the end of the story. There is another mission ahead. In this regard, the next steps are:

    • Giving contributors and coordinators the chance to find out what went well, and what we could try to improve.
    • Adding a chapter to my language services handbook (LaSH) based on the learnings from the feedback.
    • Prioritising a list of translations for publication prior to the next mission.
    • Process optimisation
      • ensuring documents for translation are saved separately and not embedded in other files.
      • Nomenclature rules for filenames (including limiting length of filenames.)
      • Handling of large files (e.g. chunking to reduce waiting time or uploading for translation during breaks etc.)
  • Outcome-based returns of human translation

    Outcome-based returns of human translation

    In my XITL presentations and discussions in recent months, I have frequently mentioned outcome-based translation. I use it to highlight why human translation is necessary. Frequently, I receive the argument “machine translation can deliver that in x seconds/minutes)”. A human translation of the same text might take hours/days.

    However, this is due to their “one size fits all” consideration of translation. As its purpose or outcome becomes clear, investing (not spending!) time on human translation is the cheaper option.

    Human translation vastly outstrips MT in terms of accountability, credibility and expertise. However, by only being reported on the expenses side of the balance sheet, the return on human translation is missed.

    How human translation beats MT.

    Here’s an example about something I encountered recently. I received an Excel table with raw MT titles of around 100 processes, each using an abbreviated name, grouped thematically. The table had been provided in German. We were informed about processes for review from the English raw MT. The raw MT had been used for that judgement call. It identified 50+ processes to examine more closely.

    We also benefitted from the synergy of my also being a process controller. This role helps me to keep abreast of changes in underlying supervisory law for banking supervision. In turn, I can determine future translation needs (e.g. for templates related to processes for reporting to the ECB). I know our processes inside out. My colleague asked me to review the list of processes, triggering my concerns about the MT output. We both believed it had misled her counterpart.

    To illustrate my concern, I cited five examples from the first block of ten processes. In all cases the MT was far enough removed to influence the request. My colleague checked with her line manager, who confirmed that I should deliver a human translated version of the table.

    While a machine translation might be ready in x seconds or minutes, sometimes the same translation, done by a professional translator in a number of hours, might lead to an outcome that results in days being saved downstream.

    What the translator did

    I compiled the Excel table to translate – due to my translation memory the 800 words took about 90 minutes. I made sure I delivered that day as I was off the following day. Once back in the office, the response was there. The number of processes highlighted had fallen from nearly 50 to around 35.

    Given that I knew what the processes achieve, I ensured in translating their titles that their context was clearer than a “word shifting” raw MT. The raw MT works at segment level only. It also only identified the combination as German > English (probably instructed). I realised that it was in fact Austrian Legal German into English.

    The positive outcome

    We managed to ensure the removal of fifteen processes from the scope for closer examination. We provided extra context on two newly identified processes, leading to their subsequent removal from the list.

    Investing three hours of time (including checking the rationale behind the processes) saved a considerable number of hours (e.g. 15 minutes / process for coordination our side, one hour / process for drafting time for the process owner in explaining the process based on questions asked, and hours of meetings involving multiple staff members). The impact in terms of reduction in unnecessary translation was around 1,500 – 2,000 Normzeilen.

    Unfortunately, other than illustrating the “Raw MT” near miss there is often not comeback on the effects of downstream expenses – unless there is resulting litigation. MT is therefore not held to account in the same way as human translation.

    Further examples of outcome-based translation jobs (non-exhaustive list)

    1. Translation of primary and secondary legislation with quality assurance by subject matter experts. The translations are subsequently used by scores of colleagues in ongoing supervision. These translations are used to confirm that EU Regulations and Directives have been duly transposed into Austrian national legislation.
    2. Translations that permit the opening of court proceedings in relation to supervisory procedures or the timely conclusion of supervisory procedures in relation to supervised entities.
    3. Notifications of supervisory measures taken, including those requiring notification to the European Supervisory Authorities (EBA, EIOPA and ESMA).
    4. Translations of documents that are reused on many occasions.
    5. Translations that assist the correct filling of forms and tables. Explanations help supervised entities to correctly notify the supervisor.