Preferences

There might be some papers or other guides out there, but their advice will be based on whatever tools happened to be available at the time they were written and on the particular types of translations the authors cared about. The technology is advancing so rapidly that you might be better off just experimenting with various LLMs and prompts for texts and language pairs you are interested in.

I started using LLMs for translation after GPT-4 came out in March 2023—not that long ago! At first, the biggest problem was the context window: it wasn’t possible to translate more than a couple of pages at a time. Also, prompt writing was in its infancy, and a lot of techniques that have since emerged were not yet widely known. Even now, I still do a lot of trial and error with my prompts, and I cannot say with confidence that my current prompting methods are the best.

But, for what it’s worth, here are some strategies I currently use when translating with LLMs:

- In the prompt, I explain where the source text came from, how the translation will be used, and how I want it to be translated. Below is a (fictional) example, prepared through some metaprompting experiments with Claude:

https://www.gally.net/temp/20250201sampletranslationprompt.h...

- I run the prompt and source text through several LLMs and glance at the results. If they are generally in the style I want, I start compiling my own translation based on them, choosing the sentences and paragraphs I like most from each. As I go along, I also make my own adjustments to the translation as I see fit.

- After I have finished compiling my draft based on the LLM versions, I check it paragraph by paragraph against the original Japanese (since I can read Japanese) to make sure that nothing is missing or mistranslated. I also continue polishing the English.

- When I am unable to think of a good English version for a particular sentence, I give the Japanese and English versions of the paragraph it is contained in to an LLM (usually, these days, Claude) and ask for ten suggestions for translations of the problematic sentence. Usually one or two of the suggestions work fine; if not, I ask for ten more. (Using an LLM as a sentence-level thesaurus on steroids is particularly wonderful.)

- I give the full original Japanese text and my polished version to one of the LLMs and ask it to compare them sentence by sentence and suggest corrections and improvements to the translation. (I have a separate prompt for this step.) I don’t adopt most of the LLM’s suggestions, but there are usually some that I agree would make the translation better. I update the translation accordingly. I then repeat this step with the updated translation and another LLM, starting a new chat each time. Often I cycle through ChatGPT --> Claude --> Gemini several times before I stop getting suggestions that I feel are worth adopting.

- I then put my final translation through a TTS engine—usually OpenAI’s—and listen to it read aloud. I often catch minor awkwardnesses that I would overlook if reading silently.

This particular workflow works for me because I am using LLMs to translate in the same language direction I did manually for many years. If I had to translate to or from a language I don’t know, I would add extra steps to have LLMs check and double-check the accuracy of the translation and the naturalness of the output.

I was asked recently by some academics I work with about how to use LLMs to translate documents related to their research into Japanese, a language they don’t know. It’s an interesting problem, and I am planning to spend some time thinking about it soon.

Please note that my translation process above is focused on quality, not on speed. If I needed to translate a large volume of text more quickly, I would write a program to do the translation, checking, and rechecking through API calls, accepting the fact that I would not be able to check and polish the translation manually as I do now.

If anyone here would like to brainstorm together about how to use LLMs for translation, please feel free to email me. My website, with my email address on the Contact page, is linked from my HN profile page.


simonw
This comment is solid gold! I will definitely be sending people to it.

Would make a great article for your own site, otherwise I'm happy to link to it here instead.

tkgally OP
Thanks! Feel free to link to the HN comment. That will encourage me to make a video or two for YouTube demonstrating for a wider audience how I use AI for translation. I hope to do that within a few weeks.
simonw
tkgally OP
I saw it. Thank you!
dpcpnry
Thanks for sharing the workflow.

I also use many LLMs to assist my translation tasks.

Recently, I have also been using Google AI Studio [1], and I find the its latest models to be smarter.

[1] https://aistudio.google.com/app/prompts/new_chat

dr_dshiv
Really appreciate the detail and contact. You’ll hear from me.

I have a large collection of Neo-Latin texts I’m trying to get translated.

My goal is to increase the accessibility of the works — not to create a perfect translation. I want to use LLMs to put text on the facing page of the source text. Errors present in the translation, I hope, can be addressed in a Wikimedia-style community editing system.

This approach could .01x lower the cost of translation—and train readers to question translations (something that is a very good thing to learn!)

Wowfunhappy
> Please note that my translation process above is focused on quality, not on speed. If I needed to translate a large volume of text more quickly, I would write a program to do the translation, checking, and rechecking through API calls, accepting the fact that I would not be able to check and polish the translation manually as I do now.

Would you still expect this to produce a better result than Deepl or other purpose-built translation software?

tkgally OP
I don’t know. I stopped using DeepL sometime last year as I found its inability to be prompted about the purpose of the translation to be too limiting for my purposes. At that time, it also had problems with things like maintaining coherent pronoun reference over multiple paragraphs—problems not seen with LLMs. Perhaps DeepL has gotten better since. In any case, I’m sure they have a lot of smart developers and understand well the problems of translation, so I have no reason to think that I would be able to produce a better fully automated translation system than they have.
idunnoman1222
This doesn’t address OP‘s concern at all about the quality degrading as the number of tokens reaches the maximum memory size or perhaps surpasses it.
learning-tr
In your experience which LLM had the best pronunciation ?

This item has no comments currently.