Suppose astronomers detect a binary radio signal, an alien message, from a star system many light years away. The message contains a large text dump (conveniently, about GPT-4 training text data sized) composed in an alien language. Let’s call it Alienese.[1]
Unfortunately we don’t understand Alienese.
Until recently, it seemed impossible to learn a language without either
correlating it to sensory experiences shared between the learner and other proficient speakers (like children learn their first language) or
having access to a dictionary which translates the unknown language into another, known language. (The Rosetta Stone served as such a dictionary, which enabled deciphering Egyptian hieroglyphs.)
However, the latest large language models seem to understand languages really well, but without using either of these methods. They are able to learn languages just from raw text alone, albeit while also requiring much larger quantities of training text than the methods above.
This poses a fundamental question:
If an LLM understands language and language , is this sufficient for it to translate between and ?[2]
Unfortunately, it is hardly possible to answer this question empirically using data from human languages. Large text dumps of, say, English and Chinese contain a lot of “Rosetta Stone” content. Bilingual documents, common expressions, translations into related third languages like Japanese, literal English-Chinese dictionaries etc. Since LLMs require a substantial amount of training text, it is not feasible to reliably filter out all this translation content.
But if we received a large text dump in Alienese, we could be certain that no dictionary-like connections to English are present. We could then train a single foundation model (a next token predictor, say a GPT-4 sized model) on both English and Alienese.
By assumption, this LLM would then be able, using adequate prompt engineering, to answer English questions with English answers, and Alienese questions with Alienese answers.
Of course we can’t simply ask any Alienese questions, as we don’t know the language. But we can create a prompt like this:
The following document contains accurate translations of text written in various languages (marked as "Original") into English.
Original: /:wYfh]%xy&v[$49F[CY1.JywUey03ei8EH:KWKY]xHRS#58JfAU:z]L4[gkf*ApjP+T!QYYVTF/F00:;(URv4vci$NU:qm2}$-!R3[BiL.RqwzP!6CCiCh%:wjzB10)xX}%Y45=kV&BFA&]ubnFz$i+9+#$(z;0FK(JjjWCxNZTPdr,v0].6G(/mKCr/J@c0[73M}{Gqi+d11aUe?J[vf4YXa4}w4]6)H]#?XBr:Wg35%)T#60B2:d+Z;jJ$9WgE?;u}uR)x1911k-CE?XhmUYMgt9(:CY7=S)[cKKLbZuU
English:
(Assume the garbled text are Alienese tokens taken from a random document in the alien text dump.)
Can we expect a prompt like this, or a similar one, to produce a reasonably adequate translation of the Alienese text into English?
- ↩︎
Perhaps the binary data dump could be identified as containing language data by testing for something like a character encoding, and whether it obeys common statistical properties of natural language, like Zipf’s Law.
- ↩︎
There is a somewhat similar question called Molyneux’s problem, which asks whether agents can identify objects between two completely unrelated sensory modalities.
Without regard to anything specific to LLMs… Math works the same for all conceivable beings. Beings that live in our universe, of sufficient advancedness, will almost certainly know about hydrogen and other elements, and fundamental constants like Planck lengths. So there will exist commonalities. And then you can build everything else on top of those. If need be, you could describe the way things looked by giving 2D pixel-grid pictures, or describe an apple by starting with elements, molecules, DNA, and so on. (See Contact and That Alien Message for explorations of this type of problem.)
It’s unlikely that any LLM resembling those of today would translate the word for an alien fruit into a description of their own DNA-equivalent and their entire biosphere… But maybe a sufficiently good LLM would have that knowledge inside it, and repeatedly querying it could draw that out.
I guess my question would then be whether the translation would work if neither language contained any information on microphysics or advanced math. Would the model be able to translate e.g. “z;0FK(JjjWCxN” into “fruit”?
The chances of the LLM being able to do this depend heavily on how similar the subjects discussed in the alien language are to things humans discuss. Removing areas where there is most likely to be similarity would reduce the chance that the LLM would find matching patterns in both. Indeed, that we’re imagining aliens for the example already probably greatly increases the difficulty for the LLM.
It’s conceivable how the characters/words are used across English and Alienese have a strong enough correspondence that you can guess matching words much better than chance. But, I’m not confident that you’d have high accuracy.
Consider encryption. If you encrypted messages by mapping the same character to the same character each time, e.g. ‘d’ always gets mapped to ‘6’, then this can be broken with decent accuracy by comparing frequency statistics of characters in your messages with the frequency statistics of characters in the English language.
If you mapped whole words to strings instead of character to character, you could use frequency statistics for whole words in the English language.
Then, between languages, this mostly gets way harder, but you might be able to make some informed guesses, based on
how often you expect certain concepts to be referred to (frequency statistics, although even between human languages, there are probably very important differences)
guesses about extremely common words like ‘a’, ‘the’, ‘of’
possible grammars
similar words being written similarly, like verb tenses of the same verb, noun and verb forms of the same word, etc..
(EDIT) Fine-grained associations between words, e.g. if a given word is used in a random sentence, how often another given word is used in that same sentence. Do this for all ordered pairs of words.
An AI might use similar facts or others, and many more, about much fine-grained and specific uses of words and associations, to guess, but I’m not sure an LLM token predictor mostly just trained on both languages in particular would do a good job.
EDIT: Unsupervised machine translation as Steven Byrnes pointed out seems to be on a better track.
Also, I would add that LLMs trained without perception of things other than text don’t really understand language. The meanings of the words aren’t grounded, and I imagine it could be possible to swap some in a way that would mostly preserve the associations (nearly isomorphic), but I’m not sure.