Given that the AI we have is prone to making things up because it “fits” according to the models it trains on, how much faith would you have in a translation done by an AI on writings made by people who lived millennia before said language models were developed?
You'd have to ask people who work in the AI field, and, alas, I'm not one of those people.
There has been a lot of language work on attempting to reconstruct the original Indo-European Language, using combinations of pattern recognition and statistical analysis of child languages. Those sorts of tools could aid in deciphering a dead written language. https://en.wikipedia.org/wiki/Proto-Indo-European_language
to expand your point, the sole job of an LLM is to, when given a sequence of words (e.g. half a sentence), predict what the next several words should be. the model has no concept of what English words mean, so instead it makes this prediction based on statistics that were derived from basically reading through hundreds of thousands of English sentences
TL;DR LLMs don’t understand languages, they’ve just memorized statistics about them
I'll have more faith once it can reliably switch back and forth between Unicode symbols and their underlying HTML entities. It understands the concept of emojis and can use them appropriately, but I can tell there's still some underlying issues in the token/object model for non-ASCII symbols.