AI Breakthrough Snaps Ancient Cuneiform Characters Into Place for Easy Translation

A new AI-powered tool is revolutionizing the study of cuneiform tablets by automatically aligning and copying complex characters, helping scholars decode ancient texts faster than ever before.

Image Credit: Fedor Selivanov / Shutterstock

Middle Eastern scholars can now use artificial intelligence to identify and copy cuneiform characters from tablet photos, allowing them to read complicated scripts easily.

Researchers from Cornell University and Tel Aviv University (TAU) have developed an approach called ProtoSnap that "snaps" into place a prototype of a character to fit the individual variations imprinted on a tablet. With the new approach, they can make an accurate copy of any character and reproduce whole tablets.

Along with Egyptian hieroglyphs, cuneiform is one of the oldest known forms of writing and consists of more than 1,000 unique characters. The appearance of these characters can vary across eras, cultures, geography, and even individual writers, making them difficult to interpret.

"When you go back to the ancient world, there's a huge variability in the character forms," said Hadar Averbuch-Elor, assistant professor of computer science at Cornell. "Even with the same character, the appearance changes across time, and so it's a very challenging problem to be able to automatically decipher what the character actually means."

Rachel Mikulinsky, a masters student and co-first author from TAU, will present "ProtoSnap: Prototype Alignment for Cuneiform Signs" in April at the International Conference on Learning Representations (ICLR).

An estimated 500,000 cuneiform tablets sit in museums, but only a fraction have been translated and published. "There's an endless amount of 2D scans of these cuneiforms, but the amount of labeled data is very scarce," Averbuch-Elor said.

To see if they could automatically decipher these scans, the team applied a diffusion model – a type of generative AI model often used for computer vision tasks, such as image generation – to calculate the similarity between each pixel in an image of a character on a tablet and a general prototype of the character. Then, they aligned the two versions and snapped the template to match the strokes of the actual character.

The snapped characters can also be used to train downstream AI models that perform optical character recognition—essentially turning images of the tablets into machine-readable text. The researchers showed that, when trained with this data, the downstream models perform far better at recognizing cuneiform characters—even rare ones or ones that show a lot of variation—compared to previous efforts using AI.

This advance could help automate the tablet-copying process, saving experts countless hours and allowing for large-scale comparisons of characters between different times, cities, and writers.

This research received funding from the TAU Center for Artificial Intelligence & Data Science and the LMU-TAU Research Cooperation Program.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of AZoAi.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.