🤖 AI Summary
This work addresses learning for moderately input-sensitive functions—those exhibiting sensitivity intermediate between image classification (low sensitivity) and arithmetic/symbolic computation (high sensitivity). Using QR code decoding as a canonical task, we propose the first end-to-end learning-based decoder: a Transformer trained under English text supervision. We find that the model implicitly focuses on data bits rather than Reed–Solomon parity bits, thereby surpassing the theoretical error-correction limit of conventional Reed–Solomon decoding. Experiments demonstrate high accuracy even under severe noise exceeding standard纠错 capabilities, alongside strong generalization to multilingual text and random strings. Our core contribution is uncovering a novel robust decoding mechanism wherein deep models leverage semantic priors—rather than purely syntactic redundancy—to recover structured symbolic outputs. This advances both the theoretical understanding and practical paradigms of learned decoding for structured data.
📝 Abstract
The hardness of learning a function that attains a target task relates to its input-sensitivity. For example, image classification tasks are input-insensitive as minor corruptions should not affect the classification results, whereas arithmetic and symbolic computation, which have been recently attracting interest, are highly input-sensitive as each input variable connects to the computation results. This study presents the first learning-based Quick Response (QR) code decoding and investigates learning functions of medium sensitivity. Our experiments reveal that Transformers can successfully decode QR codes, even beyond the theoretical error-correction limit, by learning the structure of embedded texts. They generalize from English-rich training data to other languages and even random strings. Moreover, we observe that the Transformer-based QR decoder focuses on data bits while ignoring error-correction bits, suggesting a decoding mechanism distinct from standard QR code readers.