Skip to main content
Erschienen in: International Journal on Document Analysis and Recognition (IJDAR) 3/2021

08.07.2021 | Special Issue Paper

Revealing a history: palimpsest text separation with generative networks

verfasst von: Anna Starynska, David Messinger, Yu Kong

Erschienen in: International Journal on Document Analysis and Recognition (IJDAR) | Ausgabe 3/2021

Einloggen

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

A palimpsest is a historical manuscript in which the original text (termed under-text) was erased and overwritten with another script in order to recycle the parchment. One of the main challenges in studying palimpsests is to reveal the under-text. Due to the development of multi-spectral imaging, the original text can sometimes be recovered through material differences of inks and parchment (Easton et al., in: 2011 19th European signal processing conference, IEEE, 2011). However, generally, the revealed text can be observed only partially due to the overlap with newer text and degradation of the material. In this work, we propose revealing the under-text completely using deep generative networks, by leveraging prior spatial information of the under-text script. To optimize the under-text, we mimic the process of palimpsest creation. This is done by generating the under-text from a separately trained generative network to match it to the palimpsest image after mixing it with foreground text. The mixing process is represented by a separate neural network, that is optimized with the under-text image to match the original palimpsest. We also add an additional background generative network to compensate for the unevenness of the background. We propose a novel way of training the background generative network, that does not require isolated background samples and can use any region with layers of text. This paper illustrates the first known attempt to solve palimpsest text layer separation with deep generative networks. We evaluate our method performance on artificial and real palimpsest manuscripts by measuring character recognition and pixel-wise accuracy of the reconstructed under-text.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Fußnoten
1
Sinai Greek 960, a publication of St. Catherine’s Monastery of the Sinai in collaboration with EMEL and UCLA, https://​sinai.​library.​ucla.​edu.
 
Literatur
1.
2.
Zurück zum Zitat Anirudh, R., Thiagarajan, J.J., Kailkhura, B., Bremer, T.: An unsupervised approach to solving inverse problems using generative adversarial networks. arXiv preprint arXiv:1805.07281 (2018) Anirudh, R., Thiagarajan, J.J., Kailkhura, B., Bremer, T.: An unsupervised approach to solving inverse problems using generative adversarial networks. arXiv preprint arXiv:​1805.​07281 (2018)
4.
Zurück zum Zitat Asim, M., Shamshad, F., Ahmed, A.: Blind image deconvolution using deep generative priors. In: 30th British Machine Vision Conference (2019) Asim, M., Shamshad, F., Ahmed, A.: Blind image deconvolution using deep generative priors. In: 30th British Machine Vision Conference (2019)
5.
Zurück zum Zitat Bora, A., Jalal, A., Price, E., Dimakis, A.G.: Compressed sensing using generative models. In: Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 537–546. JMLR. org (2017) Bora, A., Jalal, A., Price, E., Dimakis, A.G.: Compressed sensing using generative models. In: Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 537–546. JMLR. org (2017)
6.
Zurück zum Zitat Bora, A., Price, E., Dimakis, A.G.: Ambientgan: generative models from lossy measurements. ICLR 2, 5 (2018) Bora, A., Price, E., Dimakis, A.G.: Ambientgan: generative models from lossy measurements. ICLR 2, 5 (2018)
8.
Zurück zum Zitat Easton, R.L., Christens-Barry, W.A., Knox, K.T.: Spectral image processing and analysis of the archimedes palimpsest. In: 2011 19th European Signal Processing Conference, pp. 1440–1444. IEEE (2011) Easton, R.L., Christens-Barry, W.A., Knox, K.T.: Spectral image processing and analysis of the archimedes palimpsest. In: 2011 19th European Signal Processing Conference, pp. 1440–1444. IEEE (2011)
9.
Zurück zum Zitat Easton, R.L., Knox, K.T., Christens-Barry, W.A., Boydston, K., Toth, M.B., Emery, D., Noel, W.: Standardized system for multispectral imaging of palimpsests. In: Computer Vision and Image Analysis of Art, vol. 7531, p. 75310D. International Society for Optics and Photonics (2010) Easton, R.L., Knox, K.T., Christens-Barry, W.A., Boydston, K., Toth, M.B., Emery, D., Noel, W.: Standardized system for multispectral imaging of palimpsests. In: Computer Vision and Image Analysis of Art, vol. 7531, p. 75310D. International Society for Optics and Photonics (2010)
10.
Zurück zum Zitat Fogel, S., Averbuch-Elor, H., Cohen, S., Mazor, S., Litman, R.: Scrabblegan: semi-supervised varying length handwritten text generation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4324–4333 (2020) Fogel, S., Averbuch-Elor, H., Cohen, S., Mazor, S., Litman, R.: Scrabblegan: semi-supervised varying length handwritten text generation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4324–4333 (2020)
11.
Zurück zum Zitat Gandelsman, Y., Shocher, A., Irani, M.: Double-dip: unsupervised image decomposition via coupled deep-image-priors. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 6, p. 2 (2019) Gandelsman, Y., Shocher, A., Irani, M.: Double-dip: unsupervised image decomposition via coupled deep-image-priors. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 6, p. 2 (2019)
12.
Zurück zum Zitat Glaser, L., Deckers, D.: The basics of fast-scanning XRF element mapping for iron-gall ink palimpsests. Manuscr. Cult. 7(PUBDB–2015–06320), 104–112 (2014) Glaser, L., Deckers, D.: The basics of fast-scanning XRF element mapping for iron-gall ink palimpsests. Manuscr. Cult. 7(PUBDB–2015–06320), 104–112 (2014)
13.
Zurück zum Zitat Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672–2680 (2014) Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672–2680 (2014)
14.
Zurück zum Zitat Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: NIPS, pp. 2672–2680 (2014) Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: NIPS, pp. 2672–2680 (2014)
15.
Zurück zum Zitat Hanif, M., Tonazzini, A., Savino, P., Salerno, E.: Sparse representation based inpainting for the restoration of document images affected by bleed-through. In: Multidisciplinary Digital Publishing Institute Proceedings, vol. 2, p. 93 (2018) Hanif, M., Tonazzini, A., Savino, P., Salerno, E.: Sparse representation based inpainting for the restoration of document images affected by bleed-through. In: Multidisciplinary Digital Publishing Institute Proceedings, vol. 2, p. 93 (2018)
16.
Zurück zum Zitat He, S., Schomaker, L.: Deepotsu: document enhancement and binarization using iterative deep learning. Pattern Recognit. 91, 379–390 (2019)CrossRef He, S., Schomaker, L.: Deepotsu: document enhancement and binarization using iterative deep learning. Pattern Recognit. 91, 379–390 (2019)CrossRef
17.
Zurück zum Zitat Hollaus, F., Gau, M., Sablatnig, R., Christens-Barry, W.A., Miklas, H.: Readability enhancement and palimpsest decipherment of historical manuscripts. Kodikologie und Paläographie im Digitalen Zeitalter 3: Codicology and Palaeography in the Digital Age, vol. 3, p. 31 (2015) Hollaus, F., Gau, M., Sablatnig, R., Christens-Barry, W.A., Miklas, H.: Readability enhancement and palimpsest decipherment of historical manuscripts. Kodikologie und Paläographie im Digitalen Zeitalter 3: Codicology and Palaeography in the Digital Age, vol. 3, p. 31 (2015)
18.
Zurück zum Zitat Jacobs, B., Momoniat, E.: A novel approach to text binarization via a diffusion-based model. Appl. Math. Comput. 225, 446–460 (2013)MathSciNetMATH Jacobs, B., Momoniat, E.: A novel approach to text binarization via a diffusion-based model. Appl. Math. Comput. 225, 446–460 (2013)MathSciNetMATH
20.
Zurück zum Zitat Kong, Q., Xu, Y., Wang, W., Jackson, P.J., Plumbley, M.D.: Single-channel signal separation and deconvolution with generative adversarial networks. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) (2019) Kong, Q., Xu, Y., Wang, W., Jackson, P.J., Plumbley, M.D.: Single-channel signal separation and deconvolution with generative adversarial networks. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI-19) (2019)
21.
Zurück zum Zitat LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)CrossRef LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)CrossRef
22.
Zurück zum Zitat Lettner, M., Sablatnig, R.: Multispectral imaging for analyzing ancient manuscripts. In: 2009 17th European Signal Processing Conference, pp. 1200–1204. IEEE (2009) Lettner, M., Sablatnig, R.: Multispectral imaging for analyzing ancient manuscripts. In: 2009 17th European Signal Processing Conference, pp. 1200–1204. IEEE (2009)
23.
Zurück zum Zitat Li, J., Cui, R., Li, Y., Li, B., Du, Q., Ge, C.: Multitemporal hyperspectral image super-resolution through 3d generative adversarial network. In: 2019 10th International Workshop on the Analysis of Multitemporal Remote Sensing Images (MultiTemp), pp. 1–4. IEEE (2019) Li, J., Cui, R., Li, Y., Li, B., Du, Q., Ge, C.: Multitemporal hyperspectral image super-resolution through 3d generative adversarial network. In: 2019 10th International Workshop on the Analysis of Multitemporal Remote Sensing Images (MultiTemp), pp. 1–4. IEEE (2019)
24.
Zurück zum Zitat Li, S.C.X., Jiang, B., Marlin, B.: Misgan: learning from incomplete data with generative adversarial networks. In: ICLR (2019) Li, S.C.X., Jiang, B., Marlin, B.: Misgan: learning from incomplete data with generative adversarial networks. In: ICLR (2019)
25.
Zurück zum Zitat Lunz, S., Öktem, O., Schönlieb, C.B.: Adversarial regularizers in inverse problems. In: NIPS, pp. 8507–8516 (2018) Lunz, S., Öktem, O., Schönlieb, C.B.: Adversarial regularizers in inverse problems. In: NIPS, pp. 8507–8516 (2018)
26.
Zurück zum Zitat Mindermann, S.: Hyperspectral imaging for readability enhancement of historic manuscripts. Ph.D. thesis, Technical University of Munich (2018) Mindermann, S.: Hyperspectral imaging for readability enhancement of historic manuscripts. Ph.D. thesis, Technical University of Munich (2018)
27.
Zurück zum Zitat Moghaddam, R.F., Cheriet, M.: Low quality document image modeling and enhancement. Int. J. Doc. Anal. Recognit. (IJDAR) 11(4), 183–201 (2009)CrossRef Moghaddam, R.F., Cheriet, M.: Low quality document image modeling and enhancement. Int. J. Doc. Anal. Recognit. (IJDAR) 11(4), 183–201 (2009)CrossRef
28.
Zurück zum Zitat Pajot, A., de Bezenac, E., Gallinari, P.: Unsupervised adversarial image reconstruction. In: ICLR (2019) Pajot, A., de Bezenac, E., Gallinari, P.: Unsupervised adversarial image reconstruction. In: ICLR (2019)
29.
Zurück zum Zitat Rapantzikos, K., Balas, C.: Hyperspectral imaging: potential in non-destructive analysis of palimpsests. In: IEEE International Conference on Image Processing 2005, vol. 2, pp. II–618. IEEE (2005) Rapantzikos, K., Balas, C.: Hyperspectral imaging: potential in non-destructive analysis of palimpsests. In: IEEE International Conference on Image Processing 2005, vol. 2, pp. II–618. IEEE (2005)
30.
Zurück zum Zitat Rick Chang, J., Li, C.L., Poczos, B., Vijaya Kumar, B., Sankaranarayanan, A.C.: One network to solve them all–solving linear inverse problems using deep projection models. In: CVPR, pp. 5888–5897 (2017) Rick Chang, J., Li, C.L., Poczos, B., Vijaya Kumar, B., Sankaranarayanan, A.C.: One network to solve them all–solving linear inverse problems using deep projection models. In: CVPR, pp. 5888–5897 (2017)
32.
Zurück zum Zitat Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention, pp. 234–241. Springer (2015) Ronneberger, O., Fischer, P., Brox, T.: U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention, pp. 234–241. Springer (2015)
33.
Zurück zum Zitat Salerno, E., Tonazzini, A., Bedini, L.: Digital image analysis to enhance underwritten text in the Archimedes palimpsest. Int. J. Doc. Anal. Recognit. (IJDAR) 9(2–4), 79–87 (2007)CrossRef Salerno, E., Tonazzini, A., Bedini, L.: Digital image analysis to enhance underwritten text in the Archimedes palimpsest. Int. J. Doc. Anal. Recognit. (IJDAR) 9(2–4), 79–87 (2007)CrossRef
35.
Zurück zum Zitat Soltani, M., Jain, S., Sambasivan, A.: Learning generative models of structured signals from their superposition using gans with application to denoising and demixing. arXiv preprint arXiv:1902.04664 (2019) Soltani, M., Jain, S., Sambasivan, A.: Learning generative models of structured signals from their superposition using gans with application to denoising and demixing. arXiv preprint arXiv:​1902.​04664 (2019)
36.
Zurück zum Zitat Starynska, A., Easton Jr, R.L., Messinger, D.: Methods of data augmentation for palimpsest character recognition with deep neural network. In: Proceedings of the 4th International Workshop on Historical Document Imaging and Processing, pp. 54–58. ACM (2017) Starynska, A., Easton Jr, R.L., Messinger, D.: Methods of data augmentation for palimpsest character recognition with deep neural network. In: Proceedings of the 4th International Workshop on Historical Document Imaging and Processing, pp. 54–58. ACM (2017)
37.
Zurück zum Zitat Tonazzini, A., Bedini, L., Salerno, E.: A Markov model for blind image separation by a mean-field EM algorithm. IEEE Trans. Image Process. 15(2), 473–482 (2006)MathSciNetCrossRef Tonazzini, A., Bedini, L., Salerno, E.: A Markov model for blind image separation by a mean-field EM algorithm. IEEE Trans. Image Process. 15(2), 473–482 (2006)MathSciNetCrossRef
38.
Zurück zum Zitat Tonazzini, A., Savino, P., Salerno, E.: A non-stationary density model to separate overlapped texts in degraded documents. Signal Image Video Process. 9(1), 155–164 (2015)CrossRef Tonazzini, A., Savino, P., Salerno, E.: A non-stationary density model to separate overlapped texts in degraded documents. Signal Image Video Process. 9(1), 155–164 (2015)CrossRef
39.
Zurück zum Zitat Trier, O.D., Jain, A.K.: Goal-directed evaluation of binarization methods. IEEE Trans. Pattern Anal. Mach. Intell. 17(12), 1191–1201 (1995)CrossRef Trier, O.D., Jain, A.K.: Goal-directed evaluation of binarization methods. IEEE Trans. Pattern Anal. Mach. Intell. 17(12), 1191–1201 (1995)CrossRef
40.
Zurück zum Zitat Ulyanov, D., Vedaldi, A., Lempitsky, V.: Deep image prior. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018) Ulyanov, D., Vedaldi, A., Lempitsky, V.: Deep image prior. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)
41.
Zurück zum Zitat Valdiviezo-N, J.C., Urcid, G.: Multispectral images segmentation of ancient documents with lattice memories. In: Digital Image Processing and Analysis, p. DMD6. Optical Society of America (2010) Valdiviezo-N, J.C., Urcid, G.: Multispectral images segmentation of ancient documents with lattice memories. In: Digital Image Processing and Analysis, p. DMD6. Optical Society of America (2010)
43.
Zurück zum Zitat Van der Walt, S., Schönberger, J.L., Nunez-Iglesias, J., Boulogne, F., Warner, J.D., Yager, N., Gouillart, E., Yu, T.: scikit-image: image processing in python. PeerJ 2, e453 (2014)CrossRef Van der Walt, S., Schönberger, J.L., Nunez-Iglesias, J., Boulogne, F., Warner, J.D., Yager, N., Gouillart, E., Yu, T.: scikit-image: image processing in python. PeerJ 2, e453 (2014)CrossRef
45.
Zurück zum Zitat Wu, K., Otoo, E., Shoshani, A.: Optimizing connected component labeling algorithms. In: Medical Imaging 2005: Image Processing, vol. 5747, pp. 1965–1977. International Society for Optics and Photonics (2005) Wu, K., Otoo, E., Shoshani, A.: Optimizing connected component labeling algorithms. In: Medical Imaging 2005: Image Processing, vol. 5747, pp. 1965–1977. International Society for Optics and Photonics (2005)
46.
Zurück zum Zitat Yeh, R.A., Chen, C., Yian Lim, T., Schwing, A.G., Hasegawa-Johnson, M., Do, M.N.: Semantic image inpainting with deep generative models. In: CVPR, pp. 5485–5493 (2017) Yeh, R.A., Chen, C., Yian Lim, T., Schwing, A.G., Hasegawa-Johnson, M., Do, M.N.: Semantic image inpainting with deep generative models. In: CVPR, pp. 5485–5493 (2017)
Metadaten
Titel
Revealing a history: palimpsest text separation with generative networks
verfasst von
Anna Starynska
David Messinger
Yu Kong
Publikationsdatum
08.07.2021
Verlag
Springer Berlin Heidelberg
Erschienen in
International Journal on Document Analysis and Recognition (IJDAR) / Ausgabe 3/2021
Print ISSN: 1433-2833
Elektronische ISSN: 1433-2825
DOI
https://doi.org/10.1007/s10032-021-00379-z

Weitere Artikel der Ausgabe 3/2021

International Journal on Document Analysis and Recognition (IJDAR) 3/2021 Zur Ausgabe