Acta Informatica Pragensia 2025, 14(2), 207-214 | DOI: 10.18267/j.aip.2681365

Exploring Oral History Archives Using State-of-the-Art Artificial Intelligence Methods

Martin Bulín ORCID..., Jan ©vec ORCID..., Pavel Ircing ORCID..., Adam Frémund ORCID..., Filip Polák ORCID...
Department of Cybernetics, Faculty of Applied Sciences, University of West Bohemia in Pilsen, Pilsen, Czech Republic

Background: The preservation and analysis of spoken data in oral history archives, such as Holocaust testimonies, provide a vast and complex knowledge source. These archives pose unique challenges and opportunities for computational methods, particularly in self-supervised learning and information retrieval.

Objective: This study explores the application of state-of-the-art artificial intelligence (AI) models, particularly transformer-based architectures, to enhance navigation and engagement with large-scale oral history testimonies. The goal is to improve accessibility while preserving the authenticity and integrity of historical records.

Methods: We developed an asking questions framework utilizing a fine-tuned T5 model to generate contextually relevant questions from interview transcripts. To ensure semantic coherence, we introduced a semantic continuity model based on a BERT-like architecture trained with contrastive loss.

Results: The system successfully generated contextually relevant questions from oral history testimonies, enhancing user navigation and engagement. Filtering techniques improved question quality by retaining only semantically coherent outputs, ensuring alignment with the testimony content. The approach demonstrated effectiveness in handling spontaneous, unstructured speech, with a significant improvement in question relevance compared to models trained on structured text. Applied to real-world interview transcripts, the framework balanced enrichment of user experience with preservation of historical authenticity.

Conclusion: By integrating generative AI models with robust retrieval techniques, we enhance the accessibility of oral history archives while maintaining their historical integrity. This research demonstrates how AI-driven approaches can facilitate interactive exploration of vast spoken data repositories, benefiting researchers, historians and the general public.

Keywords: AI; Oral history archives; Transformer-based models; Machine learning in digital humanities.

Received: February 2, 2025; Revised: April 23, 2025; Accepted: May 20, 2025; Prepublished online: June 23, 2025; Published: July 26, 2025  Show citation

ACS AIP APA ASA Harvard Chicago Chicago Notes IEEE ISO690 MLA NLM Turabian Vancouver
Bulín, M., ©vec, J., Ircing, P., Frémund, A., & Polák, F. (2025). Exploring Oral History Archives Using State-of-the-Art Artificial Intelligence Methods. Acta Informatica Pragensia14(2), 207-214. doi: 10.18267/j.aip.268
Download citation

References

  1. Byrne, W., Doermann, D., Franz, M., Gustman, S., Hajic, J., Oard, D., Picheny, M., Psutka, J., Ramabhadran, B., Soergel, D., Ward, T. & Zhu, W. J. (2004). Automatic recognition of spontaneous speech for access to multilingual oral history archives. IEEE Transactions on Speech and Audio Processing, 12(4), 420-435. https://doi.org/10.1109/TSA.2004.828702 Go to original source...
  2. Gospodinov, M., MacAvaney, S., & Macdonald, C. (2023). Doc2Query-: when less is more. In European Conference on Information Retrieval, (pp. 414-422). Springer. https://doi.org/10.1007/978-3-031-28238-6_31 Go to original source...
  3. He, B., & Ounis, I. (2009). Studying query expansion effectiveness. In European conference on information retrieval, (pp. 611-619). Springer. https://doi.org/10.1007/978-3-642-00958-7_57 Go to original source...
  4. Ircing, P., & Müller, L. (2006). Benefit of proper language processing for Czech speech retrieval in the CL-SR task at CLEF 2006. In Workshop of the Cross-Language Evaluation Forum for European Languages, (pp. 759-765). Springer. https://doi.org/10.1007/978-3-540-74999-8_95 Go to original source...
  5. Ircing, P., Psutka, J., & Vavruąka, J. (2008). What can and cannot be found in Czech spontaneous speech using document-oriented IR methods-UWB at CLEF 2007 CL-SR track. In Workshop of the Cross-Language Evaluation Forum for European Languages, (pp. 712-718). Springer. https://doi.org/10.1007/978-3-540-85760-0_90 Go to original source...
  6. Khashabi, D., Min, S., Khot, T., Sabharwal, A., Tafjord, O., Clark, P., & Hajishirzi, H. (2020). UNIFIEDQA: Crossing Format Boundaries with a Single QA System. In Findings of the Association for Computational Linguistics: EMNLP 2020, (pp. 1896-1907). ACL. https://doi.org/10.18653/v1/2020.findings-emnlp.171 Go to original source...
  7. Krubiński, M., Ghadery, E., Moens, M. F., & Pecina, P. (2021). Just ask! evaluating machine translation by asking and answering questions. In Proceedings of the Sixth Conference on Machine Translation, (pp. 495-506). ACL.
  8. Lehečka, J., ©vec, J., Psutka, J. V., & Ircing, P. (2023). Transformer-based speech recognition models for oral history archives in English, German, and Czech. In Proceedings of the Interspeech 2023, (pp. 201-205). ISCA. https://doi.org/10.21437/Interspeech.2023-872 Go to original source...
  9. Mao, H. H., Li, S., McAuley, J., & Cottrell, G. (2020). Speech Recognition and Multi-Speaker Diarization of Long Conversations. In Proceedings of the Interspeech 2020, (pp.691-695). ISCA. https://doi.org/10.21437/Interspeech.2020-3039 Go to original source...
  10. Psutka, J., Ircing, P., Psutka, J. V., Hajič, J., Byrne, W. J., & Mírovský, J. (2005). Automatic transcription of Czech, Russian, and Slovak spontaneous speech in the MALACH project. In Proceedings of the Interspeech 2005, (pp. 1349-1352). ISCA. https://doi.org/10.21437/Interspeech.2005-489 Go to original source...
  11. Psutka, J., ©vec, J., Psutka, J. V., Vaněk, J., Praľák, A., ©mídl, L., & Ircing, P. (2011). System for fast lexical and phonetic spoken term detection in a Czech cultural heritage archive. EURASIP Journal on Audio Speech and Music Processing, 2011(1), Article 10. https://doi.org/10.1186/1687-4722-2011-10 Go to original source...
  12. Raffel, C., Shazeer, N., Roberts, A., Lee, K., Narang, S., Matena, M., Zhou, Y., Li, W. & Liu, P. J. (2020). Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of machine learning research, 21, 1-67.
  13. Rajpurkar, P., Jia, R., & Liang, P. (2018). Know What You Don't Know: Unanswerable Questions for SQuAD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), (pp. 784-789). ACM. https://doi.org/10.18653/v1/P18-2124 Go to original source...
  14. Reimers, N., & Gurevych, I. (2019). Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks. In Proceedings of the 2019 EMNLP-IJCNLP, (pp. 3982-3992). ACL. https://doi.org/10.18653/v1/D19-1410 Go to original source...
  15. ©vec, J., Lehečka, J., ©mídl, L. (2022a) Deep LSTM Spoken Term Detection using Wav2Vec 2.0 Recognizer. In Proceedings of the Interspeech 2022, (pp.1886-1890). ISCA. https://doi.org/10.21437/Interspeech.2022-10409 Go to original source...
  16. ©vec, J., Neduchal, P., & Hrúz, M. (2022b). Multi-modal communication system for mobile robot. IFAC-PapersOnLine, 55(4), 133-138. https://doi.org/10.1016/j.ifacol.2022.06.022 Go to original source...
  17. ©vec, J., Bulín, M., Frémund, A., & Polák, F. (2024). Asking questions framework for oral history archives. In European Conference on Information Retrieval, (pp. 167-180). Springer. https://doi.org/10.1007/978-3-031-56063-7_11 Go to original source...
  18. Vavruąka, J., ©vec, J., & Ircing, P. (2013). Phonetic spoken term detection in large audio archive using the WFST framework. In International Conference on Text, Speech and Dialogue, (pp. 402-409). Springer. https://doi.org/10.1007/978-3-642-40585-3_51 Go to original source...
  19. Wang, J., Jatowt, A., & Yoshikawa, M. (2022). Archivalqa: A large-scale benchmark dataset for open-domain question answering over historical news collections. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, (pp. 3025-3035). ACM. https://doi.org/10.1145/3477495.3531734 Go to original source...
  20. Yao, X., Tosch, E., Chen, G., Nouri, E., Artstein, R., Leuski, A., Sagae, K. & Traum, D. (2012). Creating conversational characters using question generation tools. Dialogue & Discourse, 3(2), 125-146. https://doi.org/10.5087/dad.2012.206 Go to original source...

This is an open access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC BY 4.0), which permits use, distribution, and reproduction in any medium, provided the original publication is properly cited. No use, distribution or reproduction is permitted which does not comply with these terms.