1887

Abstract

Summary

Legacy handwritten reports, often preserved in archive, hold valuable insight for geological knowledge. Despite advancements in Optical Character Recognition (OCR) technologies, which have improved the conversion of handwritten content into digital text, the resulting outputs remain unstructured and incomplete, thereby limiting their applicability to modern analytical tools. By fine-tuning LLMs specifically for geological domains, we can bridge the gap between raw OCR data and structured, actionable insights, enabling seamless access to decades of legacy knowledge. This paper explores innovative methodologies for fine-tuning LLMs to process and structure raw OCR outputs from handwritten geological reports. Our approach started with leveraging the IDEFICS model (Image-aware Decoder Enhanced a la Flamingo with Interleaved Cross-attentions) for OCR, particularly its capabilities in handling image-text. Next, we fine-tuned LLM like model Mistral-7B-Instruct using a carefully selected dataset containing prompts and matching responses. Manual correction was applied to refine the dataset where necessary. This comprehensive approach generates high-quality structured outputs, unlocking the vast potential of geological archives for modern application and analysis. As a result, the research showcases that focused fine-tuning efforts within the geology realm yield considerable advancements in the digitalization and structuration of i handwritten documents.

Loading

Article metrics loading...

/content/papers/10.3997/2214-4609.202539043
2025-03-24
2026-02-06
Loading full text...

Full text loading...

References

  1. AlsubaieR. et al., 2024. “Deep learning and Transformer-Based Optical Character Recognition Modeling for Geological Legacy Handwritten Reports”. Digitalization in Geosciences Symposium GEO4.0, Khobar, Saudi Arabia.
    [Google Scholar]
  2. Jiang, Albert Q., et al., 2023. “Mistral 7B.” arXiv preprint arXiv:2310.06825.
    [Google Scholar]
  3. XuL. et al., 2024."Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment." arXiv preprint arXiv:2312.12148.
    [Google Scholar]
  4. Hu, Edward J., et al., 2021. “Lora: Low-rank adaptation of large language models.” arXiv preprint arXiv:2106.09685.
    [Google Scholar]
  5. SaadanyH. and ConstantinO., 2021. “BLEU, METEOR, BERTScore: evaluation of metrics performance in assessing critical translation errors in sentiment-oriented text.” arXiv preprint arXiv:2109.14250.
    [Google Scholar]
/content/papers/10.3997/2214-4609.202539043
Loading
/content/papers/10.3997/2214-4609.202539043
Loading

Data & Media loading...

This is a required field
Please enter a valid email address
Approval was a Success
Invalid data
An Error Occurred
Approval was partially successful, following selected items could not be processed due to error