Full text loading...
Fossils are crucial in geology, providing evidence for stratigraphic correlation and paleoenvironmental reconstruction. However, traditional fossil records often suffer from disconnection between specimen details and collection points, limiting their scientific utility. Large amounts of valuable fossil data remain locked in legacy reports, handwritten notes, and scanned charts. To address this, a novel AI-driven framework integrates Large Vision Models (LVMs), Large Language Models (LLMs), and GIS to extract and process fossil data from over 7,000 legacy documents. The methodology employs YOLO-based object detection to identify key terms, such as formation names and faunal data, followed by Handwritten Text Recognition (HTR) and LLM-based refinement. This approach achieved significant results, resolving 2,410 fossil localities across Saudi Arabia. Quality control measures, including OCR error correction and validation against benchmarks, ensured accuracy, with a Word Error Rate of just 5%. Further, LVMs like Qwen2-VL and TrOCR enabled extraction from both handwritten and printed records, while duplication analysis reduced redundancy. A major breakthrough was achieved through AI-assisted geographic reasoning, which integrated textual descriptions with quadrangle maps, expanding locality data by 1,474 points. Ultimately, 98.5% of localities were geographically verified, demonstrating the framework’s potential to transform inaccessible fossil archives into structured, standardized datasets for future scientific research.