1887

Abstract

Summary

This paper proposes a method for rapid and interactive mapping of geological structures in 3D seismic volumes using self-supervised learning and content-based image retrieval (CBIR). The method makes it possible to label interactively, meaning that the results will appear in near-real time as the user is labelling. The method consists of a Vision Transformer (ViT) model pre-trained in a masked autoencoder setting on seismic data. We extend the standard 2D ViT to 3D by using multiple 2D slices which allows for 3D context without excessive computational requirements. The pretrained ViT is used to extract features prior to labelling. A simple metric is used to measure how close any point in the cube is to the means of positive and negative labelled points. The method is demonstrated on a complex injectite system, showing efficient mapping of the structure. Future research will focus on enhancing the pre-training and similarity search processes to generate more discriminative features and increase the precision of the labelling process.

Loading

Article metrics loading...

/content/papers/10.3997/2214-4609.202510639
2025-06-02
2026-02-12
Loading full text...

Full text loading...

References

  1. Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., and Houlsby, N. [2021]. An Image is Worth 16×16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), 1–16. https://doi.org/10.48550/arXiv.2010.11929.
    [Google Scholar]
  2. LeCun, Y., and Bengio, Y. [1995]. Convolutional Networks for Images, Speech, and Time Series. The Handbook of Brain Theory and Neural Networks, 3361(10).
    [Google Scholar]
  3. Lee, H., Homeyer, C., Herzog, R., Rexilius, J., and Rother, C. [2022]. Spatio-Temporal Outdoor Lighting Aggregation on Image Sequences Using Transformer Networks. International Journal of Computer Vision, 131, 1–12. https://doi.org/10.1007/s11263-022-01725-2.
    [Google Scholar]
  4. Nguyen, H., Larsen, E., Evans, D., Economou, D., Citraningtyas, T.M., and Alaei, B. [2024]. Integrated Machine Learning Approaches for Accelerated Understanding of Sand Injectites-Fault Relationships in Exploration and CO2 Storage. 85th EAGE Annual Conference & Exhibition, 1–5. https://doi.org/10.3997/2214-4609.2024101559
    [Google Scholar]
  5. He, K., Chen, X., Xie, S., Li, Y., Dollar, P., and Girshick, R. [2022]. Masked Autoencoders Are Scalable Vision Learners. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 15979–15988. https://doi.org/10.1109/CVPR52688.2022.01553.
    [Google Scholar]
  6. Ordoñez, A., Wade, D., Ravaut, C., and Waldeland, A.U. [2024]. Towards a Foundation Model for Seismic Interpretation. 85th EAGE Annual Conference & Exhibition, 1–5. https://doi.org/10.3997/2214-4609.2024101119.
    [Google Scholar]
  7. Ronneberger, O., Fischer, P., and Brox, T. [2015]. U- Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), 234–241. https://doi.org/10.1007/978-3-319-24574-4_28
    [Google Scholar]
  8. Rudjord, J. E., & Huuse, M. [2024]. Sinkites: km-scale sand bodies sinking into ooze. Preprint available at Research Square. https://doi.org/10.21203/rs.3.rs-5015562/v1
    [Google Scholar]
  9. Usearch, [2023], Version 2.16.6 retrieved from https://github.com/unum-cloud/usearch.
    [Google Scholar]
/content/papers/10.3997/2214-4609.202510639
Loading
/content/papers/10.3997/2214-4609.202510639
Loading

Data & Media loading...

This is a required field
Please enter a valid email address
Approval was a Success
Invalid data
An Error Occurred
Approval was partially successful, following selected items could not be processed due to error