1887

Abstract

Summary

Handling different seismic processing tasks with a single pretrained neural network model is the main advantage of StorSeismic, a Transformer-based model with a pretraining and fine-tuning framework, which has been developed recently. The pretraining stage included a mixed synthetic and label-less field data, which allowed for the storage of the features of both data, and for the direct inference on field data after a task-driven fine-tuning stage on the labeled synthetic data. Though the vanilla architecture performed well on various processing tasks, recent developments of the Transformer components, particularly the positional encoding and the attention mechanism, opened opportunities for a more effective StorSeismic model and framework. Thus, we experimented with a relative positional encoding approach and a low-rank form of the attention matrix to replace the vanilla sinusoidal positional encoding and dot-product self-attention, respectively. We show that these alternatives offer fewer pretraining iterations and competitive results on the fine-tuning tasks of denoising, demultiple, and first arrival picking on the Marmousi example compared to the vanilla model.

Loading

Article metrics loading...

/content/papers/10.3997/2214-4609.202310331
2023-06-05
2026-03-07
Loading full text...

Full text loading...

References

  1. Harsuko, R. and Alkhalifah, T.A. [2022] StorSeismic: A new paradigm in deep learning for seismic processing. IEEE Transactions on Geoscience and Remote Sensing, 60, 1–15.
    [Google Scholar]
  2. Luo, S., Li, S., Zheng, S., Liu, T.Y., Wang, L. and He, D. [2022] Your Transformer May Not be as Powerful as You Expect. arXiv preprint arXiv:2205.13401.
    [Google Scholar]
  3. Press, O., Smith, N.A. and Lewis, M. [2021] Train short, test long: Attention with linear biases enables input length extrapolation. arXiv preprint arXiv:2108.12409.
    [Google Scholar]
  4. Tay, Y, Bahri, D., Metzler, D., Juan, D.C., Zhao, Z. and Zheng, C. [2021] Synthesizer: Rethinking self-attention for transformer models. In: International conference on machine learning. PMLR, 10183–10192.
    [Google Scholar]
  5. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L. and Polosukhin, I. [2017] Attention is all you need. Advances in neural information processing systems, 30.
    [Google Scholar]
/content/papers/10.3997/2214-4609.202310331
Loading
/content/papers/10.3997/2214-4609.202310331
Loading

Data & Media loading...

This is a required field
Please enter a valid email address
Approval was a Success
Invalid data
An Error Occurred
Approval was partially successful, following selected items could not be processed due to error