Robust representations of oil wells' intervals via sparse attention
mechanism
- URL: http://arxiv.org/abs/2212.14246v3
- Date: Mon, 6 Nov 2023 10:04:49 GMT
- Title: Robust representations of oil wells' intervals via sparse attention
mechanism
- Authors: Alina Ermilova, Nikita Baramiia, Valerii Kornilov, Sergey Petrakov,
Alexey Zaytsev
- Abstract summary: We introduce the class of efficient Transformers named Regularized Transformers (Reguformers)
The focus in our experiments is on oil&gas data, namely, well logs.
To evaluate our models for such problems, we work with an industry-scale open dataset consisting of well logs of more than 20 wells.
- Score: 2.604557228169423
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Transformer-based neural network architectures achieve state-of-the-art
results in different domains, from natural language processing (NLP) to
computer vision (CV). The key idea of Transformers, the attention mechanism,
has already led to significant breakthroughs in many areas. The attention has
found their implementation for time series data as well. However, due to the
quadratic complexity of the attention calculation regarding input sequence
length, the application of Transformers is limited by high resource demands.
Moreover, their modifications for industrial time series need to be robust to
missing or noised values, which complicates the expansion of the horizon of
their application. To cope with these issues, we introduce the class of
efficient Transformers named Regularized Transformers (Reguformers). We
implement the regularization technique inspired by the dropout ideas to improve
robustness and reduce computational expenses. The focus in our experiments is
on oil&gas data, namely, well logs, a prominent example of multivariate time
series. The goal is to solve the problems of similarity and representation
learning for them. To evaluate our models for such problems, we work with an
industry-scale open dataset consisting of well logs of more than 20 wells. The
experiments show that all variations of Reguformers outperform the previously
developed RNNs, classical Transformer model, and robust modifications of it
like Informer and Performer in terms of well-intervals' classification and the
quality of the obtained well-intervals' representations. Moreover, the
sustainability to missing and incorrect data in our models exceeds that of
others by a significant margin. The best result that the Reguformer achieves on
well-interval similarity task is the mean PR~AUC score equal to 0.983, which is
comparable to the classical Transformer and outperforms the previous models.
Related papers
- PRformer: Pyramidal Recurrent Transformer for Multivariate Time Series Forecasting [82.03373838627606]
Self-attention mechanism in Transformer architecture requires positional embeddings to encode temporal order in time series prediction.
We argue that this reliance on positional embeddings restricts the Transformer's ability to effectively represent temporal sequences.
We present a model integrating PRE with a standard Transformer encoder, demonstrating state-of-the-art performance on various real-world datasets.
arXiv Detail & Related papers (2024-08-20T01:56:07Z) - Rough Transformers: Lightweight Continuous-Time Sequence Modelling with Path Signatures [46.58170057001437]
We introduce the Rough Transformer, a variation of the Transformer model that operates on continuous-time representations of input sequences.
We find that, on a variety of time-series-related tasks, Rough Transformers consistently outperform their vanilla attention counterparts.
arXiv Detail & Related papers (2024-05-31T14:00:44Z) - Rough Transformers for Continuous and Efficient Time-Series Modelling [46.58170057001437]
Time-series data in real-world medical settings typically exhibit long-range dependencies and are observed at non-uniform intervals.
We introduce the Rough Transformer, a variation of the Transformer model which operates on continuous-time representations of input sequences.
We find that Rough Transformers consistently outperform their vanilla attention counterparts while obtaining the benefits of Neural ODE-based models.
arXiv Detail & Related papers (2024-03-15T13:29:45Z) - iTransformer: Inverted Transformers Are Effective for Time Series Forecasting [62.40166958002558]
We propose iTransformer, which simply applies the attention and feed-forward network on the inverted dimensions.
The iTransformer model achieves state-of-the-art on challenging real-world datasets.
arXiv Detail & Related papers (2023-10-10T13:44:09Z) - Are Transformers Effective for Time Series Forecasting? [13.268196448051308]
Recently, there has been a surge of Transformer-based solutions for the time series forecasting (TSF) task.
This study investigates whether Transformer-based techniques are the right solutions for long-term time series forecasting.
We find that the relatively higher long-term forecasting accuracy of Transformer-based solutions has little to do with the temporal relation extraction capabilities of the Transformer architecture.
arXiv Detail & Related papers (2022-05-26T17:17:08Z) - Causal Transformer for Estimating Counterfactual Outcomes [18.640006398066188]
Estimating counterfactual outcomes over time from observational data is relevant for many applications.
We develop a novel Causal Transformer for estimating counterfactual outcomes over time.
Our model is specifically designed to capture complex, long-range dependencies among time-varying confounders.
arXiv Detail & Related papers (2022-04-14T22:40:09Z) - Transformers predicting the future. Applying attention in next-frame and
time series forecasting [0.0]
Recurrent Neural Networks were, until recently, one of the best ways to capture the timely dependencies in sequences.
With the introduction of the Transformer, it has been proven that an architecture with only attention-mechanisms without any RNN can improve on the results in various sequence processing tasks.
arXiv Detail & Related papers (2021-08-18T16:17:29Z) - Transformers Solve the Limited Receptive Field for Monocular Depth
Prediction [82.90445525977904]
We propose TransDepth, an architecture which benefits from both convolutional neural networks and transformers.
This is the first paper which applies transformers into pixel-wise prediction problems involving continuous labels.
arXiv Detail & Related papers (2021-03-22T18:00:13Z) - Long Range Arena: A Benchmark for Efficient Transformers [115.1654897514089]
Long-rangearena benchmark is a suite of tasks consisting of sequences ranging from $1K$ to $16K$ tokens.
We systematically evaluate ten well-established long-range Transformer models on our newly proposed benchmark suite.
arXiv Detail & Related papers (2020-11-08T15:53:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.