CN111815680A - An automatic horizon tracking method for deep convolutional neural networks based on identity shortcut mapping - Google Patents
An automatic horizon tracking method for deep convolutional neural networks based on identity shortcut mapping Download PDFInfo
- Publication number
- CN111815680A CN111815680A CN202010841336.6A CN202010841336A CN111815680A CN 111815680 A CN111815680 A CN 111815680A CN 202010841336 A CN202010841336 A CN 202010841336A CN 111815680 A CN111815680 A CN 111815680A
- Authority
- CN
- China
- Prior art keywords
- data
- horizon
- seismic
- convolutional neural
- neural network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
- G06T7/248—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30181—Earth observation
- G06T2207/30184—Infrastructure
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明涉及一种基于恒等快捷映射的深层卷积神经网络自动层位追踪方法。The invention relates to a deep convolutional neural network automatic layer tracking method based on identity shortcut mapping.
背景技术Background technique
地震层位解释与层序地层划分有着十分重要的相关性,准确的层位信息是地震解释基础工作,而层位的识别和跟踪是地震层位解释的重要组成部分,研究者们一直着力于解决同相轴的自动跟踪。地震数据是由炸药或地震车对地面通过爆炸或者重击产生地震波,其在遇到地表下的地层时会发生发射形成反射波并被地表的检波器接收信号,最终得到了地震振幅等反射数据。通常地震解释人员认为地震数据的剖面中的波峰或者波谷是同相轴经过的位置。常规人工层位标记是通过地震剖面的波峰和波谷确定同相轴的位置并进行连接,最终将各个剖面标记的层位连接形成同相轴的面。由于人工标记本身需要的成本及时间较大,研究者们又发展了自动层位追踪的手段。基于上述的原理,研究者们通过计算波形相关系数或者相干性等来寻找一定范围内的相似点,并将其作为同相轴经过的点,最终划定同相轴。但是在断层或分叉等情况出现时,容易出现计算误差导致最终标记出现问题。三维地震数据能够大概反映探测区域的地表下复杂地层情况,由于地下结构大多是层状分布的,在地下的不同深度有着不同年代产生的不同地层,且由于地层本身性质不同所带来的密度的不同导致地震波在此处传播的速度不同,最终使人们能够得到具有差异的地震反射数据。地层密度也叫岩石密度,其和声波在岩石上传播速度共同被称为岩石物理特性,简称岩性参数。不同地层的岩性参数差别很大,波阻抗为上述两者的乘积,这意味着波阻抗在不同地层上也会出现显著的变化。由于岩性参数的差距,声波在传播到两个地层交界处时会发生反射,最终这样的反射被地表的传感器接收生成了各种各样的波形,而变化强烈且一致并互相邻接的波形被称为同相轴,也就是层位。所以研究者们认为层位能够表达地下的层序划分,并能表示地下的岩性分布结构状况。层位一般是由多个点组成,每个点有对应的线号以及时间(深度),一般情况下勘探区域得到的三维地震体会存在多个层位,但是需要对所有层位点找到其对应所属的层位,这就体现了层位追踪的重要性以及意义。There is a very important correlation between seismic horizon interpretation and sequence stratigraphic division. Accurate horizon information is the basic work of seismic interpretation, and horizon identification and tracking are an important part of seismic horizon interpretation. Resolve automatic tracking of the event axis. Seismic data are seismic waves generated by explosives or seismic vehicles on the ground through explosions or heavy blows. When they encounter the subsurface, they will emit reflected waves and receive signals by the geophones on the surface. Finally, reflection data such as seismic amplitudes are obtained. . Seismic interpreters generally consider peaks or troughs in a profile of seismic data to be locations where the event axis passes. The conventional artificial horizon marking is to determine the position of the event axis through the peaks and troughs of the seismic section and connect them, and finally connect the horizons marked on each section to form the event axis surface. Due to the large cost and time required for manual marking, researchers have developed methods for automatic horizon tracking. Based on the above principles, researchers find similar points within a certain range by calculating the waveform correlation coefficient or coherence, and use them as the points that the event axis passes through, and finally delineate the event axis. However, when faults or bifurcations occur, calculation errors are prone to cause problems with the final marking. The 3D seismic data can roughly reflect the complex strata under the surface of the detection area. Since most of the underground structures are distributed in layers, there are different strata produced in different ages at different depths underground, and the density of the stratum itself is different due to the different nature. The difference causes the seismic waves to travel at different speeds here, which ultimately enables people to obtain different seismic reflection data. The density of the formation is also called the density of the rock, which together with the propagation speed of the sound wave on the rock is called the petrophysical property, or lithological parameter for short. The lithological parameters of different formations are very different, and the wave impedance is the product of the above two, which means that the wave impedance will also change significantly in different formations. Due to the difference in lithological parameters, the sound wave will reflect when it propagates to the junction of the two formations, and finally such reflection is received by the sensors on the surface to generate various waveforms, and the waveforms with strong changes and consistent and adjacent to each other are It is called the event axis, that is, the horizon. Therefore, researchers believe that horizons can express the subsurface sequence division and the subsurface lithology distribution structure. A horizon is generally composed of multiple points, and each point has a corresponding line number and time (depth). Generally, there are multiple horizons in the 3D seismic body obtained from the exploration area, but it is necessary to find the corresponding horizons for all horizons. The level to which it belongs, which reflects the importance and significance of level tracking.
目前主要有以下几种层位追踪方法:At present, there are mainly the following layer tracking methods:
(1)基于相关的自动层位追踪(1) Automatic horizon tracking based on correlation
它是一种传统的自动层位追踪技术,主要利用过去提出的相干算法进行施展,该种类方案通过两道或多道之间的相关性对同相轴进行追踪并标记得到层位,其鲁棒性及对噪声的抵抗性能较好,但对算力的需求相当大,而且当邻近的两个层位的波形非常近似的情况下,极易产生串层的情况,这可能导致得到的结果不够准确。It is a traditional automatic horizon tracking technology, mainly using the coherent algorithm proposed in the past. The stability and resistance to noise are good, but the demand for computing power is quite large, and when the waveforms of the two adjacent layers are very similar, it is easy to produce a situation of cascading layers, which may lead to insufficient results. precise.
(2)基于图像的自动跟踪方法(2) Image-based automatic tracking method
Hale和Naeini等人在层位追踪以及断层识别领域使用了结构张量,通过一些带方向的滤波器,得到了具体特征的方向,最后能够获得层位延伸及育成的主要方位。这种方式计算的功效较高,但是若遇到结构复杂的地层,追中的精度以及稳定程度无法得到确保。Hale and Naeini et al. used structure tensor in the field of horizon tracking and fault identification. Through some directional filters, the orientation of specific features was obtained, and finally the main orientation of horizon extension and breeding was obtained. The calculation efficiency of this method is high, but if the stratum with complex structure is encountered, the accuracy and stability of the tracking cannot be guaranteed.
(3)基于神经网络的自动跟踪方法(3) Automatic tracking method based on neural network
Huang,Lu等人发表了使用自组织映射网络的无监督自动层位跟踪方法,但是该方案的成败取决于自组织映射网络的初始化,这和结果的好坏直接相关。Alberts、Huang等人通过人工神经网络来对层位进行追踪,这种方法需要较多的训练数据,且对数据代表性有要求,但得到的效果不错。由于人工智能本身及深度学习的快速发展背后,数据仓库和图形处理器等算力相关的技术也得到快速发展,卷积神经网络的提出使其被运用到了图像识别,NLP,回归等常见的应用,在这之中Alex等人将CNN运用到图像分类中取得了良好的效果,进一步引爆了深度学习的火热。Huang, Lu et al. published an unsupervised automatic horizon tracking method using a self-organizing map network, but the success or failure of this scheme depends on the initialization of the self-organizing map network, which is directly related to the quality of the results. Alberts, Huang et al. used artificial neural networks to track horizons. This method requires more training data and requires data representativeness, but the results are good. Due to the rapid development of artificial intelligence itself and deep learning, computing power-related technologies such as data warehouses and graphics processors have also developed rapidly. The convolutional neural network has been proposed to be applied to common applications such as image recognition, NLP, and regression. , in which Alex et al. applied CNN to image classification and achieved good results, further detonating the popularity of deep learning.
(4)基于相干的同相轴自动追踪方法(4) Coherence-based automatic tracking method of event axis
目前地震勘探领域已经发展到了第三代相干技术,通过某种方式计算地震道与相邻道之间的某种内在的物理联系,或者说相似度,被称为相干性。相干程度的计算能够表现地层内的横向变化,若该值较大,则说明此处的横向变化较小,邻近的道间相似度较大,相干性较强。而出现该值较小的地方,说明可能存在断层或异常体,岩层横向变化较大,该处相干性较弱。基于相干的自动层位追踪通过使用相干技术对地震反射进行运算并最终得到相干体,该相干体能够解释岩性参数的变化。通过对相干体特定区域进行特征提取,最后再运用插值技术得到完整的层位,这就是目前商业上较为受欢迎的基于相干的自动层位追踪技术。At present, the field of seismic exploration has developed to the third-generation coherent technology, which calculates a certain intrinsic physical connection, or similarity, between the seismic trace and adjacent traces in a certain way, which is called coherence. The calculation of the degree of coherence can represent the lateral change in the formation. If the value is large, it means that the lateral change here is small, the similarity between adjacent traces is large, and the coherence is strong. Where the value is small, it indicates that there may be faults or abnormal bodies, the lateral variation of the rock formation is large, and the coherence is weak in this place. Coherence-based automatic horizon tracking operates on seismic reflections using coherent techniques and ultimately results in a coherent volume that can account for changes in lithological parameters. By extracting features from a specific area of the coherent volume, and finally using interpolation technology to obtain a complete horizon, this is the currently more popular commercial coherence-based automatic horizon tracking technology.
发明内容SUMMARY OF THE INVENTION
本发明的目的在于克服现有技术的不足,提供一种以深层卷积神经网络为骨干搭建模型,以恒定快捷连接为核心,改变了传统网络的梯度流,更有效地提取了地震反射中的特征,实现了地震相位特征的深度级联聚合,能够更精确地追踪目标层位的基于恒等快捷映射的深层卷积神经网络自动层位追踪方法。The purpose of the present invention is to overcome the deficiencies of the prior art, and to provide a model built with a deep convolutional neural network as the backbone, with constant and fast connections as the core, changing the gradient flow of the traditional network, and more effectively extracting the The deep convolutional neural network automatic horizon tracking method based on identity shortcut mapping can more accurately track the target horizon.
本发明的目的是通过以下技术方案来实现的:基于恒等快捷映射的深层卷积神经网络自动层位追踪方法,包括以下步骤:The object of the present invention is to be achieved through the following technical solutions: a deep convolutional neural network automatic horizon tracking method based on identity shortcut mapping, comprising the following steps:
S1、对地震剖面图像进行预处理;S1. Preprocess the seismic profile image;
S2、使用恒等快捷连接卷积神经网络对地震反射特征进行提取、压缩和精炼;S2. Extract, compress and refine seismic reflection features by using the identity shortcut connection convolutional neural network;
S3、对步骤S2提取的信息进一步采集局部特征以及全局特征;S3, further collect local features and global features for the information extracted in step S2;
S4、将采集的特征进行分类计算。S4, classify and calculate the collected features.
进一步地,所述步骤S1具体包括以下子步骤:Further, the step S1 specifically includes the following sub-steps:
S11、将整个地震剖面划分成多个小图像块,每个小图像块大小为32×32;S11. Divide the entire seismic section into a plurality of small image blocks, and each small image block is 32×32 in size;
S12、将标签分配给每个小图像块,具体实现方法为:判断每个小图像块的中心是否位于层位上,若是则将该图像块标记为1,否则标记为0;S12, assigning a label to each small image block, the specific implementation method is: judging whether the center of each small image block is located on the horizon, if so, marking the image block as 1, otherwise, marking it as 0;
S13、将图像块数据进行二分类训练,以判别该图像块的中心位置是否处于层位上;S13, perform two-class training on the image block data to determine whether the center position of the image block is on the horizon;
S14、对非层位数据进行处理:设定遗失概率来降低得到的非层位图像块的数量,设置遗失概率为0.5,即随机选取50%的非层位数据进行舍弃;同时在层位上设定偏差值,若一个小图像快的中心点与某个确定的层位的距离在偏差值范围内,即使该数据被人工标记标定为非层位,也认定该数据为层位数据,标记为1;S14. Process the non-horizontal data: set the loss probability to reduce the number of obtained non-horizontal image blocks, and set the loss probability to 0.5, that is, randomly select 50% of the non-horizontal data to discard; Set the deviation value. If the distance between the center point of a small image and a certain horizon is within the deviation value range, even if the data is manually marked as non-horizontal, the data will be regarded as horizon data. is 1;
S15、从层位数据集合中提取20%作为验证数据,用于对模型超参数进行调整;其余数据作为训练数据;S15. Extract 20% from the horizon data set as verification data, which is used to adjust the model hyperparameters; the rest of the data is used as training data;
S16、对数据进行归一化处理:使用Min-Max规范化方法,使数据尺度被规范化到0-1 范围内;Min-Max规范化方法通过以下公式实现:S16. Normalize the data: use the Min-Max normalization method to normalize the data scale to the range of 0-1; the Min-Max normalization method is realized by the following formula:
其中Xnor为规范化后的数据,X为原始的地震数据,Xmax和Xmin分别为地震振幅中的极大值和极小值。where X nor is the normalized data, X is the original seismic data, and X max and X min are the maximum and minimum values of the seismic amplitude, respectively.
本发明的有益效果是:The beneficial effects of the present invention are:
1、本发明以数据挖掘特征分类为出发点,提出了基于恒定快捷连接的深层卷积神经网络自动层位追踪方法。以深层卷积神经网络为骨干搭建模型,以恒定快捷连接为核心,改变了传统网络的梯度流,更有效地提取了地震反射中的特征,实现了地震相位特征的深度级联聚合。本发明对于复杂多断裂带大倾角层位以及弱地震振幅反射层位均有良好的追踪效果,提高了层位分辨的准确率。1. The present invention takes the data mining feature classification as the starting point, and proposes a deep convolutional neural network automatic horizon tracking method based on constant shortcut connections. The deep convolutional neural network is used as the backbone to build a model, with constant and fast connections as the core, which changes the gradient flow of the traditional network, extracts the features of seismic reflections more effectively, and realizes the deep cascade aggregation of seismic phase features. The invention has a good tracking effect on the high dip angle horizon and the weak seismic amplitude reflection horizon in the complex multi-fault zone, and improves the accuracy of horizon resolution.
2、本发明针对运用数据驱动的方法解决层位追踪问题易发生数据不均衡的问题,提出了数据预处理与后处理的新思路,在出现连续断层以及大倾角的情况下,能够更精确地追踪目标层位。2. The present invention proposes a new idea of data pre-processing and post-processing in order to solve the problem of data imbalance by using the data-driven method to solve the problem of horizon tracking. Track the target horizon.
3、相比普通卷积神经网络模型,本发明的模型在数据分布相较原始数据分布差异更大的情况下,拥有更强的泛化性能,具有更好的特征表达能力,更加适用于追踪复杂目标层位。3. Compared with the ordinary convolutional neural network model, the model of the present invention has stronger generalization performance and better feature expression ability when the data distribution is more different than the original data distribution, and is more suitable for tracking Complex target horizons.
附图说明Description of drawings
图1为神经网络连接方式对比;Figure 1 is a comparison of neural network connection methods;
图2为卷积过程示意图;Figure 2 is a schematic diagram of the convolution process;
图3为残差函数F(x)示意图;Figure 3 is a schematic diagram of the residual function F(x);
图4为本发明提出的模型框架结构;Fig. 4 is the model frame structure proposed by the present invention;
图5为本发明的基于恒等快捷映射的深层卷积神经网络自动层位追踪方法的流程图;5 is a flowchart of the deep convolutional neural network automatic horizon tracking method based on identity shortcut mapping of the present invention;
图6为西南某工区三维地震数据体图像;Figure 6 is a volume image of 3D seismic data in a work area in southwest China;
图7为小图像块划分示意图;7 is a schematic diagram of the division of small image blocks;
图8为层位数据和非层位数据示意图;Fig. 8 is a schematic diagram of horizon data and non-horizon data;
图9为西南某工区inline1760二维剖面图;Figure 9 is a two-dimensional cross-sectional view of inline1760 in a work area in the southwest;
图10为西南某工区inline1762二维剖面图;Figure 10 is a two-dimensional cross-sectional view of inline1762 in a work area in the southwest;
图11为西南某工区inline1762层位概率分布图;Figure 11 is the probability distribution map of the inline1762 horizon in a work area in the southwest;
图12为三种方法在西南某工区inline1762的比较图;Figure 12 is a comparison diagram of the three methods in a work area in the southwest of inline1762;
图13为图12中方框部分的放大图;Fig. 13 is an enlarged view of the block portion in Fig. 12;
图14为西南某工区inline1775二维剖面图;Figure 14 is a two-dimensional cross-sectional view of inline1775 in a work area in the southwest;
图15为西南某工区inline1775层位概率分布图;Figure 15 is the probability distribution map of the inline1775 horizon in a work area in the southwest;
图16为三种方法在西南某工区inline1775的比较图;Figure 16 is a comparison diagram of the three methods in a work area in the southwest of inline1775;
图17为图16中方框部分的放大图。FIG. 17 is an enlarged view of the block portion in FIG. 16 .
具体实施方式Detailed ways
与本发明有关的技术原理:The technical principle relevant to the present invention:
卷积神经网络Convolutional Neural Network
卷积神经网络是一种特殊的深度前向神经网络,其由带权重和偏置的神经元组成。通过采取局部连接的方式,有效避免了全连接带来的冗余问题,且有效降低了网络的参数量,进而降低了模型对数据本身的依赖性。比如每个神经元只和10×10个像素值连接。神经网络连接方式如图1所示,(a)为原始连接(全连接),(b)为新连接方式(局部连接):A convolutional neural network is a special kind of deep feed-forward neural network that consists of neurons with weights and biases. By adopting the method of partial connection, the redundancy problem caused by the full connection is effectively avoided, and the parameter quantity of the network is effectively reduced, thereby reducing the dependence of the model on the data itself. For example, each neuron is only connected to 10×10 pixel values. The neural network connection mode is shown in Figure 1, (a) is the original connection (full connection), (b) is the new connection mode (partial connection):
卷积神经网络一般由卷积层,Relu非线性激活层,池化层和全连接层组成,在这其中,根据任务本身的情况可选择是否使用池化层。卷积神经网络的默认输入的数据格式为深度×高度×宽度,若输入为图像,深度代表图像的通道数,后两者表示图像的分辨率。一般而言,使用彩色图片的通道默认为RGB三通道,使用灰度图片默认通道为1。卷积神经网络各层与前一层都采用局部连接的方式。A convolutional neural network generally consists of a convolutional layer, a Relu nonlinear activation layer, a pooling layer and a fully connected layer. Among them, the pooling layer can be selected according to the situation of the task itself. The default input data format of the convolutional neural network is depth × height × width. If the input is an image, the depth represents the number of channels of the image, and the latter two represent the resolution of the image. In general, the channels used for color images are RGB three-channel by default, and the default channel for gray-scale images is 1. Each layer of the convolutional neural network is locally connected to the previous layer.
卷积层通过卷积运算提取和收集特征,所以这一类层承担了整个神经网络的大部分计算量。在普通的卷积神经网络中,卷积层通常位于输入层或池化层之后。在进行前向传播时,使用固定尺寸的卷积核以设置的步长在输入图像上进行游走并同时卷积计算,将卷积核的权重与局部连接的对应区域相乘求和,计算输出。为更进一步的说明卷积层的工作过程,以一个实例进行说明。如图2所示,网络输入的原始图像大小5×5,使用一个3×3大小的卷积核提取图像特征,该卷积核在原始图像上的对应区域分别作点积运算,最终获得一个大小为3×3的输出。图中左边两侧和上面两行为输入的原始图像,右下角部分表示一个3×3的卷积核,右边的图像表示经过卷积核滤波后得到的网络输出。该输出值的大小由网络的步长以及填充值大小判定。可用公式计算得出,其中W为输入的原始图像的尺寸,F为卷积核尺寸,S为步长,表示卷积核每次在图像上滑动经过的像素个数。P为填充数,表示当经过卷积操作后的特征图大小与卷积核大小不满足点积运算计算要求时,将特征图外部用0进行填充补齐大小的一种方式。在卷积神经网络中,卷积的数学表达式和常规的二维卷积有些许不同,其表达如下:Convolutional layers extract and collect features through convolution operations, so this type of layer undertakes most of the computation of the entire neural network. In a normal convolutional neural network, the convolutional layer is usually placed after the input layer or the pooling layer. During forward propagation, use a fixed-size convolution kernel to walk on the input image with a set step size and perform convolution calculations at the same time. output. In order to further illustrate the working process of the convolutional layer, an example is described. As shown in Figure 2, the size of the original image input by the network is 5 × 5, and a 3 × 3 convolution kernel is used to extract image features. Output of size 3x3. The left side and the top two lines in the figure are the original images of the input, the lower right part represents a 3×3 convolution kernel, and the right image represents the network output after filtering by the convolution kernel. The size of the output value is determined by the step size of the network and the size of the padding value. Available formulas It is calculated, where W is the size of the input original image, F is the size of the convolution kernel, and S is the step size, which represents the number of pixels that the convolution kernel slides on the image each time. P is the padding number, which means that when the size of the feature map after the convolution operation and the size of the convolution kernel do not meet the calculation requirements of the dot product operation, the outside of the feature map is filled with 0 to fill the size. In the convolutional neural network, the mathematical expression of convolution is slightly different from the conventional two-dimensional convolution, which is expressed as follows:
s(i,j)=X*W(i,j)=∑m∑nx(i+m,j+n)w(m,n) (2)s(i, j)=X*W(i, j)=∑ m ∑ n x(i+m, j+n)w(m, n) (2)
当输入多通道卷积核时,函数表达式如下:When the multi-channel convolution kernel is input, the function expression is as follows:
其中inchannel为输入的通道数,Xk代表第k个输入矩阵,Wk代表卷积核的第k个子卷积核矩阵。s(i,j)即卷积核W对应的输出矩阵对应元素值。where inchannel is the number of input channels, X k represents the k-th input matrix, and W k represents the k-th sub-convolution kernel matrix of the convolution kernel. s(i, j) is the corresponding element value of the output matrix corresponding to the convolution kernel W.
若在卷积层操作后得到的特征尺寸仍然过大,不利于后续网络构建,则可以将其输入池化层进行下采样压缩。这样的操作可以避免分辨率过大带来的参数量暴涨,同时可以一定程度上降低过拟合程度。通常的做法是选定合适的窗口大小,设置合适的步长,对窗口内数据进行取极大值,平均值或随机抽取,这三种方案分别代表最大池化,平均池化和随机池化。一般来说,最大池化是最常见的池化方法,这种方法通常能在稀释数据分辨率的情况下较好地保留原始特征,且同时降低了维度,方便计算,并增强了泛化能力。If the feature size obtained after the convolutional layer operation is still too large, which is not conducive to subsequent network construction, it can be input to the pooling layer for downsampling compression. Such an operation can avoid the skyrocketing amount of parameters caused by excessive resolution, and at the same time can reduce the degree of overfitting to a certain extent. The usual practice is to select an appropriate window size, set an appropriate step size, and take the maximum value, average value or random sampling of the data in the window. These three schemes represent maximum pooling, average pooling and random pooling respectively. . Generally speaking, max pooling is the most common pooling method. This method can usually retain the original features better while diluting the data resolution, and at the same time reduces the dimension, facilitates calculation, and enhances the generalization ability. .
全连接层与随机丢失:在经过卷积层和池化层后,最终需要经过全连接层提取最终特征,并根据所需求的任务对最终维度进行整合。该层将特征压缩至向量,所以该层也被称作为特征向量层。由于不同于卷积层的局部连接,采用了全连接操作,这导致参数量的增加,可以采用dropout手段,通过设置一个丢失概率,对神经元进行随机选取丢失的操作,该方法可以降低过拟合的现象同时降低网络参数量,减少计算量提升训练效率。Dropout也可以使用在带有权重的其它层,例如卷积层等。Fully connected layer and random loss: After going through the convolutional layer and the pooling layer, the final feature needs to be extracted through the fully connected layer, and the final dimension is integrated according to the required task. This layer compresses features into vectors, so this layer is also called a feature vector layer. Since it is different from the local connection of the convolutional layer, the full connection operation is used, which leads to an increase in the amount of parameters. The dropout method can be used. By setting a loss probability, the neurons are randomly selected and lost. This method can reduce overfitting. At the same time, it reduces the amount of network parameters, reduces the amount of calculation and improves the training efficiency. Dropout can also be used in other layers with weights, such as convolutional layers.
残差网络:深度残差网络由He等人提出,解决了因网络深度加深而导致网络性能逐渐退化的问题。根据泛逼近定理(universal approximation theorem),在给定足够容量的前提下,任何函数都可以通过单层前向神经网络进行表达。然而,该前提可能导致这样的层维度超乎想象的庞大,增加了过拟合现象出现的概率。因此,学界普遍认为多层网络架构优于单层大容量网络架构。AlexNet在图像识别任务获得了良好的结果,从那之后最先进的网络架构逐步加深,相比之下,AlexNet仅有5个卷积层。在其发展后续的VGG网络和GoogleNet(代号Inception_v1);分别有19层和22层。但是,在梯度消失的局限下,网络层数并不能通过堆叠简易增加,同时由于上述原因,这样的网络极难训练。因为梯度迷倒导致反向传播时,越传到前面层的梯度可能在连乘下接近于0。这导致网络层数在加深到某个极限后,网络性能达到瓶颈甚至开始下降。ResNet的核心思想是引入一个所谓的恒等快捷连接(identity shortcut connection),直接跳过一个或多个层,如图3所示。Residual network: The deep residual network was proposed by He et al. to solve the problem of gradual degradation of network performance due to the deepening of the network depth. According to the universal approximation theorem, given sufficient capacity, any function can be expressed by a single-layer feedforward neural network. However, this premise may lead to unimaginably large dimensions of such layers, increasing the probability of overfitting. Therefore, the academic community generally believes that the multi-layer network architecture is superior to the single-layer high-capacity network architecture. AlexNet has achieved good results in image recognition tasks, and since then the state-of-the-art network architecture has gradually deepened, in contrast, AlexNet has only 5 convolutional layers. In its development, the follow-up VGG network and GoogleNet (codenamed Inception_v1); there are 19 layers and 22 layers respectively. However, under the limitation of vanishing gradients, the number of network layers cannot be easily increased by stacking, and for the above reasons, such a network is extremely difficult to train. When backpropagation is caused by gradient confusion, the gradient passed to the previous layer may be close to 0 under continuous multiplication. As a result, after the number of network layers is deepened to a certain limit, the network performance reaches a bottleneck or even begins to decline. The core idea of ResNet is to introduce a so-called identity shortcut connection, which directly skips one or more layers, as shown in Figure 3.
残差学习单元通过恒等映射的引入在输入、输出之间建立了一条直接的关联通道,从而使强大的有参层集中精力学习输入、输出之间的残差。一般用F(x)来表示残差映射,那么输出即为:y=F(x)+x。当输入、输出通道数相同时,可以直接使用x进行相加。而当它们之间的通道数目不同时,就需要考虑建立一种有效的恒等映射函数从而可以使得处理后的输入x与输出y的通道数目相同即y=F(x)+W*x。残差网络的出现,使得人们构建更深层的网络成为可能,同时使用了恒定快捷连接这样的“跳层”连接,达到了增加特征多样性,加快训练的效果。The residual learning unit establishes a direct correlation channel between input and output through the introduction of identity mapping, so that the powerful parameterized layer can concentrate on learning the residual between input and output. Generally, F(x) is used to represent the residual mapping, then the output is: y=F(x)+x. When the number of input and output channels is the same, you can directly use x for addition. When the number of channels between them is different, it is necessary to consider establishing an effective identity mapping function so that the processed input x and the output y have the same number of channels, ie y=F(x)+W*x. The emergence of residual networks makes it possible for people to build deeper networks. At the same time, the use of "jump-layer" connections such as constant shortcut connections achieves the effect of increasing feature diversity and speeding up training.
本发明提出的网络由两部分构成:前半部分使用恒等快捷连接卷积神经网络,对地震数据进行提取、压缩和精炼;得到一个256*8的数据体;后半部分将256*8的数据体变为4 个256*4的局部结构和一个256*8的全局结构,作为输入,后半部分对提取的信息进一步采集局部特征(指是对原始输入数据整体提取的特征)以及全局特征(提取小图像块的相对原始输入的尺度比较小的特征),并将采集的最终特征用于分类计算。在前半部分网络中采用了图3中的模块结构,前半部分包括一个卷积层和五个模块结构,每个模块结构内部包含卷积层和Relu操作,并伴随恒等快捷连接;每两个模块使卷积步长变为2(指卷积核的移动步长,就是说在每两个模块的计算中把卷积核的步长变为2)进行一次降维操作使维度减半;后半部分将全局结构(原始输入的一个整体数据)的1/4理解为局部结构,对每个局部结构以及全局结构进行如3所示的模块化操作,得到4个256*4的局部结构和一个256*8的全局结构;然后再进行平均池化,得到维度为1×1×256的五个向量,分别代表四个局部特征以及一个全局特征,将上述特征进行连接操作,得到最后用于分类的特征向量,输入全连接层进一步降维,并使用softmax函数进行二分类。本发明提出的模型框架结构如图4所示。The network proposed by the invention is composed of two parts: the first half uses the identity and shortcut connection convolutional neural network to extract, compress and refine the seismic data; obtain a 256*8 data body; the second half uses the 256*8 data The body becomes 4 local structures of 256*4 and a global structure of 256*8. As input, the latter part further collects local features (referring to the features extracted from the original input data as a whole) and global features ( Extract the features of small image patches whose scale is relatively small relative to the original input), and use the collected final features for classification calculations. The module structure in Figure 3 is used in the first half of the network. The first half includes a convolution layer and five module structures. Each module structure contains a convolution layer and a Relu operation, and is accompanied by an identity shortcut connection; every two The module makes the convolution step size 2 (referring to the moving step size of the convolution kernel, that is to say, the step size of the convolution kernel is changed to 2 in the calculation of every two modules) and performs a dimensionality reduction operation to halve the dimension; In the second half, 1/4 of the global structure (an overall data of the original input) is understood as a local structure, and the modular operation shown in 3 is performed on each local structure and the global structure to obtain four 256*4 local structures. and a 256*8 global structure; then average pooling is performed to obtain five vectors with a dimension of 1×1×256, representing four local features and one global feature respectively, and the above features are connected. For the feature vector of classification, input the fully connected layer to further reduce the dimension, and use the softmax function for binary classification. The model frame structure proposed by the present invention is shown in FIG. 4 .
表1是整个网络的内部架构预览。Table 1 is a preview of the internal architecture of the entire network.
表1基于恒等快捷连接的深度卷积层位追踪网络架构Table 1 Deep convolutional layer tracking network architecture based on identity shortcut connections
input输入数据维度为time×xline,逗号后为channel;Kernel内表示均为卷积核大小, block使用中括号表示连续两个卷积层;Stride为2表示该block的第一层卷积步长为2,余下为1;block6_local*表示提取对应部分的局部信息,block6_global提取整体信息; avg_pool_concat表示对所有数据体进行平均池化后连接;fc表示全连接层;softmax将输出映射到[0,1]区间内;该表未完全按照信息流顺序表示,信息流遵照图5所示。The input data dimension is time×xline, followed by the comma; the kernel is the size of the convolution kernel, and the block uses square brackets to indicate two consecutive convolution layers; Stride is 2 to indicate the block's first layer convolution step size is 2, the rest is 1; block6_local* means extracting the local information of the corresponding part, block6_global extracting the overall information; avg_pool_concat means that all data bodies are connected after average pooling; fc means the fully connected layer; softmax maps the output to [0,1 ] interval; the table is not completely represented in the order of the information flow, and the information flow follows that shown in Figure 5.
如前面所述,内部的每个block都可以用如下公式表述:As mentioned earlier, each block inside can be expressed by the following formula:
y=F(x,{Wi})+x (4)y=F(x, {W i })+x (4)
这里x和y是所考虑层的输入和输出向量。函数F(x,{Wi})表示要学习的残存映射。其中 F=W2σ(W1),σ(W1)代表激活函数,本发明使用ReLU作为激活函数,为了简略符号的使用,此处省略了偏差。F(x,{Wi})+x通过恒等快捷连接实现,同时运用元素对应相加。在对应相加后再进行二次非线性操作σ(y)。公式(4)中的恒定快捷连接既不引入额外的参数,也不引入计算复杂性,但x和F的维度必须统一,若不统一(例如当改变通道数的时候),可以通过一个线性映射来匹配通道数:Here x and y are the input and output vectors of the layer under consideration. The function F(x, {W i }) represents the residual map to be learned. Wherein F=W 2 σ(W 1 ), σ(W 1 ) represents the activation function, the present invention uses ReLU as the activation function, and for the use of abbreviated symbols, the deviation is omitted here. F(x, {W i })+x is realized by identity shortcut connection, and the corresponding addition of elements is used at the same time. The quadratic nonlinear operation σ(y) is performed after the corresponding addition. The constant shortcut connection in formula (4) does not introduce additional parameters nor computational complexity, but the dimensions of x and F must be unified. If they are not unified (for example, when changing the number of channels), a linear mapping can be used. to match the number of channels:
y=F(x,{Wi})+Wsx (5)y=F(x, {W i })+W s x (5)
本发明中使用1×1卷积来达到Ws的效果,1×1的卷积核由于大小只有1×1,所以不需要考虑像素跟周边像素的关系,它主要用于调节通道数,对不同的通道上的像素点进行线性组合后执行非线性化操作,完成了升维和降维的功能。In the present invention, 1×1 convolution is used to achieve the effect of W s . Since the size of the 1×1 convolution kernel is only 1×1, it does not need to consider the relationship between pixels and surrounding pixels. It is mainly used to adjust the number of channels. The pixels on different channels are linearly combined and then non-linearized to complete the functions of dimension raising and dimension reduction.
在本发明网络结构设计中,往往只需在某些block中才需要该项功能。可以注意到,上述公式4和公式5中的残差函数为了简单起见使用全连接层来进行表述,在实际使用中完全可以使用卷积层来进行代替。实际运用中也可以使用多个卷积层来表述F(x,{Wi}),在两个特征映射上逐通道执行元素对应相加。In the network structure design of the present invention, this function is often only required in some blocks. It can be noted that the residual functions in
将本发明模型的反向传播机制与常规模型进行对比,假设在执行一个普通的block时,损失针对权重的反向传播为:Comparing the back-propagation mechanism of the model of the present invention with the conventional model, it is assumed that when an ordinary block is executed, the back-propagation of the loss to the weight is:
而当模块使用了恒定快捷连接后,梯度的反向传播为:And when the module uses a constant shortcut connection, the back-propagation of the gradient is:
L表示损失函数,l1表示第一层网络,l2表示第二层网络,o表示整体网络的输出,o1表示网络的第一层的输出,W表示权重矩阵。这个公式就是一个简单的二层网络的公式,主要是为了对比使用恒定连接与不使用的差别。相比于无恒定连接的模块,在W处多出了1,这将梯度不衰减地传递了回去,这使得网络能够构建更深层的地震特征,提取了常规卷积神经网络不够敏感的细微特征,使得分类效果更佳精确。L represents the loss function, l 1 represents the first layer network, l 2 represents the second layer network, o represents the output of the overall network, o 1 represents the output of the first layer of the network, and W represents the weight matrix. This formula is a simple two-layer network formula, mainly to compare the difference between using a constant connection and not using it. Compared with the module without constant connection, there is an extra 1 at W, which transfers the gradient back without decay, which enables the network to build deeper seismic features and extract subtle features that conventional convolutional neural networks are not sensitive enough to. , which makes the classification effect more accurate.
下面结合附图和具体实施例进一步说明本发明的技术方案。在后续将用英文Identity Shortcut Global Local Convolutional Neural Network的简称ISGL-Net代表本发明提出的网络进行表述。The technical solutions of the present invention are further described below with reference to the accompanying drawings and specific embodiments. In the following, the abbreviation ISGL-Net of the English Identity Shortcut Global Local Convolutional Neural Network will be used to represent the network proposed by the present invention.
如图5所示,本发明的基于恒等快捷映射的深层卷积神经网络自动层位追踪方法,包括以下步骤:As shown in Figure 5, the deep convolutional neural network automatic layer tracking method based on identity shortcut mapping of the present invention includes the following steps:
S1、对地震剖面图像进行预处理;S1. Preprocess the seismic profile image;
本实施例采用西南某工区三维地震体进行实验,该数据体主测线地震解释剖面(inline) 由线号1760至线号2160共401道线号,联络线地震解释剖面(xline)由线号1780至线号2220共440道线号,该样本采样间隔为2ms,时间范围为0ms至1300ms。该三维地震体如图6所示。This example uses a 3D seismic volume in a work area in the southwest for the experiment. The seismic interpretation profile (inline) of the main survey line of the data volume has a total of 401 line numbers from
本步骤具体包括以下子步骤:This step specifically includes the following sub-steps:
S11、整个地震剖面包含了太多信息,无法直接作为输入图像。将整个地震剖面划分成多个小图像块,这样的图像块利于计算与分析;每个小图像块大小为32×32;图7展示了生成训练数据的工作流。人工标记的层位位于左侧图像的中部,在整个地震剖面上滑动一个窗口(左上角显示的小窗口)以创建大小为32×32的二维图像补丁。每幅32×32的图像的宽度和高度都为0-31,采样间隔为2ms,这代表着使用的每个图片块为32道,64ms 以及一个颜色通道。S11. The entire seismic section contains too much information to be directly used as an input image. The entire seismic section is divided into several small image blocks, which are convenient for calculation and analysis; each small image block is 32×32 in size; Figure 7 shows the workflow for generating training data. The manually marked horizons are in the middle of the left image, and a window (the small window shown in the upper left corner) is slid across the entire seismic section to create a 2D image patch of
S12、在提取图像块后,需要将标签分配给每个小图像块,使得具有相同标签的图像补丁共享特定特征;具体实现方法为:判断每个小图像块的中心(若每个小图像块置于二维坐标轴中,其一个顶点位于原点处,则将(15,15)的坐标点作为图片的中心)是否位于层位上,若是则将该图像块标记为1,否则标记为0;如图8所示,图中的小圆圈代表图像块的中心;S12. After extracting the image blocks, a label needs to be assigned to each small image block, so that the image patches with the same label share specific features; the specific implementation method is: judging the center of each small image block (if each small image block It is placed in the two-dimensional coordinate axis, and one of its vertices is located at the origin, then the coordinate point of (15, 15) is used as the center of the picture) whether it is located on the horizon, if so, the image block is marked as 1, otherwise it is marked as 0 ; As shown in Figure 8, the small circle in the figure represents the center of the image block;
S13、将图像块数据进行二分类训练,以判别该图像块的中心位置是否处于层位上;S13, perform two-class training on the image block data to determine whether the center position of the image block is on the horizon;
S14、在提取数据的过程中,由于层位数据本身的特性,会发生正负样本严重不均衡的现象,本发明通过以下方法对非层位数据进行处理:设定遗失概率来降低得到的非层位图像块的数量,设置遗失概率为0.5,即随机选取50%的非层位数据进行舍弃;使用遗失概率是为了缓解数据不均衡的现象,即将一半的非层位图像块数据进行舍弃,缓解了过多的数据块的中心不在层位上的情况。S14. In the process of extracting data, due to the characteristics of the horizon data itself, the phenomenon that the positive and negative samples are seriously unbalanced will occur. For the number of horizon image blocks, set the loss probability to 0.5, that is, randomly select 50% of the non-horizontal data to discard; the loss probability is used to alleviate the phenomenon of data imbalance, that is, to discard half of the non-horizon image block data, Alleviates the situation that the center of too many data blocks is not on the horizon.
同时在层位上设定偏差值,若一个小图像快的中心点与某个确定的层位的距离在偏差值范围内,即使该数据被人工标记标定为非层位,也认定该数据为层位数据,标记为1,本实施例所使用的偏差值为10个数据点;At the same time, the deviation value is set on the horizon. If the distance between the center point of a small image and a certain horizon is within the deviation value range, even if the data is manually marked as non-horizontal, the data will be considered as Horizon data, marked as 1, the deviation value used in this embodiment is 10 data points;
分类问题中,当不同类别的样本量差异很大,即类分布不平衡时,很容易影响分类结果。因此需要进行校正。具体方法为:对训练集里的每个类别加权,从而使模型更加关注样本数量少的类别。使用权重法,也就是样本数量多的类权重低,反之权重高,对损失进行分配。本发明中使用的权重为类对应的样本数量的倒数;这里的类就是标记为1的层位数据以及标记为0的非层位数据。例如,标记为0的数据数量为90,则其权重为1/90;标记为1的数据的数量为10,则其权重为1/10。In classification problems, when the sample sizes of different classes vary greatly, that is, the class distribution is unbalanced, it is easy to affect the classification results. Therefore correction is required. The specific method is to weight each category in the training set, so that the model pays more attention to the category with a small number of samples. The weight method is used, that is, the class with a large number of samples has a low weight, and vice versa, the loss is allocated. The weight used in the present invention is the inverse of the number of samples corresponding to the class; the class here is the horizon data marked as 1 and the non-horizontal data marked as 0. For example, if the number of data marked 0 is 90, its weight is 1/90; the number of data marked 1 is 10, its weight is 1/10.
在后续实验中发现,上述方案能够极大减弱正负样本不均衡的现象,并能够极大提升模型的准确性。In subsequent experiments, it was found that the above scheme can greatly reduce the imbalance of positive and negative samples, and can greatly improve the accuracy of the model.
S15、从层位数据集合中提取20%作为验证数据,用于对模型超参数进行调整(根据验证损失(用这20%的训练数据进行验证时得到的验证误差)来调整参数,也可以在训练的时候在checkpoint设置保留性能最好的模型);其余数据作为训练数据;S15. Extract 20% from the horizon data set as the validation data to adjust the model hyperparameters (according to the validation loss (validation error obtained when validating with the 20% of the training data) to adjust the parameters, or in When training, keep the model with the best performance in the checkpoint setting); the rest of the data is used as training data;
S16、对数据进行归一化处理:数据归一化是机器学习数据准备过程中必不可少的步骤。归一化是指将所有维度上的输入数据规范化在固定尺度内。3D地震数据的振幅属性的范围为-32767到32767,然而如果ReLU激活函数接收到任何负输入,则它返回0,这样就不会在此零值区域中进行学习。一旦单位变成负数,它们就不会在区分输入中起任何作用,并且基本上是无用的。如果输入值太大,则会影响神经网络的收敛。因此,需要对输入数据进行归一化以提高收敛速度和训练效率。本实施例使用Min-Max规范化方法,使数据尺度被规范化到0-1范围内;Min-Max规范化方法通过以下公式实现:S16, normalize the data: data normalization is an essential step in the process of machine learning data preparation. Normalization refers to normalizing the input data in all dimensions within a fixed scale. The amplitude property of 3D seismic data ranges from -32767 to 32767, however if the ReLU activation function receives any negative input it returns 0 so that no learning will take place in this zero value region. Once the units become negative, they play no role in differentiating the input and are basically useless. If the input value is too large, it will affect the convergence of the neural network. Therefore, the input data needs to be normalized to improve the convergence speed and training efficiency. This embodiment uses the Min-Max normalization method to normalize the data scale to the range of 0-1; the Min-Max normalization method is implemented by the following formula:
其中Xnor为规范化后的数据,X为原始的地震数据,Xmax和Xmin分别为地震振幅中的极大值和极小值。where X nor is the normalized data, X is the original seismic data, and X max and X min are the maximum and minimum values of the seismic amplitude, respectively.
S2、使用恒等快捷连接卷积神经网络对地震反射特征进行提取、压缩和精炼;将预处理完成后的实验数据输入模型,本发明使用的实验训练处理器为NVIDIA GeForce GTX1050(4g),相比使用中央处理器,使用图形处理器能够加速涉卷积神经网络的模型训练。S2. Extract, compress and refine the seismic reflection features by using the constant shortcut connection convolutional neural network; input the experimental data after preprocessing into the model, and the experimental training processor used in the present invention is NVIDIA GeForce GTX1050 (4g), which is similar to Using a graphics processor can speed up model training involving convolutional neural networks rather than a central processing unit.
S3、对步骤S2提取的信息进一步采集局部特征以及全局特征;S3, further collect local features and global features for the information extracted in step S2;
S4、将采集的特征进行分类计算。S4, classify and calculate the collected features.
下面使用准备好的数据进行测试,并通过结果进行分析。The following is tested with the prepared data and analyzed with the results.
本实施例选择了西南某工区受断层严重干扰以及附带大倾角特征的层位,认为该层位具有典型代表性,如图9展示了目标层位的反射模式,图9为西南某工区inline1760二维剖面,分别标记了原始层位,断层以及大倾角。In this example, a horizon that is seriously disturbed by faults and has a large dip angle in a work area in the southwest is selected, and this horizon is considered to be typical. Figure 9 shows the reflection pattern of the target horizon, and Figure 9 shows a work area in the southwest of inline1760 II Dimensional section, respectively marking the original horizon, fault and high dip.
首先使用了与训练数据较邻近的剖面inline1762的层位如图10所示,图10为西南某工区inline1762二维剖面,分别标记了原始层位,断层以及大倾角。并与普通深度卷积神经网络与传统自动层位追踪方法进行了比较,可以看到由于剖面位置较邻近,该层位与剖面inline1760十分相似。First, the horizons of the inline1762 section, which is closer to the training data, are used as shown in Figure 10. Figure 10 is a two-dimensional section of inline1762 in a work area in the southwest, with the original horizons, faults and large dips marked respectively. And compared with the ordinary deep convolutional neural network and the traditional automatic horizon tracking method, it can be seen that the horizon is very similar to the profile inline1760 due to the close proximity of the profile.
图10为CNN和ISGL-net两个方法在inline1762上的层位概率分布,可以看到两者对于两个断层以及大倾角的处理都较好,在层位最左侧开头处以及第一个断层处,常规CNN出现了部分瑕疵(图11(a)),ISGL-net(图11(b))相对常规卷积神经网络总体来看更加清晰与干净。虽然常规CNN出现了上述小部分瑕疵,实验结果仍然处于能够令人接受的状态,这样的结果是可以预见的,这是因为训练剖面与测试剖面的物理位置较接近,使训练数据与测试数据的分布较接近。Figure 10 shows the horizon probability distribution of the two methods of CNN and ISGL-net on inline1762. It can be seen that both of them handle two faults and large dip angles well, at the beginning of the leftmost horizon and the first At the fault, the conventional CNN has some defects (Fig. 11(a)), and ISGL-net (Fig. 11(b)) is generally clearer and cleaner than the conventional convolutional neural network. Although the conventional CNN has some of the above-mentioned flaws, the experimental results are still in an acceptable state, and such results are predictable, because the physical locations of the training and test profiles are close, which makes the training data and test data indistinguishable. distribution is relatively close.
接下来将两者设定阈值并去除偏差值与传统层位追踪方法的结果进行比较,本实施例所设定阈值为0.9,如图12所示。Conventional代表传统方法;Manual代表人工进行标记,可以看到,在纵向变化不大的前部区域,CNN/ISGL-net以及传统的基于波形相似同相轴追踪方法都有不错的表现。在最后部分的大倾角区域,几种方法都有不错的表现。将图12中方框部分放大,如图13所示。Next, compare the results of the two thresholds and remove the deviation with the results of the traditional horizon tracking method. The threshold set in this embodiment is 0.9, as shown in FIG. 12 . Conventional represents the traditional method; Manual represents manual marking. It can be seen that CNN/ISGL-net and the traditional waveform similarity-based event tracking method have good performance in the front area with little longitudinal change. In the region of large dip in the last part, several methods have good performance. Enlarge the block in Figure 12, as shown in Figure 13.
从图13可以看到传统层位追踪方法在第一断层处发生了严重的误判,追踪到了相邻层位,并且在第二个断层处仍然处于误判,而在该区域,CNN与ISGL-net仍然与人工标记保持不错的一致性;It can be seen from Figure 13 that the traditional horizon tracking method has a serious misjudgment at the first fault, traces the adjacent horizon, and is still misjudged at the second fault, and in this area, CNN and ISGL -net still maintains good consistency with human labeling;
接下来将各个方法以人工标记层位作为标准进行误差量化,本文使用绝对误差(Absolute Error)和平均绝对误差(Mean Absolute Error)来对距离误差进行衡量。公式如下:Next, each method uses the artificially marked horizon as the standard to quantify the error. This paper uses the absolute error (Absolute Error) and the mean absolute error (Mean Absolute Error) to measure the distance error. The formula is as follows:
其中m为测试集采样点个数。各方法以人工标记层位为准,在断层处的判定错误,给与每个判定点的惩罚值为20,对于inline1762层位的误差如表2所示。where m is the number of sampling points in the test set. Each method is based on the manually marked horizons. The judgment error at the fault is given to each judgment point with a penalty value of 20. The errors for the inline1762 horizon are shown in Table 2.
表2西南某工区inline1762层位误差比较Table 2 Comparison of inline1762 horizon errors in a work area in southwest China
综合图12以及表1可知,ISGL-net效果相对其余两者效果更加优异,CNN效果与ISGL-CNN接近,然而传统层位追踪方法在实际表现以及量化结果均不如人意。由于inline1762与训练数据inline1760物理距离相对较近,这导致两者数据分布较为接近,为了测试模型的泛化性能,本实施例选取相对原始训练数据物理实际位置更远的inline1775进行测试。inline1775及层位标定如图14所示可以看到该剖面数据分布与inline1760差异较大。Combining Figure 12 and Table 1, it can be seen that the effect of ISGL-net is better than the other two, and the effect of CNN is close to that of ISGL-CNN. However, the actual performance and quantization results of traditional horizon tracking methods are not satisfactory. Since the physical distance between inline1762 and the training data inline1760 is relatively close, the data distribution of the two is relatively close. In order to test the generalization performance of the model, this embodiment selects inline1775, which is farther from the actual physical location of the original training data, for testing. Inline1775 and horizon calibration are shown in Figure 14. It can be seen that the profile data distribution is quite different from inline1760.
图15为西南某工区inline1775层位概率分布图像,其中(a)为,CNN处理结果,(b)为ISGL-net处理结果。可以看到CNN在如图15方框所标记的两处断层处都进行了较大的错误预判,而本发明提出的ISGL-net在第一处断层进行了较为成功的断层检测,在第二处断层出现了预判失误,但从分辨效果来看仍然优于CNN。在纵向变化轻微以及大倾角部分,两者分辨效果均较为不错。Figure 15 is the inline1775 horizon probability distribution image of a work area in the southwest, where (a) is the CNN processing result, and (b) is the ISGL-net processing result. It can be seen that CNN has made a large error prediction at the two faults marked by the box in Figure 15, while the ISGL-net proposed by the present invention has performed a relatively successful fault detection on the first fault. The second fault has a prediction error, but it is still better than CNN in terms of resolution. In the slight longitudinal change and the large inclination part, the resolution effect of the two is relatively good.
同理,接下来仍然对结果进行设定阈值并去偏差值操作,并与传统层位追踪方法进行比较,结果如图16所示。在纵向变化不大的前部区域,CNN/ISGL-net以及传统层位追踪方法都有不错的表现。在最后部分的大倾角区域,三个方法都有不错的表现。In the same way, the result is still set the threshold value and de-biased value operation, and compared with the traditional horizon tracking method, the result is shown in Figure 16. In the front area with little longitudinal change, CNN/ISGL-net and traditional horizon tracking methods have good performance. In the region of large dip in the last part, all three methods perform well.
接下来将大方框框标记部分放大。如图17所示,可以观察到,在方框标记范围内,传统层位追踪方法在两个断层处出现了连续的追踪错误,均追踪到了邻近层位,包含断层在内一共出现了四处追踪错误;在第一处蓝框标记范围内,可以观察到,CNN在第一处横向断层出现了严重的预判错误,而ISGL-net在此处表现良好;在第二处蓝框标记范围内,在第二处断层,CNN与ISGL-net均出现了不同程度的追踪失误,但总体来看,ISGL-net的效果仍优于CNN。接下来对各个方法以inline 1775人工标记层位为标准量化追踪误差,误差如表3所示。Next, enlarge the marked part of the large box. As shown in Figure 17, it can be observed that within the range of the box mark, the traditional horizon tracking method has continuous tracking errors at two faults, both of which are traced to the adjacent horizons, and there are four traces including the fault. Error; within the range marked by the first blue box, it can be observed that CNN has a serious prediction error at the first transverse fault, while ISGL-net performs well here; within the range marked by the second blue box , in the second fault, both CNN and ISGL-net have different degrees of tracking errors, but overall, ISGL-net is still better than CNN. Next, for each method, the inline 1775 artificially marked horizon is used as the standard to quantify the tracking error, and the errors are shown in Table 3.
表3西南某工区inline1775层位误差较Table 3 Comparison of horizon error of inline1775 in a work area in southwest China
综上所述,在物理距离距原始训练数据较远,数据分布差距较大的情况下,ISGL-net具有比CNN更好的泛化性能。ISGL-net与CNN和传统基于波形相似的层位追踪方法相比,追踪效果更加良好。To sum up, ISGL-net has better generalization performance than CNN when the physical distance is far from the original training data and the data distribution gap is large. Compared with CNN and traditional waveform-similar-based horizon tracking methods, ISGL-net has better tracking effect.
本领域的普通技术人员将会意识到,这里所述的实施例是为了帮助读者理解本发明的原理,应被理解为本发明的保护范围并不局限于这样的特别陈述和实施例。本领域的普通技术人员可以根据本发明公开的这些技术启示做出各种不脱离本发明实质的其它各种具体变形和组合,这些变形和组合仍然在本发明的保护范围内。Those of ordinary skill in the art will appreciate that the embodiments described herein are intended to assist readers in understanding the principles of the present invention, and it should be understood that the scope of protection of the present invention is not limited to such specific statements and embodiments. Those skilled in the art can make various other specific modifications and combinations without departing from the essence of the present invention according to the technical teachings disclosed in the present invention, and these modifications and combinations still fall within the protection scope of the present invention.
Claims (2)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202010841336.6A CN111815680A (en) | 2020-08-20 | 2020-08-20 | An automatic horizon tracking method for deep convolutional neural networks based on identity shortcut mapping |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202010841336.6A CN111815680A (en) | 2020-08-20 | 2020-08-20 | An automatic horizon tracking method for deep convolutional neural networks based on identity shortcut mapping |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN111815680A true CN111815680A (en) | 2020-10-23 |
Family
ID=72860449
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202010841336.6A Pending CN111815680A (en) | 2020-08-20 | 2020-08-20 | An automatic horizon tracking method for deep convolutional neural networks based on identity shortcut mapping |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN111815680A (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112433249A (en) * | 2020-12-08 | 2021-03-02 | 中国石油天然气集团有限公司 | Horizon tracking method and device, computer equipment and computer readable storage medium |
| CN113947032A (en) * | 2021-11-17 | 2022-01-18 | 中国石油大学(北京) | Seismic horizon picking method and device based on neural network and dynamic time warping |
| CN114814947A (en) * | 2022-04-26 | 2022-07-29 | 电子科技大学 | 3D full horizon tracking method based on multi-attribute guidance |
| CN119716991A (en) * | 2023-09-26 | 2025-03-28 | 中国石油天然气集团有限公司 | Automatic layer tracking method, device and readable storage medium |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109086773A (en) * | 2018-08-29 | 2018-12-25 | 电子科技大学 | Fault plane recognition methods based on full convolutional neural networks |
| US20190064378A1 (en) * | 2017-08-25 | 2019-02-28 | Wei D. LIU | Automated Seismic Interpretation Using Fully Convolutional Neural Networks |
| CN109709603A (en) * | 2018-11-23 | 2019-05-03 | 中国石油天然气股份有限公司 | Seismic horizon identification and method for tracing, system |
| CN110516740A (en) * | 2019-08-28 | 2019-11-29 | 电子科技大学 | A Fault Recognition Method Based on Unet++ Convolutional Neural Network |
-
2020
- 2020-08-20 CN CN202010841336.6A patent/CN111815680A/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190064378A1 (en) * | 2017-08-25 | 2019-02-28 | Wei D. LIU | Automated Seismic Interpretation Using Fully Convolutional Neural Networks |
| CN109086773A (en) * | 2018-08-29 | 2018-12-25 | 电子科技大学 | Fault plane recognition methods based on full convolutional neural networks |
| CN109709603A (en) * | 2018-11-23 | 2019-05-03 | 中国石油天然气股份有限公司 | Seismic horizon identification and method for tracing, system |
| CN110516740A (en) * | 2019-08-28 | 2019-11-29 | 电子科技大学 | A Fault Recognition Method Based on Unet++ Convolutional Neural Network |
Non-Patent Citations (1)
| Title |
|---|
| 袁英淏: "基于深度学习的地震储层图像解释方法研究", 《中国优秀硕士学位论文全文数据库 基础科学辑》 * |
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112433249A (en) * | 2020-12-08 | 2021-03-02 | 中国石油天然气集团有限公司 | Horizon tracking method and device, computer equipment and computer readable storage medium |
| CN113947032A (en) * | 2021-11-17 | 2022-01-18 | 中国石油大学(北京) | Seismic horizon picking method and device based on neural network and dynamic time warping |
| CN113947032B (en) * | 2021-11-17 | 2022-08-19 | 中国石油大学(北京) | Seismic horizon picking method and device based on neural network and dynamic time warping |
| CN114814947A (en) * | 2022-04-26 | 2022-07-29 | 电子科技大学 | 3D full horizon tracking method based on multi-attribute guidance |
| CN119716991A (en) * | 2023-09-26 | 2025-03-28 | 中国石油天然气集团有限公司 | Automatic layer tracking method, device and readable storage medium |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Zhu et al. | Intelligent logging lithological interpretation with convolution neural networks | |
| CN111815680A (en) | An automatic horizon tracking method for deep convolutional neural networks based on identity shortcut mapping | |
| CN109086824B (en) | A classification method of submarine bottom sonar images based on convolutional neural network | |
| CN109709603B (en) | Seismic horizon identification and tracking method and system | |
| CN103592681B (en) | A kind of seismic image tracing of horizons method based on signal classification | |
| CN112731522B (en) | Intelligent recognition method, device and equipment for seismic stratum and storage medium | |
| CN109001801B (en) | Fault variable-scale identification method based on multiple iteration ant colony algorithm | |
| CN113516626B (en) | Side-scan sonar seafloor sand wave detection method based on multi-scale convolution and pooling strategies | |
| CN107607992B (en) | Multi-wave matching method based on convolutional neural network | |
| CN111381275A (en) | First arrival picking method and device for seismic data | |
| CN116047583A (en) | Adaptive wave impedance inversion method and system based on deep convolutional neural network | |
| CN109670533A (en) | A kind of multiple dimensioned well-log facies recognition method based on convolutional neural networks | |
| Zhang et al. | SaltISCG: Interactive salt segmentation method based on CNN and graph cut | |
| CN113762271B (en) | SAR image semantic segmentation method and system based on irregular convolution kernel neural network model | |
| CN110443801A (en) | A kind of salt dome recognition methods based on improvement AlexNet | |
| CN116299665A (en) | LSTM surface wave inversion method, device and medium | |
| CN106226816A (en) | A kind of pre-stack seismic signal waveform sorting technique | |
| CN116863293A (en) | Marine target detection method under visible light based on improved YOLOv7 algorithm | |
| CN114898160A (en) | Multi-task based fault intelligent identification method | |
| CN114814947A (en) | 3D full horizon tracking method based on multi-attribute guidance | |
| Lyu et al. | A buried hill fault detection method based on 3D U-SegNet and transfer learning | |
| Wang et al. | High-resolution seismic faults interpretation based on adversarial neural networks with a regularization technique | |
| Zhang et al. | A rule-based interpretation system for segmentation of seismic images | |
| Yuan et al. | Automatic velocity picking based on improved Mask R-CNN | |
| Wang et al. | Rotation invariant CNN using scattering transform for seismic facies classification |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201023 |