CN113537120A - Convolutional neural network based on complex coordinate attention module and target identification method - Google Patents
Convolutional neural network based on complex coordinate attention module and target identification method Download PDFInfo
- Publication number
- CN113537120A CN113537120A CN202110858271.0A CN202110858271A CN113537120A CN 113537120 A CN113537120 A CN 113537120A CN 202110858271 A CN202110858271 A CN 202110858271A CN 113537120 A CN113537120 A CN 113537120A
- Authority
- CN
- China
- Prior art keywords
- complex
- module
- output
- input
- feature map
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/02—Preprocessing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/12—Classification; Matching
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a convolutional neural network based on a complex coordinate attention module and a target identification method, and relates to the field of target identification, wherein the convolutional neural network comprises the following components: the device comprises an input layer, N basic units, a classification unit and an output layer; the processing unit is used for mapping the complex numbers into corresponding real numbers through modular operation and performing classification and identification; the N basic units include first to nth basic units, each of the N basic units including: the system comprises a first complex convolution module, a first complex batch normalization module, a first complex activation module and a first complex pooling module; wherein one of the N basic units further comprises: a plurality of coordinate attention modules; the complex coordinate attention module includes: the invention realizes the high-precision identification of the similar space cone target.
Description
Technical Field
The invention relates to the field of target identification, in particular to a convolutional neural network based on a complex coordinate attention module and a target identification method.
Background
Ballistic missiles release a bait that is very similar to the warhead during penetration, and therefore it is necessary to identify the warhead and the bait during the missile, thereby reducing the interception cost. The warhead and the bait can be regarded as similar space cone targets with the shapes consistent with the movement forms and only slight differences in movement parameters, so that the similar space cone target identification plays an important role in the fields of space resource utilization, space monitoring (surveyability) and military.
In recent years, studies for spatial object recognition by introducing a Convolutional Neural Network (CNN) have been increasing based on the idea of extracting a fine motion feature in an image domain and then using the extracted fine motion feature for recognition. Li et al, based on CNN, investigated the problem of identification of spatially diverse shaped, different precession frequency targets using a Multi-mode Fusion (Multi-mode Fusion) approach. Generating S-frequency band and X-frequency band one-dimensional distance images and time-frequency spectrograms of targets with different shapes and different precession frequencies by using an ideal point scattering model; then, the multimode data is used as input of CNN, and three targets of cone, small cone and cylinder are identified. Bai et al, Xu et al and Han et al all adopt a method of taking a time-frequency spectrogram of a target as network input to identify the target, and actually, the problem of identifying the target by utilizing the micro-motion characteristics is converted into the problem of identifying an image. Designing a CNN with three layers of depth by Bai et al, and generating a time-frequency spectrogram in three micro-motion forms of spin, precession and nutation by using an ideal point scattering model; then, the time-frequency spectrograms are properly cut and used as the input of the designed CNN, and the three common micromotion forms are identified. Xu et al, designing a CNN with the depth of six layers, and generating echo signals in four micromotion forms of spin, rolling, precession and nutation by using a scattering point model; then obtaining time-frequency spectrograms of a plurality of micro-motion periods through WVD (Wigner-Ville distribution); and finally, the time-frequency spectrogram is used as input of the CNN to finish the identification of the four micromotion forms. Han et al designed a deep learning network consisting of one-dimensional parallel structures (1-D parallel structures) and Long-Short-Term Memory (LSTM) layers. The method simulates five targets with different structural parameters and different micromotion forms by using an electromagnetic calculation method to obtain echo data; then, performing Time-frequency analysis on the echoes by Short-Time Fourier Transform (STFT) to obtain Time-frequency spectrograms of a plurality of micro-motion periods; and finally, the time-frequency spectrograms are sent to a designed network to finish the identification of the five targets. Wang et al, based on electromagnetic calculation data, obtained more than one precession cycle distance-slow time image of three targets with different geometric shapes and the same micromotion form, namely cone, cone-cylinder and cone-cylinder skirt, and send the image into the designed CNN to realize target recognition. From the prior art listed above, the main way of doing so is to retain the processing from "echo data domain" to "image domain" as preprocessing, and then replace the process of extracting the micro-motion features from "image domain" for identification with a deep convolutional neural network, which has the following problems: (1) preprocessing such as time-frequency analysis (time-frequency analysis) or distance slow time imaging is required, so that longer signal processing time is required; (2) a long time of continuous observation of the target is required to obtain a complete periodic image of the target; (3) these methods only target different shapes, different micro-motion forms, and do not achieve the identification of similar spatial cone targets.
Disclosure of Invention
In order to solve the three problems of space cone target identification based on the image domain CNN, the invention integrates the advantages of CV-CNN and an attention mechanism, introduces the coordinate attention of a real number domain into a complex number domain, constructs a convolutional neural network based on a complex number coordinate attention module and a target identification method, and aims to take radar echo complex data as input data to carry out direct operation, fully utilizes amplitude and phase information, and realizes the high-precision identification of similar space cone targets with the same geometric shape and micromotion form and the same micromotion parameters and the small difference.
To achieve the above object, the present invention provides a convolutional neural network based on a complex coordinate attention module, the convolutional neural network comprising:
the device comprises an input layer, N basic units, a classification unit and an output layer;
the processing unit is used for mapping the complex numbers into corresponding real numbers through modular operation and performing classification and identification; the N basic units comprise first to Nth basic units, the first basic unit is connected with the input layer, the output of the first basic unit is the input of the second basic unit, the input of the Nth basic unit is the output of the (N-1) th basic unit, N is an integer larger than 1, the output of the Nth basic unit is the input of the processing unit, the output of the processing unit is the input of the classifier, and the classifier is connected with the output layer; the N basic units each include: the system comprises a first complex convolution module, a first complex batch normalization module, a first complex activation module and a first complex pooling module; wherein one of the N basic units further comprises: a plurality of coordinate attention modules; the complex coordinate attention module includes: the system comprises a complex coordinate attention embedding unit and a complex coordinate attention generating unit, wherein for each channel, the complex coordinate attention embedding unit is used for encoding a first complex input feature map of the channel along the horizontal direction and the vertical direction respectively, and generating first output feature information of the first complex input feature map after the first complex input feature map is encoded along the horizontal direction and second output feature information of the first complex input feature map after the first complex input feature map is encoded along the vertical direction;
for each channel, the complex coordinate attention generating unit is to: splicing the first output characteristic information and the second output characteristic information to generate a characteristic information splicing result of the channel; performing feature dimensionality reduction on the feature information splicing result of the channel to obtain feature information after dimensionality reduction, and activating the feature information after dimensionality reduction to obtain a first complex output feature map of the channel; splitting the first complex output profile into a first tensor and a second tensor along a spatial dimension; adjusting the dimensions of the first tensor and the second tensor to be the same as the dimensions of the first complex input feature map, and obtaining a second complex output feature map of the channel in the horizontal direction and a third complex output feature map of the channel in the vertical direction; obtaining a third tensor and a fourth tensor, wherein the third tensor is the set of the second complex output characteristic maps of all the channels, and the fourth tensor is the set of the third complex output characteristic maps of all the channels;
expressing each element in the third tensor and the fourth tensor in a polar coordinate form, constraining the amplitude of the polar coordinate by using a constraint function, respectively obtaining a fourth complex output feature map and a fifth complex output feature map in the horizontal and vertical spatial directions, expanding the fourth complex output feature map and the fifth complex output feature map to generate attention weight distribution in the horizontal and vertical spatial directions, and applying the attention weight distribution to a complex input feature map of the complex coordinate attention module to obtain a complex output feature map of the complex coordinate attention module;
wherein, the complex input characteristic diagram and the complex output characteristic diagram are both complex characteristic diagrams.
When the convolutional neural network based on the complex coordinate attention module is used for target identification, preprocessing such as time-frequency analysis or distance slow time imaging is not needed, so that longer signal processing time is not needed, and the efficiency is higher; when the convolutional neural network based on the complex coordinate attention module is used for identifying the target, the target does not need to be continuously observed for a long time to obtain a complete periodic image of the target, and the efficiency is high; the convolutional neural network based on the complex coordinate attention module realizes the identification of the similar space cone target.
Preferably, in the basic unit not including the complex coordinate attention module, the output of the complex convolution module in the basic unit is the input of the complex batch normalization module, the output of the complex batch normalization module is the input of the complex activation module, and the output of the complex activation module is the input of the complex pooling module.
Preferably, in the basic unit including the complex coordinate attention module, the output of the complex convolution module in the basic unit is the input of the complex batch normalization module, the output of the complex batch normalization module is the input of the complex coordinate attention module, the output of the complex coordinate attention module is the input of the complex activation module, and the output of the complex activation module is the input of the complex pooling module.
Preferably, the classification unit includes:
the second complex convolution module, the second complex batch normalization module, the second complex activation module, the third complex convolution module and the classifier; the output of the second complex convolution module is the input of the second complex batch normalization module, the output of the second complex batch normalization module is the input of the second complex activation module, the output of the second complex activation module is the input of the third complex convolution module, and the output of the third complex convolution module is the input of the classifier.
Preferably, the convolutional neural network includes first to sixth basic units.
Preferably, the sixth base unit comprises said complex coordinate attention module.
Preferably, an optimizer is arranged in the convolutional neural network and used for updating the network weight and the bias term.
Preferably, the number of the convolution kernels of the first to sixth basic units is 64, 128, 256 and 256, respectively, the sizes of the convolution kernels are all 1 × 3, the sizes of the sampling windows of the first complex pooling module are all 1 × 2, the sliding step of the convolution is 1, and the padding number is 1.
Preferably, the complex input feature map is a complex input feature map of the spatial target identification signal, and the complex output feature map is a complex output feature map of the spatial target identification signal.
On one hand, a complex coordinate attention module CV-CA utilizes a complex convolution neural network to obtain amplitude and phase characteristics of a signal through complex real part and imaginary part correlation learning; on the other hand, spatial information and channel information in the horizontal direction and the vertical direction are concerned at the same time through the attention of complex coordinates, remote lazy relation of characteristic information is better modeled, and the characteristic characterization capability of the target object is enhanced.
In the channel attention, global pooling is usually used to encode global spatial information, but it compresses the global spatial information into a channel descriptor, so that it is difficult to maintain location information, which is particularly important for capturing spatial structure. Therefore, in the coordinate attention module, the operation of decomposing the global pooling into two one-dimensional feature codes is extended to a complex field, the complex feature maps X of each channel are encoded along two horizontal and vertical directions respectively (direction-correlation is called horizontal and vertical directions for short), and direction-correlated complex feature maps are generated, so that the features in two spatial directions are integrated respectively.
The complex coordinate attention embedding unit outputs accurate spatial position information aggregated under the global receptive field. Based on the encoding result of the complex coordinate attention embedding unit, the complex coordinate attention module designs a second transformation called a complex coordinate attention generating unit. The complex coordinate attention generating unit transformation includes three parts, respectively: (1) direction-related feature information aggregation, (2) direction-related complex feature map splitting, and (3) complex coordinate attention automatic assignment.
Preferably, in the present invention, X is a complex input feature map of the complex coordinate attention module,wherein x iscIs the complex input profile for the c-th channel,is a C x W x H dimension complex tensor,the method comprises the following steps of (1) obtaining a complex space, C being the number of channels of input feature maps, W being the width of each input feature map, and H being the height of each input feature map; y is the complex output characteristic diagram of the complex coordinate attention module,wherein, ycThe complex output characteristic diagram of the C channel is shown, C is an integer which is greater than or equal to 1 and less than or equal to C, and the dimension of X is the same as that of Y;
the output of the p channel of the complex input feature diagram X after being coded along the horizontal direction isThe output of the p channel of the complex input characteristic diagram X after being coded along the vertical direction isWherein:
wherein j represents an imaginary unit,representing the real part of the complex number,representing the imaginary part of the complex number, h being the horizontal pixel index of the input feature map, xp(h, j) is the value of the h row and j column of the p channel of the complex input characteristic diagram, i is the pixel index of the vertical direction of the input characteristic diagram, xp(i, w) is the value of the ith row and the w column of the ith channel of the complex input characteristic diagram.
Preferably, the invention willAndsplicing is carried out to obtain a characteristic information splicing result M,each tensor in M is represented asWherein, wherein [, ]]Representing splicing operation, wherein T is a transposition matrix;
the complex coordinate attention generating unit uses a convolution kernel of 1 multiplied by 1 to carry out feature dimensionality reduction on the feature information splicing result, wherein the feature dimensionality reduction can reduce the number of parameters, and meanwhile, cross-channel information interaction and integration can be realized, the complex coordinate attention generating unit uses a convolution kernel of 1 multiplied by 1 to carry out feature dimensionality reduction on the feature information splicing result, and the complex coordinate attention generating unit sets the feature dimensionality reductionA 1 x 1 convolutional kernel shared by convolutional layers, wherein,is shown ask complex convolution kernels, k being 1,2, …, C/r,in the representationThe (C) th 1 x 1 rewinding and stacking nucleus,in the representationQ is 1,2, …, C, r represents a scaling coefficient for controlling the number of channels of the convolution output feature map, s represents the step size of the convolution operation, and the k-th feature map of the convolution output is vk(i, j) wherein:
fk(i,j)=σ(vk(i,j))
wherein m isqFor the qth tensor in M,mq(i · s, j · s) are the values of the i · s row and j · s column of the q th tensor after feature information stitching, vk(i, j) represents a complex output profile of the k-th channel that is not activated,a complex output characteristic diagram representing the k-th channel, and the set of complex characteristic diagrams of each channel is denoted asfC/rIs as followsC/rA complex output profile of each channel, σ (-) representing a complex activation function; the complex activation function is a CReLU function, which is:
wherein z is a complex variable.
Preferably, in the present invention, the set of complex feature maps of each channel is divided into the first components along the spatial dimensionAnd the second tensor
Wherein,andis a complex output characteristic diagram of the k channel in the horizontal direction,the complex output characteristic diagram of the kth channel in the vertical direction, the complex output characteristic diagram of the C/r channel in the horizontal direction and the complex output characteristic diagram of the C/r channel in the vertical direction are shown.
Preferably, the present invention uses a 1 × 1 rewinding kernel to convert fhAnd f iswRestoring to the same dimension as the X to obtainAndwherein:
wherein,is a complex output characteristic diagram of the ith channel in the horizontal direction,is a complex output characteristic diagram of the ith channel in the vertical direction,is composed ofThe (o) th 1 × 1 complex convolution kernel of (a), o 1,2, ·, C/r,is a complex output characteristic diagram of the channel o in the horizontal direction,is composed ofThe (o) th 1 x 1 rewinding nucleus in (b),is a complex output characteristic diagram of the channel o in the vertical direction,second complex output bit representing the l channel in the horizontal directionSign graph, vhFor the set of the second complex output profiles for all channels,a third complex output profile, v, representing the ith channel in the vertical directionwFor the set of the third complex output profiles of all channels,
preferably, v is defined in the present inventionhAnd said vwEach element in the system is represented in a polar coordinate form, and the amplitude of the polar coordinate is constrained by a Sigmoid function, which specifically comprises the following steps:
wherein,andthe complex output characteristic graphs of the ith channel in the horizontal direction and the ith channel in the vertical direction are respectively the result of the amplitude constraint of the complex output characteristic graphs by a Sigmoid function,andthe phase of the complex output profile for the ith channel in the horizontal and vertical directions respectively,andthe magnitude of the complex output signature for the ith channel in the horizontal and vertical directions respectively,andthe phase of the complex output characteristic diagram of the ith channel in the horizontal direction and the vertical direction respectively, Sig (·) represents a Sigmoid function, and the result after constraint through Sigmoid is recorded asAnd
under a rectangular coordinate system:
preferably, in the present invention, g ishAnd said gwExpanding, generating attention weight distribution in horizontal and vertical space directions, applying the attention weight distribution on a complex input feature map of the complex coordinate attention module to obtain a complex output feature map of the complex coordinate attention module, wherein the complex output feature map of the complex coordinate attention module is yl(i,j):
Wherein x isl(i, j) is the ith row and jth column number of the ith channel complex input feature diagram.
The input and the output of the complex coordinate attention module are both complex form feature information, and the complex coordinate attention module can process the complex feature information.
The complex coordinate attention module obtains the amplitude and phase characteristics of the signal through the associated learning of a real part and an imaginary part of a complex number by using a complex convolution neural network.
According to the invention, the complex coordinate attention module focuses on the spatial information and the channel information in the horizontal and vertical directions simultaneously through the complex coordinate attention, so that the remote lazy relationship of the characteristic information is better modeled, and the characteristic representation capability of the target object is enhanced.
The invention also provides a target identification method, which comprises the following steps:
obtaining a target signal;
inputting the target information into the complex coordinate attention module-based convolutional neural network;
and the convolutional neural network outputs a target recognition result.
Preferably, the target signal is radar echo data.
One or more technical schemes provided by the invention at least have the following technical effects or advantages:
the invention integrates the advantages of CV-CNN and an attention mechanism, introduces the coordinate attention of a real number domain into a complex number domain, constructs a convolutional neural network and a target identification method based on a complex coordinate attention module, can take radar echo complex data as input data to carry out direct operation, fully utilizes amplitude and phase information, and realizes the high-precision identification of similar space cone targets with the same geometric shape and micromotion form and slightly different micromotion parameters.
When the convolutional neural network based on the complex coordinate attention module is used for target identification, preprocessing such as time-frequency analysis or distance slow time imaging is not needed, so that longer signal processing time is not needed, and the efficiency is higher; when the convolutional neural network based on the complex coordinate attention module is used for identifying the target, the target does not need to be continuously observed for a long time to obtain a complete periodic image of the target, and the efficiency is high.
The convolutional neural network based on the complex coordinate attention module can realize high-precision identification on similar space cone targets with the same micromotion form and only slightly different micromotion parameters.
According to the end-to-end similar space cone target identification method, the radar echo complex data are input and the identification result is output, so that echo signal preprocessing and phase information loss are avoided, and the time required by identification is remarkably shortened.
Drawings
The accompanying drawings, which are included to provide a further understanding of the embodiments of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the principles of the invention;
FIG. 1 is a schematic diagram of a convolutional neural network based on a complex coordinate attention module;
FIG. 2 is a schematic diagram of a plurality of coordinate attention modules;
fig. 3 is a flowchart illustrating a complex input feature map processing method.
Detailed Description
In order that the above objects, features and advantages of the present invention can be more clearly understood, a more particular description of the invention will be rendered by reference to the appended drawings. It should be noted that the embodiments of the present invention and features of the embodiments may be combined with each other without conflicting with each other.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention, however, the present invention may be practiced in other ways than those specifically described and thus the scope of the present invention is not limited by the specific embodiments disclosed below.
It is understood that the terms "a" and "an" should be interpreted as meaning that a number of one element or element is one in one embodiment, while a number of other elements is one in another embodiment, and the terms "a" and "an" should not be interpreted as limiting the number.
Example one
Referring to fig. 1, fig. 1 is a schematic structural diagram of a convolutional neural network based on a complex coordinate attention module, the convolutional neural network based on the complex coordinate attention module, and the convolutional neural network includes:
the device comprises an input layer, N basic units, a classification unit and an output layer;
the processing unit is used for mapping the complex numbers into corresponding real numbers through modular operation and performing classification and identification; the N basic units comprise first to Nth basic units, the first basic unit is connected with the input layer, the output of the first basic unit is the input of the second basic unit, the input of the Nth basic unit is the output of the (N-1) th basic unit, N is an integer larger than 1, the output of the Nth basic unit is the input of the processing unit, the output of the processing unit is the input of the classifier, and the classifier is connected with the output layer; the N basic units each include: the system comprises a first complex convolution module, a first complex batch normalization module, a first complex activation module and a first complex pooling module; wherein one of the N basic units further comprises: a plurality of coordinate attention modules; the complex coordinate attention module includes: the system comprises a complex coordinate attention embedding unit and a complex coordinate attention generating unit, wherein for each channel, the complex coordinate attention embedding unit is used for encoding a first complex input feature map of the channel along the horizontal direction and the vertical direction respectively, and generating first output feature information of the first complex input feature map after the first complex input feature map is encoded along the horizontal direction and second output feature information of the first complex input feature map after the first complex input feature map is encoded along the vertical direction;
for each channel, the complex coordinate attention generating unit is to: splicing the first output characteristic information and the second output characteristic information to generate a characteristic information splicing result of the channel; performing feature dimensionality reduction on the feature information splicing result of the channel to obtain feature information after dimensionality reduction, and activating the feature information after dimensionality reduction to obtain a first complex output feature map of the channel; splitting the first complex output profile into a first tensor and a second tensor along a spatial dimension; adjusting the dimensions of the first tensor and the second tensor to be the same as the dimensions of the first complex input feature map, and obtaining a second complex output feature map of the channel in the horizontal direction and a third complex output feature map of the channel in the vertical direction; obtaining a third tensor and a fourth tensor, wherein the third tensor is the set of the second complex output characteristic maps of all the channels, and the fourth tensor is the set of the third complex output characteristic maps of all the channels;
expressing each element in the third tensor and the fourth tensor in a polar coordinate form, constraining the amplitude of the polar coordinate by using a constraint function, respectively obtaining a fourth complex output feature map and a fifth complex output feature map in the horizontal and vertical spatial directions, expanding the fourth complex output feature map and the fifth complex output feature map to generate attention weight distribution in the horizontal and vertical spatial directions, and applying the attention weight distribution to a complex input feature map of the complex coordinate attention module to obtain a complex output feature map of the complex coordinate attention module;
wherein, the complex input characteristic diagram and the complex output characteristic diagram are both complex characteristic diagrams.
The invention integrates the advantages of CV-CNN and attention mechanism, introduces the coordinate attention of real number domain into complex number domain, and constructs complex number attention network. An attention mechanism (attention mechanism) can capture long-distance dependency relationship through global information search, automatically focus important information through weighted distribution, and ignore unimportant redundant information, which is useful for similar space cone target identification in a short observation duration. The attention mechanism undergoes the development process of spatial attention, channel attention and space-channel attention. The spatial dimension and channel dimension information play a role in improving the network identification capability, which is proved again in a newly proposed Coordinate Attention (CA) module, and the model performance is improved by embedding spatial position information into the channel Attention.
The Complex-Valued convolutional neural network (CV-CNN) can directly process echo Complex data, fully utilize amplitude and phase information, avoid echo preprocessing and reduce recognition time.
The CV-CANet is built based on a CV-CA module. The CV-CANet is an end-to-end complex convolutional neural network, and the architecture of the network is shown in FIG. 1. Each layer of basic unit of the network consists of 4 basic modules of complex convolution, complex batch normalization, complex activation and complex pooling, wherein a CV-CA module is embedded in the sixth layer. The number of convolution kernels of the first layer to the sixth layer is 64, 128, 256 and 256, respectively, the size of each convolution kernel (convolution kernel) is 1 × 3, the size of the sampling window of the pooling layer (pooling layer) is 1 × 2, the sliding step (stride) of all convolution layers is 1, and the padding (padding) is 1. The last two layers of the network replace the traditional full connection with full convolution to reduce the model parameters. The output result of the final layer of full convolution is complex number, and the class label of the target is real number. Therefore, the complex numbers are mapped into corresponding real numbers through modular operation and then sent to a Softmax classifier for classification and identification. The loss function is a cross-entropy loss function. Adaptive motion Estimation (Adam) acts as an optimizer for updating network weights and bias terms.
The number of layers of the complex convolutional neural network, the number of the basic units and the concrete CV-CA module embedded in any basic unit are not limited, and the number of layers and the number of the basic units can be flexibly adjusted according to actual requirements.
According to the similar space cone target end-to-end identification method based on CV-CANet, the identification result of the similar space cone target is directly obtained by inputting radar echo data, so that the problems of complex echo signal preprocessing and phase information loss are solved. In order to directly process radar echo complex signals, the invention provides a CV-CA module and constructs CV-CANet based on the module. The invention introduces a coordinate attention mechanism into a complex field, deduces and establishes basic structures of direction-related complex feature information aggregation, direction-related complex feature diagram splitting, complex coordinate attention automatic allocation and the like. And effective identification can be realized for similar space cone targets with the same micromotion form and only slightly different micromotion parameters.
The method is usually carried out under the condition that the observation time length does not exceed a half period, in practice, a radar cannot observe one target for a long time, or noise exists in data, or data is lost, so that the target is expected to be better identified by less data, and the real-time property is ensured.
Example two
Referring to fig. 2, fig. 2 is a schematic diagram illustrating a composition of a plurality of coordinate attention modules, in this embodiment, the plurality of coordinate attention modules include: the system comprises a complex coordinate attention embedding unit and a complex coordinate attention generating unit, wherein for each channel, the complex coordinate attention embedding unit is used for encoding a first complex input feature map of the channel along the horizontal direction and the vertical direction respectively, and generating first output feature information of the first complex input feature map after the first complex input feature map is encoded along the horizontal direction and second output feature information of the first complex input feature map after the first complex input feature map is encoded along the vertical direction;
for each channel, the complex coordinate attention generating unit is to: splicing the first output characteristic information and the second output characteristic information to generate a characteristic information splicing result of the channel; performing feature dimensionality reduction on the feature information splicing result of the channel to obtain feature information after dimensionality reduction, and activating the feature information after dimensionality reduction to obtain a first complex output feature map of the channel; splitting the first complex output profile into a first tensor and a second tensor along a spatial dimension; adjusting the dimensions of the first tensor and the second tensor to be the same as the dimensions of the first complex input feature map, and obtaining a second complex output feature map of the channel in the horizontal direction and a third complex output feature map of the channel in the vertical direction; obtaining a third tensor and a fourth tensor, wherein the third tensor is the set of the second complex output characteristic maps of all the channels, and the fourth tensor is the set of the third complex output characteristic maps of all the channels;
expressing each element in the third tensor and the fourth tensor in a polar coordinate form, constraining the amplitude of the polar coordinate by using a constraint function, respectively obtaining a fourth complex output feature map and a fifth complex output feature map in the horizontal and vertical spatial directions, expanding the fourth complex output feature map and the fifth complex output feature map to generate attention weight distribution in the horizontal and vertical spatial directions, and applying the attention weight distribution to a complex input feature map of the complex coordinate attention module to obtain a complex output feature map of the complex coordinate attention module;
wherein, the complex input characteristic diagram and the complex output characteristic diagram are both complex characteristic diagrams.
The prior art CV-CNN, which simply separates the real and imaginary parts of the complex number or uses a real convolution kernel, does not take advantage of the complex convolution kernel. Therefore, the invention follows the law of complex number calculation, and carries out detailed formula derivation and constructs a complex number coordinate attention module (CV-CA module) according to a complex number network basic unit and a Real number coordinate attention (CV-CA, RV-CA) module.
The CV-CA module proposed by the present invention includes a Complex-Coordinate Attention Information Embedding (CVCIE) and a Complex-Coordinate Attention authorization Generation (CVCAG) unit.
In practical application, the input of the CV-CA module may be any complex-form feature information, and the present invention is described by taking a radar echo signal as an example, but the input feature information in the present invention is not limited to the feature information of the radar echo signal, H is 1 for narrowband radar echo data, and the value of H in practical application may be determined according to practical situations, and the present invention is not limited specifically.
The echo signal of the radar measurement can be expressed as:
St(n)=Sth(n)+ν(n)
wherein S isthAnd (n) is a theoretical radar echo signal, v (n) represents independent and equally distributed Gaussian white noise generated by a radar receiver, and n represents a pulse sequence number.
Is provided withIs a complex input profile in whichA complex input profile representing the p-th channel,is a complex output profile in whichAnd (3) a complex output characteristic diagram of the p channel, wherein the dimension of the complex output characteristic diagram is the same as that of the X.
Global pooling is usually used for encoding global spatial information in channel attention, but it compresses the global spatial information into a channel descriptor, so that it is difficult to maintain the position information, which is particularly important for capturing the spatial structure. Therefore, in the coordinate attention module, the operation of decomposing the global pooling into two one-dimensional feature codes is extended to the complex field, the complex feature maps X of each channel are encoded along two horizontal and vertical directions (direction-correlation is abbreviated as horizontal and vertical directions), and a direction-correlation complex feature map is generated, so as to integrate the features in two spatial directions, respectively, and this operation is described mathematically as:
wherein j represents an imaginary unit,representing the real part of the complex number,representing the imaginary part of the complex number, h being the horizontal pixel index of the input feature map, xp(h, j) is the value of the h row and j column of the p channel of the complex input characteristic diagram, i is the pixel index of the vertical direction of the input characteristic diagram, xp(i, w) is the value of the ith row and the w column of the ith channel of the complex input characteristic diagram. The complex characteristic maps of each channel of X are transformed to respectively obtain two complex tensors,wherein Wherein
The above CVCIE outputs accurate spatial location information aggregated under the global receptive field. Based on the CVCIE encoding results, the CV-CA module designs a second transformation, referred to as CVCAG. The CVCAG transform includes three steps, respectively: (1) direction-related feature information aggregation, (2) direction-related complex feature map splitting, and (3) complex coordinate attention automatic assignment.
(1) Direction-related complex feature information aggregation
And (4) performing complex splicing. And splicing the results obtained by the formula (1) and the formula (2). Is provided withFor the stitched result, each tensor in M is represented as:
wherein [, ] represents a stitching operation.
And (5) reducing the dimension of the feature. The 1 x 1 convolution kernel is used for reducing the dimension of the characteristic channel, reducing the parameter quantity and realizing the information interaction and integration of the cross-channel. Is provided withA 1 x 1 rewinding kernel shared for the layer, wherein,denotes the k-th complex convolution kernel, k ═ 1,2, …, C/r,in the representationThe (C) th 1 x 1 rewinding and stacking nucleus,in the representationQ is 1,2, …, C, r is a scaling coefficient for controlling the number of channels of the convolution output feature map (r is 18 in the present invention, where r may take other values in practical applications, and the present embodiment is not specifically limited), s is a step size of convolution operation, and the kth feature map of the convolution output is vk(i,j) Wherein:
fk(i,j)=σ(vk(i,j)) (5)
wherein m isqFor the qth tensor in M,mq(i · s, j · s) are the values of the i · s row and j · s column of the q th tensor after feature information stitching, vk(i, j) represents a complex output profile of the k-th channel that is not activated,a complex output characteristic diagram representing the k-th channel, and the set of complex characteristic diagrams of each channel is denoted asfC/rFor the complex output characteristic diagram of the C/r channel, sigma (-) represents the complex activation function; the complex activation function is a CReLU function, which is:
wherein z is a complex variable.
(2) Direction-dependent complex feature map splitting
The complex signature is split. Splitting f into two independent tensors along the spatial dimensionAndnamely:
and (5) feature dimension increasing. Using a 1X 1 rewinding kernel to fhAnd fwReverting to the same dimension as the input signature X. Setting the rewinding product kernel of 1 x 1 in the convolution operation in the horizontal direction asWhereinRepresents the l (l ═ 1,2, …, C) multiple winding core,to representThe (o) th (o ═ 1,2, …, C/r)1 × 1 rewinding and stacking nucleus. In the same way, the method for preparing the composite material,is a 1 multiplied by 1 rewinding product kernel in the convolution operation in the vertical direction, wherein,denotes the l (1, 2., C) th complex convolution kernel,to representThe qth (l ═ 1,2, …, C)1 × 1 rewinding core in (a), then:
using a 1 x 1 rewinding kernel to wrap said fhAnd saidfwRestoring to the same dimension as the X to obtainAndwherein:
wherein,is a complex output characteristic diagram of the ith channel in the horizontal direction,is a complex output characteristic diagram of the ith channel in the vertical direction,is composed ofThe (o) th 1 × 1 complex convolution kernel in (a), o ═ 1,2, …, C/r,is a complex output characteristic diagram of the channel o in the horizontal direction,is composed ofThe (o) th 1 x 1 rewinding nucleus in (b),is a complex output characteristic diagram of the channel o in the vertical direction,second complex output profile, v, representing the ith channel in the horizontal directionhFor the set of the second complex output profiles for all channels,a third complex output profile, v, representing the ith channel in the vertical directionwFor the set of the third complex output profiles of all channels,
(3) automatic complex coordinate attention allocation
Direction-dependent complex attention weight coefficients are calculated. Tensor v of complex eigen-maphAnd vwEach element (complex value) in the system is written into a polar coordinate form, then the amplitude of the polar coordinate is constrained by adopting a Sigmoid function, and the amplitude is limited within a value range of 0-1, namely:
wherein,andthe complex output characteristic graphs of the ith channel in the horizontal direction and the ith channel in the vertical direction are respectively the result of the amplitude constraint of the complex output characteristic graphs by a Sigmoid function,andthe phase of the complex output profile for the ith channel in the horizontal and vertical directions respectively,andthe magnitude of the complex output signature for the ith channel in the horizontal and vertical directions respectively,andthe phase of the complex output characteristic diagram of the ith channel in the horizontal direction and the vertical direction respectively, Sig (·) represents a Sigmoid function and is used for converting the final amplitude into a numerical value between 0 and 1, and the result after being constrained by the Sigmoid is recorded as a numerical valueAndof the kind described aboveThe phase is not affected by merely converting the amplitude of the polar coordinates to the 0-1 range, i.e. the phase information is preserved.
Since the original image is in the rectangular coordinate system, the expressions of equation (11) and equation (14) are converted to the rectangular coordinate system:
under a rectangular coordinate system:
complex coordinate attention is automatically assigned. G output for horizontal and vertical spatial directionshAnd gwAnd expanding, generating attention weight distribution in each space direction, and acting on the complex input feature diagram to realize automatic distribution of complex coordinate attention. The output of the complex coordinate attention module is obtained as:
wherein x isl(i, j) is the ith row and jth column number of the ith channel complex input feature diagram.
According to the above CV-CA construction process, on one hand, the CV-CA obtains amplitude and phase characteristics of a target signal, such as a radar echo signal, through complex real part and imaginary part associated learning by using a complex convolutional neural network; on the other hand, spatial information and channel information in the horizontal direction and the vertical direction are concerned at the same time through the attention of complex coordinates, remote lazy relation of characteristic information is better modeled, and the characteristic characterization capability of the target object is enhanced.
The CV-CA module provided by the invention comprises two parts, wherein the first part is complex coordinate attention embedding, and the second part is complex coordinate attention generation. The physical significance of each part is explained in detail below.
The first part is complex coordinate attention embedding. In the field of computer vision, the position information on the feature map has an important influence on acquiring the structural features of the space. Since the targets to be resolved by the invention are very similar space cone targets, the invention considers that the space structure information is beneficial to the resolution and identification of the targets. Therefore, in order for the proposed complex coordinate attention module to retain the position information and further capture the long distance dependency on the space by using the position information, the present invention decomposes the global pooling in CNN into pooling operation in the horizontal direction and pooling operation in the vertical direction.
The second part is complex coordinate attention generation, which is done in three sub-steps. For this section, the general design principle of the present invention has three points: 1) the modules should be as simple and light-weight as possible. 2) The module should make full use of the spatial location information obtained in the first part. 3) The module should take into account the interrelationship between the channels in order to take advantage of the channel attention.
Direction-related feature information aggregation. The first part has obtained spatial position information in both the horizontal and vertical directions. Under the principle that the designed module is as simple as possible and the parameter quantity is as small as possible, the invention firstly splices the spatial position information in the horizontal direction and the vertical direction, and aims to simultaneously retain the information in the two directions. Then, the invention uses 1 × 1 convolution kernel to convolute the splicing result for dimension reduction. By the design, the characteristic information among the channels is considered, and the parameter quantity is reduced.
The direction-related complex feature information is split. The present invention expects that the horizontal weight and the vertical weight should be applied to the horizontal direction and the vertical direction of the input feature map, respectively, and the number of channels of the weights should be consistent with the number of channels of the input feature map. Therefore, in the first sub-step, the weights of the spatial position information and the channel information are taken into consideration, and the weight in the horizontal direction and the weight in the vertical direction of the channel information are taken into consideration. Then, the two directions are subjected to dimensionality raising by using a convolution kernel of 1 × 1 respectively.
A plurality of attentions are automatically assigned. After the above-mentioned operation steps, the present invention obtains complex weights considering both spatial position information and channel information, and on one hand, the present invention is expected to retain phase information of the complex weights, and on the other hand, the magnitude of the weights is limited to the range of 0-1. Finally, the attention weight is respectively applied to each element of the input feature vector and each channel, so that the complex coordinate attention provided by the invention is realized. The method can weight each channel and pay attention to the important channel; spatial information is also considered; regions that facilitate target recognition are focused.
In addition, the CV-CA module in the embodiment can obtain the optimal feature recognition capability with the increase of fewer parameters under the condition of ensuring the operation efficiency of the model, and can improve the recognition capability of the model and reduce the target misjudgment probability.
EXAMPLE III
A third embodiment of the present invention provides a complex input feature map processing method, please refer to fig. 3, where fig. 3 is a schematic flow chart of the complex input feature map processing method, where the method includes:
obtaining a complex input characteristic diagram to be processed;
coding the first complex input characteristic diagram of the channel along the horizontal direction and the vertical direction respectively, and generating first output characteristic information of the first complex input characteristic diagram coded along the horizontal direction and second output characteristic information coded along the vertical direction in the channel respectively;
for each channel, splicing the first output characteristic information and the second output characteristic information to generate a characteristic information splicing result of the channel; performing feature dimensionality reduction on the feature information splicing result of the channel to obtain feature information after dimensionality reduction, and activating the feature information after dimensionality reduction to obtain a first complex output feature map of the channel; splitting the first complex output profile into a first tensor and a second tensor along a spatial dimension; adjusting the dimensions of the first tensor and the second tensor to be the same as the dimensions of the first complex input feature map, and obtaining a second complex output feature map of the channel in the horizontal direction and a third complex output feature map of the channel in the vertical direction; obtaining a third tensor and a fourth tensor, wherein the third tensor is the set of the second complex output characteristic maps of all the channels, and the fourth tensor is the set of the third complex output characteristic maps of all the channels;
expressing each element in the third tensor and the fourth tensor in a polar coordinate form, constraining the amplitude of the polar coordinate by using a constraint function, respectively obtaining a fourth complex output feature diagram and a fifth complex output feature diagram in the horizontal and vertical spatial directions, expanding the fourth complex output feature diagram and the fifth complex output feature diagram to generate attention weight distribution in the horizontal and vertical spatial directions, and applying the attention weight distribution to the to-be-processed complex input feature diagram to obtain a processed complex output feature diagram;
wherein, the complex input characteristic diagram and the complex output characteristic diagram are both complex characteristic diagrams.
The method can be used for processing complex output characteristic information, amplitude and phase characteristics of signals are obtained through complex convolution neural network through complex real part and imaginary part correlation learning, spatial information and channel information in the horizontal direction and the vertical direction are concerned by the method through complex coordinate attention, remote lazy relation of the characteristic information is better modeled, and characteristic characterization capability of a target object is enhanced.
While preferred embodiments of the present invention have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all such alterations and modifications as fall within the scope of the invention.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.
Claims (10)
1. A convolutional neural network based on a complex coordinate attention module, said convolutional neural network comprising:
the device comprises an input layer, N basic units, a classification unit and an output layer;
the processing unit is used for mapping the complex numbers into corresponding real numbers through modular operation and performing classification and identification; the N basic units comprise first to Nth basic units, the first basic unit is connected with the input layer, the output of the first basic unit is the input of the second basic unit, the input of the Nth basic unit is the output of the (N-1) th basic unit, N is an integer larger than 1, the output of the Nth basic unit is the input of the processing unit, the output of the processing unit is the input of the classifier, and the classifier is connected with the output layer; the N basic units each include: the system comprises a first complex convolution module, a first complex batch normalization module, a first complex activation module and a first complex pooling module; wherein one of the N basic units further comprises: a plurality of coordinate attention modules; the complex coordinate attention module includes: the system comprises a complex coordinate attention embedding unit and a complex coordinate attention generating unit, wherein for each channel, the complex coordinate attention embedding unit is used for encoding a first complex input feature map of the channel along the horizontal direction and the vertical direction respectively, and generating first output feature information of the first complex input feature map after the first complex input feature map is encoded along the horizontal direction and second output feature information of the first complex input feature map after the first complex input feature map is encoded along the vertical direction;
for each channel, the complex coordinate attention generating unit is to: splicing the first output characteristic information and the second output characteristic information to generate a characteristic information splicing result of the channel; performing feature dimensionality reduction on the feature information splicing result of the channel to obtain feature information after dimensionality reduction, and activating the feature information after dimensionality reduction to obtain a first complex output feature map of the channel; splitting the first complex output profile into a first tensor and a second tensor along a spatial dimension; adjusting the dimensions of the first tensor and the second tensor to be the same as the dimensions of the first complex input feature map, and obtaining a second complex output feature map of the channel in the horizontal direction and a third complex output feature map of the channel in the vertical direction; obtaining a third tensor and a fourth tensor, wherein the third tensor is the set of the second complex output characteristic maps of all the channels, and the fourth tensor is the set of the third complex output characteristic maps of all the channels;
expressing each element in the third tensor and the fourth tensor in a polar coordinate form, constraining the amplitude of the polar coordinate by using a constraint function, respectively obtaining a fourth complex output feature map and a fifth complex output feature map in the horizontal and vertical spatial directions, expanding the fourth complex output feature map and the fifth complex output feature map to generate attention weight distribution in the horizontal and vertical spatial directions, and applying the attention weight distribution to a complex input feature map of the complex coordinate attention module to obtain a complex output feature map of the complex coordinate attention module;
wherein, the complex input characteristic diagram and the complex output characteristic diagram are both complex characteristic diagrams.
2. The complex coordinate attention module based convolutional neural network of claim 1, wherein in the basic unit not comprising the complex coordinate attention module, the output of the complex convolution module in the basic unit is the input of the complex batch normalization module, the output of the complex batch normalization module is the input of the complex activation module, and the output of the complex activation module is the input of the complex pooling module.
3. The complex coordinate attention module based convolutional neural network of claim 1, wherein in the basic unit comprising the complex coordinate attention module, the output of the complex convolution module in the basic unit is the input of the complex batch normalization module, the output of the complex batch normalization module is the input of the complex coordinate attention module, the output of the complex coordinate attention module is the input of the complex activation module, and the output of the complex activation module is the input of the complex pooling module.
4. The complex coordinate attention module based convolutional neural network of claim 1, wherein said classification unit comprises:
the second complex convolution module, the second complex batch normalization module, the second complex activation module, the third complex convolution module and the classifier; the output of the second complex convolution module is the input of the second complex batch normalization module, the output of the second complex batch normalization module is the input of the second complex activation module, the output of the second complex activation module is the input of the third complex convolution module, and the output of the third complex convolution module is the input of the classifier.
5. The complex coordinate attention module based convolutional neural network of claim 1, comprising first through sixth base units.
6. The complex coordinate attention module based convolutional neural network of claim 5, wherein a sixth base unit comprises the complex coordinate attention module.
7. The complex coordinate attention module based convolutional neural network of claim 1, wherein an optimizer is provided in the convolutional neural network for updating network weights and bias terms.
8. The complex coordinate attention module based convolutional neural network of claim 1, wherein the complex input feature map is a complex input feature map of a spatial target recognition signal, and the complex output feature map is a complex output feature map of the spatial target recognition signal.
9. An object recognition method, characterized in that the method comprises:
obtaining a target signal;
inputting the target information into the complex coordinate attention module based convolutional neural network of any one of claims 1-5;
and the convolutional neural network outputs a target recognition result.
10. The method of claim 9, wherein the target signal is radar echo data.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202110858271.0A CN113537120B (en) | 2021-07-28 | 2021-07-28 | Complex convolution neural network target identification method based on complex coordinate attention |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202110858271.0A CN113537120B (en) | 2021-07-28 | 2021-07-28 | Complex convolution neural network target identification method based on complex coordinate attention |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN113537120A true CN113537120A (en) | 2021-10-22 |
| CN113537120B CN113537120B (en) | 2023-04-07 |
Family
ID=78121256
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202110858271.0A Active CN113537120B (en) | 2021-07-28 | 2021-07-28 | Complex convolution neural network target identification method based on complex coordinate attention |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN113537120B (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114972280A (en) * | 2022-06-07 | 2022-08-30 | 重庆大学 | Fine Coordinate Attention Module and Its Application in Surface Defect Detection |
| CN116050474A (en) * | 2022-12-29 | 2023-05-02 | 上海天数智芯半导体有限公司 | A convolution calculation method, SOC chip, electronic equipment and storage medium |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9373059B1 (en) * | 2014-05-05 | 2016-06-21 | Atomwise Inc. | Systems and methods for applying a convolutional network to spatial data |
| CN111340186A (en) * | 2020-02-17 | 2020-06-26 | 之江实验室 | Compressed representation learning method based on tensor decomposition |
| CN112329538A (en) * | 2020-10-10 | 2021-02-05 | 杭州电子科技大学 | A Target Classification Method Based on Microwave Vision |
| CN112965062A (en) * | 2021-02-09 | 2021-06-15 | 西安电子科技大学 | Radar range profile target identification method based on LSTM-DAM network |
-
2021
- 2021-07-28 CN CN202110858271.0A patent/CN113537120B/en active Active
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9373059B1 (en) * | 2014-05-05 | 2016-06-21 | Atomwise Inc. | Systems and methods for applying a convolutional network to spatial data |
| CN111340186A (en) * | 2020-02-17 | 2020-06-26 | 之江实验室 | Compressed representation learning method based on tensor decomposition |
| CN112329538A (en) * | 2020-10-10 | 2021-02-05 | 杭州电子科技大学 | A Target Classification Method Based on Microwave Vision |
| CN112965062A (en) * | 2021-02-09 | 2021-06-15 | 西安电子科技大学 | Radar range profile target identification method based on LSTM-DAM network |
Non-Patent Citations (3)
| Title |
|---|
| YAXIN LI ET.AL: "Multi-mode Fusion and Classification Method for Space Targets Based on Convolutional Neural Network" * |
| 周剑: "面向复杂场景的目标分类方法" * |
| 王光光: "基于深度学习的PolSAR分类和视频行为识别" * |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114972280A (en) * | 2022-06-07 | 2022-08-30 | 重庆大学 | Fine Coordinate Attention Module and Its Application in Surface Defect Detection |
| CN114972280B (en) * | 2022-06-07 | 2023-11-17 | 重庆大学 | Fine coordinate attention module and application thereof in surface defect detection |
| CN116050474A (en) * | 2022-12-29 | 2023-05-02 | 上海天数智芯半导体有限公司 | A convolution calculation method, SOC chip, electronic equipment and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| CN113537120B (en) | 2023-04-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Wang et al. | Review of image fusion based on pulse-coupled neural network | |
| Wang et al. | Optimization-based post-training quantization with bit-split and stitching | |
| CN112818764A (en) | Low-resolution image facial expression recognition method based on feature reconstruction model | |
| CN113537120B (en) | Complex convolution neural network target identification method based on complex coordinate attention | |
| CN113569735B (en) | Complex input feature graph processing method and system based on complex coordinate attention module | |
| CN118230758A (en) | Underwater sound target identification method and system based on combination of encoder and convolution | |
| CN114724245A (en) | CSI-based incremental learning human body action identification method | |
| CN116612364B (en) | SAR image target generation method based on information maximization generation countermeasure network | |
| El-Bana et al. | Evaluating the potential of wavelet pooling on improving the data efficiency of light-weight cnns | |
| CN117725466A (en) | Unmanned aerial vehicle radio frequency signal identification method based on pulse neural network | |
| CN116888605A (en) | Calculation methods, training methods and devices of neural network models | |
| Liu et al. | Radar-based human motion recognition using semisupervised triple-GAN | |
| CN116012877A (en) | An Attention Mechanism Based Human Pose Recognition Method Based on Millimeter Wave Radar 3D Point Cloud | |
| CN117994672A (en) | Remote sensing image on-board target identification method | |
| Korti et al. | Advanced Human Activity Recognition through Data Augmentation and Feature Concatenation of Micro-Doppler Signatures | |
| Zhang et al. | No-reference image quality assessment using independent component analysis and convolutional neural network | |
| CN119049479B (en) | Voice spoofing attack detection method and device | |
| Jiang et al. | Gesture Recognition Using sEMG Based on Multi-Scale Fusion Convolution and Channel Attention | |
| CN110781764A (en) | Intelligent microwave sign language identification method | |
| CN120612561B (en) | Small target visual detection method under geological exploration scene | |
| CN116058852B (en) | Classification system, method, electronic device and storage medium for MI-EEG signals | |
| Liu et al. | KENet: Distilling Convolutional Networks via Knowledge Enhancement | |
| Chen et al. | Tensor-Based Chaotic Convolutional Neural Network for Remote Sensing Data Classification | |
| CN118311530A (en) | Multi-channel radar human behavior recognition method and system based on scattering separation | |
| Dave et al. | Insect recognition: A visionary step towards smart agriculture |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |