US20210215481A1 - Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network - Google Patents
Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network Download PDFInfo
- Publication number
- US20210215481A1 US20210215481A1 US16/652,346 US201916652346A US2021215481A1 US 20210215481 A1 US20210215481 A1 US 20210215481A1 US 201916652346 A US201916652346 A US 201916652346A US 2021215481 A1 US2021215481 A1 US 2021215481A1
- Authority
- US
- United States
- Prior art keywords
- antenna
- target
- image
- downtilt angle
- semantic segmentation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B21/00—Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant
- G01B21/22—Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant for measuring angles or tapers; for testing the alignment of axes
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G06N3/0454—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/17—Terrestrial scenes taken from planes or by drones
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/176—Urban or other man-made structures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/70—Labelling scene content, e.g. deriving syntactic or semantic representations
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Definitions
- the present disclosure relates to the field of mobile communication, and in particular, to a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network.
- an azimuth angle and a downtilt angle of an antenna affect the coverage of signals and the interference between the signals, so the antenna needs to be timely calculated and adjusted strictly to improve the quality of network signals.
- the first one is climbing to an antenna base station manually and using a measuring instrument (a compass, a slope meter, or the like) for measurement; and the second one is installing an angle sensor on the antenna to return data.
- the antenna is susceptible to wind, snow and other factors, resulting in a change in the downtilt angle, so it needs to be measured regularly.
- the first method as the base station is high and the number of the antennas is larger, the manual safety hazard and workload are larger, and the practicability is low.
- the installation time is long, and the antenna models are different, so the installation cost of the instruments is high and the practicability is not high. Both the two methods consume a lot of manpower and material resources and are not suitable for large-scale measurement today.
- the present disclosure aims at providing a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network.
- the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
- An antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network including:
- base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
- predicting a target bounding box a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
- target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified;
- the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
- the collecting image data includes:
- the predicting a target bounding box includes:
- each grid includes five predictors (x, y, w, h, confidence) and a c class
- the output of the network is of a size of S*S*(5*B+C)
- B is the number of the bounding boxes in each grid
- C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:
- performing target recognition and semantic segmentation includes:
- antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:
- object)*Pr(object)*Pr(object)*IOU pred truth Pr(object)*IOU pred truth (6)
- object) is an object class probability
- a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
- dilated convolutional network After the target image is input, it first goes through feature extraction by a dilated convolutional network; and after a feature image is input, dilated convolution is calculated:
- an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal;
- pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
- the calculating an antenna downtilt angle includes:
- the downtilt angle of the base station antenna being an angle ⁇ between the base station antenna and a vertical plane:
- the present disclosure has the following beneficial effects: the present disclosure adopts an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network.
- the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
- FIG. 1 is a schematic diagram of a downtilt angle of a base station antenna
- FIG. 2 is a flowchart of an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 3 is a schematic diagram of border prediction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 4 is a schematic diagram of a network structure for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 5 is a schematic diagram of a bottleneck block for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 6 is a schematic diagram of standard convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 7 is a schematic diagram of high-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 8 is a schematic diagram of one-dimensional low-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
- FIG. 9 is a schematic diagram of dilated convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure.
- FIG. 10 is a view of a random field for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure.
- an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network including:
- base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
- predicting a target bounding box a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
- target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified;
- the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
- the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
- the step of collecting image data includes:
- the step of predicting a target bounding box includes:
- each grid includes 5 predictors (x, y, w, h, confidence) and a c class
- the output of the network is of a size of S*S*(5*B+C)
- B is the number of the bounding boxes in each grid
- C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:
- multi-scale prediction is used. There is no need to fix the size of an input image, so different step sizes can be used to detect feature maps of different sizes.
- Three different detection layers are used to detect the antenna image for the target antenna, and different detection layers are realized by controlling the step size.
- the first detection layer is down-sampled with a step size of 32 to reduce the feature dimension. In order to connect with the previous identical feature graph, the layer is up-sampled, and a high resolution can be obtained at this point.
- the second detection layer with a step size of 16 is used, and the remaining feature processing is consistent with that of the first layer.
- the step size is set to 8 in the third layer, feature prediction is performed thereon, and finally, the detection accuracy of the target antenna is greater.
- step of performing target recognition and semantic segmentation includes:
- antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:
- object)*Pr(object)*Pr(object)*IOU pred truth Pr(object)*IOU pred truth (6)
- object) is an object class probability
- a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
- layers 75-105 are feature interaction layers of a neural convolutional network, which can be divided into three scales; local feature interaction is realized by means of convolution kernels, and its network structure is as shown in FIG. 4 .
- Semantic image segmentation is performed on the antenna image by using a deep convolutional network, and the pixel points of the target image and the background are classified.
- the target image After the target image is input, it first goes through feature extraction by a dilated convolutional network. Since the measured boundary precision is not high enough, the pixel of the target image cannot be well separated from the background pixel, and the pixel classification of the image boundary can be improved by combining a fully connected conditional random field, so that the segmentation effect can be better.
- FIG. 6 is a schematic diagram of one-dimensional low-resolution feature map extraction.
- FIG. 9 is a schematic diagram of dilated convolution.
- the convolution with holes can effectively increase a receptive field of a filter and integrate multi-scale information.
- dilated convolution is calculated:
- an output corresponding to each position i is y
- w is a filter
- the detour rate r is a step size for sampling the input signal.
- the receptive field of the filter can be improved, and the convolution with holes can enlarge the convolution kernel.
- a residual module of multi-scale feature learning is used in the feature network extraction, while the bottleneck block is used in the present disclosure.
- each convolution is processed by normalization and processed by an activation function.
- contextual information of the context is enriched, and the bottleneck block is as shown in FIG. 5 .
- pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
- FIG. 10 A view of a random field is as shown in FIG. 10 .
- Each circle represents a pixel point
- xi white circle
- yi black circle
- xi the classification of the labeled pixel points is determined by the reference value yi.
- I) is an energy function
- An image function output through a dilated convolutional network is a unary potential function:
- a binary potential function is
- the function the relationship between pixels, and will assign the same symbols to the same prime points.
- the unary potential function extracts feature vectors of a node in different feature maps, and the binary function connects the nodes extracted by the unitary potential function to learn its edges. All the nodes are connected to form a conditional random field of a fully connected layer, and an image finally output by the function is more accurate.
- step of calculating an antenna downtilt angle includes:
- the downtilt angle of the base station antenna being an angle ⁇ between the base station antenna and a vertical plane:
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Computational Linguistics (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Remote Sensing (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
Description
- This application is a national stage application under 35 U.S.C. 371 of PCT Application No. PCT/CN2019/076718, filed on 1 Mar. 2019, which PCT application claimed the benefit of Chinese Patent Application No. 2018113384154, filed on 9 Nov. 2018, the entire disclosure of each of which are hereby incorporated herein by reference.
- The present disclosure relates to the field of mobile communication, and in particular, to a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network.
- Nowadays, in the era of network information, the quality of mobile communication networks is extremely important. In GSM-R construction and planning, as shown in
FIG. 1 , an azimuth angle and a downtilt angle of an antenna affect the coverage of signals and the interference between the signals, so the antenna needs to be timely calculated and adjusted strictly to improve the quality of network signals. - There are two traditional methods to measure the antenna downtilt angle: the first one is climbing to an antenna base station manually and using a measuring instrument (a compass, a slope meter, or the like) for measurement; and the second one is installing an angle sensor on the antenna to return data. The antenna is susceptible to wind, snow and other factors, resulting in a change in the downtilt angle, so it needs to be measured regularly. For the first method, as the base station is high and the number of the antennas is larger, the manual safety hazard and workload are larger, and the practicability is low. For the second method, the installation time is long, and the antenna models are different, so the installation cost of the instruments is high and the practicability is not high. Both the two methods consume a lot of manpower and material resources and are not suitable for large-scale measurement today.
- To solve the above problems, the present disclosure aims at providing a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network. The method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
- The technical scheme adopted by the present disclosure to solve the problems is as follows:
- An antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network, including:
- collecting image data: base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
- predicting a target bounding box: a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
- performing target recognition and semantic segmentation: target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and
- calculating an antenna downtilt angle: the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
- Further, the collecting image data includes:
- locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set.
- Further, the predicting a target bounding box includes:
- positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein 4 coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy), box heights of the bounding boxes are px, py respectively, and the network predicts their values as:
-
b x=σ(t x)+c x (1) -
b y=σ(t y)+c y (2) -
bw=pwetw (3) -
bw=phetw (4) - where the input antenna image is divided into N*N grids, each grid includes five predictors (x, y, w, h, confidence) and a c class, and the output of the network is of a size of S*S*(5*B+C); B is the number of the bounding boxes in each grid, C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:
-
confidence=Pr(object)*IOUprd truth (5) - where a threshold is set to 0.5 when Pr(Object)=1; the target antenna falls in the center of the grid, that is, the bounding box currently predicted coincides with an actual background box object better than before; if the predicted bounding box is not the best currently, the bounding box is not predicted when the threshold is smaller than 0.5, and it is determined that the target antenna does not fall into the grid.
- Further, the performing target recognition and semantic segmentation includes:
- performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:
-
Pr(Classi|object)*Pr(object)*Pr(object)*IOUpred truth=Pr(object)*IOUpred truth (6) - where Pr(Classi|object) is an object class probability;
- then applying the activation function by logistic regression:
-
- a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
- then performing semantic image segmentation on the antenna image by using a deep convolutional network, and classifying the pixel points of the target image and the background:
- after the target image is input, it first goes through feature extraction by a dilated convolutional network; and after a feature image is input, dilated convolution is calculated:
-
y[i]=Σ k x[i+r*k]*w[k] (8) - for a two-dimensional signal, an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal;
- after the input image is processed by the convolutional network for output, pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
- Further, the calculating an antenna downtilt angle includes:
- obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane:
-
- The present disclosure has the following beneficial effects: the present disclosure adopts an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network. The method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
- The present disclosure is further described below with reference to the accompanying drawings and examples.
-
FIG. 1 is a schematic diagram of a downtilt angle of a base station antenna; -
FIG. 2 is a flowchart of an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 3 is a schematic diagram of border prediction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 4 is a schematic diagram of a network structure for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 5 is a schematic diagram of a bottleneck block for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 6 is a schematic diagram of standard convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 7 is a schematic diagram of high-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 8 is a schematic diagram of one-dimensional low-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; -
FIG. 9 is a schematic diagram of dilated convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; and -
FIG. 10 is a view of a random field for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure. - Referring to
FIG. 2 , an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network is provided in an embodiment of the present disclosure, including: - collecting image data: base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
- predicting a target bounding box: a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
- performing target recognition and semantic segmentation: target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and
- calculating an antenna downtilt angle: the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
- In the embodiment, the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
- Further, the step of collecting image data includes:
- locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set.
- Further, the step of predicting a target bounding box includes:
- positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein 4 coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy), box heights of the bounding boxes are px, py respectively, box prediction is as shown in
FIG. 3 , and the network predicts their values as: -
b x=σ(t x)+c x (1) -
b y=σ(t y)+c y (2) -
bw=pwetw (3) -
bw=phetw (4) - where the input antenna image is divided into N*N grids, each grid includes 5 predictors (x, y, w, h, confidence) and a c class, and the output of the network is of a size of S*S*(5*B+C); B is the number of the bounding boxes in each grid, C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:
-
confidence=Pr(object)*IOUprd truth (5) - where a threshold is set to 0.5 when Pr(Object)=1; the target antenna falls in the center of the grid, that is, the bounding box currently predicted coincides with an actual background box object better than before; if the predicted bounding box is not the best currently, the bounding box is not predicted when the threshold is smaller than 0.5, and it is determined that the target antenna does not fall into the grid.
- In the accuracy of a target, multi-scale prediction is used. There is no need to fix the size of an input image, so different step sizes can be used to detect feature maps of different sizes. Three different detection layers are used to detect the antenna image for the target antenna, and different detection layers are realized by controlling the step size. The first detection layer is down-sampled with a step size of 32 to reduce the feature dimension. In order to connect with the previous identical feature graph, the layer is up-sampled, and a high resolution can be obtained at this point. The second detection layer with a step size of 16 is used, and the remaining feature processing is consistent with that of the first layer. The step size is set to 8 in the third layer, feature prediction is performed thereon, and finally, the detection accuracy of the target antenna is greater.
- Further, the step of performing target recognition and semantic segmentation includes:
- performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:
-
Pr(Classi|object)*Pr(object)*Pr(object)*IOUpred truth=Pr(object)*IOUpred truth (6) - wherein Pr(Classi|object) is an object class probability;
- then applying the activation function by logistic regression:
-
- a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
- in a network layer structure, there are 53 convolutional layers and 22 residual layers among layers 0-74; layers 75-105 are feature interaction layers of a neural convolutional network, which can be divided into three scales; local feature interaction is realized by means of convolution kernels, and its network structure is as shown in
FIG. 4 . - In the production of the data set, only the antenna is detected, so the class is 1. Therefore, in the training, the output of the last convolutional layer is 3*(1+4+1)=18.
- Semantic image segmentation is performed on the antenna image by using a deep convolutional network, and the pixel points of the target image and the background are classified.
- After the target image is input, it first goes through feature extraction by a dilated convolutional network. Since the measured boundary precision is not high enough, the pixel of the target image cannot be well separated from the background pixel, and the pixel classification of the image boundary can be improved by combining a fully connected conditional random field, so that the segmentation effect can be better.
- It is first feature-extracted by using a dilated convolutional network. The feature extraction of the network convolutional layer can be divided into two cases: a low-resolution input image is feature-extracted by a standard convolutional layer, as shown in
FIG. 6 . Dense features of a high-resolution input image are extracted by a detour convolution at a rate of 2, as shown inFIG. 7 , and its step size is set to 2 to thus reduce the feature dimension. In the convolutional network layer, the convolution kernel is set to 3, the stride length is 1, and the step size is 1.FIG. 8 is a schematic diagram of one-dimensional low-resolution feature map extraction.FIG. 9 is a schematic diagram of dilated convolution. - In a network structure of a serial module and a spatial pyramid pooling layer module, the convolution with holes can effectively increase a receptive field of a filter and integrate multi-scale information. After a feature image is input, dilated convolution is calculated:
-
y[i]=Σ k x[i+r*k]*w[k] (8) - For a two-dimensional signal, an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal. The receptive field of the filter can be improved, and the convolution with holes can enlarge the convolution kernel. A residual module of multi-scale feature learning is used in the feature network extraction, while the bottleneck block is used in the present disclosure. In the bottleneck block, each convolution is processed by normalization and processed by an activation function. Thus, contextual information of the context is enriched, and the bottleneck block is as shown in
FIG. 5 . - After the input image is processed by the convolutional network for output, pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
- A view of a random field is as shown in
FIG. 10 . Each circle represents a pixel point, xi (white circle) is a labeled pixel point (node), two connected pixel points are edges of the pixel, yi (black circle) is a reference value of xi, and the classification of the labeled pixel points is determined by the reference value yi. According to the Gibbs distribution function, -
- where y is the reference value of xi, E(y|I) is an energy function.
-
- An image function output through a dilated convolutional network is a unary potential function: A binary potential function is
-
- The function the relationship between pixels, and will assign the same symbols to the same prime points. The unary potential function extracts feature vectors of a node in different feature maps, and the binary function connects the nodes extracted by the unitary potential function to learn its edges. All the nodes are connected to form a conditional random field of a fully connected layer, and an image finally output by the function is more accurate.
- Further, the step of calculating an antenna downtilt angle includes:
- obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane:
-
- The above are merely preferred embodiments of the present disclosure. The present disclosure is not limited to the above implementations. As long as the implementations can achieve the technical effect of the present disclosure with the same means, they are all encompassed in the protection scope of the present disclosure.
Claims (6)
b x=σ(t x)+c x
b y=σ(t y)+c y
bw=pwet
bw=phet
confidence=Pr(object)*IOUprd truth
Pr(Classi|object)*Pr(object)*Pr(object)*IOUpred truth=Pr(object)*IOUpred truth
y[i]=Σ k x[i+r*k]*w[k]
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201811338415.4 | 2018-11-09 | ||
| CN201811338415.4A CN109685762A (en) | 2018-11-09 | 2018-11-09 | A kind of Downtilt measurement method based on multiple dimensioned deep semantic segmentation network |
| PCT/CN2019/076718 WO2020093630A1 (en) | 2018-11-09 | 2019-03-01 | Antenna downward inclination angle measurement method based on multi-scale deep semantic segmentation network |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20210215481A1 true US20210215481A1 (en) | 2021-07-15 |
| US11561092B2 US11561092B2 (en) | 2023-01-24 |
Family
ID=66184673
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/652,346 Active 2040-01-27 US11561092B2 (en) | 2018-11-09 | 2019-03-01 | Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network |
Country Status (4)
| Country | Link |
|---|---|
| US (1) | US11561092B2 (en) |
| EP (1) | EP3680609A4 (en) |
| CN (1) | CN109685762A (en) |
| WO (1) | WO2020093630A1 (en) |
Cited By (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11145082B2 (en) * | 2018-11-06 | 2021-10-12 | Wuyi University | Method for measuring antenna downtilt angle based on deep instance segmentation network |
| CN113538476A (en) * | 2021-07-20 | 2021-10-22 | 大连民族大学 | Deep learning image segmentation method and system based on edge feature extraction |
| CN114322793A (en) * | 2022-03-16 | 2022-04-12 | 科大天工智能装备技术(天津)有限公司 | Workpiece size measurement method, device and storage medium based on global segmentation network |
| CN114549958A (en) * | 2022-02-24 | 2022-05-27 | 四川大学 | A nighttime and camouflaged target detection method based on contextual information perception mechanism |
| CN114677510A (en) * | 2022-03-22 | 2022-06-28 | 中南大学 | Feature map up-sampling method, small target semantic segmentation method and imaging method |
| CN114743010A (en) * | 2022-06-13 | 2022-07-12 | 山东科技大学 | Ultrahigh voltage power transmission line point cloud data semantic segmentation method based on deep learning |
| CN114993243A (en) * | 2022-08-04 | 2022-09-02 | 深圳粤讯通信科技有限公司 | Antenna attitude monitoring and early warning system based on Internet of things |
| CN115187959A (en) * | 2022-07-14 | 2022-10-14 | 清华大学 | Method and system for landing flying vehicle in mountainous region based on binocular vision |
| US20220351497A1 (en) * | 2019-03-22 | 2022-11-03 | Nec Corporation | Image processing system, image processing device, image processing method, and computer-readable medium |
| CN115393725A (en) * | 2022-10-26 | 2022-11-25 | 西南科技大学 | A Bridge Crack Identification Method Based on Feature Enhancement and Semantic Segmentation |
| US20230124572A1 (en) * | 2020-01-08 | 2023-04-20 | Google Llc | Translation of text depicted in images |
| CN116052003A (en) * | 2023-02-07 | 2023-05-02 | 中科星图数字地球合肥有限公司 | Method and device for measuring antenna angle information and related equipment |
| CN116258817A (en) * | 2023-02-16 | 2023-06-13 | 浙江大学 | A method and system for constructing an autonomous driving digital twin scene based on multi-view 3D reconstruction |
| CN116612282A (en) * | 2023-05-26 | 2023-08-18 | 东风悦享科技有限公司 | A new fast semantic segmentation method, system and storage medium for unmanned public transport system |
| US20230281819A1 (en) * | 2022-02-18 | 2023-09-07 | Techcyte, Inc. | Image analysis for identifying objects and classifying background exclusions |
| US20230334876A1 (en) * | 2022-04-14 | 2023-10-19 | Toyota Research Institute, Inc. | End-to-end learned lane boundary detection based on a transformer |
| CN117078760A (en) * | 2023-09-18 | 2023-11-17 | 北方民族大学 | Valve body center positioning method based on image processing |
| CN117994797A (en) * | 2024-04-02 | 2024-05-07 | 杭州海康威视数字技术股份有限公司 | Water gauge reading method and device, storage medium and electronic equipment |
| CN118823551A (en) * | 2024-09-19 | 2024-10-22 | 干将未来科技产业(杭州)有限公司 | An AI algorithm optimization system and method based on deep learning technology |
| CN119741293A (en) * | 2024-12-26 | 2025-04-01 | 中国舰船研究设计中心 | A rapid assessment method for ship corrosion based on image recognition |
Families Citing this family (21)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110647794B (en) * | 2019-07-12 | 2023-01-03 | 五邑大学 | Attention mechanism-based multi-scale SAR image recognition method and device |
| CN110688904A (en) * | 2019-08-30 | 2020-01-14 | 中通服建设有限公司 | Method and device for base station antenna engineering parameter survey based on 5G UAV |
| CN110930417B (en) * | 2019-11-26 | 2023-08-08 | 腾讯科技(深圳)有限公司 | Training method and device for image segmentation model, and image segmentation method and device |
| CN111340139B (en) * | 2020-03-27 | 2024-03-05 | 中国科学院微电子研究所 | A method and device for judging the complexity of image content |
| CN111539983B (en) * | 2020-04-15 | 2023-10-20 | 上海交通大学 | Moving object segmentation method and system based on depth image |
| CN111768392B (en) * | 2020-06-30 | 2022-10-14 | 创新奇智(广州)科技有限公司 | Target detection method and device, electronic equipment and storage medium |
| CN111860297A (en) * | 2020-07-17 | 2020-10-30 | 厦门理工学院 | A SLAM loopback detection method applied to indoor fixed space |
| CN112171668A (en) * | 2020-09-21 | 2021-01-05 | 河南颂达信息技术有限公司 | Rail-mounted robot anti-jamming detection method and device based on artificial intelligence |
| CN112329808A (en) * | 2020-09-25 | 2021-02-05 | 武汉光谷信息技术股份有限公司 | A Optimization Method and System of Deeplab Semantic Segmentation Algorithm |
| CN112132965B (en) * | 2020-09-25 | 2024-03-26 | 中国矿业大学 | Multi-scale characterization method for rock-soil body hole-crack structure |
| CN112965369B (en) * | 2021-01-29 | 2022-02-01 | 电子科技大学 | Distributed deployment method of three-dimensional wireless communication network of unmanned aerial vehicle |
| CN112784857B (en) * | 2021-01-29 | 2022-11-04 | 北京三快在线科技有限公司 | Model training and image processing method and device |
| CN113239815B (en) * | 2021-05-17 | 2022-09-06 | 广东工业大学 | Remote sensing image classification method, device and equipment based on real semantic full-network learning |
| CN113450311B (en) * | 2021-06-01 | 2023-01-13 | 国网河南省电力公司漯河供电公司 | Pin screw defect detection method and system based on semantic segmentation and spatial relationship |
| CN114413855B (en) * | 2022-01-25 | 2023-04-07 | 湖南大学 | Method and system for detecting downtilt of mobile base station antenna based on image recognition |
| CN116758147B (en) * | 2022-03-02 | 2025-11-11 | 上海交通大学 | Plate-shaped antenna downward inclination angle measurement method and system based on mobile phone information fusion |
| CN114638991A (en) * | 2022-03-18 | 2022-06-17 | 聊城大学 | A smart retail settlement method based on YOLOv4-MobileNetv3 and PyQt5 |
| CN115170667B (en) * | 2022-07-15 | 2025-06-17 | 浙江大学 | A water quality transparency detection method based on deep learning using a waterless Secchi disk |
| CN116051843A (en) * | 2023-02-07 | 2023-05-02 | 智洋创新科技股份有限公司 | A method and system for quantitative analysis of molten stream tilt |
| CN120121470B (en) * | 2025-05-15 | 2025-07-08 | 中南大学 | Method and device for identifying slurry fluidity image suitable for non-vertical shooting |
| CN120947588A (en) * | 2025-10-15 | 2025-11-14 | 浙江点创信息科技有限公司 | Methods, devices, computer equipment, and storage media for verifying base station antenna attitude. |
Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150278632A1 (en) * | 2014-03-28 | 2015-10-01 | Xerox Corporation | Extending data-driven detection to the prediction of object part locations |
| US9596617B2 (en) * | 2015-04-14 | 2017-03-14 | ETAK Systems, LLC | Unmanned aerial vehicle-based systems and methods associated with cell sites and cell towers |
| US20170077586A1 (en) * | 2014-05-27 | 2017-03-16 | Huawei Technologies Co., Ltd. | Method and Device for Obtaining Antenna Engineering Parameter and System |
| US20180089505A1 (en) * | 2016-09-23 | 2018-03-29 | Samsung Electronics Co., Ltd. | System and method for deep network fusion for fast and robust object detection |
| US20180137642A1 (en) * | 2016-11-15 | 2018-05-17 | Magic Leap, Inc. | Deep learning system for cuboid detection |
| US20180218351A1 (en) * | 2017-01-31 | 2018-08-02 | Focal Systems, Inc. | Automated checkout system through mobile shopping units |
| US20180260415A1 (en) * | 2017-03-10 | 2018-09-13 | Xerox Corporation | Instance-level image retrieval with a region proposal network |
| US20190015059A1 (en) * | 2017-07-17 | 2019-01-17 | Siemens Healthcare Gmbh | Semantic segmentation for cancer detection in digital breast tomosynthesis |
| US20190043003A1 (en) * | 2017-08-07 | 2019-02-07 | Standard Cognition, Corp | Predicting inventory events using foreground/background processing |
| US20190130189A1 (en) * | 2017-10-30 | 2019-05-02 | Qualcomm Incorporated | Suppressing duplicated bounding boxes from object detection in a video analytics system |
| US20190213438A1 (en) * | 2018-01-05 | 2019-07-11 | Irobot Corporation | Mobile Cleaning Robot Artificial Intelligence for Situational Awareness |
| US20200090519A1 (en) * | 2018-09-13 | 2020-03-19 | Volvo Car Corporation | System and method for camera or sensor-based parking spot detection and identification |
| US10872228B1 (en) * | 2017-09-27 | 2020-12-22 | Apple Inc. | Three-dimensional object detection |
| US20220004770A1 (en) * | 2018-10-31 | 2022-01-06 | Arcus Holding A/S | Object detection using a combination of deep learning and non-deep learning techniques |
Family Cites Families (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE60317625T2 (en) * | 2002-01-24 | 2008-11-27 | Telecom Italia S.P.A. | Measurement of antenna parameters of a base station for cellular telephony |
| US8374979B2 (en) * | 2009-11-18 | 2013-02-12 | Nec Laboratories America, Inc. | Fast image parsing by graph adaptive dynamic programming (GADP) performing classification, detection, and segmentation simultaneously |
| US20110150317A1 (en) * | 2009-12-17 | 2011-06-23 | Electronics And Telecommunications Research Institute | System and method for automatically measuring antenna characteristics |
| US9109889B2 (en) * | 2011-06-24 | 2015-08-18 | Trimble Navigation Limited | Determining tilt angle and tilt direction using image processing |
| AT511191B1 (en) * | 2011-07-01 | 2012-10-15 | Thomas Dr Neubauer | METHOD AND DEVICE FOR DETERMINING AND STORING THE POSITION AND ORIENTATION OF ANTENNA STRUCTURES |
| CN103630107B (en) | 2012-08-23 | 2015-09-16 | 北京交通大学 | A kind of antenna for base station inclination angle measurement method and data processing method |
| CN104504381B (en) | 2015-01-09 | 2017-12-26 | 新智认知数据服务有限公司 | Non-rigid object detection method and its system |
| US9855658B2 (en) * | 2015-03-19 | 2018-01-02 | Rahul Babu | Drone assisted adaptive robot control |
| CN104978580B (en) * | 2015-06-15 | 2018-05-04 | 国网山东省电力公司电力科学研究院 | A kind of insulator recognition methods for unmanned plane inspection transmission line of electricity |
| US9918235B2 (en) | 2015-11-24 | 2018-03-13 | Verizon Patent And Licensing Inc. | Adaptive antenna operation for UAVs using terrestrial cellular networks |
| CN106851665A (en) * | 2015-12-07 | 2017-06-13 | 上海无线通信研究中心 | The downdip adjusting method of antenna and base station |
| CN107664491B (en) * | 2016-07-28 | 2020-06-05 | 中国电信股份有限公司 | Method, device and system for measuring downward inclination angle of base station antenna |
| CN106709568B (en) * | 2016-12-16 | 2019-03-22 | 北京工业大学 | The object detection and semantic segmentation method of RGB-D image based on deep layer convolutional network |
| CN106683091B (en) * | 2017-01-06 | 2019-09-24 | 北京理工大学 | A kind of target classification and attitude detecting method based on depth convolutional neural networks |
| US10565787B1 (en) * | 2017-01-27 | 2020-02-18 | NHIAE Group, LLC | Systems and methods for enhanced 3D modeling of a complex object |
| US10402689B1 (en) * | 2017-04-04 | 2019-09-03 | Snap Inc. | Generating an image mask using machine learning |
| US11257198B1 (en) * | 2017-04-28 | 2022-02-22 | Digimarc Corporation | Detection of encoded signals and icons |
| CN110809784B (en) * | 2017-09-27 | 2021-04-20 | 谷歌有限责任公司 | End-to-end network model for high-resolution image segmentation |
| CN107830846B (en) * | 2017-09-30 | 2020-04-10 | 杭州艾航科技有限公司 | Method for measuring angle of communication tower antenna by using unmanned aerial vehicle and convolutional neural network |
| US10649459B2 (en) * | 2018-04-26 | 2020-05-12 | Zoox, Inc. | Data segmentation using masks |
-
2018
- 2018-11-09 CN CN201811338415.4A patent/CN109685762A/en active Pending
-
2019
- 2019-03-01 WO PCT/CN2019/076718 patent/WO2020093630A1/en not_active Ceased
- 2019-03-01 EP EP19856443.7A patent/EP3680609A4/en not_active Withdrawn
- 2019-03-01 US US16/652,346 patent/US11561092B2/en active Active
Patent Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150278632A1 (en) * | 2014-03-28 | 2015-10-01 | Xerox Corporation | Extending data-driven detection to the prediction of object part locations |
| US20170077586A1 (en) * | 2014-05-27 | 2017-03-16 | Huawei Technologies Co., Ltd. | Method and Device for Obtaining Antenna Engineering Parameter and System |
| US9596617B2 (en) * | 2015-04-14 | 2017-03-14 | ETAK Systems, LLC | Unmanned aerial vehicle-based systems and methods associated with cell sites and cell towers |
| US20180089505A1 (en) * | 2016-09-23 | 2018-03-29 | Samsung Electronics Co., Ltd. | System and method for deep network fusion for fast and robust object detection |
| US20180137642A1 (en) * | 2016-11-15 | 2018-05-17 | Magic Leap, Inc. | Deep learning system for cuboid detection |
| US20180218351A1 (en) * | 2017-01-31 | 2018-08-02 | Focal Systems, Inc. | Automated checkout system through mobile shopping units |
| US20180260415A1 (en) * | 2017-03-10 | 2018-09-13 | Xerox Corporation | Instance-level image retrieval with a region proposal network |
| US20190015059A1 (en) * | 2017-07-17 | 2019-01-17 | Siemens Healthcare Gmbh | Semantic segmentation for cancer detection in digital breast tomosynthesis |
| US20190043003A1 (en) * | 2017-08-07 | 2019-02-07 | Standard Cognition, Corp | Predicting inventory events using foreground/background processing |
| US10872228B1 (en) * | 2017-09-27 | 2020-12-22 | Apple Inc. | Three-dimensional object detection |
| US20190130189A1 (en) * | 2017-10-30 | 2019-05-02 | Qualcomm Incorporated | Suppressing duplicated bounding boxes from object detection in a video analytics system |
| US20190213438A1 (en) * | 2018-01-05 | 2019-07-11 | Irobot Corporation | Mobile Cleaning Robot Artificial Intelligence for Situational Awareness |
| US20200090519A1 (en) * | 2018-09-13 | 2020-03-19 | Volvo Car Corporation | System and method for camera or sensor-based parking spot detection and identification |
| US20220004770A1 (en) * | 2018-10-31 | 2022-01-06 | Arcus Holding A/S | Object detection using a combination of deep learning and non-deep learning techniques |
Cited By (23)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11145082B2 (en) * | 2018-11-06 | 2021-10-12 | Wuyi University | Method for measuring antenna downtilt angle based on deep instance segmentation network |
| US20220351497A1 (en) * | 2019-03-22 | 2022-11-03 | Nec Corporation | Image processing system, image processing device, image processing method, and computer-readable medium |
| US11748977B2 (en) * | 2019-03-22 | 2023-09-05 | Nec Corporation | Image processing system, image processing device, image processing method, and computer-readable medium |
| US12217017B2 (en) * | 2020-01-08 | 2025-02-04 | Google Llc | Translation of text depicted in images |
| US20230124572A1 (en) * | 2020-01-08 | 2023-04-20 | Google Llc | Translation of text depicted in images |
| CN113538476A (en) * | 2021-07-20 | 2021-10-22 | 大连民族大学 | Deep learning image segmentation method and system based on edge feature extraction |
| US20230281819A1 (en) * | 2022-02-18 | 2023-09-07 | Techcyte, Inc. | Image analysis for identifying objects and classifying background exclusions |
| CN114549958A (en) * | 2022-02-24 | 2022-05-27 | 四川大学 | A nighttime and camouflaged target detection method based on contextual information perception mechanism |
| CN114322793A (en) * | 2022-03-16 | 2022-04-12 | 科大天工智能装备技术(天津)有限公司 | Workpiece size measurement method, device and storage medium based on global segmentation network |
| CN114677510A (en) * | 2022-03-22 | 2022-06-28 | 中南大学 | Feature map up-sampling method, small target semantic segmentation method and imaging method |
| US20230334876A1 (en) * | 2022-04-14 | 2023-10-19 | Toyota Research Institute, Inc. | End-to-end learned lane boundary detection based on a transformer |
| US12159465B2 (en) * | 2022-04-14 | 2024-12-03 | Toyota Research Institute, Inc. | End-to-end learned lane boundary detection based on a transformer |
| CN114743010A (en) * | 2022-06-13 | 2022-07-12 | 山东科技大学 | Ultrahigh voltage power transmission line point cloud data semantic segmentation method based on deep learning |
| CN115187959A (en) * | 2022-07-14 | 2022-10-14 | 清华大学 | Method and system for landing flying vehicle in mountainous region based on binocular vision |
| CN114993243A (en) * | 2022-08-04 | 2022-09-02 | 深圳粤讯通信科技有限公司 | Antenna attitude monitoring and early warning system based on Internet of things |
| CN115393725A (en) * | 2022-10-26 | 2022-11-25 | 西南科技大学 | A Bridge Crack Identification Method Based on Feature Enhancement and Semantic Segmentation |
| CN116052003A (en) * | 2023-02-07 | 2023-05-02 | 中科星图数字地球合肥有限公司 | Method and device for measuring antenna angle information and related equipment |
| CN116258817A (en) * | 2023-02-16 | 2023-06-13 | 浙江大学 | A method and system for constructing an autonomous driving digital twin scene based on multi-view 3D reconstruction |
| CN116612282A (en) * | 2023-05-26 | 2023-08-18 | 东风悦享科技有限公司 | A new fast semantic segmentation method, system and storage medium for unmanned public transport system |
| CN117078760A (en) * | 2023-09-18 | 2023-11-17 | 北方民族大学 | Valve body center positioning method based on image processing |
| CN117994797A (en) * | 2024-04-02 | 2024-05-07 | 杭州海康威视数字技术股份有限公司 | Water gauge reading method and device, storage medium and electronic equipment |
| CN118823551A (en) * | 2024-09-19 | 2024-10-22 | 干将未来科技产业(杭州)有限公司 | An AI algorithm optimization system and method based on deep learning technology |
| CN119741293A (en) * | 2024-12-26 | 2025-04-01 | 中国舰船研究设计中心 | A rapid assessment method for ship corrosion based on image recognition |
Also Published As
| Publication number | Publication date |
|---|---|
| US11561092B2 (en) | 2023-01-24 |
| EP3680609A4 (en) | 2021-02-24 |
| EP3680609A1 (en) | 2020-07-15 |
| CN109685762A (en) | 2019-04-26 |
| WO2020093630A1 (en) | 2020-05-14 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11561092B2 (en) | Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network | |
| CN114488194B (en) | A method for target detection and recognition on structured roads for intelligent driving vehicles | |
| Yan et al. | Automated asphalt highway pavement crack detection based on deformable single shot multi-box detector under a complex environment | |
| EP3997616B1 (en) | Object-based change detection using a neural network | |
| US12214809B2 (en) | Instance segmentation imaging system | |
| Wang et al. | A center location algorithm for tropical cyclone in satellite infrared images | |
| KR102507501B1 (en) | Artificial Intelligence-based Water Quality Contaminant Monitoring System and Method | |
| CN108052940A (en) | SAR remote sensing images waterborne target detection methods based on deep learning | |
| CN114252868B (en) | Laser radar calibration method, device, computer equipment and storage medium | |
| CN111553280A (en) | Target part identification method based on deep learning | |
| CN116168246B (en) | A method, device, equipment and medium for identifying abandoned slag sites in railway engineering | |
| CN114252883A (en) | Target detection method, apparatus, computer device and medium | |
| CN114252884A (en) | Method and device for positioning and monitoring roadside radar, computer equipment and storage medium | |
| CN116246096A (en) | A Point Cloud 3D Object Detection Method Based on Foreground Enhanced Knowledge Distillation | |
| Wu et al. | Multimodal collaboration networks for geospatial vehicle detection in dense, occluded, and large-scale events | |
| CN114255264B (en) | Multi-base-station registration method and device, computer equipment and storage medium | |
| Shi et al. | Fine object change detection based on vector boundary and deep learning with high-resolution remote sensing images | |
| CN114252859B (en) | Method, device, computer equipment and storage medium for determining target area | |
| Han et al. | Targets mask U-Net for wind turbines detection in remote sensing images | |
| CN113298042A (en) | Method and device for processing remote sensing image data, storage medium and computer equipment | |
| Yang et al. | Sardet-cl: Self-supervised contrastive learning with feature enhancement and imaging mechanism constraints for sar target detection | |
| CN119716839A (en) | CN-YOLOv-based high-precision detection method for multi-mode data fusion anti-unmanned aerial vehicle | |
| CN119964098A (en) | A road surface abnormality detection method, device, storage medium and electronic device | |
| CN118799827A (en) | An intelligent detection and positioning method for sea surface ship targets using dual-polarization SAR images | |
| CN115082786A (en) | Remote sensing image target detection method under condition of small amount of marks |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| AS | Assignment |
Owner name: WUYI UNIVERSITY, CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHAI, YIKUI;ZHOU, JIHUA;WU, YUETING;AND OTHERS;REEL/FRAME:056108/0041 Effective date: 20200312 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |