[go: up one dir, main page]

US20210215481A1 - Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network - Google Patents

Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network Download PDF

Info

Publication number
US20210215481A1
US20210215481A1 US16/652,346 US201916652346A US2021215481A1 US 20210215481 A1 US20210215481 A1 US 20210215481A1 US 201916652346 A US201916652346 A US 201916652346A US 2021215481 A1 US2021215481 A1 US 2021215481A1
Authority
US
United States
Prior art keywords
antenna
target
image
downtilt angle
semantic segmentation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US16/652,346
Other versions
US11561092B2 (en
Inventor
Yikui ZHAI
Jihua Zhou
Yueting WU
Yu Zheng
Ying Xu
Junying GAN
Junying ZENG
Wenbo DENG
Qirui KE
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuyi University Fujian
Original Assignee
Wuyi University Fujian
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuyi University Fujian filed Critical Wuyi University Fujian
Assigned to WUYI UNIVERSITY reassignment WUYI UNIVERSITY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DENG, Wenbo, GAN, Junying, KE, Qirui, WU, Yueting, XU, YING, ZENG, Junying, ZHAI, Yikui, ZHENG, YU, ZHOU, JIHUA
Publication of US20210215481A1 publication Critical patent/US20210215481A1/en
Application granted granted Critical
Publication of US11561092B2 publication Critical patent/US11561092B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B21/00Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant
    • G01B21/22Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant for measuring angles or tapers; for testing the alignment of axes
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • G06N3/0454
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/17Terrestrial scenes taken from planes or by drones
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/176Urban or other man-made structures
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/70Labelling scene content, e.g. deriving syntactic or semantic representations
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • the present disclosure relates to the field of mobile communication, and in particular, to a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network.
  • an azimuth angle and a downtilt angle of an antenna affect the coverage of signals and the interference between the signals, so the antenna needs to be timely calculated and adjusted strictly to improve the quality of network signals.
  • the first one is climbing to an antenna base station manually and using a measuring instrument (a compass, a slope meter, or the like) for measurement; and the second one is installing an angle sensor on the antenna to return data.
  • the antenna is susceptible to wind, snow and other factors, resulting in a change in the downtilt angle, so it needs to be measured regularly.
  • the first method as the base station is high and the number of the antennas is larger, the manual safety hazard and workload are larger, and the practicability is low.
  • the installation time is long, and the antenna models are different, so the installation cost of the instruments is high and the practicability is not high. Both the two methods consume a lot of manpower and material resources and are not suitable for large-scale measurement today.
  • the present disclosure aims at providing a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network.
  • the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
  • An antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network including:
  • base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
  • predicting a target bounding box a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
  • target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified;
  • the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
  • the collecting image data includes:
  • the predicting a target bounding box includes:
  • each grid includes five predictors (x, y, w, h, confidence) and a c class
  • the output of the network is of a size of S*S*(5*B+C)
  • B is the number of the bounding boxes in each grid
  • C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:
  • performing target recognition and semantic segmentation includes:
  • antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:
  • object)*Pr(object)*Pr(object)*IOU pred truth Pr(object)*IOU pred truth (6)
  • object) is an object class probability
  • a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
  • dilated convolutional network After the target image is input, it first goes through feature extraction by a dilated convolutional network; and after a feature image is input, dilated convolution is calculated:
  • an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal;
  • pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
  • the calculating an antenna downtilt angle includes:
  • the downtilt angle of the base station antenna being an angle ⁇ between the base station antenna and a vertical plane:
  • the present disclosure has the following beneficial effects: the present disclosure adopts an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network.
  • the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
  • FIG. 1 is a schematic diagram of a downtilt angle of a base station antenna
  • FIG. 2 is a flowchart of an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 3 is a schematic diagram of border prediction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 4 is a schematic diagram of a network structure for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 5 is a schematic diagram of a bottleneck block for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 6 is a schematic diagram of standard convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 7 is a schematic diagram of high-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 8 is a schematic diagram of one-dimensional low-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure
  • FIG. 9 is a schematic diagram of dilated convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure.
  • FIG. 10 is a view of a random field for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure.
  • an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network including:
  • base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
  • predicting a target bounding box a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
  • target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified;
  • the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
  • the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
  • the step of collecting image data includes:
  • the step of predicting a target bounding box includes:
  • each grid includes 5 predictors (x, y, w, h, confidence) and a c class
  • the output of the network is of a size of S*S*(5*B+C)
  • B is the number of the bounding boxes in each grid
  • C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:
  • multi-scale prediction is used. There is no need to fix the size of an input image, so different step sizes can be used to detect feature maps of different sizes.
  • Three different detection layers are used to detect the antenna image for the target antenna, and different detection layers are realized by controlling the step size.
  • the first detection layer is down-sampled with a step size of 32 to reduce the feature dimension. In order to connect with the previous identical feature graph, the layer is up-sampled, and a high resolution can be obtained at this point.
  • the second detection layer with a step size of 16 is used, and the remaining feature processing is consistent with that of the first layer.
  • the step size is set to 8 in the third layer, feature prediction is performed thereon, and finally, the detection accuracy of the target antenna is greater.
  • step of performing target recognition and semantic segmentation includes:
  • antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:
  • object)*Pr(object)*Pr(object)*IOU pred truth Pr(object)*IOU pred truth (6)
  • object) is an object class probability
  • a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
  • layers 75-105 are feature interaction layers of a neural convolutional network, which can be divided into three scales; local feature interaction is realized by means of convolution kernels, and its network structure is as shown in FIG. 4 .
  • Semantic image segmentation is performed on the antenna image by using a deep convolutional network, and the pixel points of the target image and the background are classified.
  • the target image After the target image is input, it first goes through feature extraction by a dilated convolutional network. Since the measured boundary precision is not high enough, the pixel of the target image cannot be well separated from the background pixel, and the pixel classification of the image boundary can be improved by combining a fully connected conditional random field, so that the segmentation effect can be better.
  • FIG. 6 is a schematic diagram of one-dimensional low-resolution feature map extraction.
  • FIG. 9 is a schematic diagram of dilated convolution.
  • the convolution with holes can effectively increase a receptive field of a filter and integrate multi-scale information.
  • dilated convolution is calculated:
  • an output corresponding to each position i is y
  • w is a filter
  • the detour rate r is a step size for sampling the input signal.
  • the receptive field of the filter can be improved, and the convolution with holes can enlarge the convolution kernel.
  • a residual module of multi-scale feature learning is used in the feature network extraction, while the bottleneck block is used in the present disclosure.
  • each convolution is processed by normalization and processed by an activation function.
  • contextual information of the context is enriched, and the bottleneck block is as shown in FIG. 5 .
  • pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
  • FIG. 10 A view of a random field is as shown in FIG. 10 .
  • Each circle represents a pixel point
  • xi white circle
  • yi black circle
  • xi the classification of the labeled pixel points is determined by the reference value yi.
  • I) is an energy function
  • An image function output through a dilated convolutional network is a unary potential function:
  • a binary potential function is
  • the function the relationship between pixels, and will assign the same symbols to the same prime points.
  • the unary potential function extracts feature vectors of a node in different feature maps, and the binary function connects the nodes extracted by the unitary potential function to learn its edges. All the nodes are connected to form a conditional random field of a fully connected layer, and an image finally output by the function is more accurate.
  • step of calculating an antenna downtilt angle includes:
  • the downtilt angle of the base station antenna being an angle ⁇ between the base station antenna and a vertical plane:

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Remote Sensing (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

A method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network is disclosed, including: collecting base station antenna data by using an unmanned aerial vehicle, and labeling an acquired antenna image with a labeling tool to make a data set; calling the data set for training and debugging a model; recognizing and detecting a target antenna, performing semantic segmentation on an output image, finally obtaining a target image finally segmented, and calculating a downtilt angle of the target image. The method is highly applicable, cost-effective, and safe.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a national stage application under 35 U.S.C. 371 of PCT Application No. PCT/CN2019/076718, filed on 1 Mar. 2019, which PCT application claimed the benefit of Chinese Patent Application No. 2018113384154, filed on 9 Nov. 2018, the entire disclosure of each of which are hereby incorporated herein by reference.
  • TECHNICAL FIELD
  • The present disclosure relates to the field of mobile communication, and in particular, to a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network.
  • BACKGROUND
  • Nowadays, in the era of network information, the quality of mobile communication networks is extremely important. In GSM-R construction and planning, as shown in FIG. 1, an azimuth angle and a downtilt angle of an antenna affect the coverage of signals and the interference between the signals, so the antenna needs to be timely calculated and adjusted strictly to improve the quality of network signals.
  • There are two traditional methods to measure the antenna downtilt angle: the first one is climbing to an antenna base station manually and using a measuring instrument (a compass, a slope meter, or the like) for measurement; and the second one is installing an angle sensor on the antenna to return data. The antenna is susceptible to wind, snow and other factors, resulting in a change in the downtilt angle, so it needs to be measured regularly. For the first method, as the base station is high and the number of the antennas is larger, the manual safety hazard and workload are larger, and the practicability is low. For the second method, the installation time is long, and the antenna models are different, so the installation cost of the instruments is high and the practicability is not high. Both the two methods consume a lot of manpower and material resources and are not suitable for large-scale measurement today.
  • SUMMARY
  • To solve the above problems, the present disclosure aims at providing a method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network. The method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
  • The technical scheme adopted by the present disclosure to solve the problems is as follows:
  • An antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network, including:
  • collecting image data: base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
  • predicting a target bounding box: a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
  • performing target recognition and semantic segmentation: target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and
  • calculating an antenna downtilt angle: the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
  • Further, the collecting image data includes:
  • locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set.
  • Further, the predicting a target bounding box includes:
  • positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein 4 coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy), box heights of the bounding boxes are px, py respectively, and the network predicts their values as:

  • b x=σ(t x)+c x  (1)

  • b y=σ(t y)+c y  (2)

  • bw=pwet w   (3)

  • bw=phet w   (4)
  • where the input antenna image is divided into N*N grids, each grid includes five predictors (x, y, w, h, confidence) and a c class, and the output of the network is of a size of S*S*(5*B+C); B is the number of the bounding boxes in each grid, C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:

  • confidence=Pr(object)*IOUprd truth  (5)
  • where a threshold is set to 0.5 when Pr(Object)=1; the target antenna falls in the center of the grid, that is, the bounding box currently predicted coincides with an actual background box object better than before; if the predicted bounding box is not the best currently, the bounding box is not predicted when the threshold is smaller than 0.5, and it is determined that the target antenna does not fall into the grid.
  • Further, the performing target recognition and semantic segmentation includes:
  • performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:

  • Pr(Classi|object)*Pr(object)*Pr(object)*IOUpred truth=Pr(object)*IOUpred truth  (6)
  • where Pr(Classi|object) is an object class probability;
  • then applying the activation function by logistic regression:
  • f ( x ) = 1 1 + e - x ( 7 )
  • a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
  • then performing semantic image segmentation on the antenna image by using a deep convolutional network, and classifying the pixel points of the target image and the background:
  • after the target image is input, it first goes through feature extraction by a dilated convolutional network; and after a feature image is input, dilated convolution is calculated:

  • y[i]=Σ k x[i+r*k]*w[k]  (8)
  • for a two-dimensional signal, an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal;
  • after the input image is processed by the convolutional network for output, pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
  • Further, the calculating an antenna downtilt angle includes:
  • obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane:
  • θ = arctan x y . ( 12 )
  • The present disclosure has the following beneficial effects: the present disclosure adopts an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network. The method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present disclosure is further described below with reference to the accompanying drawings and examples.
  • FIG. 1 is a schematic diagram of a downtilt angle of a base station antenna;
  • FIG. 2 is a flowchart of an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 3 is a schematic diagram of border prediction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 4 is a schematic diagram of a network structure for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 5 is a schematic diagram of a bottleneck block for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 6 is a schematic diagram of standard convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 7 is a schematic diagram of high-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 8 is a schematic diagram of one-dimensional low-resolution feature extraction for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure;
  • FIG. 9 is a schematic diagram of dilated convolution for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure; and
  • FIG. 10 is a view of a random field for an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network according to an embodiment of the present disclosure.
  • DETAILED DESCRIPTION
  • Referring to FIG. 2, an antenna downtilt angle measuring method based on a multi-scale deep semantic segmentation network is provided in an embodiment of the present disclosure, including:
  • collecting image data: base station antenna data is collected by using an unmanned aerial vehicle and antenna images collected are taken as a data set;
  • predicting a target bounding box: a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
  • performing target recognition and semantic segmentation: target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and
  • calculating an antenna downtilt angle: the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
  • In the embodiment, the method for measuring a downtilt angle of a mobile base station antenna by calling a target detection algorithm and a semantic segmentation algorithm and using an unmanned aerial vehicle as a carrier is highly applicable, cost-effective, and safe.
  • Further, the step of collecting image data includes:
  • locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set.
  • Further, the step of predicting a target bounding box includes:
  • positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein 4 coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy), box heights of the bounding boxes are px, py respectively, box prediction is as shown in FIG. 3, and the network predicts their values as:

  • b x=σ(t x)+c x  (1)

  • b y=σ(t y)+c y  (2)

  • bw=pwet w   (3)

  • bw=phet w   (4)
  • where the input antenna image is divided into N*N grids, each grid includes 5 predictors (x, y, w, h, confidence) and a c class, and the output of the network is of a size of S*S*(5*B+C); B is the number of the bounding boxes in each grid, C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid includes two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:

  • confidence=Pr(object)*IOUprd truth  (5)
  • where a threshold is set to 0.5 when Pr(Object)=1; the target antenna falls in the center of the grid, that is, the bounding box currently predicted coincides with an actual background box object better than before; if the predicted bounding box is not the best currently, the bounding box is not predicted when the threshold is smaller than 0.5, and it is determined that the target antenna does not fall into the grid.
  • In the accuracy of a target, multi-scale prediction is used. There is no need to fix the size of an input image, so different step sizes can be used to detect feature maps of different sizes. Three different detection layers are used to detect the antenna image for the target antenna, and different detection layers are realized by controlling the step size. The first detection layer is down-sampled with a step size of 32 to reduce the feature dimension. In order to connect with the previous identical feature graph, the layer is up-sampled, and a high resolution can be obtained at this point. The second detection layer with a step size of 16 is used, and the remaining feature processing is consistent with that of the first layer. The step size is set to 8 in the third layer, feature prediction is performed thereon, and finally, the detection accuracy of the target antenna is greater.
  • Further, the step of performing target recognition and semantic segmentation includes:
  • performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:

  • Pr(Classi|object)*Pr(object)*Pr(object)*IOUpred truth=Pr(object)*IOUpred truth  (6)
  • wherein Pr(Classi|object) is an object class probability;
  • then applying the activation function by logistic regression:
  • f ( x ) = 1 1 + e - x ( 7 )
  • a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
  • in a network layer structure, there are 53 convolutional layers and 22 residual layers among layers 0-74; layers 75-105 are feature interaction layers of a neural convolutional network, which can be divided into three scales; local feature interaction is realized by means of convolution kernels, and its network structure is as shown in FIG. 4.
  • In the production of the data set, only the antenna is detected, so the class is 1. Therefore, in the training, the output of the last convolutional layer is 3*(1+4+1)=18.
  • Semantic Segmentation
  • Semantic image segmentation is performed on the antenna image by using a deep convolutional network, and the pixel points of the target image and the background are classified.
  • After the target image is input, it first goes through feature extraction by a dilated convolutional network. Since the measured boundary precision is not high enough, the pixel of the target image cannot be well separated from the background pixel, and the pixel classification of the image boundary can be improved by combining a fully connected conditional random field, so that the segmentation effect can be better.
  • It is first feature-extracted by using a dilated convolutional network. The feature extraction of the network convolutional layer can be divided into two cases: a low-resolution input image is feature-extracted by a standard convolutional layer, as shown in FIG. 6. Dense features of a high-resolution input image are extracted by a detour convolution at a rate of 2, as shown in FIG. 7, and its step size is set to 2 to thus reduce the feature dimension. In the convolutional network layer, the convolution kernel is set to 3, the stride length is 1, and the step size is 1. FIG. 8 is a schematic diagram of one-dimensional low-resolution feature map extraction. FIG. 9 is a schematic diagram of dilated convolution.
  • In a network structure of a serial module and a spatial pyramid pooling layer module, the convolution with holes can effectively increase a receptive field of a filter and integrate multi-scale information. After a feature image is input, dilated convolution is calculated:

  • y[i]=Σ k x[i+r*k]*w[k]  (8)
  • For a two-dimensional signal, an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal. The receptive field of the filter can be improved, and the convolution with holes can enlarge the convolution kernel. A residual module of multi-scale feature learning is used in the feature network extraction, while the bottleneck block is used in the present disclosure. In the bottleneck block, each convolution is processed by normalization and processed by an activation function. Thus, contextual information of the context is enriched, and the bottleneck block is as shown in FIG. 5.
  • After the input image is processed by the convolutional network for output, pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
  • A view of a random field is as shown in FIG. 10. Each circle represents a pixel point, xi (white circle) is a labeled pixel point (node), two connected pixel points are edges of the pixel, yi (black circle) is a reference value of xi, and the classification of the labeled pixel points is determined by the reference value yi. According to the Gibbs distribution function,
  • P ( Y = y | I ) = 1 Z ( I ) exp ( - E ( y | I ) ) ( 9 )
  • where y is the reference value of xi, E(y|I) is an energy function.
  • E ( y | I ) = i Ψ u ( y i ) + i < y Ψ p ( y i , y j ) ( 10 )
  • An image function output through a dilated convolutional network is a unary potential function: A binary potential function is
  • Ψ P ( y i , y j ) = u ( y i , y j ) m = 1 M w ( m ) k G ( m ) ( f i , f j ) ( 11 )
  • The function the relationship between pixels, and will assign the same symbols to the same prime points. The unary potential function extracts feature vectors of a node in different feature maps, and the binary function connects the nodes extracted by the unitary potential function to learn its edges. All the nodes are connected to form a conditional random field of a fully connected layer, and an image finally output by the function is more accurate.
  • Further, the step of calculating an antenna downtilt angle includes:
  • obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane:
  • θ = arctan x y . ( 12 )
  • The above are merely preferred embodiments of the present disclosure. The present disclosure is not limited to the above implementations. As long as the implementations can achieve the technical effect of the present disclosure with the same means, they are all encompassed in the protection scope of the present disclosure.

Claims (6)

1. A method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network, comprising:
collecting image data: wherein base station antenna data is collected by using an unmanned aerial vehicle, and antenna images collected are taken as a data set;
predicting a target bounding box: wherein a target antenna in the data set is positioned, and a bounding box is predicted by logistic regression;
performing target recognition and semantic segmentation: wherein target features of the target antenna in the data set are extracted, the target features are learned and processed by an activation function, a target image is output for semantic image segmentation, and pixel points of the target image and the background are classified; and
calculating an antenna downtilt angle: wherein the width and height of an antenna box are obtained according to a border of the target image to calculate the antenna downtilt angle.
2. The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 1, wherein the step of collecting image data comprises:
locating the unmanned aerial vehicle on the top of a pole of a base station antenna, and recording the longitude and latitude (L0, W0) of the pole in the vertical direction; causing the unmanned aerial vehicle to fly around a point of the base station antenna, setting a flight radius of the unmanned aerial vehicle, and the unmanned aerial vehicle moving around the pole along the radius on the same horizontal plane to acquire antenna images with different attitudes and angles of a mobile base station antenna as a data set.
3. The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 2, wherein the step of predicting a target bounding box comprises:
positioning a target antenna in the antenna image, predicting a bounding box by logistic regression, first dividing the entire antenna image into N*N grids, predicting the entire antenna image after the antenna image is input, scanning each grid at a time, and starting to predict the target antenna when the center of the grid where the target antenna is located is positioned, wherein four coordinate values predicted for each bounding box are tx, ty, tw, and th, respectively, an upper-left offset of each target cell is (cx, cy), box heights of the bounding boxes are px, py respectively, and the network predicts their values as:

b x=σ(t x)+c x

b y=σ(t y)+c y

bw=pwet w

bw=phet w
where the input antenna image is divided into N*N grids, each grid comprises 5 predictors (x, y, w, h, confidence) and a c class, and the output of the network is of a size of S*S*(5*B+C); B is the number of the bounding boxes in each grid, C means the class is only antenna in the present disclosure, and thus is 1, and confidence represents that the predicted grid comprises two pieces of information, i.e., confidence of the target antenna and prediction accuracy of the bounding box:

confidence=Pr(object)*IOUprd truth
4. The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 6, wherein the step of performing target recognition and semantic segmentation comprises:
performing target recognition by using a network convolutional layer for feature extraction: antenna image pixel 416*416 is input, the channel number is 3, there are 32 layers of convolution kernels, each kernel has a size of 3*3, 32 layers of convolution kernels are used to learn 32 feature maps, and for color differences of the target antenna, features of the target antenna are learned by using different convolution kernels; convolutional layer up-sampling is performed during feature extraction, and a prediction formula for object classes is as follows:

Pr(Classi|object)*Pr(object)*Pr(object)*IOUpred truth=Pr(object)*IOUpred truth
wherein Pr(Classi|object) is an object class probability;
then applying the activation function by logistic regression:
f ( x ) = 1 1 + e - x
a predicted target output range is made between 0 and 1, the antenna image is processed by the activation function after feature extraction, and when the output value is greater than 0.5, the object is determined as an antenna;
then performing semantic image segmentation on the antenna image by using a deep convolutional network, and classifying the pixel points of the target image and the background:
after the target image is input, the feature extraction goes through by a dilated convolutional network; and after a feature image is input, dilated convolution is calculated:

y[i]=Σ k x[i+r*k]*w[k]
for a two-dimensional signal, an output corresponding to each position i is y, w is a filter, and the detour rate r is a step size for sampling the input signal;
after the input image is processed by the convolutional network for output, pixel points of the output target image are classified by a fully connected conditional random field, and the classification is mainly performed for the target image and the background boundary.
5. The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 4, wherein the step of calculating an antenna downtilt angle comprises:
obtaining the width x and the height y of the antenna box according to the border of the target image, and calculating a downtilt angle of the base station antenna according to a geometric relation, the downtilt angle of the base station antenna being an angle θ between the base station antenna and a vertical plane:
θ = arctan x y .
6. The method for measuring an antenna downtilt angle based on a multi-scale deep semantic segmentation network according to claim 3, wherein:
a threshold is set to 0.5 when Pr(Object)=1; the target antenna falls in the center of the grid, that is, the bounding box currently predicted coincides with an actual background box object better than before; if the predicted bounding box is not the best currently, the bounding box is not predicted when the threshold is smaller than 0.5, so as to determine that the target antenna does not fall into the grid.
US16/652,346 2018-11-09 2019-03-01 Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network Active 2040-01-27 US11561092B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201811338415.4 2018-11-09
CN201811338415.4A CN109685762A (en) 2018-11-09 2018-11-09 A kind of Downtilt measurement method based on multiple dimensioned deep semantic segmentation network
PCT/CN2019/076718 WO2020093630A1 (en) 2018-11-09 2019-03-01 Antenna downward inclination angle measurement method based on multi-scale deep semantic segmentation network

Publications (2)

Publication Number Publication Date
US20210215481A1 true US20210215481A1 (en) 2021-07-15
US11561092B2 US11561092B2 (en) 2023-01-24

Family

ID=66184673

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/652,346 Active 2040-01-27 US11561092B2 (en) 2018-11-09 2019-03-01 Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network

Country Status (4)

Country Link
US (1) US11561092B2 (en)
EP (1) EP3680609A4 (en)
CN (1) CN109685762A (en)
WO (1) WO2020093630A1 (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11145082B2 (en) * 2018-11-06 2021-10-12 Wuyi University Method for measuring antenna downtilt angle based on deep instance segmentation network
CN113538476A (en) * 2021-07-20 2021-10-22 大连民族大学 Deep learning image segmentation method and system based on edge feature extraction
CN114322793A (en) * 2022-03-16 2022-04-12 科大天工智能装备技术(天津)有限公司 Workpiece size measurement method, device and storage medium based on global segmentation network
CN114549958A (en) * 2022-02-24 2022-05-27 四川大学 A nighttime and camouflaged target detection method based on contextual information perception mechanism
CN114677510A (en) * 2022-03-22 2022-06-28 中南大学 Feature map up-sampling method, small target semantic segmentation method and imaging method
CN114743010A (en) * 2022-06-13 2022-07-12 山东科技大学 Ultrahigh voltage power transmission line point cloud data semantic segmentation method based on deep learning
CN114993243A (en) * 2022-08-04 2022-09-02 深圳粤讯通信科技有限公司 Antenna attitude monitoring and early warning system based on Internet of things
CN115187959A (en) * 2022-07-14 2022-10-14 清华大学 Method and system for landing flying vehicle in mountainous region based on binocular vision
US20220351497A1 (en) * 2019-03-22 2022-11-03 Nec Corporation Image processing system, image processing device, image processing method, and computer-readable medium
CN115393725A (en) * 2022-10-26 2022-11-25 西南科技大学 A Bridge Crack Identification Method Based on Feature Enhancement and Semantic Segmentation
US20230124572A1 (en) * 2020-01-08 2023-04-20 Google Llc Translation of text depicted in images
CN116052003A (en) * 2023-02-07 2023-05-02 中科星图数字地球合肥有限公司 Method and device for measuring antenna angle information and related equipment
CN116258817A (en) * 2023-02-16 2023-06-13 浙江大学 A method and system for constructing an autonomous driving digital twin scene based on multi-view 3D reconstruction
CN116612282A (en) * 2023-05-26 2023-08-18 东风悦享科技有限公司 A new fast semantic segmentation method, system and storage medium for unmanned public transport system
US20230281819A1 (en) * 2022-02-18 2023-09-07 Techcyte, Inc. Image analysis for identifying objects and classifying background exclusions
US20230334876A1 (en) * 2022-04-14 2023-10-19 Toyota Research Institute, Inc. End-to-end learned lane boundary detection based on a transformer
CN117078760A (en) * 2023-09-18 2023-11-17 北方民族大学 Valve body center positioning method based on image processing
CN117994797A (en) * 2024-04-02 2024-05-07 杭州海康威视数字技术股份有限公司 Water gauge reading method and device, storage medium and electronic equipment
CN118823551A (en) * 2024-09-19 2024-10-22 干将未来科技产业(杭州)有限公司 An AI algorithm optimization system and method based on deep learning technology
CN119741293A (en) * 2024-12-26 2025-04-01 中国舰船研究设计中心 A rapid assessment method for ship corrosion based on image recognition

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110647794B (en) * 2019-07-12 2023-01-03 五邑大学 Attention mechanism-based multi-scale SAR image recognition method and device
CN110688904A (en) * 2019-08-30 2020-01-14 中通服建设有限公司 Method and device for base station antenna engineering parameter survey based on 5G UAV
CN110930417B (en) * 2019-11-26 2023-08-08 腾讯科技(深圳)有限公司 Training method and device for image segmentation model, and image segmentation method and device
CN111340139B (en) * 2020-03-27 2024-03-05 中国科学院微电子研究所 A method and device for judging the complexity of image content
CN111539983B (en) * 2020-04-15 2023-10-20 上海交通大学 Moving object segmentation method and system based on depth image
CN111768392B (en) * 2020-06-30 2022-10-14 创新奇智(广州)科技有限公司 Target detection method and device, electronic equipment and storage medium
CN111860297A (en) * 2020-07-17 2020-10-30 厦门理工学院 A SLAM loopback detection method applied to indoor fixed space
CN112171668A (en) * 2020-09-21 2021-01-05 河南颂达信息技术有限公司 Rail-mounted robot anti-jamming detection method and device based on artificial intelligence
CN112329808A (en) * 2020-09-25 2021-02-05 武汉光谷信息技术股份有限公司 A Optimization Method and System of Deeplab Semantic Segmentation Algorithm
CN112132965B (en) * 2020-09-25 2024-03-26 中国矿业大学 Multi-scale characterization method for rock-soil body hole-crack structure
CN112965369B (en) * 2021-01-29 2022-02-01 电子科技大学 Distributed deployment method of three-dimensional wireless communication network of unmanned aerial vehicle
CN112784857B (en) * 2021-01-29 2022-11-04 北京三快在线科技有限公司 Model training and image processing method and device
CN113239815B (en) * 2021-05-17 2022-09-06 广东工业大学 Remote sensing image classification method, device and equipment based on real semantic full-network learning
CN113450311B (en) * 2021-06-01 2023-01-13 国网河南省电力公司漯河供电公司 Pin screw defect detection method and system based on semantic segmentation and spatial relationship
CN114413855B (en) * 2022-01-25 2023-04-07 湖南大学 Method and system for detecting downtilt of mobile base station antenna based on image recognition
CN116758147B (en) * 2022-03-02 2025-11-11 上海交通大学 Plate-shaped antenna downward inclination angle measurement method and system based on mobile phone information fusion
CN114638991A (en) * 2022-03-18 2022-06-17 聊城大学 A smart retail settlement method based on YOLOv4-MobileNetv3 and PyQt5
CN115170667B (en) * 2022-07-15 2025-06-17 浙江大学 A water quality transparency detection method based on deep learning using a waterless Secchi disk
CN116051843A (en) * 2023-02-07 2023-05-02 智洋创新科技股份有限公司 A method and system for quantitative analysis of molten stream tilt
CN120121470B (en) * 2025-05-15 2025-07-08 中南大学 Method and device for identifying slurry fluidity image suitable for non-vertical shooting
CN120947588A (en) * 2025-10-15 2025-11-14 浙江点创信息科技有限公司 Methods, devices, computer equipment, and storage media for verifying base station antenna attitude.

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150278632A1 (en) * 2014-03-28 2015-10-01 Xerox Corporation Extending data-driven detection to the prediction of object part locations
US9596617B2 (en) * 2015-04-14 2017-03-14 ETAK Systems, LLC Unmanned aerial vehicle-based systems and methods associated with cell sites and cell towers
US20170077586A1 (en) * 2014-05-27 2017-03-16 Huawei Technologies Co., Ltd. Method and Device for Obtaining Antenna Engineering Parameter and System
US20180089505A1 (en) * 2016-09-23 2018-03-29 Samsung Electronics Co., Ltd. System and method for deep network fusion for fast and robust object detection
US20180137642A1 (en) * 2016-11-15 2018-05-17 Magic Leap, Inc. Deep learning system for cuboid detection
US20180218351A1 (en) * 2017-01-31 2018-08-02 Focal Systems, Inc. Automated checkout system through mobile shopping units
US20180260415A1 (en) * 2017-03-10 2018-09-13 Xerox Corporation Instance-level image retrieval with a region proposal network
US20190015059A1 (en) * 2017-07-17 2019-01-17 Siemens Healthcare Gmbh Semantic segmentation for cancer detection in digital breast tomosynthesis
US20190043003A1 (en) * 2017-08-07 2019-02-07 Standard Cognition, Corp Predicting inventory events using foreground/background processing
US20190130189A1 (en) * 2017-10-30 2019-05-02 Qualcomm Incorporated Suppressing duplicated bounding boxes from object detection in a video analytics system
US20190213438A1 (en) * 2018-01-05 2019-07-11 Irobot Corporation Mobile Cleaning Robot Artificial Intelligence for Situational Awareness
US20200090519A1 (en) * 2018-09-13 2020-03-19 Volvo Car Corporation System and method for camera or sensor-based parking spot detection and identification
US10872228B1 (en) * 2017-09-27 2020-12-22 Apple Inc. Three-dimensional object detection
US20220004770A1 (en) * 2018-10-31 2022-01-06 Arcus Holding A/S Object detection using a combination of deep learning and non-deep learning techniques

Family Cites Families (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE60317625T2 (en) * 2002-01-24 2008-11-27 Telecom Italia S.P.A. Measurement of antenna parameters of a base station for cellular telephony
US8374979B2 (en) * 2009-11-18 2013-02-12 Nec Laboratories America, Inc. Fast image parsing by graph adaptive dynamic programming (GADP) performing classification, detection, and segmentation simultaneously
US20110150317A1 (en) * 2009-12-17 2011-06-23 Electronics And Telecommunications Research Institute System and method for automatically measuring antenna characteristics
US9109889B2 (en) * 2011-06-24 2015-08-18 Trimble Navigation Limited Determining tilt angle and tilt direction using image processing
AT511191B1 (en) * 2011-07-01 2012-10-15 Thomas Dr Neubauer METHOD AND DEVICE FOR DETERMINING AND STORING THE POSITION AND ORIENTATION OF ANTENNA STRUCTURES
CN103630107B (en) 2012-08-23 2015-09-16 北京交通大学 A kind of antenna for base station inclination angle measurement method and data processing method
CN104504381B (en) 2015-01-09 2017-12-26 新智认知数据服务有限公司 Non-rigid object detection method and its system
US9855658B2 (en) * 2015-03-19 2018-01-02 Rahul Babu Drone assisted adaptive robot control
CN104978580B (en) * 2015-06-15 2018-05-04 国网山东省电力公司电力科学研究院 A kind of insulator recognition methods for unmanned plane inspection transmission line of electricity
US9918235B2 (en) 2015-11-24 2018-03-13 Verizon Patent And Licensing Inc. Adaptive antenna operation for UAVs using terrestrial cellular networks
CN106851665A (en) * 2015-12-07 2017-06-13 上海无线通信研究中心 The downdip adjusting method of antenna and base station
CN107664491B (en) * 2016-07-28 2020-06-05 中国电信股份有限公司 Method, device and system for measuring downward inclination angle of base station antenna
CN106709568B (en) * 2016-12-16 2019-03-22 北京工业大学 The object detection and semantic segmentation method of RGB-D image based on deep layer convolutional network
CN106683091B (en) * 2017-01-06 2019-09-24 北京理工大学 A kind of target classification and attitude detecting method based on depth convolutional neural networks
US10565787B1 (en) * 2017-01-27 2020-02-18 NHIAE Group, LLC Systems and methods for enhanced 3D modeling of a complex object
US10402689B1 (en) * 2017-04-04 2019-09-03 Snap Inc. Generating an image mask using machine learning
US11257198B1 (en) * 2017-04-28 2022-02-22 Digimarc Corporation Detection of encoded signals and icons
CN110809784B (en) * 2017-09-27 2021-04-20 谷歌有限责任公司 End-to-end network model for high-resolution image segmentation
CN107830846B (en) * 2017-09-30 2020-04-10 杭州艾航科技有限公司 Method for measuring angle of communication tower antenna by using unmanned aerial vehicle and convolutional neural network
US10649459B2 (en) * 2018-04-26 2020-05-12 Zoox, Inc. Data segmentation using masks

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150278632A1 (en) * 2014-03-28 2015-10-01 Xerox Corporation Extending data-driven detection to the prediction of object part locations
US20170077586A1 (en) * 2014-05-27 2017-03-16 Huawei Technologies Co., Ltd. Method and Device for Obtaining Antenna Engineering Parameter and System
US9596617B2 (en) * 2015-04-14 2017-03-14 ETAK Systems, LLC Unmanned aerial vehicle-based systems and methods associated with cell sites and cell towers
US20180089505A1 (en) * 2016-09-23 2018-03-29 Samsung Electronics Co., Ltd. System and method for deep network fusion for fast and robust object detection
US20180137642A1 (en) * 2016-11-15 2018-05-17 Magic Leap, Inc. Deep learning system for cuboid detection
US20180218351A1 (en) * 2017-01-31 2018-08-02 Focal Systems, Inc. Automated checkout system through mobile shopping units
US20180260415A1 (en) * 2017-03-10 2018-09-13 Xerox Corporation Instance-level image retrieval with a region proposal network
US20190015059A1 (en) * 2017-07-17 2019-01-17 Siemens Healthcare Gmbh Semantic segmentation for cancer detection in digital breast tomosynthesis
US20190043003A1 (en) * 2017-08-07 2019-02-07 Standard Cognition, Corp Predicting inventory events using foreground/background processing
US10872228B1 (en) * 2017-09-27 2020-12-22 Apple Inc. Three-dimensional object detection
US20190130189A1 (en) * 2017-10-30 2019-05-02 Qualcomm Incorporated Suppressing duplicated bounding boxes from object detection in a video analytics system
US20190213438A1 (en) * 2018-01-05 2019-07-11 Irobot Corporation Mobile Cleaning Robot Artificial Intelligence for Situational Awareness
US20200090519A1 (en) * 2018-09-13 2020-03-19 Volvo Car Corporation System and method for camera or sensor-based parking spot detection and identification
US20220004770A1 (en) * 2018-10-31 2022-01-06 Arcus Holding A/S Object detection using a combination of deep learning and non-deep learning techniques

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11145082B2 (en) * 2018-11-06 2021-10-12 Wuyi University Method for measuring antenna downtilt angle based on deep instance segmentation network
US20220351497A1 (en) * 2019-03-22 2022-11-03 Nec Corporation Image processing system, image processing device, image processing method, and computer-readable medium
US11748977B2 (en) * 2019-03-22 2023-09-05 Nec Corporation Image processing system, image processing device, image processing method, and computer-readable medium
US12217017B2 (en) * 2020-01-08 2025-02-04 Google Llc Translation of text depicted in images
US20230124572A1 (en) * 2020-01-08 2023-04-20 Google Llc Translation of text depicted in images
CN113538476A (en) * 2021-07-20 2021-10-22 大连民族大学 Deep learning image segmentation method and system based on edge feature extraction
US20230281819A1 (en) * 2022-02-18 2023-09-07 Techcyte, Inc. Image analysis for identifying objects and classifying background exclusions
CN114549958A (en) * 2022-02-24 2022-05-27 四川大学 A nighttime and camouflaged target detection method based on contextual information perception mechanism
CN114322793A (en) * 2022-03-16 2022-04-12 科大天工智能装备技术(天津)有限公司 Workpiece size measurement method, device and storage medium based on global segmentation network
CN114677510A (en) * 2022-03-22 2022-06-28 中南大学 Feature map up-sampling method, small target semantic segmentation method and imaging method
US20230334876A1 (en) * 2022-04-14 2023-10-19 Toyota Research Institute, Inc. End-to-end learned lane boundary detection based on a transformer
US12159465B2 (en) * 2022-04-14 2024-12-03 Toyota Research Institute, Inc. End-to-end learned lane boundary detection based on a transformer
CN114743010A (en) * 2022-06-13 2022-07-12 山东科技大学 Ultrahigh voltage power transmission line point cloud data semantic segmentation method based on deep learning
CN115187959A (en) * 2022-07-14 2022-10-14 清华大学 Method and system for landing flying vehicle in mountainous region based on binocular vision
CN114993243A (en) * 2022-08-04 2022-09-02 深圳粤讯通信科技有限公司 Antenna attitude monitoring and early warning system based on Internet of things
CN115393725A (en) * 2022-10-26 2022-11-25 西南科技大学 A Bridge Crack Identification Method Based on Feature Enhancement and Semantic Segmentation
CN116052003A (en) * 2023-02-07 2023-05-02 中科星图数字地球合肥有限公司 Method and device for measuring antenna angle information and related equipment
CN116258817A (en) * 2023-02-16 2023-06-13 浙江大学 A method and system for constructing an autonomous driving digital twin scene based on multi-view 3D reconstruction
CN116612282A (en) * 2023-05-26 2023-08-18 东风悦享科技有限公司 A new fast semantic segmentation method, system and storage medium for unmanned public transport system
CN117078760A (en) * 2023-09-18 2023-11-17 北方民族大学 Valve body center positioning method based on image processing
CN117994797A (en) * 2024-04-02 2024-05-07 杭州海康威视数字技术股份有限公司 Water gauge reading method and device, storage medium and electronic equipment
CN118823551A (en) * 2024-09-19 2024-10-22 干将未来科技产业(杭州)有限公司 An AI algorithm optimization system and method based on deep learning technology
CN119741293A (en) * 2024-12-26 2025-04-01 中国舰船研究设计中心 A rapid assessment method for ship corrosion based on image recognition

Also Published As

Publication number Publication date
US11561092B2 (en) 2023-01-24
EP3680609A4 (en) 2021-02-24
EP3680609A1 (en) 2020-07-15
CN109685762A (en) 2019-04-26
WO2020093630A1 (en) 2020-05-14

Similar Documents

Publication Publication Date Title
US11561092B2 (en) Method for measuring antenna downtilt angle based on multi-scale deep semantic segmentation network
CN114488194B (en) A method for target detection and recognition on structured roads for intelligent driving vehicles
Yan et al. Automated asphalt highway pavement crack detection based on deformable single shot multi-box detector under a complex environment
EP3997616B1 (en) Object-based change detection using a neural network
US12214809B2 (en) Instance segmentation imaging system
Wang et al. A center location algorithm for tropical cyclone in satellite infrared images
KR102507501B1 (en) Artificial Intelligence-based Water Quality Contaminant Monitoring System and Method
CN108052940A (en) SAR remote sensing images waterborne target detection methods based on deep learning
CN114252868B (en) Laser radar calibration method, device, computer equipment and storage medium
CN111553280A (en) Target part identification method based on deep learning
CN116168246B (en) A method, device, equipment and medium for identifying abandoned slag sites in railway engineering
CN114252883A (en) Target detection method, apparatus, computer device and medium
CN114252884A (en) Method and device for positioning and monitoring roadside radar, computer equipment and storage medium
CN116246096A (en) A Point Cloud 3D Object Detection Method Based on Foreground Enhanced Knowledge Distillation
Wu et al. Multimodal collaboration networks for geospatial vehicle detection in dense, occluded, and large-scale events
CN114255264B (en) Multi-base-station registration method and device, computer equipment and storage medium
Shi et al. Fine object change detection based on vector boundary and deep learning with high-resolution remote sensing images
CN114252859B (en) Method, device, computer equipment and storage medium for determining target area
Han et al. Targets mask U-Net for wind turbines detection in remote sensing images
CN113298042A (en) Method and device for processing remote sensing image data, storage medium and computer equipment
Yang et al. Sardet-cl: Self-supervised contrastive learning with feature enhancement and imaging mechanism constraints for sar target detection
CN119716839A (en) CN-YOLOv-based high-precision detection method for multi-mode data fusion anti-unmanned aerial vehicle
CN119964098A (en) A road surface abnormality detection method, device, storage medium and electronic device
CN118799827A (en) An intelligent detection and positioning method for sea surface ship targets using dual-polarization SAR images
CN115082786A (en) Remote sensing image target detection method under condition of small amount of marks

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

AS Assignment

Owner name: WUYI UNIVERSITY, CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ZHAI, YIKUI;ZHOU, JIHUA;WU, YUETING;AND OTHERS;REEL/FRAME:056108/0041

Effective date: 20200312

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE