WO2022052032A1 - 图像的分割方法、装置和图像的三维重建方法、装置 - Google Patents
图像的分割方法、装置和图像的三维重建方法、装置 Download PDFInfo
- Publication number
- WO2022052032A1 WO2022052032A1 PCT/CN2020/114752 CN2020114752W WO2022052032A1 WO 2022052032 A1 WO2022052032 A1 WO 2022052032A1 CN 2020114752 W CN2020114752 W CN 2020114752W WO 2022052032 A1 WO2022052032 A1 WO 2022052032A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- pixel
- image
- sub
- color gamut
- pixels
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/28—Quantising the image, e.g. histogram thresholding for discrimination between background and foreground patterns
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/75—Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
- G06V10/751—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
Definitions
- the present disclosure relates to the field of computer technology, and in particular, to an image segmentation method, an image segmentation device, an image three-dimensional reconstruction method, an image three-dimensional reconstruction device, an electronic device, a wearable device, and a non-volatile computer-readable storage medium.
- Image segmentation is a fundamental problem in image processing and computer vision because it is a key process in many applications of region-specific extraction.
- an image segmentation method including: dividing each pixel in an image to be segmented into different pixel sets according to the color gamut range to which the pixel value belongs; Matching situation between pixels in each pixel set; image segmentation is performed on the to-be-segmented image according to the matching situation.
- the method further includes: in a coordinate system with the red component, green component and blue component of the pixel value as variables, dividing the color gamut cube formed by the red component, the green component and the blue component into multiple A color gamut sub-cube as each color gamut range.
- the method further includes: determining one of the vertexes of the color gamut cube, the center point of each color gamut sub-cube, and the mean point of each color gamut sub-cube contained in each color gamut sub-cube as the corresponding color
- the characteristic pixel value of the gamut range according to the characteristic pixel value, the color gamut range to which the pixel value of each pixel in the to-be-segmented image belongs is determined.
- determining the matching situation between the pixels in each pixel set includes: selecting a pixel in any pixel set as a seed pixel; calculating the other pixels in the pixel set. The difference between the pixel value and the pixel value of the seed pixel; according to the difference, it is determined whether the other pixels match the seed pixel.
- the determining, according to the difference, whether the other pixel matches the seed pixel includes: using a membership function to determine, according to the difference, a fuzzy set to which the other pixel belongs; and according to the determined fuzzy set Set to determine whether the other pixels match the seed pixel.
- the pixel value includes a red component, a green component and a blue component
- using a membership function to determine the fuzzy set to which the other pixel belongs includes: according to the difference of the red component, The difference of the green component and the difference of the blue component determines the fuzzy set to which the red component, the green component and the blue component of the other pixels belong.
- selecting a pixel in any pixel set as a seed pixel includes: according to the difference between the pixel value of each pixel in any pixel set and the characteristic pixel value of the color gamut range to which the pixel set belongs, to The pixels in the pixel set are sorted, and the characteristic pixel value is the vertex of the color gamut cube contained in the color gamut sub-cube corresponding to the color gamut range to which it belongs, the center point of the corresponding color gamut sub-cube, One of the mean points of the corresponding color gamut sub-cubes; according to the sorting result, each pixel in the pixel set is sequentially selected as the seed pixel.
- performing image segmentation on the to-be-segmented image according to the matching situation includes: generating a plurality of sub-images according to the pixels and their matching pixels; Perform merging processing on the plurality of sub-images; and determine an image segmentation result according to the merging result.
- the combining the multiple sub-images according to the overlap between the sub-images includes: calculating the number of pixels included in the intersection of the first sub-image and the second sub-image; The ratio of the number of pixels contained in the first sub-image to the number of pixels contained in the first sub-image is used to determine an overlap parameter for judging the overlap situation; when the overlap parameter is greater than a threshold, compare the first sub-image with the The second sub-image is merged.
- the method further includes: determining an interference pixel according to the pixel value distribution of each pixel in the original image; determining a matching pixel of the interference pixel according to the pixel value of each pixel in the original image; In the image, the interference pixels and their matching pixels are removed to obtain the to-be-segmented image.
- the to-be-segmented image is a two-dimensional image generated according to acquired underwater sonar data.
- a 3D reconstruction method of an image including: according to the segmentation method of any one of the above embodiments, performing segmentation processing on an image to be segmented; and performing 3D reconstruction according to the segmentation processing result to obtain a 3D image.
- an image segmentation apparatus including at least one processor, the processor is configured to perform the following steps: The pixels are divided into different pixel sets; according to the pixel value, the matching situation between each pixel in each pixel set is determined respectively; according to the matching situation, image segmentation is performed on the to-be-segmented image.
- the method further includes: in a coordinate system with the red component, green component and blue component of the pixel value as variables, dividing the color gamut cube formed by the red component, the green component and the blue component into multiple A color gamut sub-cube as each color gamut range.
- the processor is further configured to perform the step of: converting the vertexes of the color gamut cubes, the center points of the color gamut sub-cubes, and the mean points of the color gamut sub-cubes contained in the color gamut sub-cubes into the following steps: One is determined as the characteristic pixel value of the corresponding color gamut range; according to the characteristic pixel value, the color gamut range to which the pixel value of each pixel in the image to be divided belongs is determined.
- determining the matching situation between the pixels in each pixel set includes: selecting a pixel in any pixel set as a seed pixel; calculating the other pixels in the pixel set. The difference between the pixel value and the pixel value of the seed pixel; according to the difference, it is determined whether the other pixels match the seed pixel.
- the determining, according to the difference, whether the other pixel matches the seed pixel includes: using a membership function to determine, according to the difference, a fuzzy set to which the other pixel belongs; and according to the determined fuzzy set Set to determine whether the other pixels match the seed pixel.
- the pixel value includes a red component, a green component and a blue component
- using a membership function to determine the fuzzy set to which the other pixel belongs includes: according to the difference of the red component, The difference of the green component and the difference of the blue component determines the fuzzy set to which the red component, the green component and the blue component of the other pixels belong.
- selecting a pixel in any pixel set as a seed pixel includes: according to the difference between the pixel value of each pixel in any pixel set and the characteristic pixel value of the color gamut range to which the pixel set belongs, to The pixels in the pixel set are sorted, and the characteristic pixel value is the vertex of the color gamut cube contained in the color gamut sub-cube corresponding to the color gamut range to which it belongs, the center point of the corresponding color gamut sub-cube, One of the mean points of the corresponding color gamut sub-cubes; according to the sorting result, each pixel in the pixel set is sequentially selected as the seed pixel.
- performing image segmentation on the to-be-segmented image according to the matching situation includes: generating a plurality of sub-images according to the pixels and their matching pixels; Perform merging processing on the plurality of sub-images; and determine an image segmentation result according to the merging result.
- the combining the multiple sub-images according to the overlap between the sub-images includes: calculating the number of pixels included in the intersection of the first sub-image and the second sub-image; The ratio of the number of pixels contained in the first sub-image to the number of pixels contained in the first sub-image is used to determine an overlap parameter for judging the overlap situation; when the overlap parameter is greater than a threshold, compare the first sub-image with the The second sub-image is merged.
- the processor is further configured to perform the steps of: determining interfering pixels according to the pixel value distribution of each pixel in the original image; determining the interfering pixel according to the pixel value of each pixel in the original image In the original image, the interference pixels and their matching pixels are removed, and the image to be segmented is obtained.
- the to-be-segmented image is a two-dimensional image generated according to acquired underwater sonar data.
- an apparatus for three-dimensional reconstruction of an image comprising at least one processor configured to perform the steps of: segmenting an image to be segmented according to the segmentation method of any one of the foregoing embodiments Processing; 3D reconstruction is performed according to the segmentation processing result, and a 3D image is obtained.
- a wearable device comprising: a three-dimensional reconstruction apparatus for an image in any one of the above embodiments; a display screen for displaying a three-dimensional image acquired by the three-dimensional reconstruction apparatus.
- the three-dimensional reconstruction device generates the image to be segmented according to the acquired underwater sonar data, and reconstructs the three-dimensional image according to the segmentation result of the image to be segmented.
- an electronic device comprising: a memory; and a processor coupled to the memory, the processor configured to execute the above based on instructions stored in the memory device.
- a non-volatile computer-readable storage medium on which a computer program is stored, and when the program is executed by a processor, implements the image segmentation method according to any of the foregoing embodiments or 3D reconstruction of images.
- FIG. 1 shows a flowchart of some embodiments of a segmentation method of an image according to the present disclosure
- FIG. 2 shows a flowchart of other embodiments of the image segmentation method according to the present disclosure
- FIG. 3 shows a schematic diagram of some embodiments of an image segmentation method according to the present disclosure
- FIG. 4 shows a flowchart of some embodiments of step 120 in FIG. 1;
- FIG. 5 shows a schematic diagram of other embodiments of the image segmentation method according to the present disclosure.
- FIG. 6 shows a flowchart according to some embodiments of step 130 in FIG. 1;
- FIG. 7 shows a schematic diagram of further embodiments of the image segmentation method according to the present disclosure.
- FIG. 8 shows a schematic diagram of some embodiments of a wearable device according to the present disclosure
- FIG. 9 illustrates a block diagram of some embodiments of wearable devices according to the present disclosure.
- Figure 10 shows a block diagram of some embodiments of the electronic device of the present disclosure
- FIG. 11 shows a block diagram of further embodiments of the electronic device of the present disclosure.
- Image segmentation methods using binarization methods or based on RGB (Red, Green, Blue, red, green, blue) color space are based on a segmentation principle, that is, each pixel is classified into a unique subset.
- RGB Red, Green, Blue, red, green, blue
- the accuracy of image segmentation is not high, which can lead to poorer effect of subsequent processing.
- the seawater, the seafloor and the target have completely different characteristics.
- the three also have different opacities and colors in the 3D visualization imaging processing after image segmentation. In this way, the above-mentioned segmentation method often causes the target at the segmentation line of each region to be covered, resulting in poor three-dimensional imaging effect.
- the present disclosure proposes a technical solution for image segmentation.
- the technical solution divides each pixel whose pixel value belongs to the same color gamut range into the same pixel set, and performs pixel value matching processing in each pixel set, so as to determine the image segmentation result.
- the division of the color gamut space can be refined, the recognition rate of different image areas can be improved, and the accuracy of image segmentation can be improved.
- the technical solutions of the present disclosure can be implemented through the following embodiments.
- FIG. 1 shows a flowchart of some embodiments of a method of segmentation of an image according to the present disclosure.
- the method includes: step 110 , dividing different sets of pixels; step 120 , determining the matching situation in the set; and step 130 , segmenting the image according to the matching situation.
- each pixel in the image to be segmented is divided into different pixel sets according to the color gamut range to which the pixel value belongs.
- the image to be segmented is a two-dimensional image generated according to the acquired underwater sonar data.
- the pixel set to which each pixel belongs can be determined according to the difference between the pixel value of each pixel and the characteristic pixel of each color gamut range, so as to realize the classification of the pixels.
- similar pixels can be classified into one category to achieve preliminary image segmentation; further matching of pixel values within each category can improve the accuracy of image segmentation.
- the entire color gamut may be modeled, and the color gamut may be divided into multiple color gamut ranges based on the modeling. On this basis, the color gamut range to which the pixel value of each pixel in the to-be-segmented image belongs can be determined. For example, color gamut modeling and division can be achieved by the embodiment in FIG. 2 .
- FIG. 2 shows a flowchart of further embodiments of the image segmentation method according to the present disclosure.
- the method further includes: step 210 , modeling a color gamut cube; step 220 , dividing the color gamut range; step 230 , determining characteristic pixel values; and step 240 , determining the color gamut range to which the pixel values belong.
- step 210 according to the value ranges of the red component, the green component and the blue component of the pixel value, in the coordinate system with the red component, the green component and the blue component as variables, the entire color gamut is modeled as a color Domain cube.
- the color gamut cube is divided into a plurality of color gamut sub-cubes as each color gamut range.
- a gamut cube can be modeled by the embodiment in FIG. 3 and divided into gamut sub-cubes.
- FIG. 3 shows a schematic diagram of some embodiments of a segmentation method of an image according to the present disclosure.
- the three coordinate values of the coordinate system respectively represent the values of the three components R (red component), G (green component) and B (blue component) of the pixel value.
- the value range of each component is [0, 255], and the pixel value (0, 0, 0) at the origin P5 represents black.
- the cubes with P 1 to P 8 as vertices are the color gamut cubes corresponding to the entire color gamut, that is, the color gamut space.
- the color gamut cube may be divided into 8 color gamut sub-cubes including vertices P 1 to P 8 respectively in the direction of 3 components according to the pixel value interval of 127, that is, the color gamut subspace.
- the vertices P 1 to P 8 represent blue, pink, white, cyan, black, red, yellow, and green, respectively.
- each color gamut sub-cube represents a different color gamut range, that is, pixel values within the same color gamut range have similar color information.
- the color gamut range to which the pixel value of each pixel in the to-be-segmented image belongs may be determined by using the embodiment in FIG. 2 .
- the vertices of the color gamut cubes included in each color gamut sub-cube are determined as characteristic pixel values of the corresponding color gamut range.
- any pixel value in each color gamut sub-cube that can represent the corresponding color gamut range can be determined as a characteristic pixel value, such as a vertex, a center point, a mean point, and the like.
- the color gamut range to which the pixel value of each pixel in the image to be divided belongs is determined according to the characteristic pixel value. For example, it can be determined which color gamut range the pixel belongs to according to the difference between the pixel and each characteristic pixel.
- the pixel value of the pixel can be used as a multi-dimensional feature vector, and the difference between the pixels can be determined by calculating the similarity between the feature vectors (such as Euclidean distance, Mingshi distance, Mahalanobis distance, etc.).
- the distances between the point corresponding to each pixel in the image to be segmented and the 8 vertices of the color gamut cube are calculated respectively.
- the distance is calculated by the following formula:
- p(n,m) RGB is the point defined by any pixel in the color image to be segmented in the RGB coordinate system
- v i RGB is the vertex i of the color gamut cube
- d i (n, m) is the distance between the point and the vertex the distance.
- the pixels in the image to be segmented are spatially classified one by one.
- the amount of pixels each subspace contains can be clearly located. For example, no post-processing is done for gamut subcubes that do not contain any pixels. Moreover, after the spatial classification is performed, repeated calculations for the same pixels can be avoided in subsequent processing (such as blurred color extraction).
- image segmentation can be continued through the embodiment in FIG. 1 .
- step 120 according to the pixel value, the matching situation between each pixel in each pixel set is determined respectively.
- only purposeful matching of pixels belonging to the same color gamut range (as if belonging to one color type) can improve the efficiency and accuracy of matching, thereby improving the matching results based on The efficiency and accuracy of image segmentation.
- step 120 may be implemented by the embodiment in FIG. 4 .
- FIG. 4 shows a flowchart of some embodiments of step 120 in FIG. 1 .
- step 120 includes: step 1210, selecting seed pixels; step 1220, calculating pixel value differences; step 1230, determining the fuzzy set to which the pixel belongs; and step 1240, determining whether the pixels match.
- a pixel is selected from any pixel set as a seed pixel.
- the pixels in the pixel set are sorted; according to the sorting result, the pixels are sorted in sequence. Each pixel in the pixel set is selected as a seed pixel.
- the pixels in each pixel set are sorted from small to large according to their distances from the corresponding gamut sub-cube containing vertices. Starting from the closest point, each pixel is used as a seed pixel for fuzzy color extraction in a round-robin manner.
- the sub-pixel if a sub-pixel cannot find a matching pixel, the sub-pixel is discarded, and the next pixel within the corresponding gamut sub-cube is used as the sub-pixel.
- the sub-pixels are selected for matching in turn. For example, the selection and matching of seed pixels can be performed simultaneously within 8 gamut subcubes.
- step 1220 the difference between the pixel values of other pixels in the pixel set and the pixel values of the seed pixels is calculated.
- fuzzy color extraction is performed through steps 1230 and 1240 to determine the matching condition.
- step 1230 the fuzzy set to which the difference belongs is determined using membership functions and fuzzy logic.
- the pixel value includes a red component, a green component, and a blue component
- the red, green, and blue components of other pixels are determined based on the difference in the red component, the difference in the green component, and the difference in the blue component, respectively.
- an FCE Fuzzy Color Extractor
- the components of its sub-pixels in the RGB space are p(n,m) R , p(n,m) G and p(n,m) B .
- the current pixel to be processed is seed, and its RGB components are seed R , seed G and seed B .
- the selection of the seed can be selected according to the needs of the algorithm, or determined according to the pixels in the image to be processed.
- M and N represent the size of the image (positive integer).
- the fuzzy set to which the color component difference belongs can be calculated by using a preset membership function according to the color component difference.
- the corresponding membership function of each fuzzy set can be determined by the embodiment in FIG. 5 .
- FIG. 5 shows a schematic diagram of other embodiments of the image segmentation method according to the present disclosure.
- the fuzzy sets to which the color component differences belong include the Zero set, the Negative set, and the Positive set.
- the three function curves correspond to the membership functions of the three fuzzy sets respectively.
- ⁇ 1 and ⁇ 2 are adjustable blur thresholds set according to the actual situation and prior knowledge.
- the matching situation of the pixels can be determined through step 1240 in FIG. 4 .
- step 1240 according to the determined fuzzy set, it is determined whether other pixels match the seed pixel.
- fuzzy logic can be:
- the language method is used to configure the fuzzy logic
- the input and output functions are relatively simple, and an accurate mathematical model is not required, thereby optimizing the amount of calculation.
- the fuzzy matching method has strong robustness and is suitable for solving the problems of nonlinearity, strong coupling and time-varying, and lag in the classification process, thereby improving the accuracy of image segmentation.
- Fuzzy matching method has strong fault tolerance ability and can adapt to the changes of the controlled object's own characteristics and environmental characteristics.
- the fuzzy color extraction algorithm is suitable for image segmentation processing in complex environments (such as underwater sonar data images), and can improve the accuracy of image segmentation.
- image segmentation can be performed through step 130 in FIG. 1 .
- step 130 image segmentation is performed on the image to be segmented according to the matching situation.
- image segmentation may be performed by the embodiment in FIG. 6 .
- FIG. 6 shows a flowchart according to some embodiments of step 130 in FIG. 1 .
- step 130 may include: step 1310, generating a plurality of sub-images; step 1320, merging the sub-images; and step 1330, determining a segmentation result.
- a plurality of sub-images are generated according to each pixel and its matching pixel. For example, after each fuzzy color extraction, a sub-image can be obtained based on a seed pixel and its matching pixels.
- sub-images corresponding to seed pixels with similar pixel values generally have overlapping parts. In cases where multiple sub-images share a common color region in RGB space, it may even occur that one sub-image completely covers another.
- step 1320 the multiple sub-images are merged according to the overlap between the sub-images.
- Two sub-images are considered to share a common area if they have spatial similarity and color similarity, and they can be joined together to form an image partition.
- the number of pixels included in the intersection of the first sub-image and the second sub-image is calculated; according to the ratio of the number of pixels included in the intersection to the number of pixels included in the first sub-image, an overlap parameter is determined as an overlap condition; When the parameter is greater than the threshold, the first sub-image is merged with the second sub-image.
- the overlap parameter can be determined using the following formula:
- NUM() is the number of pixels in parentheses.
- the overlap parameter the size of the common area of the two sub-images in the RGB space can be detected.
- the overlap parameter is greater than the threshold, the sub-images I SAMPLE (i) and I SAMPLE (l) are considered to be similar and can be merged.
- the threshold can be set in the algorithm. In order to make the image segmentation more accurate, the threshold can be set larger, such as 90, 100 and so on.
- step 1330 the image segmentation result is determined according to the merging result.
- the extraction and image segmentation of different regions in the image can be realized.
- the imaging environment in some cases where the imaging environment is more complicated, there are many unknown factors in the original image.
- the seabed environment is complex, the dynamic range of sonar data obtained by scanning is very small, and there will be a lot of interference in underwater sonar images. Therefore, the image can be preprocessed using logarithmic transformation to expand the dynamic range of the data and reduce interference.
- preprocessing such as denoising and contrast enhancement can also be performed on the original image.
- the interference in the image may be removed by the embodiment in FIG. 7 .
- FIG. 7 shows a schematic diagram of further embodiments of the image segmentation method according to the present disclosure.
- the method may further include: step 710 , determining interfering pixels; step 720 , determining matching pixels of the interfering pixels; and step 730 , removing interference.
- interference pixels are determined according to the pixel value distribution of each pixel in the original image. For example, interference pixels can be selected according to prior knowledge and actual needs.
- interfering pixels may be selected based on a priori knowledge. If the color range (eg, red domain, etc.) of the interference factor in the image has been determined, the pixels in the color range can be determined as the interference pixels.
- the color range eg, red domain, etc.
- a matching pixel of the interference pixel is determined according to the pixel value of each pixel in the original image.
- matching can be performed by the method in any of the above embodiments (eg, fuzzy color extraction).
- step 730 in the original image, the interference pixels and their matching pixels are removed to obtain the image to be segmented.
- an interference image I INT composed of interference pixels and matching pixels can be determined.
- the color image (matching pixel) close to the interference pixel in the original image can be removed through I SOURCE -I INT to obtain the required image I SAMPLE to be segmented.
- segmentation processing is performed on the image to be segmented; 3D reconstruction is performed according to the segmentation processing result to obtain a 3D image.
- a two-dimensional image obtained by performing image segmentation on an underwater sonar image different regions such as oceans, formations, and objects can be identified.
- the 3D structures can be reconstructed from the segmented 2D images (such as Unity 3D tools).
- the volume rendering technology can be used to realize the upper three-dimensional visualization effect.
- volume rendering process there is no need to construct the geometric image of the intermediate process, and only the 3D data volume needs to be processed to reveal its internal details.
- the operation of such three-dimensional reconstruction processing is simple and the conversion is quick.
- three-dimensional visualization can be realized by VTK (Visualization Toolkit).
- the water body can be separated from the bottom layer more effectively and the target objects (such as underwater buried mines, bombs, etc.) can be extracted more effectively, and the accuracy rate is high.
- the target objects such as underwater buried mines, bombs, etc.
- it can well solve the uncertainty and ambiguity in practical applications, and adapt to the different emphasis on color of different observers in different color spaces.
- the Unity platform is used to construct the 3D scene. Users can construct very complex 3D images and scenes within a short understanding time, greatly improving work efficiency.
- the processed 3D data is represented by appropriate geometry, color and brightness, and mapped to the 2D image plane.
- three-dimensional images can be rendered in a VR (Virtual Reality, virtual reality) head-mounted device, allowing users to watch sonar images in a virtual environment to enhance immersion.
- VR Virtual Reality, virtual reality
- the image segmentation apparatus includes at least one processor configured to perform the image segmentation method in any of the above embodiments.
- the apparatus for three-dimensional reconstruction of an image includes at least one processor, and the processor is configured to perform: the segmentation method of any one of the foregoing embodiments, to perform segmentation processing on the image to be segmented; image.
- FIG. 8 shows a schematic diagram of some embodiments of a wearable device according to the present disclosure.
- the wearable device can adopt a VR split machine structure, including: a PC (Personal Computer, personal computer) part (such as an image reconstruction device) and a VR head display part (such as a display screen).
- a PC Personal Computer, personal computer
- a VR head display part such as a display screen
- processing processes such as image preprocessing, image segmentation, and volume rendering may be completed in the PC part first, and then the obtained three-dimensional image is rendered into the VR head-mounted display part through a DP (Display Port, display interface).
- DP Display Port, display interface
- image preprocessing may include denoising, contrast enhancement and other processing
- image segmentation may include FCE processing in any of the above embodiments
- using Unity 3D to construct three-dimensional images and scenes use VTK to perform three-dimensional image visualization processing.
- image segmentation processing is performed on the image of the sonar data, it can be displayed in the virtual reality head-mounted device through volume rendering technology. In this way, users can observe the three-dimensional images of underwater sonar data water-strata-target objects in the VR scene.
- FIG. 9 illustrates a block diagram of some embodiments of a wearable device according to the present disclosure.
- the wearable device 9 includes: a three-dimensional reconstruction apparatus 91 for images in any of the foregoing embodiments; a display screen 92 for displaying a three-dimensional image acquired by the three-dimensional reconstruction apparatus 91 .
- the three-dimensional reconstruction device 91 generates the image to be segmented according to the acquired underwater sonar data, and reconstructs the three-dimensional image according to the segmentation result of the image to be segmented.
- FIG. 10 illustrates a block diagram of some embodiments of the electronic device of the present disclosure.
- the electronic device 10 of this embodiment includes: a memory U11 and a processor U12 coupled to the memory U11, the processor U12 is configured to execute any one of the present disclosure based on instructions stored in the memory U11 The image segmentation method or the three-dimensional reconstruction method of the image in the embodiment.
- the memory U11 may include, for example, a system memory, a fixed non-volatile storage medium, and the like.
- the system memory stores, for example, an operating system, an application program, a boot loader (Boot Loader), a database, and other programs.
- FIG. 11 shows a block diagram of further embodiments of the electronic device of the present disclosure.
- the electronic device 11 of this embodiment includes: a memory U10 and a processor U20 coupled to the memory U10, and the processor U20 is configured to execute any one of the foregoing embodiments based on instructions stored in the memory U10 The segmentation method of the image or the three-dimensional reconstruction method of the image.
- the memory U10 may include, for example, a system memory, a fixed non-volatile storage medium, and the like.
- the system memory stores, for example, an operating system, an application program, a boot loader (Boot Loader), and other programs. )
- the electronic device 6 may further include an input/output interface U30, a network interface U40, a storage interface U50, and the like. These interfaces U30, U40, U50 and the memory U10 and the processor U20 can be connected, for example, through a bus U60.
- the input and output interface U30 provides a connection interface for input and output devices such as a display, a mouse, a keyboard, a touch screen, a microphone, and a speaker.
- the network interface U40 provides connection interfaces for various networked devices.
- the storage interface U50 provides a connection interface for external storage devices such as SD cards and U disks.
- embodiments of the present disclosure may be provided as a method, system, or computer program product. Accordingly, the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product embodied on one or more computer-usable non-transitory storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein .
- computer-usable non-transitory storage media including, but not limited to, disk storage, CD-ROM, optical storage, etc.
- the methods and systems of the present disclosure may be implemented in many ways.
- the methods and systems of the present disclosure may be implemented in software, hardware, firmware, or any combination of software, hardware, and firmware.
- the above-described order of steps of the method is for illustration only, and the steps of the method of the present disclosure are not limited to the order specifically described above unless specifically stated otherwise.
- the present disclosure can also be implemented as programs recorded in a recording medium, the programs including machine-readable instructions for implementing methods according to the present disclosure.
- the present disclosure also covers a recording medium storing a program for executing the method according to the present disclosure.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (18)
- 一种图像的分割方法,包括:根据像素值所属的色域范围,将待分割图像中的各像素划分到不同的像素集合中;根据像素值,分别确定每个像素集合内的各像素之间的匹配情况;根据所述匹配情况,对所述待分割图像进行图像分割。
- 根据权利要求1所述的分割方法,还包括:在以像素值的红色分量、绿色分量和蓝色分量为变量的坐标系中,将由红色分量、绿色分量和蓝色分量构成的色域立方体划分为多个色域子立方体,作为各色域范围。
- 根据权利要求2所述的分割方法,还包括:将各色域子立方体包含的所述色域立方体的顶点、各色域子立方体的中心点、各色域子立方体的均值点中的一个,确定为相应的色域范围的特征像素值;根据所述特征像素值,确定所述待分割图像中各像素的像素值所属的色域范围。
- 根据权利要求1所述的分割方法,其中,所述根据像素值,分别确定每个像素集合内的各像素之间的匹配情况包括:在任一像素集合中选取一个像素,作为种子像素;计算该像素集合中其他像素的像素值与所述种子像素的像素值的差异;根据所述差异,确定所述其他像素是否与所述种子像素匹配。
- 根据权利要求4所述的分割方法,其中,所述根据所述差异,确定所述其他像素是否与所述种子像素匹配包括:利用隶属函数,确定所述差异属于的模糊集合;根据确定的模糊集合和模糊逻辑,确定所述其他像素是否与所述种子像素匹配。
- 根据权利要求5所述的分割方法,其中,所述像素值包括红色分量、绿色分量和蓝色分量,所述根据所述差异,利用隶属函数,确定所述其他像素属于的模糊集合包括:分别根据红色分量的差异、绿色分量的差异和蓝色分量的差异,确定所述其他像素的红色分量、绿色分量和蓝色分量属于的模糊集合。
- 根据权利要求4所述的分割方法,其中,所述在任一像素集合中选取一个像素,作为种子像素包括:根据任一像素集合中各像素的像素值与该像素集合所属的色域范围的特征像素值的差异,对该像素集合中各像素进行排序,所述特征像素值为所述所属的色域范围对应的色域子立方体包含的所述色域立方体的顶点、所述对应的色域子立方体的中心点、所述对应的各色域子立方体的均值点中的一个;根据排序结果,依次将该像素集合中各像素选取为所述种子像素。
- 根据权利要求1所述的分割方法,其中,所述根据所述匹配情况,对所述待分割图像进行图像分割包括:根据所述各像素及其匹配像素,生成多个子图像;根据各子图像之间的重叠情况,对所述多个子图像进行合并处理;根据合并结果,确定图像分割结果。
- 根据权利要求8所述的分割方法,其中,所述根据各子图像之间的重叠情况,对所述多个子图像进行合并处理包括:计算第一子图像与第二子图像的交集包含的像素数量;根据所述交集包含的像素数量与所述第一子图像包含的像素数量的比值,确定重叠参数用于判断所述重叠情况;在所述重叠参数大于阈值的情况下,将所述第一子图像与所述第二子图像合并。
- 根据权利要求1-9任一项所述的分割方法,还包括:根据原始图像中各像素的像素值分布,确定干扰像素;根据所述原始图像中各像素的像素值,确定所述干扰像素的匹配像素;在所述原始图像中,去除所述干扰像素及其匹配像素,获取所述待分割图像。
- 根据权利要求1-9任一项所述的分割方法,其中,所述待分割图像为根据获取的水下声呐数据生成的二维图像。
- 一种图像的三维重建方法,包括:根据权利要求1-11任一项所述的分割方法,对待分割图像进行分割处理;根据分割处理结果进行三维重建,获取三维图像。
- 一种图像的分割装置,包括至少一个处理器,所述处理器被配置为执行如下步骤:根据像素值所属的色域范围,将待分割图像中的各像素划分到不同的像素集合中;根据像素值,分别确定每个像素集合内的各像素之间的匹配情况;根据所述匹配情况,对所述待分割图像进行图像分割。
- 一种图像的三维重建装置,包括至少一个处理器,所述处理器被配置为执行如下步骤:根据权利要求1-11任一项所述的分割方法,对待分割图像进行分割处理;根据分割处理结果进行三维重建,获取三维图像。
- 一种电子设备,包括:存储器;和耦接至所述存储器的处理器,所述处理器被配置为基于存储在所述存储器装置中的指令,执行权利要求1-11任一项所述的图像的分割方法,或者权利要求12所述的图像的三维重建方法。
- 一种非易失性计算机可读存储介质,其上存储有计算机程序,该程序被处理器执行时实现如权利要求1-11任一项所述的图像的分割方法,或者权利要求12所述的图像的三维重建方法。
- 一种可穿戴设备,包括:权利要求14所述的图像的三维重建装置;和显示屏,用于显示所述三维重建装置获取的三维图像。
- 根据权利要求17所述的可穿戴设备,其中,所述三维重建装置根据获取的水下声呐数据生成待分割图像,并根据所述待分割图像的分割结果重建所述三维图像。
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202080001926.6A CN115053257A (zh) | 2020-09-11 | 2020-09-11 | 图像的分割方法、装置和图像的三维重建方法、装置 |
| US17/312,156 US12125207B2 (en) | 2020-09-11 | 2020-09-11 | Image segmentation method and apparatus and image three-dimensional reconstruction method and apparatus |
| PCT/CN2020/114752 WO2022052032A1 (zh) | 2020-09-11 | 2020-09-11 | 图像的分割方法、装置和图像的三维重建方法、装置 |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2020/114752 WO2022052032A1 (zh) | 2020-09-11 | 2020-09-11 | 图像的分割方法、装置和图像的三维重建方法、装置 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2022052032A1 true WO2022052032A1 (zh) | 2022-03-17 |
Family
ID=80630212
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2020/114752 Ceased WO2022052032A1 (zh) | 2020-09-11 | 2020-09-11 | 图像的分割方法、装置和图像的三维重建方法、装置 |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US12125207B2 (zh) |
| CN (1) | CN115053257A (zh) |
| WO (1) | WO2022052032A1 (zh) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116434065A (zh) * | 2023-04-19 | 2023-07-14 | 北京卫星信息工程研究所 | 全色几何校正遥感影像的水体分割方法 |
| CN116704152A (zh) * | 2022-12-09 | 2023-09-05 | 荣耀终端有限公司 | 图像处理方法和电子设备 |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117911631B (zh) * | 2024-03-19 | 2024-05-28 | 广东石油化工学院 | 一种基于异源图像匹配的三维重建方法 |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103337064A (zh) * | 2013-04-28 | 2013-10-02 | 四川大学 | 图像立体匹配中的一种误匹配点剔除方法 |
| CN104967761A (zh) * | 2015-06-26 | 2015-10-07 | 深圳市华星光电技术有限公司 | 色域匹配方法 |
| CN105122306A (zh) * | 2013-03-29 | 2015-12-02 | 欧姆龙株式会社 | 区域分割方法以及检查装置 |
| US9299009B1 (en) * | 2013-05-13 | 2016-03-29 | A9.Com, Inc. | Utilizing color descriptors to determine color content of images |
| CN105869177A (zh) * | 2016-04-20 | 2016-08-17 | 内蒙古农业大学 | 一种图像分割方法及装置 |
| CN109377490A (zh) * | 2018-10-31 | 2019-02-22 | 深圳市长隆科技有限公司 | 水质检测方法、装置及计算机终端 |
| CN110751660A (zh) * | 2019-10-18 | 2020-02-04 | 南京林业大学 | 一种彩色图像分割方法 |
Family Cites Families (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105427272A (zh) * | 2014-09-17 | 2016-03-23 | 富士通株式会社 | 图像处理设备、图像处理方法以及电子装置 |
| CN106340023B (zh) * | 2016-08-22 | 2019-03-05 | 腾讯科技(深圳)有限公司 | 图像分割的方法和装置 |
| CN108681994B (zh) * | 2018-05-11 | 2023-01-10 | 京东方科技集团股份有限公司 | 一种图像处理方法、装置、电子设备及可读存储介质 |
| CN109872374A (zh) * | 2019-02-19 | 2019-06-11 | 江苏通佑视觉科技有限公司 | 一种图像语义分割的优化方法、装置、存储介质及终端 |
| US10957049B2 (en) * | 2019-07-31 | 2021-03-23 | Intel Corporation | Unsupervised image segmentation based on a background likelihood estimation |
| US11455485B2 (en) * | 2020-06-29 | 2022-09-27 | Adobe Inc. | Content prediction based on pixel-based vectors |
-
2020
- 2020-09-11 CN CN202080001926.6A patent/CN115053257A/zh active Pending
- 2020-09-11 WO PCT/CN2020/114752 patent/WO2022052032A1/zh not_active Ceased
- 2020-09-11 US US17/312,156 patent/US12125207B2/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105122306A (zh) * | 2013-03-29 | 2015-12-02 | 欧姆龙株式会社 | 区域分割方法以及检查装置 |
| CN103337064A (zh) * | 2013-04-28 | 2013-10-02 | 四川大学 | 图像立体匹配中的一种误匹配点剔除方法 |
| US9299009B1 (en) * | 2013-05-13 | 2016-03-29 | A9.Com, Inc. | Utilizing color descriptors to determine color content of images |
| CN104967761A (zh) * | 2015-06-26 | 2015-10-07 | 深圳市华星光电技术有限公司 | 色域匹配方法 |
| CN105869177A (zh) * | 2016-04-20 | 2016-08-17 | 内蒙古农业大学 | 一种图像分割方法及装置 |
| CN109377490A (zh) * | 2018-10-31 | 2019-02-22 | 深圳市长隆科技有限公司 | 水质检测方法、装置及计算机终端 |
| CN110751660A (zh) * | 2019-10-18 | 2020-02-04 | 南京林业大学 | 一种彩色图像分割方法 |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116704152A (zh) * | 2022-12-09 | 2023-09-05 | 荣耀终端有限公司 | 图像处理方法和电子设备 |
| CN116704152B (zh) * | 2022-12-09 | 2024-04-19 | 荣耀终端有限公司 | 图像处理方法和电子设备 |
| CN116434065A (zh) * | 2023-04-19 | 2023-07-14 | 北京卫星信息工程研究所 | 全色几何校正遥感影像的水体分割方法 |
| CN116434065B (zh) * | 2023-04-19 | 2023-12-19 | 北京卫星信息工程研究所 | 全色几何校正遥感影像的水体分割方法 |
Also Published As
| Publication number | Publication date |
|---|---|
| CN115053257A (zh) | 2022-09-13 |
| US12125207B2 (en) | 2024-10-22 |
| US20220327710A1 (en) | 2022-10-13 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12423849B2 (en) | Systems and methods for hybrid depth regularization | |
| Li et al. | Multi-angle head pose classification when wearing the mask for face recognition under the COVID-19 coronavirus epidemic | |
| US8994723B2 (en) | Recognition and pose determination of 3D objects in multimodal scenes | |
| TWI395145B (zh) | 手勢辨識系統及其方法 | |
| WO2019169884A1 (zh) | 基于深度信息的图像显著性检测方法和装置 | |
| CN105809651A (zh) | 基于边缘非相似性对比的图像显著性检测方法 | |
| WO2022052032A1 (zh) | 图像的分割方法、装置和图像的三维重建方法、装置 | |
| CN117593618B (zh) | 基于神经辐射场和深度图的点云生成方法 | |
| CN108388901B (zh) | 基于空间-语义通道的协同显著目标检测方法 | |
| Peng et al. | Ocm3d: Object-centric monocular 3d object detection | |
| Akizuki et al. | ASM-Net: Category-level Pose and Shape Estimation Using Parametric Deformation. | |
| Chochia | Image segmentation based on the analysis of distances in an attribute space | |
| Xiao et al. | Sgcnerf: Few-shot neural rendering via sparse geometric consistency guidance | |
| JP3514373B2 (ja) | 3次元情報抽出方法、装置、および3次元情報抽出プログラムを記録した記録媒体 | |
| CN118967913A (zh) | 一种基于直线约束的神经辐射场渲染方法 | |
| CN114817595B (zh) | 基于草图的三维模型检索方法、装置、设备及介质 | |
| Abello et al. | A graph-based method for joint instance segmentation of point clouds and image sequences | |
| CN113129351A (zh) | 一种基于光场傅里叶视差分层的特征检测方法 | |
| Wang et al. | Geodesic-HOF: 3D reconstruction without cutting corners | |
| CN112749713A (zh) | 一种基于人工智能的大数据图像识别系统及方法 | |
| CN116012430B (zh) | 一种弱对齐rgb-d图像引导的深度图补全方法及系统 | |
| He | Research on outdoor garden scene reconstruction based on PMVS Algorithm | |
| Lee et al. | Occlusion-Robust and Efficient 6D Pose Estimation with Scene-Level Segmentation Refinement and 3D Partial-to-6D Full Point Cloud Transformation | |
| Kowdle et al. | Scribble based interactive 3d reconstruction via scene co-segmentation | |
| Garces et al. | Light-field surface color segmentation with an application to intrinsic decomposition |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20952820 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 20952820 Country of ref document: EP Kind code of ref document: A1 |
|
| 32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 20.11.2023) |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 20952820 Country of ref document: EP Kind code of ref document: A1 |