WO2015197026A1 - 一种获取目标物体体征数据的方法、装置及终端 - Google Patents
一种获取目标物体体征数据的方法、装置及终端 Download PDFInfo
- Publication number
- WO2015197026A1 WO2015197026A1 PCT/CN2015/082539 CN2015082539W WO2015197026A1 WO 2015197026 A1 WO2015197026 A1 WO 2015197026A1 CN 2015082539 W CN2015082539 W CN 2015082539W WO 2015197026 A1 WO2015197026 A1 WO 2015197026A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target object
- model
- pixel
- depth
- contour
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating three-dimensional [3D] models or images for computer graphics
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/145—Illumination specially adapted for pattern recognition, e.g. using gratings
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/147—Details of sensors, e.g. sensor lenses
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/273—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion removing elements interfering with the pattern to be recognised
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/30—Noise filtering
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
- G06V10/422—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation for representing the structure of the pattern or shape of an object therefor
- G06V10/426—Graphical representations
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
- G06V10/431—Frequency domain transformation; Autocorrelation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/60—Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional [3D] objects
- G06V20/647—Three-dimensional [3D] objects by matching two-dimensional images to three-dimensional objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional [3D] objects
- G06V20/653—Three-dimensional [3D] objects by matching three-dimensional models, e.g. conformal mapping of Riemann surfaces
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/103—Static body considered as a whole, e.g. static pedestrian or occupant recognition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20072—Graph-based image processing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/12—Acquisition of 3D measurements of objects
- G06V2201/121—Acquisition of 3D measurements of objects using special illumination
Definitions
- the present invention relates to the field of data acquisition, and in particular, to a method, device and terminal for acquiring physical sign data of a target object.
- the embodiment of the invention provides a method, a device and a terminal for acquiring the target object's vital sign data, and recovering the captured object image by the image and the skeleton of the target object, and then combining the local or cloud search application to identify the target object.
- the data is presented to the user to achieve a "what you see is what you get" experience for the majority of users.
- the present invention provides a method for obtaining vital sign data of a target object, comprising:
- the 3D depth image is a two-dimensional image with distance information, the distance information including a distance between the target object and the imaging device;
- the depth value is a point on the target object acquired according to the distance information to the imaging device the distance between;
- the obtaining target includes:
- Obtaining the graphic contour of the target object according to the depth value of the pixel in the 3D depth image of the target object includes:
- performing the difference calculation on the depth value of the pixel in the 3D depth image includes:
- the neighboring pixels corresponding to the at least one first depth difference value are marked as contour positions;
- a graphical outline of the target object is acquired based on the pixel marked as a contour position.
- the method further includes:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the first 3D target depth image, Obtaining a graphical outline of the target object.
- the performing the background denoising processing on the 3D depth image acquiring the first 3D target depth image, including :
- the method further includes:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the second 3D target depth image, Obtaining a graphical outline of the target object.
- the performing the edge denoising processing on the first 3D target depth image to obtain the second 3D target depth includes:
- the 3D depth image according to the target object Obtaining the skeleton parameters of the target object by the depth value of the pixel includes:
- the area defined by the first line and the second line constitutes a skeleton of the target object, and the corresponding lateral thickness and the longitudinal thickness are skeleton parameters of the target object.
- the 3D model library retrieves a 3D that matches a graphical contour and a skeleton parameter of the target object a model that obtains a parameter ratio of the 3D model, including:
- a viewing angle parameter of the 3D model Calculating a viewing angle parameter of the 3D model according to a graphic contour of the 3D model and a front view contour of the 3D model, the viewing angle parameter being a visual angle of the graphic contour of the 3D model based on a front view contour of the 3D model ;
- the 3D model is the graphic wheel of the target object a 3D model with a matching profile and skeleton parameters
- a parameter ratio of the 3D model is obtained by the 3D model.
- the 3D model library includes each view graphic contour of the 3D model, where at least the 3D is included The forward view graphic outline of the model.
- the graphical contour of the target object and the graphical contour of the 3D model in the 3D model library are performed Matching, obtaining the graphical outline of the 3D model with the highest matching degree includes:
- the acquiring at least one of the target objects Real size including:
- the transmission time is a time difference between the emitted sound wave signal and the received sound wave signal
- At least one true size of the target object is calculated by the distance from the imaging device.
- the present invention provides an apparatus for acquiring physical sign data of a target object, including: a module for acquiring a 3D depth image of the target object; the 3D depth image being a two-dimensional image with distance information, the distance information including a distance between the target object and the imaging device;
- a graphic contour and skeleton parameter obtaining module configured to acquire a graphic contour and a skeleton parameter of the target object according to a depth value of a pixel in a 3D depth image of the target object; the depth value is the obtained according to the distance information The distance from a point on the target object to the imaging device;
- a parameter ratio obtaining module configured to retrieve a 3D model matching the graphic contour and the skeleton parameter of the target object in the 3D model library, and obtain a parameter proportion of the 3D model;
- a real size acquisition module configured to acquire at least one real size of the target object
- the vital sign data obtaining module is configured to acquire the vital sign data of the target object according to the parameter ratio of the 3D model and the at least one real size.
- the imaging module comprises:
- a transmitting unit configured to transmit a reference pattern to the target object
- a receiving unit configured to receive a secondary pattern that is reflected by the reference object by the target object
- a calculating unit configured to calculate an offset value of the secondary pattern relative to the reference pattern
- an image acquiring unit configured to perform Fourier transform on the offset value to obtain the distance information, and obtain the 3D depth image from the distance information.
- the graphic contour and skeleton parameter obtaining module is specifically configured to: view the 3D depth image Performing a difference calculation on the depth value of the pixel to obtain a graphic outline of the target object;
- performing the difference calculation on the depth value of the pixel in the 3D depth image includes:
- the neighboring pixels corresponding to the at least one first depth difference value are marked as contour positions;
- a graphical outline of the target object is acquired based on the pixel marked as a contour position.
- the device further includes: a noise canceling module;
- the denoising module is used to:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the first 3D target depth image, Obtaining a graphical outline of the target object.
- the denoising module is specifically configured to:
- the denoising module is further configured to:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: entering a depth value of the pixel in the second 3D target depth image A line difference calculation is performed to obtain a graphic outline of the target object.
- the denoising module is specifically configured to:
- the graphic contour and skeleton parameter obtaining module is specifically used to:
- the area defined by the first line and the second line constitutes a skeleton of the target object, and the corresponding lateral thickness and the longitudinal thickness are skeleton parameters of the target object.
- the parameter ratio obtaining module is specifically configured to:
- a viewing angle parameter of the 3D model Calculating a viewing angle parameter of the 3D model according to a graphic contour of the 3D model and a front view contour of the 3D model, the viewing angle parameter being a visual angle of the graphic contour of the 3D model based on a front view contour of the 3D model ;
- the 3D model is the graphic contour of the target object And a 3D model matching the skeleton parameters;
- a parameter ratio of the 3D model is obtained by the 3D model.
- the 3D model library includes each view graphic contour of the 3D model, where the at least the 3D is included The forward view graphic outline of the model.
- the parameter ratio obtaining module is specifically configured to:
- the real size obtaining module is specifically configured to:
- the transmission time is the sound wave signal and the sound wave Describe the time difference of receiving the acoustic signal
- At least one true size of the target object is calculated by the distance from the imaging device.
- the present invention provides a terminal for acquiring physical condition data of a target object, including:
- a 3D sensor for acquiring a 3D depth image of the target object;
- the 3D depth image being a two-dimensional image with distance information, the distance information including a distance between the target object and the imaging device;
- a processor configured to acquire a graphic contour and a skeleton parameter of the target object according to a depth value of a pixel in a 3D depth image of the target object; the depth value is a point on the target object acquired according to the distance information a distance to the imaging device, the processor is further configured to retrieve a 3D model matching the graphic contour and the skeleton parameter of the target object in a 3D model library, and obtain a parameter ratio of the 3D model.
- the processor is further configured to acquire at least one real size of the target object, and acquire the vital sign data of the target object according to the parameter ratio of the 3D model and the at least one real size.
- the 3D sensor is specifically configured to:
- Transmitting a reference pattern to the target object and receiving a secondary pattern obtained by the reference pattern being reflected by the target object, calculating an offset value of the secondary pattern relative to the reference pattern, and calculating the offset
- the value is subjected to Fourier transform to obtain the distance information, and the 3D depth image is obtained from the distance information.
- the processor is specifically configured to perform a depth value of a pixel in the 3D depth image. Calculating a difference, acquiring a graphic outline of the target object;
- performing the difference calculation on the depth value of the pixel in the 3D depth image includes:
- the neighboring pixels corresponding to the at least one first depth difference value are marked as contour positions;
- a graphical outline of the target object is acquired based on the pixel marked as a contour position.
- the processor is further configured to: Depth image performing background denoising processing to obtain a first 3D target depth image;
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the first 3D target depth image, Obtaining a graphical outline of the target object.
- the processor is specifically configured to:
- the processor is further configured to:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the second 3D target depth image, Obtaining a graphical outline of the target object.
- the processor is specifically configured to:
- the processor is specifically configured to:
- the area defined by the first line and the second line constitutes a skeleton of the target object, and the corresponding lateral thickness and the longitudinal thickness are skeleton parameters of the target object.
- the processor is specifically configured to:
- a viewing angle parameter of the 3D model Calculating a viewing angle parameter of the 3D model according to a graphic contour of the 3D model and a front view contour of the 3D model, the viewing angle parameter being a visual angle of the graphic contour of the 3D model based on a front view contour of the 3D model ;
- the 3D model is the graphic contour of the target object And a 3D model matching the skeleton parameters;
- a parameter ratio of the 3D model is obtained by the 3D model.
- the 3D model library includes each view graphic contour of the 3D model, where the at least the 3D is included The forward view graphic outline of the model.
- the processor is specifically configured to: use the Zernike moment descriptor and the Fourier descriptor to target the target Describe the graphic outline of the object to obtain the first description information;
- the processor is specifically configured to:
- the transmission time is a time difference between the emitted sound wave signal and the received sound wave signal
- At least one true size of the target object is calculated by the distance from the imaging device.
- the 3D depth image of the target object is acquired, and the graphic contour and the skeleton parameter of the target object are obtained according to the depth value of the pixel in the 3D depth image of the target object, and the target object is retrieved in the 3D model library.
- a 3D model in which the contour of the graphic and the skeleton parameters are matched obtaining a parameter ratio of the 3D model, and acquiring at least one real size of the target object, thereby obtaining according to a parameter ratio of the 3D model and the at least one real size
- the vital sign data of the target object combines the instant acquisition of the terminal and the local or cloud 3D model library to instantly present the vital signs data of the target object, and realizes the dream of “what you see is what you get”.
- FIG. 1 is a schematic flowchart diagram of a method for acquiring physical sign data of a target object according to an embodiment of the present invention
- FIG. 2 is a reference diagram of specific steps of a method for acquiring target object vital sign data according to an embodiment of the present invention
- FIG. 3 is another reference diagram of specific steps of a method for acquiring target object vital sign data according to an embodiment of the present invention
- FIG. 4 is a further reference diagram of a specific step of a method for acquiring physical sign data of a target object according to an embodiment of the present invention
- FIG. 5 is a first schematic diagram of an apparatus for acquiring target object sign data according to an embodiment of the present invention
- FIG. 6 is a second schematic diagram of an apparatus for acquiring target object sign data according to an embodiment of the present invention.
- FIG. 7 is a schematic diagram of acquiring a target object vital sign data terminal according to an embodiment of the present invention.
- Embodiments of the present invention are directed to a method for acquiring body sign data of a target object, which is obtained by restoring the image of the captured target object to the figure and skeleton of the target object, and then combining the local or cloud type search application with the physical condition data of the target object. Presented to users to realize the dream of “what you see is what you get”.
- FIG. 1 is a method 100 for acquiring target object vitals data according to an embodiment of the present invention, including:
- the 3D depth image is a two-dimensional image with distance information, and the distance information includes a distance between the target object and the imaging device.
- the target camera may be used to capture a target object to obtain a 3D depth image.
- the acquisition process may be implemented as follows:
- the terminal transmits a reference pattern to the target object, where the reference pattern can have various options, such as a square grid pattern, a honeycomb grid pattern, or a pattern formed by scattered light spots.
- the light generating light source of the reference pattern is an infrared beam generator capable of performing beam power control, so when the target object is a human body or an animal body, the light source of the reference pattern is not correct The human or animal body causes damage.
- the terminal transmits a reference pattern to the target object, and the feature size parameters of the reference patterns are preset.
- the reference pattern is a grid pattern
- the shape and spacing of the grid can be preset. of.
- Receiving may be performed by the terminal by the built-in or external camera of the terminal, thereby obtaining a secondary pattern reflected by the target object by the target object, and acquiring the target image together Two-dimensional images, because the camera acquires a two-dimensional planar image of an object is a relatively mature technology, so the implementation method of the technology is not laid out and limited.
- the offset value which may also be referred to as a deformation value, by which the shape variable generated by the secondary pattern relative to the reference pattern is reflected.
- the distance value is subjected to Fourier transform to obtain the distance information, and the 3D depth image is obtained from the distance information.
- the Fourier transform is used to process the distance information, and the distance information is used to describe the distance between the imaged camera and the captured target object, which may be embodied as the target object.
- the distance of each pixel on the two-dimensional image from the imaging camera based on this distance information combined with the two-dimensional image of the target object acquired above, can obtain a 3D depth image with distance information.
- the 3D depth image of the target object contains all the pixels constituting the image, and the distance information acquired according to S101 can be used to describe the distance information of the pixels from the imaging camera, that is, the pixel depth value here, and further, according to the pixel depth.
- the value obtains the contour of the target object and the skeleton parameters. In the specific implementation process, obtaining the contour of the target object according to the pixel depth value can be implemented as follows:
- An exclusive value; the first pixel herein may be any one of the 3D depth images.
- the positional relationship between the four connected neighboring pixels and the first pixel may be left, right, up, and down, respectively.
- a depth difference threshold can be pre-set according to experience.
- the second depth difference threshold may be preset according to experience.
- the second pixel herein may also be any pixel in the 3D depth image, when its eight When the pixels marked as the contour position are displayed in the adjacent pixels, the difference between the pixels marked as the outline pixels and the pixel depth values of the second pixel is calculated, and the calculated difference result is calculated. Greater than the second difference threshold marks this second pixel as the contour position.
- the graphical outline of the target object is acquired according to the pixel marked as the contour position.
- the pixels in the above-described 3D depth image are divided into pixels marked as contour positions and pixels not marked as contour positions, and all pixels marked as contour positions constitute a graphic outline of the target object.
- a skeleton parameter of the target object is acquired according to a depth value of a pixel in a 3D depth image of the target object.
- the first thing to understand is the principle of the skeleton parameters. Take the human skeleton parameters as an example. The human anatomy skeleton structure satisfies certain natural proportions, whether it is Westerners or Orientals, men and women, minors and adults.
- the 18-node human skeleton model for example, if any two human skeleton models, all the same human body parts include the forearm, upper arm, torso, head, waist, hips, thighs, calves,
- the distance between adjacent and interconnected nodes is the same as the ratio of the skeleton reference (middle axis).
- the two human skeletons are identical.
- FIG. 2 it is assumed that the ratio of the distance between the nodes adjacent to each other and connected to each other in the human skeleton model 1601 and the skeleton reference (middle axis) are reconstructed from the standard human skeleton model 1602.
- the skeleton 1601 and the skeleton 1602 are the same 3D model. which is:
- Z refers to the skeleton length of the target object
- Base refers to the skeleton reference length of the target object
- Zi refers to the skeleton length of the model numbered i in the 3D model library
- Basei refers to the number i in the 3D model library.
- the skeleton reference length of the model, ⁇ is an allowable difference, and the allowable difference value may be obtained according to an empirical value or may be selected according to a specific situation.
- acquiring the skeleton parameter of the target object according to the depth value of the pixel in the 3D depth image of the target object may be specifically implemented by the following steps:
- the central axis of the target object is obtained by a linear least squares method according to pixel depth values of all pixels in the 3D depth image of the target object.
- the central axis is usually the spine of the human body.
- the lateral thickness of the graphical outline of the target object is calculated along a plurality of first lines perpendicular to the central axis.
- the central axis is the spine of the human body
- a plurality of first lines perpendicular to the spine extend the skeleton of the lateral direction of the human body.
- a longitudinal thickness of the graphical outline of the target object is calculated along a plurality of second lines parallel to the central axis.
- the central axis is the spine of the human body
- a plurality of second lines parallel to the spine extend the skeleton of the longitudinal direction of the human body.
- the area defined by the first line and the second line constitutes a skeleton of the target object, and the corresponding lateral thickness and the longitudinal thickness are skeleton parameters of the target object.
- a 3D model matching the graphic contour and the skeleton parameter of the target object is retrieved in the 3D model library.
- the 3D model library here can be a standard model library stored by the cloud server, or a standard model library stored locally.
- the model with the highest matching degree of the contour and skeleton parameters of the target object obtained above is retrieved from the standard model library.
- These pre-stored model library data can be derived from socially-derived human 3D data provided by third-party data providers, which typically includes typical body-type 3D data from various countries, regions, and races around the world.
- the pre-stored model library data may also be derived from the self-learning results of the machine.
- the owner of the intelligent terminal may measure, calculate, and adjust the self-learning 3D model data for a specific target for a period of time. It can be understood that in addition to the human body 3D data, there may be 3D data of an animal body or another visible target body, which will not be described herein.
- the pre-stored 3D model includes at least the graphic contour and the skeleton parameter. These two parameters.
- the example is a 2D contour parameter of a cow's 3D model, which respectively represents a graphic outline of a recording target from each orientation of the target, including 1 side, 2 front and 4 left front side, 5 left rear side, 6
- the projection profile of the right front side, the right rear side of 7, and the projection outline just above 3 are usually not necessary.
- the graphic contour may also include the front left front side, the left front side, and the right front side. Projection pattern outline (not shown) of the target recorded at each angle, such as the side, the right front side, and the like.
- the chen algorithm that is, the 2D Zernke moment descriptor and the FD-Fourier Descriptor are used to compare the graphic contour of one or more target objects of the same target and the various direction graphics of the 3D model library.
- the similarity between the contours, the contour contour with the highest similarity is retrieved, and the visual angle value of the projection of the 3D standard model corresponding to the contour of the graphic is returned.
- the contour of a target object and the contour of the 3D model of the cow in the 3D database have the highest similarity.
- the returned observation is that the contour of the target object is the X coordinate of the 3D model of the cow.
- the 3D model that matches the graphic contour and the skeleton parameter of the target object is retrieved in the 3D model library, and the parameter proportion of the 3D model is obtained, which specifically includes:
- a viewing angle parameter of the 3D model Calculating a viewing angle parameter of the 3D model according to a graphic contour of the 3D model and a front view contour of the 3D model, the viewing angle parameter being a visual angle of the graphic contour of the 3D model based on a front view contour of the 3D model ;
- the 3D model is the graphic contour of the target object And a 3D model matching the skeleton parameters;
- a parameter ratio of the 3D model is obtained by the 3D model.
- the 3D model library includes various view graphic outlines of the 3D model, including at least a positive view graphic outline of the 3D model.
- the matching the graphic contour of the target object with the graphic contour of the 3D model in the 3D model library to obtain the graphic contour of the 3D model with the highest matching degree includes:
- the skeleton data in the standard 3D model library needs to be rotated to the negative direction of the projection as the X-axis to achieve accurate skeleton similarity retrieval and matching.
- the search method can also obtain better results when the environment changes.
- the different clothing and postures worn by the human target affect the contour of the human body calculated by the 3D depth image, such as wearing thin clothes and winter society in summer. Wearing heavy clothing, the calculated contour of the human body will be greatly different.
- the skeleton parameters are introduced here. Graphic outline and skeleton parameters In the case of certainty, the accuracy of the obtained 3D model is relatively high.
- the parameter ratio of the 3D model of the target object retrieved in step S103 is a standard unit model, and it is also required to multiply at least one real geometric parameter of the target (such as the actual height or arm length of the user), and the same scale is enlarged before being able to A 3D model in which the target object is perfectly matched.
- at least one true size of the target object can be obtained by the following steps:
- the transmission time is a time difference between the emitted sound wave signal and the received sound wave signal
- At least one true size of the target object is calculated by the distance from the imaging device.
- An optional method is to measure the height of the target by the method of recording the image of the target object by the camera.
- the speaker component of the mobile phone periodically transmits an acoustic signal to the target.
- the transmitting action can be synchronized with the detecting action of the 3D sensor.
- the acoustic signal can be above the human hearing frequency range (20HZ ⁇ 20KHZ) to avoid interference with the user and the human target.
- the acoustic signal When the acoustic signal encounters the return of the target object, it will be received by the microphone component.
- the image of the target object will be recorded on the camera component, and the height h of the image pixel of the target object can be calculated by the image contour recognition technology.
- the image distance D2 of the camera assembly is the only determined hardware parameter, so that the true height H of the target object can be calculated according to the following formula.
- the mobile terminal When the user operates the mobile terminal to take a picture and 3D measurement of the target object, the mobile terminal simultaneously records two or more photos of the target object, and the photos are captured by camera components with different focal length parameters.
- the terminal camera obtains photos with three different image distances at three different focal lengths, two of which are taken as an example: the two photos respectively correspond to the distance D21, D22, and are imaged.
- the height is h1, h2, and the lens distance of the two sets of shooting parameters changes to ⁇ . Since the true height of the target object is uniquely determined, the photo clearly satisfies the following geometric formula:
- D21 and D22 are known parameters, and h1 and h2 can be calculated by the pixel method. Then the height H of the target object can be calculated according to the following formula.
- the true skeletal model of the target can be obtained by multiplying the retrieved 3D model parameter proportion with the highest matching degree by the target real height H.
- a real geometric parameter of the target can also be directly input through the touch screen of the mobile phone I/O interface, which can be one of data such as height, arm length, shoulder width, etc., to further calculate Signs data of human target objects.
- the user can also dynamically adjust the true geometric parameters of the input target to calculate the relative accuracy of the target object's vital data.
- the body weight, the size of the measurements, the arm length, the shoulder width and other physical parameters of the human body target 420 can be calculated according to specific parameters (such as density) in the physical characteristic parameter library, and the user input and output interfaces are input. Displayed on the display, combined with various business databases and user habits, can also be used for clothing size, matching suggestions and advertising push.
- the embodiment can be applied not only to the measurement of the vital sign data in the process of photographing the target object, but also to the process of capturing the 3D measurement and the vital sign data of the moving target object by video.
- the embodiment of the present invention obtains a 3D depth image of the target object, and restores a graphic contour and a skeleton parameter of the target object according to the 3D depth image, thereby retrieving a 3D model corresponding to the target object based on the graphic contour and the skeleton parameter, and Further, the 3D model obtains the vital sign data of the target object, so that the majority of users can perform virtual reconstruction through the terminal at any time and place to obtain the physical parameters of the object seen, thereby realizing the "what you see is what you get" user experience.
- the main target is a human body
- the result of this calculation is relatively accurate, but if two or more people appear in the real scene and overlap or obscure each other, Different human body objects need to be processed separately during processing.
- a simple method is to use image depth filtering algorithm to process overlapping human targets in the background noise mode.
- Other feasible methods are to separate and superimpose overlapping images. Need to explain is in this implementation In the technical solutions involved in the examples, the single target object or the independent target object that has been separated is mainly involved.
- the present embodiment performs background denoising on the 3D depth image based on the background denoising process to obtain the first 3D target depth image, thereby obtaining an independent
- the target object graphic contour specifically includes: performing difference calculation on the depth value of the pixel in the first 3D target depth image, and acquiring a graphic contour of the target object.
- performing background denoising processing on the 3D depth image to obtain a first 3D target depth image including:
- the obtained first 3D target depth graph may be further processed, namely:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the second 3D target depth image, Obtaining a graphical outline of the target object.
- performing edge denoising processing on the first 3D target depth image, and acquiring the second 3D target depth image includes:
- FIG. 5 is a device 300 for acquiring target object vital sign data according to an embodiment of the present invention, including:
- the imaging module 302 is configured to acquire a 3D depth image of the target object; the 3D depth image is a two-dimensional image with distance information, and the distance information includes a distance between the target object and the imaging device.
- a graphic contour and skeleton parameter obtaining module 304 configured to acquire a graphic contour and a skeleton parameter of the target object according to a depth value of a pixel in the 3D depth image of the target object; the depth value is obtained according to the distance information The distance from a point on the target object to the imaging device.
- the parameter ratio obtaining module 306 is configured to retrieve a 3D model matching the graphic contour and the skeleton parameter of the target object in the 3D model library, and obtain a parameter proportion of the 3D model.
- the real size acquisition module 308 is configured to acquire at least one real size of the target object.
- the vital sign data obtaining module 310 is configured to acquire the vital sign data of the target object according to the parameter ratio of the 3D model and the at least one real size.
- a 3D depth image of the target object is acquired by the imaging device, and the graphic contour and skeleton parameter acquiring module restores the graphic contour and the skeleton parameter of the target object according to the 3D depth image, and the parameter proportional acquisition module is based on the graphic contour and
- the skeleton parameter retrieves the 3D model corresponding to the target object, and the vital sign data acquisition module acquires the vital sign data of the target object according to the 3D model, so that the majority of users can obtain the object by virtual reconstruction through the terminal anytime and anywhere.
- the physical parameters of the body to achieve a "what you see is what you get" user experience.
- the imaging module 302 may specifically include:
- a transmitting unit configured to transmit a reference pattern to the target object
- a receiving unit configured to receive a secondary pattern that is reflected by the reference object by the target object
- a calculating unit configured to calculate an offset value of the secondary pattern relative to the reference pattern
- an image acquiring unit configured to perform Fourier transform on the offset value to obtain the distance information, and obtain the 3D depth image from the distance information.
- the graphic contour and skeleton parameter obtaining module is specifically configured to: perform a difference calculation on a depth value of a pixel in the 3D depth image, and acquire the target object Graphic outline
- performing the difference calculation on the depth value of the pixel in the 3D depth image includes:
- the neighboring pixels corresponding to the at least one first depth difference value are marked as contour positions;
- a graphical outline of the target object is acquired based on the pixel marked as a contour position.
- the apparatus 300 further includes: a noise canceling module 312;
- the denoising module 312 is configured to:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the first 3D target depth image, Obtaining a graphical outline of the target object.
- the denoising module 312 is specifically configured to:
- the denoising module 312 can also be used to:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the second 3D target depth image, Obtaining a graphical outline of the target object.
- the denoising module 312 is specifically configured to:
- the graphic contour and skeleton parameter obtaining module 304 is specifically configured to:
- the area defined by the first line and the second line constitutes a skeleton of the target object, and the corresponding lateral thickness and the longitudinal thickness are skeleton parameters of the target object.
- the parameter ratio obtaining module 306 is specifically configured to:
- a viewing angle parameter of the 3D model Calculating a viewing angle parameter of the 3D model according to a graphic contour of the 3D model and a front view contour of the 3D model, the viewing angle parameter being a visual angle of the graphic contour of the 3D model based on a front view contour of the 3D model ;
- the 3D model is the graphic contour of the target object And a 3D model matching the skeleton parameters;
- a parameter ratio of the 3D model is obtained by the 3D model.
- the 3D model library includes each view graphic outline of the 3D model, wherein at least a positive view graphic outline of the 3D model is included.
- the parameter ratio obtaining module 306 is specifically configured to:
- the real size obtaining module 308 is specifically configured to:
- the transmission time is a time difference between the emitted sound wave signal and the received sound wave signal
- At least one true size of the target object is calculated by the distance from the imaging device.
- FIG. 7 is a terminal 400 for acquiring target object vital sign data according to an embodiment of the present invention, including:
- a 3D sensor 402 configured to acquire a 3D depth image of the target object;
- the 3D depth image is a two-dimensional image with distance information, and the distance information includes a distance between the target object and the imaging device;
- the processor 404 is configured to acquire a graphic contour and a skeleton parameter of the target object according to a depth value of the pixel in the 3D depth image of the target object; the depth value is a certain target object acquired according to the distance information. a point to a distance between the imaging devices, the processor is further configured to retrieve a 3D model matching the graphic contour and the skeleton parameter of the target object in a 3D model library, and obtain a parameter ratio of the 3D model, The processor is further configured to acquire at least one of the target objects Real size, and obtaining the vital sign data of the target object according to the parameter ratio of the 3D model and the at least one real size.
- the 3D sensor 402 can be specifically used for:
- Transmitting a reference pattern to the target object and receiving a secondary pattern obtained by the reference pattern being reflected by the target object, calculating an offset value of the secondary pattern relative to the reference pattern, and calculating the offset
- the value is subjected to Fourier transform to obtain the distance information, and the 3D depth image is obtained from the distance information.
- the processor 404 is further configured to perform a difference calculation on the depth value of the pixel in the 3D depth image to obtain a graphic contour of the target object.
- performing the difference calculation on the depth value of the pixel in the 3D depth image includes:
- the neighboring pixels corresponding to the at least one first depth difference value are marked as contour positions;
- a graphical outline of the target object is acquired based on the pixel marked as a contour position.
- the processor 404 may be further configured to perform background denoising processing on the 3D depth image to obtain a first 3D target depth image.
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the first 3D target depth image, Obtaining a graphical outline of the target object.
- processor 404 can be specifically configured to:
- the processor 404 is further specifically configured to:
- performing a difference calculation on the depth value of the pixel in the 3D depth image, and acquiring the graphic contour of the target object includes: performing a difference calculation on the depth value of the pixel in the second 3D target depth image, Obtaining a graphical outline of the target object.
- processor 404 can be specifically configured to:
- processor 404 can be specifically configured to:
- An area defined by the first line and the second line constitutes a skeleton of the target object, corresponding to The lateral thickness and the longitudinal thickness are skeleton parameters of the target object.
- processor 404 can be specifically configured to:
- a viewing angle parameter of the 3D model Calculating a viewing angle parameter of the 3D model according to a graphic contour of the 3D model and a front view contour of the 3D model, the viewing angle parameter being a visual angle of the graphic contour of the 3D model based on a front view contour of the 3D model ;
- the 3D model is the graphic contour of the target object And a 3D model matching the skeleton parameters;
- a parameter ratio of the 3D model is obtained by the 3D model.
- the 3D model library includes each view graphic outline of the 3D model, wherein at least a positive view graphic outline of the 3D model is included.
- the processor 404 is specifically configured to: describe a graphic contour of the target object by using a Zernike moment descriptor and a Fourier descriptor to obtain first description information;
- processor 404 can be specifically configured to:
- the transmission time is a time difference between the emitted sound wave signal and the received sound wave signal
- At least one true size of the target object is calculated by the distance from the imaging device.
- the program may be stored in a computer readable storage medium, and the storage medium may include: Flash disk, Read-Only Memory (ROM), Random Access Memory (RAM), disk or optical disk.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Geometry (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Computer Graphics (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Vascular Medicine (AREA)
- Artificial Intelligence (AREA)
- Medical Informatics (AREA)
- Quality & Reliability (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Processing Or Creating Images (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
Description
Claims (36)
- 一种获取目标物体体征数据的方法,其特征在于,包括:获取目标物体的3D深度图像;所述3D深度图像为带有距离信息的二维图像,所述距离信息包括所述目标物体到成像设备之间的距离;根据所述目标物体的3D深度图像中像素的深度值获取所述目标物体的图形轮廓和骨架参数;所述深度值为根据所述距离信息获取的所述目标物体上某一点到所述成像设备之间的距离;在3D模型库中检索与所述目标物体的图形轮廓及骨架参数相匹配的3D模型,获取所述3D模型的参数比例;获取所述目标物体的至少一个真实尺寸;根据所述3D模型的参数比例和所述的至少一个真实尺寸获取所述目标物体的体征数据。
- 根据权利要求1所述的方法,其特征在于,所述获取目标物体的3D深度图像包括:向所述目标物体发射参考图案;接收所述参考图案经所述目标物体反射得到的二次图案;计算所述二次图案相对于所述参考图案的偏移值;对所述偏移值进行傅立叶变换获取所述距离信息,由所述距离信息得到所述3D深度图像。
- 根据权利要求1或2所述的方法,其特征在于,所述根据所述目标物体的3D深度图像中像素的深度值获取所述目标物体的图形轮廓包括:对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓;具体地,所述对所述3D深度图像中像素的深度值进行差值计算包括:分别计算所述3D深度图像中第一像素的像素深度值与所述第一像素的四个相连的邻近像素的像素深度值之间的深度值差异,获取四个第一深度差异值;当所述四个第一深度差异值中至少一个第一深度差异值大于第一差异阈值时,将所述的至少一个第一深度差异值对应的邻近像素标记为轮廓位置;查询所述3D深度图像中第二像素的八个相连的邻近像素中是否有被标记为轮廓位置的像素;若有,则将所述八个相连的邻近像素中非轮廓位置的像素的像素深度值分别与所述第二像素的像素深度值进行差值计算,获取第二深度差异值;当至少一个所述第二深度差异值大于第二差异阈值时,将所述第二像素标记为轮廓位置;根据所述标记为轮廓位置的像素获取所述目标物体的图形轮廓。
- 根据权利要求1至3所述的任一项方法,其特征在于,在所述获取目标物体的3D深度图像之后,所述方法还包括:对所述3D深度图像进行背景去噪处理,获取第一3D目标深度图像;对应地,所述对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓包括:对所述第一3D目标深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓。
- 根据权利要求4所述的方法,其特征在于,所述对所述3D深度图像进行背景去噪处理,获取第一3D目标深度图像,包括:设置深度阈值;比较所述3D深度图像中的各个像素深度值与所述深度阈值的大小,将所述3D深度图像中像素深度值大于所述深度阈值的像素滤除,获取剩余像素形成所述第一3D目标深度图像。
- 根据权利要求5所述的方法,其特征在于,在所述获取所述第一3D目标深度图像之后,所述方法还包括:对所述第一3D目标深度图像进行边缘去噪处理,获取第二3D目标深度图像;对应地,所述对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓包括:对所述第二3D目标深度图像中像素的深度值进 行差值计算,获取所述目标物体的图形轮廓。
- 根据权利要求6所述的方法,其特征在于,所述对所述第一3D目标深度图像进行边缘去噪处理,获取第二3D目标深度图像包括:将所述第一3D目标深度图像分割成多个像素块;设置像素深度分段区间;分别对每块所述像素块内所有像素的像素深度值做均值处理,获取所述每块像素块的像素均值;将所述像素均值映射至所述像素深度分段区间中的对应区间,并将同一区间内的所有像素均值对应的像素块进行合并,获取所述第二3D目标深度图像。
- 根据权利要求1至7所述的任一项方法,其特征在于,所述根据所述目标物体的3D深度图像中像素的深度值获取所述目标物体的骨架参数包括:根据所述目标物体的3D深度图像中所有像素的像素深度值,用线性最小二乘法获取所述目标物体的中轴;沿着垂直于所述中轴的多个第一线计算所述目标物体的图形轮廓的横向厚度;沿着平行于所述中心轴的多个第二线计算所述目标物体的图形轮廓的纵向厚度;由所述第一线和所述第二线限定的区域构成所述目标物体的骨架,对应的所述横向厚度和所述纵向厚度为所述目标物体的骨架参数。
- 根据权利要求8所述的方法,其特征在于,所述在3D模型库中检索与所述目标物体的图形轮廓及骨架参数相匹配的3D模型,获取所述3D模型的参数比例,包括:将所述目标物体的图形轮廓与所述3D模型库中3D模型的图形轮廓进行匹配,获取匹配度最高的3D模型的图形轮廓;当所述3D模型的图形轮廓不是所述3D模型的正视图形轮廓时,则根据所述3D模型的图形轮廓获取所述3D模型的正视图形轮廓;根据所述3D模型的图形轮廓与所述3D模型的正视图形轮廓计算所述3D模型的视角参数,所述视角参数为所述3D模型的图形轮廓基于所述3D模型的正视图形轮廓的视觉角度;将所述3D模型的正视图形轮廓基于所述视角参数旋转,获取所述3D模型的骨架参数;将所述目标物体的骨架参数与所述3D模型的骨架参数进行相似度比较,当所述相似度小于预设值时,则所述3D模型为所述与所述目标物体的图形轮廓及骨架参数相匹配的3D模型;通过所述3D模型获取所述3D模型的参数比例。
- 根据权利要求9所述的方法,其特征在于,所述3D模型库中包括所述3D模型的各视角图形轮廓,其中至少包括所述3D模型的正视角图形轮廓。
- 根据权利要求10所述的方法,其特征在于,所述将所述目标物体的图形轮廓与所述3D模型库中3D模型的图形轮廓进行匹配,获取匹配度最高的3D模型的图形轮廓包括:采用Zernike矩描述子和傅里叶描述子对所述目标物体的图形轮廓进行描述,获取第一描述信息;采用Zernike矩描述子和傅里叶描述子对所述3D模型库中3D模型的图形轮廓进行描述,获取第二描述信息;比较所述第一描述信息与所述第二描述信息,将与所述第一描述信息相差预设阈值的第二描述信息对应的3D模型的图形轮廓作为所述匹配度最高的3D模型的图形轮廓。
- 根据权利要求1至11所述任一项方法,其特征在于,所述获取所述目标物体的至少一个真实尺寸,包括:向所述目标物体发射声波信号;接收被所述目标物体反射回来的声波信号;获取所述声波信号的传输时间;所述传输时间为所述发射声波信号和所述接收声波信号的时间差;利用所述传输时间和所述声波信号的传播速率,计算所述目标物体表面到所述成像设备的距离;通过所述距离和所述成像设备的相距,计算所述目标物体的至少一个真实尺寸。
- 一种获取目标物体体征数据的装置,其特征在于,包括:成像模块,用于获取目标物体的3D深度图像;所述3D深度图像为带有距离信息的二维图像,所述距离信息包括所述目标物体到成像设备之间的距离;图形轮廓和骨架参数获取模块,用于根据所述目标物体的3D深度图像中像素的深度值获取所述目标物体的图形轮廓和骨架参数;所述深度值为根据所述距离信息获取的所述目标物体上某一点到所述成像设备之间的距离;参数比例获取模块,用于在3D模型库中检索与所述目标物体的图形轮廓及骨架参数相匹配的3D模型,获取所述3D模型的参数比例;真实尺寸获取模块,用于获取所述目标物体的至少一个真实尺寸;体征数据获取模块,用于根据所述3D模型的参数比例和所述的至少一个真实尺寸获取所述目标物体的体征数据。
- 根据权利要求13所述的装置,其特征在于,所述成像模块包括:发射单元,用于向所述目标物体发射参考图案;接收单元,用于接收所述参考图案经所述目标物体反射得到的二次图案;计算单元,用于计算所述二次图案相对于所述参考图案的偏移值;图像获取单元,用于对所述偏移值进行傅立叶变换获取所述距离信息,由所述距离信息得到所述3D深度图像。
- 根据权利要求13或14所述的装置,其特征在于,所述图形轮廓和骨架参数获取模块具体用于:对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓;具体地,所述对所述3D深度图像中像素的深度值进行差值计算包括:分别计算所述3D深度图像中第一像素的像素深度值与所述第一像素的 四个相连的邻近像素的像素深度值之间的深度值差异,获取四个第一深度差异值;当所述四个第一深度差异值中至少一个第一深度差异值大于第一差异阈值时,将所述的至少一个第一深度差异值对应的邻近像素标记为轮廓位置;查询所述3D深度图像中第二像素的八个相连的邻近像素中是否有被标记为轮廓位置的像素;若有,则将所述八个相连的邻近像素中非轮廓位置的像素的像素深度值分别与所述第二像素的像素深度值进行差值计算,获取第二深度差异值;当至少一个所述第二深度差异值大于第二差异阈值时,将所述第二像素标记为轮廓位置;根据所述标记为轮廓位置的像素获取所述目标物体的图形轮廓。
- 根据权利要求13至15所述的任一装置,其特征在于,所述装置还包括:去噪模块;所述去噪模块用于:对所述3D深度图像进行背景去噪处理,获取第一3D目标深度图像;对应地,所述对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓包括:对所述第一3D目标深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓。
- 根据权利要求16所述的装置,其特征在于,所述去噪模块具体用于:设置深度阈值;比较所述3D深度图像中的各个像素深度值与所述深度阈值的大小,将所述3D深度图像中像素深度值大于所述深度阈值的像素滤除,获取剩余像素形成所述第一3D目标深度图像。
- 根据权利要求17所述的装置,其特征在于,所述去噪模块还用于:对所述第一3D目标深度图像进行边缘去噪处理,获取第二3D目标深度图像;对应地,所述对所述3D深度图像中像素的深度值进行差值计算,获取所 述目标物体的图形轮廓包括:对所述第二3D目标深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓。
- 根据权利要求18所述的装置,其特征在于,所述去噪模块具体用于:将所述第一3D目标深度图像分割成多个像素块;设置像素深度分段区间;分别对每块所述像素块内所有像素的像素深度值做均值处理,获取所述每块像素块的像素均值;将所述像素均值映射至所述像素深度分段区间中的对应区间,并将同一区间内的所有像素均值对应的像素块进行合并,获取所述第二3D目标深度图像。
- 根据权利要求13至19所述的任一装置,其特征在于,所述图形轮廓和骨架参数获取模块具体用于:根据所述目标物体的3D深度图像中所有像素的像素深度值,用线性最小二乘法获取所述目标物体的中轴;沿着垂直于所述中轴的多个第一线计算所述目标物体的图形轮廓的横向厚度;沿着平行于所述中心轴的多个第二线计算所述目标物体的图形轮廓的纵向厚度;由所述第一线和所述第二线限定的区域构成所述目标物体的骨架,对应的所述横向厚度和所述纵向厚度为所述目标物体的骨架参数。
- 根据权利要求20所述的装置,其特征在于,所述参数比例获取模块具体用于:将所述目标物体的图形轮廓与所述3D模型库中3D模型的图形轮廓进行匹配,获取匹配度最高的3D模型的图形轮廓;当所述3D模型的图形轮廓不是所述3D模型的正视图形轮廓时,则根据所述3D模型的图形轮廓获取所述3D模型的正视图形轮廓;根据所述3D模型的图形轮廓与所述3D模型的正视图形轮廓计算所述3D 模型的视角参数,所述视角参数为所述3D模型的图形轮廓基于所述3D模型的正视图形轮廓的视觉角度;将所述3D模型的正视图形轮廓基于所述视角参数旋转,获取所述3D模型的骨架参数;将所述所述目标物体的骨架参数与所述3D模型的骨架参数进行相似度比较,当所述相似度小于预设值时,则所述3D模型为所述与所述目标物体的图形轮廓及骨架参数相匹配的3D模型;通过所述3D模型获取所述3D模型的参数比例。
- 根据权利要求21所述的装置,其特征在于,所述3D模型库中包括所述3D模型的各视角图形轮廓,其中至少包括所述3D模型的正视角图形轮廓。
- 根据权利要求22所述的装置,其特征在于,所述参数比例获取模块具体用于:采用Zernike矩描述子和傅里叶描述子对所述目标物体的图形轮廓进行描述,获取第一描述信息;采用Zernike矩描述子和傅里叶描述子对所述3D模型库中3D模型的图形轮廓进行描述,获取第二描述信息;比较所述第一描述信息与所述第二描述信息,将与所述第一描述信息相差预设阈值的第二描述信息对应的3D模型的图形轮廓作为所述匹配度最高的3D模型的图形轮廓。
- 根据权利要求13至23所述的任一装置,其特征在于,所述真实尺寸获取模块具体用于:向所述目标物体发射声波信号;接收被所述目标物体反射回来的声波信号;获取所述声波信号的传输时间;所述传输时间为所述发射声波信号和所述接收声波信号的时间差;利用所述传输时间和所述声波信号的传播速率,计算所述目标物体表面 到所述成像设备的距离;通过所述距离和所述成像设备的相距,计算所述目标物体的至少一个真实尺寸。
- 一种获取目标物体体征数据的终端,其特征在于,包括:3D传感器,用于获取目标物体的3D深度图像;所述3D深度图像为带有距离信息的二维图像,所述距离信息包括所述目标物体到成像设备之间的距离;处理器,用于根据所述目标物体的3D深度图像中像素的深度值获取所述目标物体的图形轮廓和骨架参数;所述深度值为根据所述距离信息获取的所述目标物体上某一点到所述成像设备之间的距离,所述处理器还用于在3D模型库中检索与所述目标物体的图形轮廓及骨架参数相匹配的3D模型,获取所述3D模型的参数比例,所述处理器还用于获取所述目标物体的至少一个真实尺寸,并根据所述3D模型的参数比例和所述的至少一个真实尺寸获取所述目标物体的体征数据。
- 根据权利要求25所述的终端,其特征在于,所述3D传感器具体用于:向所述目标物体发射参考图案,并接收所述参考图案经所述目标物体反射得到的二次图案,计算所述二次图案相对于所述参考图案的偏移值,并对所述偏移值进行傅立叶变换获取所述距离信息,由所述距离信息得到所述3D深度图像。
- 根据权利要求25或26所述的终端,其特征在于,所述处理器具体用于对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓;具体地,所述对所述3D深度图像中像素的深度值进行差值计算包括:分别计算所述3D深度图像中第一像素的像素深度值与所述第一像素的四个相连的邻近像素的像素深度值之间的深度值差异,获取四个第一深度差异值;当所述四个第一深度差异值中至少一个第一深度差异值大于第一差异阈值时,将所述的至少一个第一深度差异值对应的邻近像素标记为轮廓位置;查询所述3D深度图像中第二像素的八个相连的邻近像素中是否有被标记为轮廓位置的像素;若有,则将所述八个相连的邻近像素中非轮廓位置的像素的像素深度值分别与所述第二像素的像素深度值进行差值计算,获取第二深度差异值;当至少一个所述第二深度差异值大于第二差异阈值时,将所述第二像素标记为轮廓位置;根据所述标记为轮廓位置的像素获取所述目标物体的图形轮廓。
- 根据权利要求25至27所述的任一终端,其特征在于,所述处理器还用于:对所述3D深度图像进行背景去噪处理,获取第一3D目标深度图像;对应地,所述对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓包括:对所述第一3D目标深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓。
- 根据权利要求28所述的终端,其特征在于,所述处理器具体用于:设置深度阈值;比较所述3D深度图像中的各个像素深度值与所述深度阈值的大小,将所述3D深度图像中像素深度值大于所述深度阈值的像素滤除,获取剩余像素形成所述第一3D目标深度图像。
- 根据权利要求29所述的终端,其特征在于,所述处理器还用于:对所述第一3D目标深度图像进行边缘去噪处理,获取第二3D目标深度图像;对应地,所述对所述3D深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓包括:对所述第二3D目标深度图像中像素的深度值进行差值计算,获取所述目标物体的图形轮廓。
- 根据权利要求30所述的终端,其特征在于,所述处理器具体用于:将所述第一3D目标深度图像分割成多个像素块;设置像素深度分段区间;分别对每块所述像素块内所有像素的像素深度值做均值处理,获取所述每块像素块的像素均值;将所述像素均值映射至所述像素深度分段区间中的对应区间,并将同一区间内的所有像素均值对应的像素块进行合并,获取所述第二3D目标深度图像。
- 根据权利要求25至31所述的终端,其特征在于,所述处理器具体用于:根据所述目标物体的3D深度图像中所有像素的像素深度值,用线性最小二乘法获取所述目标物体的中轴;沿着垂直于所述中轴的多个第一线计算所述目标物体的图形轮廓的横向厚度;沿着平行于所述中心轴的多个第二线计算所述目标物体的图形轮廓的纵向厚度;由所述第一线和所述第二线限定的区域构成所述目标物体的骨架,对应的所述横向厚度和所述纵向厚度为所述目标物体的骨架参数。
- 根据权利要求32所述的终端,其特征在于,所述处理器具体用于:将所述目标物体的图形轮廓与所述3D模型库中3D模型的图形轮廓进行匹配,获取匹配度最高的3D模型的图形轮廓;当所述3D模型的图形轮廓不是所述3D模型的正视图形轮廓时,则根据所述3D模型的图形轮廓获取所述3D模型的正视图形轮廓;根据所述3D模型的图形轮廓与所述3D模型的正视图形轮廓计算所述3D模型的视角参数,所述视角参数为所述3D模型的图形轮廓基于所述3D模型的正视图形轮廓的视觉角度;将所述3D模型的正视图形轮廓基于所述视角参数旋转,获取所述3D模型的骨架参数;将所述所述目标物体的骨架参数与所述3D模型的骨架参数进行相似度 比较,当所述相似度小于预设值时,则所述3D模型为所述与所述目标物体的图形轮廓及骨架参数相匹配的3D模型;通过所述3D模型获取所述3D模型的参数比例。
- 根据权利要求33所述的终端,其特征在于,所述3D模型库中包括所述3D模型的各视角图形轮廓,其中至少包括所述3D模型的正视角图形轮廓。
- 根据权利要求34所述的终端,其特征在于,所述处理器具体用于:采用Zernike矩描述子和傅里叶描述子对所述目标物体的图形轮廓进行描述,获取第一描述信息;采用Zernike矩描述子和傅里叶描述子对所述3D模型库中3D模型的图形轮廓进行描述,获取第二描述信息;比较所述第一描述信息与所述第二描述信息,将与所述第一描述信息相差预设阈值的第二描述信息对应的3D模型的图形轮廓作为所述匹配度最高的3D模型的图形轮廓。
- 根据权利要求25至35任一项所述的终端,其特征在于,所述处理器具体用于:向所述目标物体发射声波信号;接收被所述目标物体反射回来的声波信号;获取所述声波信号的传输时间;所述传输时间为所述发射声波信号和所述接收声波信号的时间差;利用所述传输时间和所述声波信号的传播速率,计算所述目标物体表面到所述成像设备的距离;通过所述距离和所述成像设备的相距,计算所述目标物体的至少一个真实尺寸。
Priority Applications (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2016575462A JP6549620B2 (ja) | 2014-06-27 | 2015-06-26 | ターゲットオブジェクトのサインデータを取得するための方法、装置、および端末 |
| BR112016030027-0A BR112016030027B1 (pt) | 2014-06-27 | 2015-06-26 | Método e terminal para obter dados de sinal de um objeto alvo |
| EP15810950.4A EP3144900B1 (en) | 2014-06-27 | 2015-06-26 | Method and terminal for acquiring sign data of target object |
| KR1020177001223A KR101922039B1 (ko) | 2014-06-27 | 2015-06-26 | 목표 물체의 증상 데이터를 획득하기 위한 방법, 장치, 및 단말 |
| US15/391,068 US9984461B2 (en) | 2014-06-27 | 2016-12-27 | Method, apparatus, and terminal for obtaining vital sign data of target object |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201410301461.2A CN105336005B (zh) | 2014-06-27 | 2014-06-27 | 一种获取目标物体体征数据的方法、装置及终端 |
| CN201410301461.2 | 2014-06-27 |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/391,068 Continuation US9984461B2 (en) | 2014-06-27 | 2016-12-27 | Method, apparatus, and terminal for obtaining vital sign data of target object |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2015197026A1 true WO2015197026A1 (zh) | 2015-12-30 |
Family
ID=54936983
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2015/082539 Ceased WO2015197026A1 (zh) | 2014-06-27 | 2015-06-26 | 一种获取目标物体体征数据的方法、装置及终端 |
Country Status (7)
| Country | Link |
|---|---|
| US (1) | US9984461B2 (zh) |
| EP (1) | EP3144900B1 (zh) |
| JP (1) | JP6549620B2 (zh) |
| KR (1) | KR101922039B1 (zh) |
| CN (1) | CN105336005B (zh) |
| BR (1) | BR112016030027B1 (zh) |
| WO (1) | WO2015197026A1 (zh) |
Families Citing this family (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| IT201600122898A1 (it) * | 2016-12-02 | 2018-06-02 | Ecole Polytechnique Fed Lausanne Epfl | Metodi e apparati per codificare e decodificare immagini o flussi video digitali |
| JP2018136896A (ja) * | 2017-02-24 | 2018-08-30 | キヤノン株式会社 | 情報処理装置、システム、情報処理方法、および物品の製造方法 |
| CN107270829B (zh) * | 2017-06-08 | 2020-06-19 | 南京华捷艾米软件科技有限公司 | 一种基于深度图像的人体三围测量方法 |
| CN109658515B (zh) * | 2017-10-11 | 2022-11-04 | 阿里巴巴集团控股有限公司 | 点云网格化方法、装置、设备及计算机存储介质 |
| CN108229332B (zh) * | 2017-12-08 | 2020-02-14 | 华为技术有限公司 | 骨骼姿态确定方法、装置及计算机可读存储介质 |
| FR3076028B1 (fr) * | 2017-12-21 | 2021-12-24 | Commissariat Energie Atomique | Methode de reconnaissance d'objets dans une scene observee en trois dimensions |
| TWI672937B (zh) * | 2018-02-05 | 2019-09-21 | 廣達電腦股份有限公司 | 三維影像處理之裝置及方法 |
| CN112639731A (zh) | 2018-07-24 | 2021-04-09 | 奇跃公司 | 应用共享 |
| CN110910393B (zh) * | 2018-09-18 | 2023-03-24 | 北京市商汤科技开发有限公司 | 数据处理方法及装置、电子设备及存储介质 |
| CN115516364B (zh) * | 2020-02-14 | 2024-04-26 | 奇跃公司 | 工具桥 |
| CN111768878A (zh) * | 2020-06-30 | 2020-10-13 | 杭州依图医疗技术有限公司 | 可视化指引病灶的方法及计算机可读存储介质 |
| CN115130077A (zh) * | 2021-03-25 | 2022-09-30 | 上海肇观电子科技有限公司 | 用于身份识别的方法、电子电路、电子设备和介质 |
| CN114842147A (zh) * | 2022-05-16 | 2022-08-02 | 聚好看科技股份有限公司 | 一种测量人体参数的方法及电子设备 |
| KR102772935B1 (ko) * | 2023-06-01 | 2025-02-28 | 주식회사 케이씨 | 모바일 로봇을 포함하는 자동화 가스 공급 시스템 |
| EP4509057A1 (en) * | 2023-08-17 | 2025-02-19 | Koninklijke Philips N.V. | Method and system for visualizing change in position |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103294996A (zh) * | 2013-05-09 | 2013-09-11 | 电子科技大学 | 一种3d手势识别方法 |
| CN103718175A (zh) * | 2011-07-27 | 2014-04-09 | 三星电子株式会社 | 检测对象姿势的设备、方法和介质 |
| CN103745218A (zh) * | 2014-01-26 | 2014-04-23 | 清华大学 | 一种深度图像中的姿势识别方法及装置 |
Family Cites Families (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US4954962A (en) * | 1988-09-06 | 1990-09-04 | Transitions Research Corporation | Visual navigation and obstacle avoidance structured light system |
| US7298415B2 (en) * | 2001-07-13 | 2007-11-20 | Xenogen Corporation | Structured light imaging apparatus |
| US7113217B2 (en) * | 2001-07-13 | 2006-09-26 | Xenogen Corporation | Multi-view imaging apparatus |
| KR100488366B1 (ko) | 2002-03-13 | 2005-05-11 | 주식회사 에스알에스텍 | 초음파를 이용한 수소유기균열 및 부식 측정 시스템과안전성 평가 방법 |
| US7128024B2 (en) * | 2003-07-15 | 2006-10-31 | Doyle Ii John Conan | System and method for measuring animals |
| EP1934945A4 (en) | 2005-10-11 | 2016-01-20 | Apple Inc | METHOD AND SYSTEM FOR OBJECT RECONSTRUCTION |
| CN103778635B (zh) * | 2006-05-11 | 2016-09-28 | 苹果公司 | 用于处理数据的方法和装置 |
| KR101640039B1 (ko) * | 2009-09-22 | 2016-07-18 | 삼성전자주식회사 | 영상 처리 장치 및 방법 |
| US9098873B2 (en) | 2010-04-01 | 2015-08-04 | Microsoft Technology Licensing, Llc | Motion-based interactive shopping environment |
| US20120050480A1 (en) | 2010-08-27 | 2012-03-01 | Nambi Seshadri | Method and system for generating three-dimensional video utilizing a monoscopic camera |
| US8994792B2 (en) | 2010-08-27 | 2015-03-31 | Broadcom Corporation | Method and system for creating a 3D video from a monoscopic 2D video and corresponding depth information |
| US8810565B2 (en) | 2010-08-27 | 2014-08-19 | Broadcom Corporation | Method and system for utilizing depth information as an enhancement layer |
| JP5597087B2 (ja) * | 2010-10-04 | 2014-10-01 | パナソニック株式会社 | 仮想物体操作装置 |
| CN104395929B (zh) * | 2012-06-21 | 2017-10-03 | 微软技术许可有限责任公司 | 使用深度相机的化身构造 |
| CN104854623A (zh) * | 2012-08-02 | 2015-08-19 | 微软技术许可有限责任公司 | 基于化身的虚拟试衣室 |
-
2014
- 2014-06-27 CN CN201410301461.2A patent/CN105336005B/zh active Active
-
2015
- 2015-06-26 EP EP15810950.4A patent/EP3144900B1/en active Active
- 2015-06-26 BR BR112016030027-0A patent/BR112016030027B1/pt active IP Right Grant
- 2015-06-26 WO PCT/CN2015/082539 patent/WO2015197026A1/zh not_active Ceased
- 2015-06-26 KR KR1020177001223A patent/KR101922039B1/ko active Active
- 2015-06-26 JP JP2016575462A patent/JP6549620B2/ja active Active
-
2016
- 2016-12-27 US US15/391,068 patent/US9984461B2/en active Active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN103718175A (zh) * | 2011-07-27 | 2014-04-09 | 三星电子株式会社 | 检测对象姿势的设备、方法和介质 |
| CN103294996A (zh) * | 2013-05-09 | 2013-09-11 | 电子科技大学 | 一种3d手势识别方法 |
| CN103745218A (zh) * | 2014-01-26 | 2014-04-23 | 清华大学 | 一种深度图像中的姿势识别方法及装置 |
Non-Patent Citations (1)
| Title |
|---|
| See also references of EP3144900A4 * |
Also Published As
| Publication number | Publication date |
|---|---|
| KR101922039B1 (ko) | 2018-11-26 |
| BR112016030027A2 (pt) | 2017-08-22 |
| CN105336005B (zh) | 2018-12-14 |
| KR20170019430A (ko) | 2017-02-21 |
| EP3144900A1 (en) | 2017-03-22 |
| US20170109885A1 (en) | 2017-04-20 |
| CN105336005A (zh) | 2016-02-17 |
| JP2017525029A (ja) | 2017-08-31 |
| US9984461B2 (en) | 2018-05-29 |
| BR112016030027B1 (pt) | 2023-10-10 |
| EP3144900B1 (en) | 2023-04-05 |
| JP6549620B2 (ja) | 2019-07-24 |
| EP3144900A4 (en) | 2017-07-26 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| WO2015197026A1 (zh) | 一种获取目标物体体征数据的方法、装置及终端 | |
| CN110599540B (zh) | 多视点相机下的实时三维人体体型与姿态重建方法及装置 | |
| CN107820593B (zh) | 一种虚拟现实交互方法、装置及系统 | |
| US10068344B2 (en) | Method and system for 3D capture based on structure from motion with simplified pose detection | |
| JP2024056955A (ja) | 光学式捕捉によるパーソナライズされたhrtf | |
| CN111649690A (zh) | 一种能够手持的3d信息采集的设备及方法 | |
| WO2022012019A1 (zh) | 身高测量方法、身高测量装置和终端 | |
| WO2020063987A1 (zh) | 三维扫描方法、装置、存储介质和处理器 | |
| CN103697882A (zh) | 一种基于图像识别的地理三维空间定位方法和装置 | |
| WO2018075053A1 (en) | Object pose based on matching 2.5d depth information to 3d information | |
| CN112423191A (zh) | 一种视频通话设备和音频增益方法 | |
| CN112419388A (zh) | 深度检测方法、装置、电子设备和计算机可读存储介质 | |
| WO2018028152A1 (zh) | 一种图像采集设备、虚拟现实设备 | |
| CN112270709A (zh) | 地图构建方法及装置、计算机可读存储介质和电子设备 | |
| CN112270702A (zh) | 体积测量方法及装置、计算机可读介质和电子设备 | |
| Wu et al. | mmhpe: Robust multiscale 3-d human pose estimation using a single mmwave radar | |
| CN113424522A (zh) | 使用半球形或球形可见光深度图像进行三维跟踪 | |
| Liang et al. | Reduced-complexity data acquisition system for image-based localization in indoor environments | |
| CN113196283A (zh) | 使用射频信号的姿态估计 | |
| WO2024083010A1 (zh) | 一种视觉定位方法及相关装置 | |
| CN114663917A (zh) | 基于多视角的多人三维人体位姿估计方法及装置 | |
| CN115546829B (zh) | 基于zed立体相机的行人空间信息感知方法及装置 | |
| CN112465890A (zh) | 深度检测方法、装置、电子设备和计算机可读存储介质 | |
| CN107478227B (zh) | 交互式大型空间的定位算法 | |
| CN113424524A (zh) | 使用半球形或球形可见光深度图像的三维建模 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15810950 Country of ref document: EP Kind code of ref document: A1 |
|
| REEP | Request for entry into the european phase |
Ref document number: 2015810950 Country of ref document: EP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2015810950 Country of ref document: EP |
|
| ENP | Entry into the national phase |
Ref document number: 2016575462 Country of ref document: JP Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112016030027 Country of ref document: BR |
|
| ENP | Entry into the national phase |
Ref document number: 20177001223 Country of ref document: KR Kind code of ref document: A |
|
| ENP | Entry into the national phase |
Ref document number: 112016030027 Country of ref document: BR Kind code of ref document: A2 Effective date: 20161220 |