Disclosure of Invention
In view of the above-mentioned drawbacks or shortcomings in the prior art, it is desirable to provide a 3D recognition apparatus, a 3D recognition system and a recognition method thereof, which can improve a 3D display effect.
In a first aspect, the present application provides a 3D recognition apparatus, comprising:
the light emitting module is configured for emitting light to the surface of the object to be measured;
the first receiving module is configured for receiving the light reflected by at least part of the surface of the object to be measured;
the second receiving module is configured to receive at least part of the light emitted from the surface of the object to be measured;
the first receiving module is provided with a first polaroid on one side close to the object to be measured, and/or the second receiving module is provided with a second polaroid on one side close to the object to be measured.
Optionally, the polarization angles of the first polarizer and the second polarizer are the same.
Optionally, the first polarizer and the second polarizer have different polarization angles.
Optionally, the polarization angle of the first polarizer is 0-180 °, and the polarization angle of the second polarizer is 0-180 °
Optionally, the first receiving module and the second receiving module are located on the same side of the light emitting module or the first receiving module and the second receiving module are located on different sides of the light emitting module.
Optionally, the light emitting module includes a light source and a diffractive optical element, and the light source includes one or more of a high-contrast vertical cavity surface emitting laser, a single-aperture wide-area vertical cavity surface emitting laser, an array vertical cavity surface emitting laser, a laser diode, and an LED light source.
In a second aspect, the present application provides a 3D identification system comprising a 3D identification device as described in any of the above, wherein the 3D identification system is a 3D structured light identification system,
the light emitting module is configured for emitting the coded structured light to the surface of the measured object;
the first receiving module is configured to receive the structured light reflected by at least part of the surface of the object to be measured so as to acquire a first image related to the object to be measured;
the second receiving module is configured to receive the structured light reflected by at least part of the surface of the object to be measured so as to acquire a second image related to the object to be measured;
and the data processing unit is configured to acquire the first image and the second image and establish 3D depth information of the measured object based on the first image and the second image.
In a third aspect, the present application provides a 3D identification system comprising a 3D identification apparatus as described in any of the above, wherein the 3D identification system is a 3D TOF identification system,
the light emitting module is configured to emit TOF light to the surface of the measured object;
the first receiving module is configured to receive TOF light reflected by at least part of the surface of the object to be measured to acquire a first image related to the object to be measured;
the second receiving module is configured to receive the TOF light reflected by at least part of the surface of the object to be measured to acquire a second image related to the object to be measured;
and the data processing unit is configured to acquire the first image and the second image and establish 3D depth information of the measured object based on the first image and the second image.
In a fourth aspect, the present application provides a 3D identification system, the 3D identification system is a binocular vision identification system, comprising an active binocular vision unit and a passive binocular vision unit, wherein the active binocular vision unit comprises the 3D identification device as described in any one of the above, a third receiving module and a fourth receiving module are provided in the passive binocular vision unit, the third receiving module is provided with a third polarizer on a side close to the object to be tested, and/or the fourth receiving module is provided with a fourth polarizer on a side close to the object to be tested, wherein,
the light emitting module is configured to emit the coded structured light or TOF light to the surface of the measured object;
the first receiving module is configured to receive the structured light or the TOF light reflected by at least part of the surface of the object to be measured to acquire a first image related to the object to be measured;
the second receiving module is configured to receive the structured light or the TOF light reflected by at least part of the surface of the object to be measured to acquire a second image related to the object to be measured;
the third receiving module is configured to receive the structured light or the TOF light reflected by at least part of the surface of the object to be measured to acquire a third image related to the object to be measured;
the fourth receiving module is configured to receive the structured light or the TOF light reflected by at least part of the surface of the object to be measured to acquire a fourth image related to the object to be measured;
the data processing unit is configured to acquire the first image and the second image, and establish active binocular display 3D depth information of the object to be measured based on the first image and the second image; and the image acquisition unit is also configured to acquire the third image and the fourth image and establish passive binocular display 3D depth information of the object to be measured based on the third image and the fourth image.
In a fifth aspect, the present application provides a 3D identification method applied to the 3D identification system as described in any one of the above, including:
emitting light to the surface of the measured object through the light emitting module;
acquiring a first image related to the measured object through a first receiving module;
acquiring a second image related to the measured object through a second receiving module;
comparing the first image with the second image through a data processing unit, and selecting one of the images as a standard image, or processing the first image and the second image through the data processing unit to obtain the standard image;
and carrying out post-processing on the standard image to obtain the 3D depth information of the measured object.
The technical scheme provided by the embodiment of the application can have the following beneficial effects:
the 3D recognition device that this application embodiment provided includes a light emission module and two receiving module, through the different combination collocation mode of the polaroid of receiving module front portion installation, makes can be through the different luminance of the image of receiving in some scenes, selects the better image of output or carries out some rear end synthesis through two images and then exports, makes 3D recognition system's expansion to more use scenes.
Detailed Description
The present application will be described in further detail with reference to the following drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the relevant invention and not restrictive of the invention. It should be noted that, for convenience of description, only the portions related to the present invention are shown in the drawings.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
Referring to fig. 1-4 in detail, in a first aspect, the present application provides a 3D identification apparatus, comprising:
the light emitting module 10 is configured to emit light to the surface of the object to be measured;
the first receiving module 20 is configured to receive the light reflected by at least part of the surface of the object to be measured;
a second receiving module 30 configured to receive at least a part of the light emitted from the surface of the object to be measured;
the first receiving module 20 is provided with a first polarizer 201 on a side close to the object to be measured, and/or the second receiving module 30 is provided with a second polarizer 301 on a side close to the object to be measured.
In addition, the 3D identification system provided by the embodiment of the present application further includes a color module 40 configured to obtain a color image of the surface of the measured object. The color image and the texture information of the object can be rendered and output as a three-dimensional image of the scene by the data processor unit after being matched with the depth image.
In some embodiments, the polarization angles of the first polarizer 201 and the second polarizer 301 are the same, the polarization angle of the first polarizer 201 is 0 to 180 °, and the polarization angle of the second polarizer 301 is 0 to 180 °.
In some embodiments, the polarization angles of the first polarizer 201 and the second polarizer 301 are different, the polarization angle of the first polarizer 201 is 0 to 180 °, and the polarization angle of the second polarizer 301 is 0 to 180 °.
Various arrangements of the polarizing plates can be applied: 1) one module is provided with a polaroid and the other module is not provided with the polaroid; 2) the two modules are provided with polaroids, and the angles of the light rays passing through the polaroids are consistent, as shown in FIG. 3, the polarization angles of the first polaroid 201 and the second polaroid 301 are consistent); 3) the two modules are provided with polarizers, and the angles of the light rays passing through the polarizers are different, as shown in fig. 4, the polarization angle of the first polarizer 201 is 0 ° and the polarization angle of the second polarizer 301 is 90 °; 4) the two modules are provided with polaroids, the angles of light rays passing through the polaroids are different, and the polarization angles of the two polaroids can be random and can be between 0 and 180 degrees.
In addition, in a specific setting, the first receiving module 20 and the second receiving module 30 are located on the same side of the light emitting module 10 or the first receiving module 20 and the second receiving module 30 are located on different sides of the light emitting module 10. In the embodiment of the present application, the positions of the first receiving module 20 and the second receiving module 30 are not limited, and different identification devices may be applied according to different application scenarios.
The light emitting module 10 includes, but is not limited to, a light source including one or more of a high contrast VCSEL, a single aperture broad area VCSEL, an array VCSEL, a laser diode, and an LED light source, and a diffractive optical element. The light emitting module 10 may further include optical elements with other functions, and when the light emitting module is specifically configured, the light source may be appropriately adjusted according to different application scenarios, such as a micro lens array, a grating, and the like.
Wherein the light source (not shown) is used for emitting a light beam, and the diffractive optical element (not shown) is arranged on the light emitting side of the light source and used for receiving the light beam and projecting the light to the surface of the measured object. Diffractive Optical Elements (DOE, full name: Diffraction Optical Elements) realize multiple Optical outputs.
The light source may be an infrared light beam, and the light source may also be a vcsel (vertical Cavity Surface Emitting laser), and the light source emits infrared structured light and/or floodlight as an object identification signal through the diffractive optical element.
The light beam emitted by the light source may be a pattern light beam, and the pattern light beam is, for example, in a pattern of an irregular lattice, a grid, a stripe, or an encoded pattern. The diffractive optical element may split and expand the light beam emitted from the light source into a plurality of light beams as the object recognition signal. The diffractive optical element can also expand the light beam emitted by the light source into a plurality of infrared floodlights substantially covering the space to be illuminated, with a substantially uniform light intensity distribution in each region of the space to be illuminated.
It should be noted that, in the embodiment of the present application, the object identification signal is, for example, but not limited to, a pattern beam and an image. The pattern light beam is in a pattern of, for example, an irregular lattice, a grid, a stripe, or an encoded pattern. The 3D recognition device in the embodiment of the application can be used in various 3D technologies, and the 3D technologies in the industry at present mainly comprise structured light, TOF (time of flight), binocular and the like.
Example one
The application provides a 3D identification system, wherein, 3D identification system is 3D structure light identification system, as shown in fig. 1, includes:
the light emitting module 10 is configured to emit the coded structured light to the surface of the measured object;
the first receiving module 20 is configured to receive the structured light reflected by at least part of the surface of the object to be measured to obtain a first image related to the object to be measured;
the second receiving module 30 is configured to receive the structured light reflected by at least part of the surface of the object to be measured to obtain a second image related to the object to be measured;
and the data processing unit is configured to acquire the first image and the second image and establish 3D depth information of the measured object based on the first image and the second image.
When the 3D structured light shoots an object, because of the fact that glare and other conditions appear in partial images due to some special reasons, the shot object cannot be seen clearly, on the basis, the polarizing films of the two receiving modules have different effects, and it can be seen that the right image received by the receiving modules can be distinguished as a light identification point, and the rear-end algorithm calculation with better imaging effect is selected to output better depth information and depth images.
As shown in fig. 5, the present application provides a 3D identification method applied to the 3D identification system as described above, including:
ST11, emitting light to the surface of the measured object through the light emitting module 10;
ST12, acquiring a first image related to the object to be measured through the first receiving module 20;
ST13, acquiring a second image related to the object to be measured through the second receiving module 30;
ST14, comparing the first image and the second image through a data processing unit, and selecting one of the images as a standard image, or processing the first image and the second image through the data processing unit to obtain the standard image;
ST15, carrying out post-processing on the standard image to obtain the 3D depth information of the measured object.
In step ST11, the light emitted by the light emitting module 10 to the surface of the object to be measured is a pattern beam, which is, for example, in the form of an irregular lattice, a grid, a stripe, or a code pattern. In the embodiments of the present application, an irregular lattice is taken as an exemplary illustration.
In step ST14, one of the first image and the second image, which is superior in image effect, may be selected as a standard image for post-processing; or the first image and the second image are superposed and synthesized to form a new image as a standard image for post-processing.
It should be noted that, in the embodiment of the present application, the data processing unit processes the first image and the second image, including but not limited to image transformation, image enhancement and restoration, image segmentation, image coding, image description, and image recognition.
For the image processing domain, image transformation: indirect processing techniques such as fourier transform, walsh transform, discrete cosine transform, etc., convert the processing in the spatial domain into processing in the transform domain, which not only reduces the amount of computation, but also allows for more efficient processing. Image coding: compressed image coding compression techniques may reduce the amount of data (i.e., the number of bits) describing an image in order to save image transmission, processing time, and reduce the amount of memory occupied. Image enhancement and restoration: the purpose of image enhancement and restoration is to improve the quality of an image, such as removing noise, improving the sharpness of an image, and the like. Image segmentation: image segmentation is one of the key techniques in digital image processing. The image segmentation is to extract a meaningful characteristic part in the image, wherein the meaningful characteristic is an edge, a region and the like in the image, and the meaningful characteristic is a basis for further image recognition, analysis and understanding. Image description: image description is a necessary prerequisite for image recognition and understanding. As the simplest binary image, the geometric characteristics of the binary image can be used for describing the characteristics of an object, and a general image description method adopts two-dimensional shape description which has two types of methods of boundary description and region description. Image recognition: the image recognition belongs to the category of pattern recognition, and the main content of the image recognition is that after certain preprocessing (enhancement, restoration and compression), the image is subjected to image segmentation and feature extraction, so that judgment and classification are performed.
The projected set of beams in known spatial directions is called structured light (structured light). 3D structured light principle, as shown in fig. 6: the method is characterized in that invisible infrared laser with a specific wavelength is used as a light source, emitted light forms an image with a certain encoding rule through an optical diffraction element (DOE) and is projected on an object, meanwhile, a receiving module receives the image with the encoding rule on the surface of the object, and then the distortion of a returned encoding pattern is calculated based on an optical triangulation method measuring principle to obtain the position and depth information of the object. Namely, the 3D structured light module can set a reference picture in a burning way on the reference surface under a specific distance, the reference picture contains some specific information, and the picture of the actually shot object surface is compared with the previously determined reference picture, and the difference between the picture and the previously determined reference picture is confirmed and then converted into a depth map for displaying after algorithm calculation.
It should be noted that, in the embodiment of the present application, the standard image is post-processed by using other processing methods in the prior art to obtain the image depth map, which is not described herein again, and any processing method belongs to the protection scope of the present application without departing from the inventive concept of the present application.
The use scenes of payment equipment and the like in the consumer payment industry are gradually expanded from indoor to outdoor, but the coded image of the surface of an object is often submerged in external ambient light under the outdoor high illumination condition of structured light, so that the depth information of the object cannot be effectively identified and acquired, the depth image imaging of the object is lost, and the payment failure is caused.
In a normal environment, the matching of the light emitting module 10 and a normal receiving module (for example, a receiving module without a polarizer) can generate a depth map in real time, and no problem occurs; if the situation of the overexposure of the ambient light as shown in fig. 7 occurs, the whole brightness is reduced through the collocation of the transmitting module and the receiving module with the polaroid, so that the spot points can be distinguished from the ambient light, as shown in fig. 8, the depth map can be ensured to be output normally, and the effect of the depth map under the outdoor high-illumination situation can be improved.
Example two
The application provides a 3D identification system, 3D identification system is 3D TOF identification system, as shown in FIG. 2, includes:
a light emitting module 10 configured to emit TOF light toward the surface of the object to be measured;
a first receiving module 20 configured to receive TOF light reflected by at least a portion of the surface of the object to be measured to obtain a first image related to the object to be measured;
a second receiving module 30 configured to receive TOF light reflected by at least a portion of the surface of the object to be measured to obtain a second image related to the object to be measured;
and the data processing unit is configured to acquire the first image and the second image and establish 3D depth information of the measured object based on the first image and the second image.
When shooting an object in the 3D TOF, due to the fact that glare and other conditions occur in partial images caused by some special reasons, the object cannot be seen clearly, on the basis, the polarizing films of the two receiving modules have different effects, it can be seen that the right image received by the receiving modules can be distinguished from the object, the rear-end algorithm calculation is carried out by selecting the receiving modules with better imaging effect, and better depth information and depth images are output.
As shown in fig. 9, the present application provides a 3D identification method applied to the 3D identification system as described in any one of the above, including:
ST21, emitting TOF light to the surface of the measured object through the light emitting module 10;
ST22, acquiring a first image related to the object to be measured through the first receiving module 20;
ST23, acquiring a second image related to the object to be measured through the second receiving module 30;
ST24, comparing the first image and the second image through a data processing unit, and selecting one of the images as a standard image, or processing the first image and the second image through the data processing unit to obtain the standard image;
ST25, carrying out post-processing on the standard image to obtain the 3D depth information of the measured object.
TOF (time of flight), transliterated as time of flight. The working principle of the TOF imaging device is that a laser emission module 10 continuously transmits light pulses to a measured object, then a TOF camera receives light signals returned from the object, and finally a control system of the TOF imaging device calculates the flight (round trip) time of the emitted and received light signals to obtain the distance of the measured object. Different from ultrasonic ranging, ultrasonic ranging has higher requirement on a reflecting object, objects with small areas, such as lines and conical objects, can not be basically measured, and TOF imaging devices have lower requirements on the size, the area and the like of the measured objects, and are high in ranging precision, long in ranging distance and fast in response.
TOF can be classified into:
1) pulse Modulation (DTOF): measuring and calculating the distance directly according to the time difference between the pulse transmission and the pulse reception; as shown in fig. 10.
2) Continuous Wave Modulation (ITOF): sinusoidal modulation is commonly employed. Since the phase shift of the sinusoidal waves of the receiving end and the transmitting end is proportional to the distance of the object from the camera, the distance can be measured by using the phase shift. As shown in fig. 11.
It should be noted that, in the embodiment of the present application, the standard image is post-processed by using other methods in the prior art to obtain the image depth map, which is not described herein again, and any processing method is within the protection scope of the present application without departing from the inventive concept of the present application.
EXAMPLE III
The application provides a 3D identification system, the 3D identification system is binocular vision identification system, including initiative binocular vision unit 100 and passive binocular vision unit 200, wherein, as shown in fig. 14, initiative binocular vision unit 100 includes as above arbitrary 3D recognition device, as shown in fig. 15, be provided with third receiving module 50 and fourth receiving module 60 in the passive binocular vision unit 200, third receiving module 50 is provided with third polaroid 501 in the one side of being close to the measured object, and/or fourth receiving module 60 is provided with the fourth polaroid in the one side of being close to the measured object, wherein,
the light emitting module 10 is configured to emit the coded structured light or TOF light to the surface of the measured object;
the first receiving module 20 is configured to receive the structured light or the TOF light reflected by at least a part of the surface of the object to be measured to obtain a first image related to the object to be measured;
the second receiving module 30 is configured to receive the structured light or the TOF light reflected by at least a part of the surface of the object to be measured to obtain a second image related to the object to be measured;
the third receiving module 50 is configured to receive the structured light or the TOF light reflected by at least a part of the surface of the object to be measured to obtain a third image related to the object to be measured;
the fourth receiving module 60 is configured to receive the structured light or the TOF light reflected by at least a part of the surface of the object to be measured to obtain a fourth image related to the object to be measured;
the data processing unit is configured to acquire the first image and the second image, and establish active binocular display 3D depth information of the object to be measured based on the first image and the second image; and the image acquisition unit is also configured to acquire the third image and the fourth image and establish passive binocular display 3D depth information of the object to be measured based on the third image and the fourth image.
In the 3D binocular vision system provided by the application, the light source used in active binocular may be the light emitting module 10 of one of the multiple light sources explained above, polarizing plates which can pass through the same or different angles are arranged in front of the two receiving modules, and 3D depth information of an object can be better output after images respectively received by the two receiving modules are calculated through a rear-end algorithm. And a light source is not used in the passive binocular, polarizing films which can pass through the same or different angles are arranged at the front parts of the two receiving modules, and 3D depth information of an object can be better output after the images respectively received by the two receiving modules are calculated through a rear-end algorithm.
The application provides a 3D identification method, applied to the 3D identification system as described in any of the above, comprising:
ST31, emitting light to the surface of the measured object through the light emitting module 10;
ST32, acquiring a first image related to the object to be measured through the first receiving module 20;
ST33, acquiring a second image related to the object to be measured through the second receiving module 30;
ST34, comparing the first image and the second image through a data processing unit, and selecting one of the images as a standard image, or processing the first image and the second image through the data processing unit to obtain the standard image;
ST35, carrying out post-processing on the standard image to obtain the 3D depth information of the measured object.
As shown in fig. 16, binocular stereoscopic vision is a method for acquiring three-dimensional geometric information of an object by acquiring two images of the object to be measured from different positions by using an imaging device based on a parallax principle and calculating a position deviation between corresponding points of the images, and the binocular vision only depends on the images for feature matching, and the positions of two cameras in the binocular vision must be accurately calibrated before using a binocular vision camera.
In the embodiment of the present application, different working modes can be executed according to different application scenarios and environmental conditions, for example: when the target object is close to the application scene with extremely high requirements on the precision of 3D imaging, such as face authentication, face payment and the like, a monocular mode can be executed; it can be understood that, in the monocular mode, the measurement range is often relatively small, and therefore, if a depth image with a wider measurement range is to be measured, the binocular mode can be performed, and although the projection field angle of the projection module is not changed, since the total collection field angle at which the two receiving modules collect the structured light image is larger than the collection field angle at which the single receiving module collects the structured light image, an image with a wider field angle than that of the single receiving module can be obtained by using the two receiving modules.
It should be noted that, in the embodiment of the present application, the light emitting module 10 in the active binocular mode may adopt a structured light unit structure, and may also adopt a TOF unit structure, which is not limited in this application. For the processing modes of the structured light and the TOF light, different modes in the prior art can be adopted, which is not described herein again, and any processing mode is within the protection scope of the present application without departing from the inventive concept of the present application.
It will be understood that the terms "length," "width," "upper," "lower," "front," "rear," "left," "right," "vertical," "horizontal," "top," "bottom," "inner," "outer," and the like, as used herein, refer to an orientation or positional relationship indicated in the drawings that is solely for the purpose of facilitating the description and simplifying the description, and do not indicate or imply that the device or element being referred to must have a particular orientation, be constructed and operated in a particular orientation, and is therefore not to be construed as limiting the invention.
Furthermore, the terms "first", "second" and "first" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include one or more of that feature. In the description of the present invention, "a plurality" means two or more unless specifically defined otherwise.
Unless defined otherwise, technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. Terms such as "disposed" and the like, as used herein, may refer to one element being directly attached to another element or one element being attached to another element through intervening elements. Features described herein in one embodiment may be applied to another embodiment, either alone or in combination with other features, unless the feature is otherwise inapplicable or otherwise stated in the other embodiment.
The present invention has been described in terms of the above embodiments, but it should be understood that the above embodiments are for purposes of illustration and description only and are not intended to limit the invention to the scope of the described embodiments. It will be appreciated by those skilled in the art that many variations and modifications may be made to the teachings of the invention, which fall within the scope of the invention as claimed.