[go: up one dir, main page]

CN114821479B - Image recognition method, device and equipment - Google Patents

Image recognition method, device and equipment Download PDF

Info

Publication number
CN114821479B
CN114821479B CN202210491488.7A CN202210491488A CN114821479B CN 114821479 B CN114821479 B CN 114821479B CN 202210491488 A CN202210491488 A CN 202210491488A CN 114821479 B CN114821479 B CN 114821479B
Authority
CN
China
Prior art keywords
information
target
recognition
recognition target
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210491488.7A
Other languages
Chinese (zh)
Other versions
CN114821479A (en
Inventor
李劼
邬浩
王益安
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Datamesh Technology Co ltd
Original Assignee
Beijing Datamesh Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Datamesh Technology Co ltd filed Critical Beijing Datamesh Technology Co ltd
Priority to CN202210491488.7A priority Critical patent/CN114821479B/en
Publication of CN114821479A publication Critical patent/CN114821479A/en
Application granted granted Critical
Publication of CN114821479B publication Critical patent/CN114821479B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

本申请实施例涉及计算机领域,公开了一种图像识别方法、装置及设备。获取识别目标的图片信息,每个图片信息包括:每个识别目标的颜色信息;依据识别目标的颜色信息得到识别目标的第一坐标信息;获取识别目标的深度图信息;得到识别目标对应的空间变换特征;依据识别目标对应的第一坐标信息以及空间变换特征,得到识别目标对应的第二坐标信息;依据识别目标对应的第二坐标信息得到维度信息。可见,本技术方案通过利用目标对象的图片信息以及相应的空间特征变换,将原有的3D识别对象的识别过程转换为2D图片信息的识别。这样,可以使目标对象的识别过程更简易,降低应用难度;另一方面,实现了采用2D图片信息对3D识别对象的训练,提高训练效率。

The embodiments of the present application relate to the field of computers, and disclose an image recognition method, device, and equipment. Obtain image information of the recognition target, each image information includes: color information of each recognition target; obtain first coordinate information of the recognition target based on the color information of the recognition target; obtain depth map information of the recognition target; obtain spatial transformation features corresponding to the recognition target; obtain second coordinate information corresponding to the recognition target based on the first coordinate information and spatial transformation features corresponding to the recognition target; obtain dimension information based on the second coordinate information corresponding to the recognition target. It can be seen that the technical solution converts the original 3D recognition object recognition process into 2D image information recognition by utilizing the image information of the target object and the corresponding spatial feature transformation. In this way, the recognition process of the target object can be made simpler and the application difficulty can be reduced; on the other hand, the training of 3D recognition objects using 2D image information is realized, thereby improving training efficiency.

Description

Image recognition method, device and equipment
Technical Field
The embodiment of the invention relates to the field of computers, and relates to an image identification method, device and equipment.
Background
Image recognition technology refers to technology that processes, analyzes, and understands images with a computer to recognize recognition objects of various different modes. The recognition objects of the existing image recognition technology are generally divided into two-dimensional (2D) objects and three-dimensional (3D) objects according to dimension division. In some scenes identified by detection, a computer is generally required to identify and distinguish a 2D object from a 3D object according to picture information.
For the recognition of the 2D object, the 2D picture information is usually combined with a corresponding learning method to train the computer, the method is mature, and the recognition process is simple. For 3D object recognition, depth information of the 3D object is generally acquired by a plurality of image acquisition devices arranged at different angles, and corresponding spatial characteristics of the 3D object are obtained by combining a point cloud technology, so that a computer can complete 3D object recognition through the corresponding spatial characteristics.
On the one hand, the current recognition of the 3D object is complicated in the process of obtaining depth information and applying the point cloud technology, so that the actual application difficulty is high, and on the other hand, the training object is an actual object, so that the available training set resources are less than those of the training set resources available for recognizing the 2D object, and the training efficiency is low.
Disclosure of Invention
The embodiment of the application provides an image recognition method, device and equipment, which are used for solving the problems of high difficulty in distinguishing and recognizing 3D images and low training efficiency in the existing detection and recognition scene.
In a first aspect, an embodiment of the present application provides an image recognition method, where the method includes:
Acquiring picture information of at least one identification target, wherein each identification target corresponds to one piece of picture information, and each piece of picture information comprises color information corresponding to each identification target;
Obtaining first coordinate information of each identification target in corresponding picture information according to the color information corresponding to each identification target;
Acquiring depth map information of at least one identification target, wherein each identification target corresponds to one depth map information;
Obtaining the space transformation characteristics corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target;
Obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristic corresponding to each recognition target, wherein the second coordinate information acts on the depth map information;
and obtaining the dimension information of the identification target according to the second coordinate information corresponding to the identification target.
Therefore, the spatial transformation characteristic relation is established through the picture information and the depth map information of the identification target, the dimension information of the identification target can be obtained through calculation through the pixel coordinates obtained through the picture information, and the identification difficulty of the 3D identification target is reduced.
In some possible embodiments, the picture information is acquired by a picture acquisition device and the depth map information is acquired by a grey-scale map acquisition device.
Therefore, the acquired picture information and depth map information can be mutually independent, and the subsequent establishment of the space transformation characteristic is facilitated.
In some possible implementations, the first coordinate information of each recognition target in the corresponding picture information includes four pixel coordinate information of an outline of each recognition target in the corresponding picture information.
Therefore, the rectangular region where the outer contour of the identification target is located can be obtained through the four pixel coordinate information, and the second coordinate information can be obtained through the space transformation characteristic conveniently.
In some possible embodiments, the spatial transformation characteristics are derived from a distortion correction matrix.
In some possible implementations, the second coordinate information corresponding to each recognition target includes four depth value information of an outer contour of each recognition target in a corresponding gray scale map, and the pixel coordinate information of the outer contour of each recognition target corresponds to the depth value information of the outer contour of each recognition target one by one.
In this way, the first coordinate information can be converted into the second coordinate information by the spatial transformation feature, and the feature of the identification target in the picture information can be represented in the gray scale map.
In some possible embodiments, the obtaining the dimension information of the identification target according to the second coordinate information corresponding to the identification target includes:
if the second coordinate information is larger than or equal to a preset threshold value, judging that the dimension information corresponding to the identification target is a three-dimensional target;
and if the second coordinate information is smaller than a preset threshold value, judging that the dimension information corresponding to the identification target is a two-dimensional target.
In some possible embodiments, the obtaining the first coordinate information of each recognition target in the corresponding picture information according to the color information corresponding to each recognition target is implemented by a target detection neural network, R-CNN.
In a second aspect, an embodiment of the present application further provides an image recognition apparatus, including:
the first acquisition module is used for acquiring picture information of at least one identification target, wherein each identification target corresponds to one piece of picture information, and each piece of picture information comprises color information corresponding to each identification target;
The first processing module is used for obtaining first coordinate information of each identification target in the corresponding picture information according to the color information corresponding to each identification target;
the second acquisition module is used for acquiring the depth map information of the at least one identification target, wherein each identification target corresponds to one depth map information;
The feature construction module is used for obtaining the space transformation feature corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target;
the second processing module is used for obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristic corresponding to each recognition target, and the second coordinate information acts on the depth map information;
And the judging module is used for obtaining the dimension information of the identification target according to the second coordinate information corresponding to the identification target.
In a third aspect, the embodiment of the application also provides an electronic device, which comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete communication with each other through the communication bus;
The memory is configured to store executable instructions that, when executed, cause the processor to perform the test method of any of the possible implementations of the first or second aspect.
In a fourth aspect, embodiments of the present application also provide a computer-readable storage medium having stored therein executable instructions that, when executed, cause a computing device to perform the test method of any of the possible implementations of the first or second aspects.
The embodiment of the application provides a technical scheme of an image recognition method, which comprises the steps of firstly obtaining picture information of at least one recognition target, wherein each recognition target corresponds to one picture information, each picture information comprises color information corresponding to each recognition target, obtaining first coordinate information of each recognition target in the corresponding picture information according to the color information corresponding to each recognition target, obtaining depth map information of the at least one recognition target, wherein each recognition target corresponds to one depth map information, obtaining space transformation characteristics corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target, obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristics corresponding to each recognition target, and acting on the depth map information according to the second coordinate information corresponding to each recognition target. The color information of the identification target is used for obtaining the coordinates of the position of the identification target in the picture information in the process, the second space coordinates corresponding to the identification target in the gray level diagram can be obtained through calculation according to the space feature transformation between the picture information and the depth diagram information, the size of the second space coordinates is combined with the threshold value preset in the corresponding gray level diagram for comparison, and the final judging result is used as confirmation of the dimension information of the identification target. Therefore, the technical scheme of the application converts the original identification process of the 3D identification object into the identification process of the 2D picture information by utilizing the picture information of the target object and the corresponding spatial feature transformation. Therefore, on one hand, the recognition process of the target object is simpler, the application difficulty is reduced, and on the other hand, the training of recognizing the 3D recognition object by adopting the 2D picture information is realized, and the training efficiency is improved.
Drawings
FIG. 1 is a schematic flow chart of an image processing method according to an embodiment of the present application;
fig. 2 is a schematic diagram illustrating an exemplary composition of an image processing apparatus according to an embodiment of the present application;
fig. 3 is a schematic diagram of an exemplary structure of an image processing apparatus according to an embodiment of the present application.
Detailed Description
The terminology used in the following examples of the application is for the purpose of describing alternative embodiments only and is not intended to be limiting of the application. As used in the specification of the present application and the appended claims, the singular forms "a," "an," "the," and "the" are intended to include the plural forms as well. It should also be understood that, although the terms first, second, etc. may be used in the following embodiments to describe some type of object, the object is not limited to these terms. These terms are used to distinguish between specific objects of that class of objects. For example, other classes of objects that may be described in the following embodiments using the terms first, second, etc. are not described in detail herein.
The embodiment of the application provides a technical scheme of an image recognition method, which comprises the steps of firstly obtaining picture information of at least one recognition target, wherein each recognition target corresponds to one picture information, each picture information comprises color information corresponding to each recognition target, obtaining first coordinate information of each recognition target in the corresponding picture information according to the color information corresponding to each recognition target, obtaining depth map information of the at least one recognition target, wherein each recognition target corresponds to one depth map information, obtaining space transformation characteristics corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target, obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristics corresponding to each recognition target, and acting on the depth map information according to the second coordinate information corresponding to each recognition target. The color information of the identification target is used for obtaining the coordinates of the position of the identification target in the picture information in the process, the second space coordinates corresponding to the identification target in the gray level diagram can be obtained through calculation according to the space feature transformation between the picture information and the depth diagram information, the size of the second space coordinates is combined with the threshold value preset in the corresponding gray level diagram for comparison, and the final judging result is used as confirmation of the dimension information of the identification target. Therefore, the technical scheme of the application converts the original identification process of the 3D identification object into the identification process of the 2D picture information by utilizing the picture information of the target object and the corresponding spatial feature transformation. Therefore, on one hand, the recognition process of the target object is simpler, the application difficulty is reduced, and on the other hand, the training of recognizing the 3D recognition object by adopting the 2D picture information is realized, and the training efficiency is improved.
Any of the electronic devices to which embodiments of the present application relate may be electronic devices such as cell phones, tablet computers, wearable devices (e.g., smartwatches, smartbracelets, etc.), notebook computers, desktop computers, and vehicle-mounted devices. The electronic device is pre-installed with a software deployment application. It will be appreciated that embodiments of the present application are not limited in any way by the particular type of electronic device.
Image recognition technology refers to technology that processes, analyzes, and understands images with a computer to recognize recognition objects of various different modes. The recognition objects of the existing image recognition technology are generally divided into two-dimensional (2D) objects and three-dimensional (3D) objects according to dimension division. In some scenes identified by detection, a computer is generally required to identify and distinguish a 2D object from a 3D object according to picture information.
For the recognition of the 2D object, the 2D picture information is usually combined with a corresponding learning method to train the computer, the method is mature, and the recognition process is simple. For 3D object recognition, depth information of the 3D object is generally acquired by a plurality of image acquisition devices arranged at different angles, and corresponding spatial characteristics of the 3D object are obtained by combining a point cloud technology, so that a computer can complete 3D object recognition through the corresponding spatial characteristics.
On the one hand, the current recognition of the 3D object is complicated in the process of obtaining depth information and applying the point cloud technology, so that the actual application difficulty is high, and on the other hand, the training object is an actual object, so that the available training set resources are less than those of the training set resources available for recognizing the 2D object, and the training efficiency is low.
The following is a description of several exemplary embodiments, and describes technical solutions of the embodiments of the present application and technical effects produced by the technical solutions of the present application.
In a first aspect of the present application, an image recognition method is provided, referring to fig. 1, fig. 1 illustrates a flowchart of an image processing method provided by an embodiment of the present application, including the following steps:
Acquiring picture information of at least one identification target, wherein each identification target corresponds to one piece of picture information, and each piece of picture information comprises color information corresponding to each identification target;
Obtaining first coordinate information of each identification target in corresponding picture information according to the color information corresponding to each identification target;
Acquiring depth map information of at least one identification target, wherein each identification target corresponds to one depth map information;
Obtaining the space transformation characteristics corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target;
Obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristic corresponding to each recognition target, wherein the second coordinate information acts on the depth map information;
and obtaining the dimension information of the identification target according to the second coordinate information corresponding to the identification target.
Optionally, the picture information is acquired by a picture acquisition device (e.g. a high definition camera), and the depth map information is acquired by a depth map (e.g. a depth camera) acquisition device.
Optionally, for the obtained depth map information, the expression form may be presented in a gray scale form, and taking a K device of a certain platform as an example, the number of the obtained depth map data information is the same as the number of pixels, and the value range may be 0-65535 frames. And according to different equipment precision of practical application, performing corresponding mapping processing on the value range to 0-1 frame to obtain a corresponding gray level map.
Optionally, the first coordinate information of each identification target in the corresponding picture information comprises four pixel coordinate information of the outline of each identification target in the corresponding picture information.
Alternatively, the four pixel coordinate information may be the abscissa information corresponding to two pixel points, for example, the coordinate (x 1, y 1) corresponding to the pixel point a, the coordinate (x 2, y 2) corresponding to the pixel point B, and the four pixel coordinate information may be the x1, y1, x2 and y2.
Optionally, the four pixel coordinate information includes x (the abscissa of the rectangular plane coordinate system where the object is located), y (the ordinate of the rectangular plane coordinate system where the object is located), width (the horizontal width corresponding to the object), and height (the vertical height corresponding to the object).
Optionally, the spatial transformation characteristic is obtained from a distortion correction matrix.
Optionally, the second coordinate information corresponding to each identification target includes four depth value information of the outer contour of each identification target in the corresponding gray scale map, and the pixel coordinate information of the outer contour of each identification target corresponds to the depth value information of the outer contour of each identification target one by one.
Optionally, the obtaining the dimension information of the identification target according to the second coordinate information corresponding to the identification target includes:
if the second coordinate information is larger than or equal to a preset threshold value, judging that the dimension information corresponding to the identification target is a three-dimensional target;
and if the second coordinate information is smaller than a preset threshold value, judging that the dimension information corresponding to the identification target is a two-dimensional target.
Optionally, the first coordinate information of each recognition target in the corresponding picture information is obtained according to the color information corresponding to each recognition target, and is implemented through a target detection neural network, and R-CNN.
For example, a company needs to identify a specific beverage quantity on a shelf (i.e. three-dimensional identification objects), but needs to distinguish beverage bottle pictures (i.e. two-dimensional identification objects) on a billboard placed near the shelf, so as to avoid interference during quantity comparison, i.e. realize the distinction of two-dimensional information and three-dimensional information of identification targets.
First, the image capturing device and the gray-scale image capturing device may be used to obtain color information and depth map information for the beverage bottle image on the billboard. Thus, color information and depth map information corresponding to the two-dimensional recognition target can be obtained;
Obtaining pixel coordinates (namely first coordinate information) corresponding to the outline of the beverage bottle in the billboard or four pixel vertex coordinates (also can be used as the first coordinate information) of a rectangular frame of the area of the beverage bottle in the billboard by training the color information and combining with a corresponding neural network (for example, the neural network can identify the position of the beverage bottle in the picture according to the color information);
According to the obtained picture information and depth map information, a distortion correction matrix (namely a space transformation characteristic) corresponding to the beverage bottle in the billboard is obtained, and the relation of the gray value corresponding to the identification object (namely the beverage bottle in the billboard) in the picture information pixel by pixel can be obtained through the matrix;
Further, by acquiring the depth map information of the two-dimensional recognition target (including the second coordinate information and the corresponding depth value information), a threshold value for distinguishing the dimension information of the recognition target may be obtained and set, for example, when the depth value information of the recognition target is smaller than the threshold value, the recognition target is determined to be a two-dimensional target (i.e., the dimension information of the recognition target is a two-dimensional target), and when the depth value information of the recognition target is greater than or equal to the threshold value, the recognition target is determined to be a three-dimensional target (i.e., the dimension information of the recognition target is a three-dimensional target). In the process, two-dimensional and three-dimensional distinguishing and identification are not carried out on the identification target by means of a plurality of depth image acquisition devices and in a point cloud mode, so that the distinguishing process of the two-dimensional and three-dimensional identification target is simplified. And the process can obtain a training set of the two-dimensional target, and the training set is used for training and identifying the two-dimensional target by the system.
Likewise, the three-dimensional recognition target (i.e., the beverage bottle on the shelf) is recognized, and the above process is not described in detail herein. Finally, a training set of the three-dimensional target is obtained, and the training set of the three-dimensional target is also a picture set because the picture information and the gray picture information are adopted for the three-dimensional target. Therefore, training of the three-dimensional target by using the picture set (two-dimensional training set) can be achieved, and training efficiency is improved.
The embodiment of the application provides a technical scheme of an image recognition method, which comprises the steps of firstly obtaining picture information of at least one recognition target, wherein each recognition target corresponds to one picture information, each picture information comprises color information corresponding to each recognition target, obtaining first coordinate information of each recognition target in the corresponding picture information according to the color information corresponding to each recognition target, obtaining depth map information of the at least one recognition target, wherein each recognition target corresponds to one depth map information, obtaining space transformation characteristics corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target, obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristics corresponding to each recognition target, and acting on the depth map information according to the second coordinate information corresponding to each recognition target. The color information of the identification target is used for obtaining the coordinates of the position of the identification target in the picture information in the process, the second space coordinates corresponding to the identification target in the gray level diagram can be obtained through calculation according to the space feature transformation between the picture information and the depth diagram information, the size of the second space coordinates is combined with the threshold value preset in the corresponding gray level diagram for comparison, and the final judging result is used as confirmation of the dimension information of the identification target. It can be seen that the original process of recognizing the 3D recognition object is converted into a process of recognizing the 2D picture information by using the picture information of the target object and the corresponding spatial feature transformation. Therefore, on one hand, the recognition process of the target object is simpler, the application difficulty is reduced, and on the other hand, the training of recognizing the 3D recognition object by adopting the 2D picture information is realized, and the training efficiency is improved.
The above embodiments describe each implementation of the image processing method provided by the embodiment of the present application from aspects of at least one image information of an identification target, obtaining first coordinate information, obtaining depth map information of at least one identification target, obtaining a spatial transformation feature, obtaining second coordinate information, obtaining target dimension information, and the like. It should be understood that the above-mentioned functions may be implemented in hardware or a combination of hardware and computer software in the embodiments of the present application by the processing steps of acquiring the picture information of at least one recognition target, acquiring the first coordinate information, acquiring the depth map information of at least one recognition target, acquiring the spatial transformation feature, acquiring the second coordinate information, acquiring the dimension information of the target, and so on. Whether a function is implemented as hardware or computer software driven hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
For example, if the above implementation steps are implemented by software modules, the corresponding functions are implemented. As shown in fig. 2, the image recognition apparatus may include a first acquisition module, a first processing module, a second acquisition module, a feature construction module, a second processing module, and a determination module. The image processing apparatus may be configured to perform part or all of the operations of the image processing method described above.
For example:
the first acquisition module is used for acquiring picture information of at least one identification target, wherein each identification target corresponds to one piece of picture information, and each piece of picture information comprises color information corresponding to each identification target;
The first processing module is used for obtaining first coordinate information of each identification target in the corresponding picture information according to the color information corresponding to each identification target;
the second acquisition module is used for acquiring the depth map information of the at least one identification target, wherein each identification target corresponds to one depth map information;
The feature construction module is used for obtaining the space transformation feature corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target;
the second processing module is used for obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristic corresponding to each recognition target, and the second coordinate information acts on the depth map information;
And the judging module is used for obtaining the dimension information of the identification target according to the second coordinate information corresponding to the identification target.
It can be seen that in this way,
The embodiment of the application provides a technical scheme of an image recognition method, which comprises the steps of firstly obtaining picture information of at least one recognition target, wherein each recognition target corresponds to one picture information, each picture information comprises color information corresponding to each recognition target, obtaining first coordinate information of each recognition target in the corresponding picture information according to the color information corresponding to each recognition target, obtaining depth map information of the at least one recognition target, wherein each recognition target corresponds to one depth map information, obtaining space transformation characteristics corresponding to each recognition target according to the picture information corresponding to each recognition target and the depth map information corresponding to each recognition target, obtaining second coordinate information corresponding to each recognition target according to the first coordinate information corresponding to each recognition target and the space transformation characteristics corresponding to each recognition target, and acting on the depth map information according to the second coordinate information corresponding to each recognition target. The color information of the identification target is used for obtaining the coordinates of the position of the identification target in the picture information in the process, the second space coordinates corresponding to the identification target in the gray level diagram can be obtained through calculation according to the space feature transformation between the picture information and the depth diagram information, the size of the second space coordinates is combined with the threshold value preset in the corresponding gray level diagram for comparison, and the final judging result is used as confirmation of the dimension information of the identification target. Therefore, the technical scheme of the application converts the original identification process of the 3D identification object into the identification process of the 2D picture information by utilizing the picture information of the target object and the corresponding spatial feature transformation. Therefore, on one hand, the recognition process of the target object is simpler, the application difficulty is reduced, and on the other hand, the training of recognizing the 3D recognition object by adopting the 2D picture information is realized, and the training efficiency is improved.
It will be appreciated that the functions of the above modules may be integrated into a hardware entity implementation, for example, the acquisition module may be integrated into a transceiver implementation, the generation module, the selection module, and the construction module may be integrated into a processor implementation, and programs and instructions implementing the functions of the above modules may be maintained in a memory. As shown in fig. 3, there is provided an electronic device including a processor, a transceiver for performing image information and object picture acquisition in the image generation method, and a memory for storing the program/code preloaded by the aforementioned deployment apparatus, or for storing the code for execution by the processor, or the like. The processor, when executing the code stored in the memory, causes the electronic device to perform some or all of the operations of the software deployment method in the method.
The specific procedures are described in detail in the examples of the above method and are not described in detail here.
In a specific implementation, corresponding to the foregoing electronic device, the embodiment of the present application further provides a computer storage medium, where the computer storage medium provided in the electronic device may store a program, and when the program is executed, part or all of the steps in each embodiment including the foregoing software deployment method may be implemented. The storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM), a random-access memory (random access memory, RAM), or the like.
One or more of the above modules or units may be implemented in software, hardware, or a combination of both. When any of the above modules or units are implemented in software, the software exists in the form of computer program instructions and is stored in a memory, a processor can be used to execute the program instructions and implement the above method flows. The processor may include, but is not limited to, at least one of a central processing unit (central processing unit, CPU), microprocessor, digital Signal Processor (DSP), microcontroller (microcontroller unit, MCU), or artificial intelligence processor, etc. running various types of software computing devices, each of which may include one or more cores for executing software instructions to perform operations or processes. The processor may be built into a SoC (system on a chip) or an application-specific integrated circuit (ASIC), or may be a separate semiconductor chip. The processor may further include necessary hardware accelerators, such as field programmable gate arrays (field programmable GATE ARRAY, FPGAs), PLDs (programmable logic devices), or logic circuits implementing dedicated logic operations, in addition to the cores for executing software instructions for operation or processing.
When the above modules or units are implemented in hardware, the hardware may be any one or any combination of a CPU, microprocessor, DSP, MCU, artificial intelligence processor, ASIC, soC, FPGA, PLD, dedicated digital circuitry, hardware accelerator, or non-integrated discrete device that may run the necessary software or that is independent of the software to perform the above method flows.
Further, a bus interface may be included in FIG. 3, which may include any number of interconnected buses and bridges, with various circuits of the memory, in particular, represented by one or more of the processors and the memory. The bus interface may also link together various other circuits such as peripheral devices, voltage regulators, power management circuits, etc., which are well known in the art and, therefore, will not be described further herein. The bus interface provides an interface. The transceiver provides a means for communicating with various other apparatus over a transmission medium. The processor is responsible for managing the bus architecture and general processing, and the memory may store data used by the processor in performing operations.
When the above modules or units are implemented in software, they may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, produces a flow or function in accordance with embodiments of the present invention, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer instructions may be stored in or transmitted from one computer-readable storage medium to another, for example, by wired (e.g., coaxial cable, optical fiber, digital Subscriber Line (DSL)), or wireless (e.g., infrared, wireless, microwave, etc.). The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., floppy disk, hard disk, tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., solid state disk Solid STATE DISK (SSD)), etc.
It should be understood that, in various embodiments of the present application, the size of the sequence number of each process does not mean that the execution sequence of each process should be determined by its functions and internal logic, and should not constitute any limitation on the implementation process of the embodiments.
All parts of the specification are described in a progressive manner, and all parts of the embodiments which are the same and similar to each other are referred to each other, and each embodiment is mainly described as being different from other embodiments. In particular, for apparatus and system embodiments, the description is relatively simple, as it is substantially similar to method embodiments, with reference to the description of the method embodiments section.
While alternative embodiments of the present application have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. It is therefore intended that the following claims be interpreted as including the preferred embodiments and all such alterations and modifications as fall within the scope of the application.
The foregoing embodiments have been provided for the purpose of illustrating the general principles of the present application in further detail, and are not to be construed as limiting the scope of the application, but are merely intended to cover any modifications, equivalents, improvements, etc. based on the teachings of the application.

Claims (9)

1.一种图像识别方法,其特征在于,所述方法包括:1. An image recognition method, characterized in that the method comprises: 获取至少一个识别目标的图片信息,其中,每个识别目标对应一个图片信息,每个所述图片信息包括:所述每个识别目标对应的颜色信息;Acquire image information of at least one recognition target, wherein each recognition target corresponds to one image information, and each of the image information includes: color information corresponding to each recognition target; 依据所述每个识别目标对应的颜色信息得到所述每个识别目标在对应图片信息中的第一坐标信息;Obtaining first coordinate information of each recognition target in the corresponding image information according to the color information corresponding to each recognition target; 获取所述至少一个识别目标的深度图信息,其中,所述每个识别目标对应一个深度图信息;Acquire depth map information of the at least one recognition target, wherein each recognition target corresponds to one depth map information; 依据所述每个识别目标对应的图片信息以及所述每个识别目标对应的深度图信息得到所述每个识别目标对应的空间变换特征;Obtaining a spatial transformation feature corresponding to each recognition target according to the image information corresponding to each recognition target and the depth map information corresponding to each recognition target; 依据所述每个识别目标对应的第一坐标信息以及所述每个识别目标对应的空间变换特征,得到每个识别目标对应的第二坐标信息,所述第二坐标信息作用于所述深度图信息中;According to the first coordinate information corresponding to each identified target and the spatial transformation feature corresponding to each identified target, second coordinate information corresponding to each identified target is obtained, and the second coordinate information acts on the depth map information; 依据所述识别目标对应的第二坐标信息得到所述识别目标维度信息;Obtaining dimension information of the identified target according to the second coordinate information corresponding to the identified target; 所述每个识别目标对应的第二坐标信息包括:所述每个识别目标在对应灰度图中的外轮廓的四个深度值信息,每个所述每个识别目标的外轮廓的像素坐标信息与每个所述每个识别目标的外轮廓的深度值信息一一对应。The second coordinate information corresponding to each identified target includes: four depth value information of the outer contour of each identified target in the corresponding grayscale image, and the pixel coordinate information of the outer contour of each identified target corresponds one-to-one to the depth value information of the outer contour of each identified target. 2.根据权利要求1所述的图像识别方法,其特征在于,所述图片信息通过图片采集设备获取,所述深度图信息通过深度图采集设备获取。2. The image recognition method according to claim 1 is characterized in that the image information is obtained through an image acquisition device, and the depth map information is obtained through a depth map acquisition device. 3.根据权利要求1所述的图像识别方法,其特征在于,所述每个识别目标在对应图片信息中的第一坐标信息包括:所述每个识别目标在对应图片信息中的外轮廓的四个像素坐标信息。3. The image recognition method according to claim 1 is characterized in that the first coordinate information of each recognized target in the corresponding picture information includes: four pixel coordinate information of the outer contour of each recognized target in the corresponding picture information. 4.根据权利要求1所述的图像识别方法,其特征在于,所述空间变换特征由畸变矫正矩阵得到。4. The image recognition method according to claim 1 is characterized in that the spatial transformation features are obtained by a distortion correction matrix. 5.根据权利要求1所述的图像识别方法,其特征在于,所述依据所述识别目标对应的第二坐标信息得到所述识别目标维度信息,实现方式包括:5. The image recognition method according to claim 1, characterized in that the dimension information of the recognition target is obtained according to the second coordinate information corresponding to the recognition target, and the implementation method includes: 若所述深度值信息大于或等于预设阈值,则判定所述识别目标对应的维度信息为三维目标;If the depth value information is greater than or equal to a preset threshold, it is determined that the dimension information corresponding to the identified target is a three-dimensional target; 若所述深度值信息小于预设阈值,则判定所述识别目标对应的维度信息为二维目标。If the depth value information is less than a preset threshold, the dimension information corresponding to the identified target is determined to be a two-dimensional target. 6.根据权利要求1所述的图像识别方法,其特征在于,所述依据所述每个识别目标对应的颜色信息得到所述每个识别目标在对应图片信息中的第一坐标信息通过目标检测神经网络,R-CNN实现。6. The image recognition method according to claim 1 is characterized in that the first coordinate information of each identified target in the corresponding image information is obtained based on the color information corresponding to each identified target through a target detection neural network, R-CNN. 7.一种图像识别装置,其特征在于,所述装置包括:7. An image recognition device, characterized in that the device comprises: 第一获取模块,用于获取至少一个识别目标的图片信息,其中,每个识别目标对应一个图片信息,每个所述图片信息包括:所述每个识别目标对应的颜色信息;A first acquisition module is used to acquire image information of at least one recognition target, wherein each recognition target corresponds to a piece of image information, and each piece of image information includes: color information corresponding to each recognition target; 第一处理模块,用于依据所述每个识别目标对应的颜色信息得到所述每个识别目标在对应图片信息中的第一坐标信息;A first processing module, used for obtaining first coordinate information of each recognition target in the corresponding image information according to the color information corresponding to each recognition target; 第二获取模块,用于获取所述至少一个识别目标的深度图信息,其中,所述每个识别目标对应一个深度图信息;A second acquisition module is used to acquire the depth map information of the at least one recognition target, wherein each recognition target corresponds to a depth map information; 特征构建模块,用于依据所述每个识别目标对应的图片信息以及所述每个识别目标对应的深度图信息得到所述每个识别目标对应的空间变换特征;A feature construction module, used to obtain the spatial transformation feature corresponding to each recognition target according to the image information corresponding to each recognition target and the depth map information corresponding to each recognition target; 第二处理模块,用于依据所述每个识别目标对应的第一坐标信息以及所述每个识别目标对应的空间变换特征,得到每个识别目标对应的第二坐标信息,所述第二坐标信息作用于所述深度图信息中;A second processing module, configured to obtain second coordinate information corresponding to each identified target according to the first coordinate information corresponding to each identified target and the spatial transformation feature corresponding to each identified target, wherein the second coordinate information acts on the depth map information; 判定模块,用于依据所述识别目标对应的第二坐标信息得到所述识别目标维度信息;所述每个识别目标对应的第二坐标信息包括:所述每个识别目标在对应灰度图中的外轮廓的四个深度值信息,每个所述每个识别目标的外轮廓的像素坐标信息与每个所述每个识别目标的外轮廓的深度值信息一一对应。A determination module is used to obtain the dimension information of the identified target based on the second coordinate information corresponding to the identified target; the second coordinate information corresponding to each identified target includes: four depth value information of the outer contour of each identified target in the corresponding grayscale image, and the pixel coordinate information of the outer contour of each identified target corresponds one-to-one to the depth value information of the outer contour of each identified target. 8.一种电子设备,其特征在于,所述电子设备包括:存储器和处理器,所述存储器和所述处理器之间互相通信连接,所述存储器中存储有计算机指令,所述处理器通过执行所述计算机指令执行权利要求1-6中任一项所述的方法。8. An electronic device, characterized in that the electronic device comprises: a memory and a processor, the memory and the processor are communicatively connected to each other, the memory stores computer instructions, and the processor executes the method according to any one of claims 1 to 6 by executing the computer instructions. 9.一种计算机可读存储介质,其特征在于,所述计算机可读存储介质存储有计算机指令,所述计算机指令用于使所述计算机执行权利要求1-6中任一项所述的方法。9. A computer-readable storage medium, characterized in that the computer-readable storage medium stores computer instructions, and the computer instructions are used to enable the computer to execute the method according to any one of claims 1 to 6.
CN202210491488.7A 2022-05-07 2022-05-07 Image recognition method, device and equipment Active CN114821479B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210491488.7A CN114821479B (en) 2022-05-07 2022-05-07 Image recognition method, device and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210491488.7A CN114821479B (en) 2022-05-07 2022-05-07 Image recognition method, device and equipment

Publications (2)

Publication Number Publication Date
CN114821479A CN114821479A (en) 2022-07-29
CN114821479B true CN114821479B (en) 2025-04-01

Family

ID=82510578

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210491488.7A Active CN114821479B (en) 2022-05-07 2022-05-07 Image recognition method, device and equipment

Country Status (1)

Country Link
CN (1) CN114821479B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107748869A (en) * 2017-10-26 2018-03-02 深圳奥比中光科技有限公司 3D face identity authentications and device
CN111310528A (en) * 2018-12-12 2020-06-19 马上消费金融股份有限公司 Image detection method, identity verification method, payment method and device

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2050067A1 (en) * 2006-08-03 2009-04-22 Alterface S.A. Method and device for identifying and extracting images of multiple users, and for recognizing user gestures
US10956714B2 (en) * 2018-05-18 2021-03-23 Beijing Sensetime Technology Development Co., Ltd Method and apparatus for detecting living body, electronic device, and storage medium
CN109242903B (en) * 2018-09-07 2020-08-07 百度在线网络技术(北京)有限公司 Three-dimensional data generation method, device, equipment and storage medium
CN111815754B (en) * 2019-04-12 2023-05-30 Oppo广东移动通信有限公司 A method for determining three-dimensional information, a device for determining three-dimensional information, and a terminal device
CN110706267B (en) * 2019-09-04 2022-11-15 五邑大学 Mining process-based ore three-dimensional coordinate acquisition method and device
CN113469036A (en) * 2021-06-30 2021-10-01 北京市商汤科技开发有限公司 Living body detection method and apparatus, electronic device, and storage medium

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107748869A (en) * 2017-10-26 2018-03-02 深圳奥比中光科技有限公司 3D face identity authentications and device
CN111310528A (en) * 2018-12-12 2020-06-19 马上消费金融股份有限公司 Image detection method, identity verification method, payment method and device

Also Published As

Publication number Publication date
CN114821479A (en) 2022-07-29

Similar Documents

Publication Publication Date Title
US11182917B2 (en) Stereo camera depth determination using hardware accelerator
US10970864B2 (en) Method and apparatus for recovering point cloud data
CN111353512B (en) Obstacle classification method, device, storage medium and computer equipment
US10410354B1 (en) Method and apparatus for multi-model primitive fitting based on deep geometric boundary and instance aware segmentation
WO2019011249A1 (en) Method, apparatus, and device for determining pose of object in image, and storage medium
WO2021114776A1 (en) Object detection method, object detection device, terminal device, and medium
WO2022206517A1 (en) Target detection method and apparatus
KR102386444B1 (en) Image depth determining method and living body identification method, circuit, device, and medium
CN109657538B (en) Scene segmentation method and system based on context information guidance
CN114359377B (en) A real-time 6D pose estimation method and computer-readable storage medium
Wang et al. Deep learning‐based vehicle detection with synthetic image data
CN113627478A (en) Target detection method, target detection device and robot
CN114998610A (en) Target detection method, device, equipment and storage medium
CN115861595A (en) A Multi-Scale Domain Adaptive Heterogeneous Image Matching Method Based on Deep Learning
CN116663650A (en) Training method of deep learning model, target object detection method and device
CN115049717B (en) A depth estimation method and device
CN116503460A (en) Depth map acquisition method, device, electronic equipment and storage medium
CN111626241A (en) Face detection method and device
CN116030206A (en) Map generation method, training method, device, electronic device and storage medium
WO2025185158A1 (en) Target detection method based on three-dimensional data, device, and medium
KR102299902B1 (en) Apparatus for providing augmented reality and method therefor
CN114821479B (en) Image recognition method, device and equipment
CN118135419A (en) SAR ship detection method based on lightweight neural network and storage medium
CN114913213B (en) Method and device for learning aerial view characteristics
CN116642490A (en) Visual positioning navigation method based on hybrid map, robot and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant