Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

2671 results about "Rgb image" patented technology

Uterine cervical cancer computer-aided-diagnosis (CAD)

Uterine cervical cancer Computer-Aided-Diagnosis (CAD) according to this invention consists of a core processing system that automatically analyses data acquired from the uterine cervix and provides tissue and patient diagnosis, as well as adequacy of the examination. The data can include, but is not limited to, color still images or video, reflectance and fluorescence multi-spectral or hyper-spectral imagery, coherent optical tomography imagery, and impedance measurements, taken with and without the use of contrast agents like 3-5% acetic acid, Lugol's iodine, or 5-aminolevulinic acid. The core processing system is based on an open, modular, and feature-based architecture, designed for multi-data, multi-sensor, and multi-feature fusion. The core processing system can be embedded in different CAD system realizations. For example: A CAD system for cervical cancer screening could in a very simple version consist of a hand-held device that only acquires one digital RGB image of the uterine cervix after application of 3-5% acetic acid and provides automatically a patient diagnosis. A CAD system used as a colposcopy adjunct could provide all functions that are related to colposcopy and that can be provided by a computer, from automation of the clinical workflow to automated patient diagnosis and treatment recommendation.
Owner:STI MEDICAL SYST

Kinect-based robot self-positioning method

The invention discloses a Kinect-based robot self-positioning method. The method includes the following steps that: the RGB image and depth image of an environment are acquired through the Kinect, and the relative motion of a robot is estimated through the information of visual fusion and a physical speedometer, and pose tracking can be realized according to the pose of the robot at a last time point; depth information is converted into three-dimensional point cloud, and a ground surface is extracted from the point cloud, and the height and pitch angle of the Kinect relative to the ground surface are automatically calibrated according to the ground surface, so that the three-dimensional point cloud can be projected to the ground surface, and therefore, two-dimensional point cloud similar to laser data can be obtained, and the two-dimensional point cloud is matched with pre-constructed environment raster map, and thus, accumulated errors in a robot tracking process can be corrected, and the pose of the robot can be estimated accurately. According to the Kinect-based robot self-positioning method of the invention, the Kinect is adopted to replace laser to perform positioning, and therefore, cost is low; image and depth information is fused, so that the method can have high precision; and the method is compatible with a laser map, and the mounting height and pose of the Kinect are not required to be calibrated in advance, and therefore, the method is convenient to use, and requirements for autonomous positioning and navigation of the robot can be satisfied.
Owner:HANGZHOU JIAZHI TECH CO LTD

Full convolution neural network (FCN)-based monocular image depth estimation method

The invention discloses a full convolution neural network (FCN)-based monocular image depth estimation method. The method comprises the steps of acquiring training image data; inputting the training image data into a full convolution neural network (FCN), and sequentially outputting through pooling layers to obtain a characteristic image; subjecting each characteristic image outputted by a last pooling layer sequentially to amplification treatment to obtain a new characteristic image the same with the dimension of a characteristic image outputted by a previous pooling layer, and fusing the twocharacteristic images; sequentially fusing the outputted characteristic image of each pooling layer from back to front so as to obtain a final prediction depth image; training the parameters of the full convolution neural network (FCN) by utilizing a random gradient descent method (SGD) during training; acquiring an RGB image required for depth prediction, and inputting the RGB image into the well trained full convolution neural network (FCN) so as to obtain a corresponding prediction depth image. According to the method, the problem that the resolution of an output image is low in the convolution process can be solved. By adopting the form of the full convolution neural network, a full-connection layer is removed. The number of parameters in the network is effectively reduced.
Owner:NANJING UNIV OF POSTS & TELECOMM

Bidirectional long short-term memory unit-based behavior identification method for video

The invention discloses a bidirectional long short-term memory unit-based behavior identification method for a video. The method comprises the steps of (1) inputting a video sequence and extracting an RGB (Red, Green and Blue) frame sequence and an optical flow image from the video sequence; (2) respectively training a deep convolutional network of an RGB image and a deep convolutional network of the optical flow image; (3) extracting multilayer characteristics of the network, wherein characteristics of a third convolutional layer, a fifth convolutional layer and a seventh fully connected layer are at least extracted, and the characteristics of the convolutional layers are pooled; (4) training a recurrent neural network constructed by use of a bidirectional long short-term memory unit to obtain a probability matrix of each frame of the video; and (5) averaging the probability matrixes, finally fusing the probability matrixes of an optical flow frame and an RGB frame, taking a category with a maximum probability as a final classification result, and thus realizing behavior identification. According to the method, the conventional artificial characteristics are replaced with multi-layer depth learning characteristics, the depth characteristics of different layers represent different pieces of information, and the combination of multi-layer characteristics can improve the accuracy rate of classification; and the time information is captured by use of the bidirectional long short-term memory, many pieces of time domain structural information are obtained and a behavior identification effect is improved.
Owner:SUZHOU UNIV

Method and device for correcting color based on RGBIR (red, green and blue, infra red) image sensor

The invention discloses a method and a device for correcting color based on an RGBIR (red, green and blue, infra red) image sensor. The method comprises the following steps of entering an RGBIR image; interpolating the image, and calculating the channel values of an RGB (red, green and blue) channel and an IR (infra red) channel on each pixel position; calculating the channel statistical values of an RGB channel and an IR channel in a particular area; according to the statistical values, looking up the table to obtain the correction coefficients of the current scene; configuring the correction coefficients into a correction matrix, and separating the RGB channel and the IR channel; carrying out the color correction on the separated RGB channel, to obtain the corrected RGB image. A pre-correcting unit eliminates the effect of invisible light on the RGB channel of the original image, and the output utilizes a back-correcting unit to eliminate the effect of wavelength change of visible light on the RGB channel. The method not only can correct the color of the RBGIR image sensor, but also can divide the color correction into two relatively independent parts, and can be well compatible with the color correction method in the existing image processing module.
Owner:SHANGHAI FULLHAN MICROELECTRONICS

Dynamic gesture recognition method and system based on deep neural network

The invention discloses a dynamic gesture recognition method and system based on a deep neural network. The dynamic gesture recognition method comprises the steps of collecting dynamic gesture video clips with different gesture meanings to generate a training sample data set, wherein the sample data includes RGB images and depth information; designing a dynamic gesture recognition network model based on the deep neural network, and training the model by using the training samples; and performing dynamic gesture testing and recognition by using the trained dynamic gesture recognition model. Thedynamic gesture recognition network model is composed of a feature extraction network, a front and back frame association network and a classification recognition network, wherein the front and backframe association network is used for performing front and back time frame association mapping on feature vectors obtained through the feature extraction network of the samples of each gesture meaningand merging the feature vectors into a fusion feature vector of the gesture meaning. According to the invention, a bidirectional LSTM model is introduced into the network model to understand the correlation between continuous gesture postures, thereby greatly improving the recognition rate of dynamic gestures.
Owner:广州智能装备研究院有限公司

Three-dimensional target detection method and device based on multi-sensor information fusion

The invention discloses a three-dimensional target detection method, apparatus and device based on multi-sensor information fusion, and a computer readable storage medium. The three-dimensional targetdetection method comprises the steps: fusing 3D point cloud and an RGB image collected by a laser radar and a camera sensor, and generating an RGB-I image; generating a multi-channel aerial view according to the 3D point cloud so as to determine a region of interest; respectively extracting and fusing region-of-interest features of the RGB-I image and the aerial view based on a convolutional neural network; utilizing a multi-layer perceptron to fuse the confidence coefficient, the approximate position and the size of the image prediction target based on the features of the region of interest,and determining a candidate box; adaptively endowing different pixel weights to different sensor candidate box feature maps based on an attention mechanism, and carrying out skip fusion; and processing the candidate frame feature fusion image by using a multi-layer perceptron, and outputting a three-dimensional detection result. According to the three-dimensional target detection method, apparatus and device, and the computer readable storage medium provided by the invention, the target recognition rate is improved, and the target can be accurately positioned.
Owner:CHANGCHUN INST OF OPTICS FINE MECHANICS & PHYSICS CHINESE ACAD OF SCI
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products