Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

436 results about "Visual saliency" patented technology

Visual salience (or visual saliency) is the distinct subjective perceptual quality which makes some items in the world stand out from their neighbors and immediately grab our attention.

HEVC (High Efficiency Video Coding) code rate control method based on region-of-interest

The invention provides an HEVC (High Efficiency Video Coding) code rate control method based on region-of-interest. The HEVC code rate control method based on region-of-interest comprises the steps: generating a spatial domain significance graph of the current frame according to a GBVS (Graph Based Visual Saliency) model; generating a time domain significance graph of the current frame through motion vector information; integrating the time domain significance graph and the spatial domain significance graph by using the consistency normalization method to obtain the eventual significance graph; performing regional division for the current frame image by using the significance graph so as to divide the region into a region-of-interest and a non-region-of-interest; performing bit distribution for the region-of-interest and the non-region-of-interest respectively; performing bit distribution for each LCU in the current frame according to the significance; calculating lambda and a QP (Quantization Parameter) value according to the distributed code rate, and clipping and correcting the lambda and the QP value; and performing coding by using the eventually obtained lambda and the QP value. The HEVC code rate control method based on region-of-interest can improve the subjective quality for video coding and accurately control the output bit at the same time.
Owner:SHANGHAI UNIV

Target tracking method based on frequency domain saliency

The invention relates to a target tracking method based on frequency domain saliency, which comprises the steps of S1-S4, establishing direction feature maps, color feature maps, gray feature maps and motion feature maps; S5-S6, establishing static and moving polynomials and performing Fourier transform to the static and moving polynomials; S7, performing Gaussian low-pass filtration and inverse Fourier transform to magnitude spectra to obtain static saliency maps and moving saliency maps; S8, multiplying the moving saliency maps by the static saliency maps with the corresponding scales to obtain saliency multi-scale detection result saliency map; S9, calculating the one-dimensional entropy function of the histogram of the saliency map and extracting a time domain saliency map corresponding to a minimum information entropy as an optimal saliency map at the moment t; S10, using products of average weight of t-1 and t-2 frame saliency maps and the optimal saliency map at the moment t as visual saliency maps; S11, calculating difference of central positions of the visual saliency maps of adjacent frames, judging whether the tracking is failed or not and recording a failure saliency map; and S12, comparing the visual saliency map of the current frame with the failure saliency map and judging whether a target returns back to a visual field or not.
Owner:INST OF OPTICS & ELECTRONICS - CHINESE ACAD OF SCI

Rice disease recognition method based on principal component analysis and neural network and rice disease recognition system thereof

The invention relates to a rice disease recognition method based on principal component analysis and a neural network. The method comprises the steps that rice disease image data are acquired and image preprocessing is performed; visual saliency detection is performed, and rice disease images of ideal disease spot outlines are searched from salient map sequences; features are extracted from the rice disease images from the aspects of color, shape and texture, and difference analysis and principal component analysis are performed so that different feature combinations are found; and construction of a machine learning model is performed on different feature combinations and a prediction result is fed back to a client side. The invention also discloses a rice disease recognition system based on principal component analysis and the neural network. Image information is acquired and the images are transmitted to a server side through the network. Preprocessing and disease spot detection are performed on the acquired tissue culturing images through the server side, and management personnel are prompted through a mobile phone short message and a signal lamp and a PC side according to the detection result.
Owner:WUXI CAS INTELLIGENT AGRI DEV

Visual saliency model based automatic detecting and tracking method

The invention discloses a visual saliency model based automatic detecting and tracking method. The visual saliency model based automatic detecting and tracking method comprises the steps of: calculating a color, brightness and direction saliency graph of an input video image by using a visual saliency model, and defining a simple scene and a complex scene based on the weighted saliency graph; establishing a rectangular frame to serve as a tracking target to be tracked by using a saliency region when the simple scene is detected; correcting a manually selected tracking frame based on different weights when the complex scene is detected; tracking the tracking frame by utilizing a tracking studying and detecting algorithm, and detecting that the tracking is failed; detecting the image of each frame after the failure by using the visual saliency model, performing histogram matching on each region in the saliency graph and the online model before the tracking failure, and tracking a region with the highest similarity; and sending multiple regions with similar similarity into a target detector at the same time for detection, repeating tracking detection for the image target of the next frame, using a histogram comparison step until a target is detected again and tracking.
Owner:INST OF OPTICS & ELECTRONICS - CHINESE ACAD OF SCI

Visual saliency detection method combined with image classification

The invention provides a visual saliency detection method combined with image classification. The method comprises the steps of utilizing a visual saliency detecting model which comprises an image coding network, an image decoding network and an image identification model, using a multidirectional image as an input of the image coding network, and extracting an image characteristic on the condition of multiple resolution as a coding characteristic vector F; fixing a weight except for the last two layers in the image coding network, and training network parameters for obtaining a visual saliency picture of an original image; using the F as the input of the image decoding network, and performing normalization processing on the saliency picture which corresponds with the original image; for the input F of the image decoding network, finally obtaining a generated visual saliency picture through an upsampling layer and a nonlinear sigmoid layer; by means of the image identification network, using the visual saliency picture of the original image and the generated visual saliency picture as the input, performing characteristic extraction by means of a convolutional layer with a small convolution kernel and performing pooling processing, and finally outputting probability distribution of the generated picture and probability distribution of classification labels by means of three total connecting layers. The method provided by the invention realizes quick and effective image analysis and determining and furthermore realizes good effects such as saving manpower and physical resource costs and remarkably improving accuracy in practices such as image marking, supervising and behavior predicating.
Owner:以萨技术股份有限公司

Video foreground object extracting method based on visual saliency and superpixel division

The invention discloses a video foreground object extracting method based on visual saliency and superpixel division. The video foreground object extracting method includes steps: a, dividing multiple layers of superpixels of video: dividing the superpixels of the video used as a three-dimensional video body, and grouping elements of the video body into body areas; b, detecting visual saliency areas of key frames of the video and extracting foreground objects of the key frames: analyzing the visual saliency areas in images of the key frames of the video by a visual saliency detecting method, then using the visual saliency areas as initial values and obtaining the foreground objects of the key frames by an image foreground extracting method; and c, matching the foreground objects of the key frames with a dividing result of the superpixels of the video and transmitting foreground object extracting results of the key frames among the frames: diffusing areas, covered by the foreground objects of the key frames, of the video body, and further continuously transmitting the foreground object extracting results among the frames. The video foreground object extracting method is high in efficiency, accurate in result and little in manual intervention and is robust.
Owner:TSINGHUA UNIV +1

Method for image processing and video compression

A method for video compression through image processing and object detection, to be carried out by an electronic processing unit, based either on images or on a digital video stream of images, the images being defined by a single frame or by sequences of frames of said video stream, with the aim of enhancing and then isolating the frequency domain signals representing a content to be identified, and decreasing or ignoring the frequency domain noise with respect to the content within the images or the video stream, comprises the steps of: obtaining a digital image or a sequence of digital images from either a corresponding single frame or a corresponding sequence of frames of said video stream, all the digital images being defined in a spatial domain; selecting one or more pairs of sparse zones, each covering at least a portion of said single frame or at least two frames of said sequence of frames, each pair of sparse zones generating a selected feature, each zone being defined by two sequences of spatial data; transforming the selected features into frequency domain data by combining, for each zone, said two sequences of spatial data through a 2D variation of an L-transformation, varying the transfer function, shape and direction of the frequency domain data for each zone, thus generating a normalized complex vector for each of said selected features; combining all said normalized complex vectors to define a model of the content to be identified; and inputting that model from said selected features in a classifier, therefore obtaining the data for object detection or visual saliency to use for video compression.
Owner:INTEL CORP

All-reference three-dimensional image quality objective evaluation method based on visual salient feature extraction

The invention discloses an all-reference three-dimensional image quality objective evaluation method based on visual salient feature extraction. According to the method, a left view and a right view of a three-dimensional image pair are processed to obtain a corresponding disparity map; image fusion is performed on the left view and the right view of the three-dimensional image pair to obtain an intermediate reference image and an intermediate distortion image; a spectral residual visual saliency model is utilized to obtain a reference saliency map and a distortion saliency map, and a visual saliency map is obtained through integration; visual information features are extracted from the intermediate reference image and the intermediate distortion image, and depth information features are extracted from the disparity map of the three-dimensional image pair; similarity measurement is performed to obtain measurement indexes of all the visual information features of vision saliency enhancement; and support vector machine training prediction is performed, an objective quality score is obtained, mapping of three-dimensional image quality is realized, and measurement and evaluation of three-dimensional image quality are completed. Through the method, image quality objective evaluation and subjective evaluation have good consistency, and the performance is superior to that of existingthree-dimensional image quality evaluation methods.
Owner:ZHEJIANG UNIV

Sparse coding and visual saliency-based method for detecting airport through infrared remote sensing image

ActiveCN102831402AAdd screening strategyEfficient captureCharacter and pattern recognitionSupport vector machineLeast significant difference
The invention relates to a sparse coding and visual saliency-based method for detecting an airport through an infrared remote sensing image. The method comprises the following steps of: firstly, down-sampling an original remote sensing image, linearly detecting the down-sampled remote sensing image via using an LSD (Least Significant Difference) algorithm, calculating the saliency of the image via using an FT algorithm; then detecting the airport by utilizing a sliding window target detector, judging whether a linear section exists in the sliding window, if not, sliding the window continuously, if so, carrying out the sparse coding on the window by utilizing a dictionary constructed by an airport target image of the remote sensing image, and screening sparse codes in a way of combining the sparse codes with salient values of the window, so as to obtain sparse expression characteristics of the window; finally, discriminating the sparse code characteristics of the sliding window via an SVM (Support Vector Machine) binary classifier, judging whether the airport exists in the window, and realizing the detection of an airport target ultimately. Compared with other invented technologies, the method has the advantages of high airport detection accuracy and low false alarm rate.
Owner:NORTHWESTERN POLYTECHNICAL UNIV

Method for detecting degree of visual saliency of image in different regions

The invention discloses a method for detecting the degree of visual saliency of an image in different regions, which comprises the following steps: segmenting the input image into non-overlapping image blocks, and carrying out vectorization on each image block; carrying out dimensionality reduction on all vectors obtained in the step 1 through the PCA principle component analytical method for reducing noise and redundant information in the image; calculating the non-similarity degree between each image block and all the other image blocks by utilizing the vectors after the dimensionality reduction, calculating the degree of visual saliency of each image block by further combining with the distance between the image blocks and obtaining a saliency map; imposing central bias on the saliency map, and obtaining the saliency map after imposing the central bias; and smoothing the saliency map after imposing the central bias through a two-dimensional Gaussian smoothing operator, and obtaining a final result image which reflects the degree of saliency of the image in all the regions. Compared with the prior art, the method does not need to extract visual features, such as color, orientation, texture and the like and can avoid the step of selecting the features. The method has the advantages of simpleness and high efficiency.
Owner:BEIJING UNIV OF TECH

Visual saliency detection method with fusion of region color and HoG (histogram of oriented gradient) features

The invention relates to a visual saliency detection method with fusion of region color and HoG (histogram of oriented gradient) features. At present, the existing method is generally based on a pure calculation model of the region color feature and is insensitive to salient difference of texture. The method disclosed by the invention comprises the following steps of: firstly calculating a color saliency value of each pixel by analyzing color contrast and distribution feature of a superpixel region on a CIELAB (CIE 1976 L*, a*, b*) space color component diagram of an original image; then extracting an HoG-based local rectangular region texture feature on an RGB (red, green and blue) space color component diagram of the original image, and calculating a texture saliency value of each pixel by analyzing texture contrast and distribution feature of a local rectangular region; and finally fusing the color saliency value and the texture saliency value of each pixel into a final saliency value of the pixel by adopting a secondary non-linear fusion method. According to the method disclosed by the invention, a full-resolution saliency image which is in line with sense of sight of human eyes can be obtained, and the distinguishing capability against a saliency object is further stronger.
Owner:海宁鼎丞智能设备有限公司
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products