Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

No-reference image quality evaluation method based on deep forest classification

An image quality and quality evaluation technology, applied in the field of no-reference image quality evaluation based on deep forest classification, can solve the problem of ignoring the overall quality characteristics of the image

Active Publication Date: 2019-11-01
LANZHOU UNIVERSITY OF TECHNOLOGY
View PDF6 Cites 4 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

This method achieves image quality evaluation through regional mutual information, but it takes more into account the local features of the image block, and ignores the overall quality features of the image.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • No-reference image quality evaluation method based on deep forest classification
  • No-reference image quality evaluation method based on deep forest classification
  • No-reference image quality evaluation method based on deep forest classification

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0067] The present invention will be further described below in conjunction with the drawings and specific embodiments, but the embodiments described by the drawings are exemplary and are only used to explain the present invention and cannot limit the scope of the present invention.

[0068] A no-reference image quality evaluation method based on deep forest classification in the present invention is as follows: figure 1 As shown, the main steps are as follows:

[0069] Step 1. Image Classification

[0070] Since people tend to give qualitative descriptions rather than quantitative values ​​for subjective evaluation of image quality, the present invention first classifies the images in the quality evaluation database. The images are sorted according to the subjective scores of the images in the quality evaluation database. If the subjective scores are MOS values, they are sorted from large to small; if the subjective scores are DMOS values, the images are sorted from small to...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a no-reference image quality evaluation method based on deep forest classification. The method comprises the following steps: step 1, image classification; step 2, extracting color quality characteristics of the image; step 3, extracting texture quality characteristics of the image; step 4, simulating the difference of different people on image quality cognition by utilizing the difference of decision tree extraction features in the deep forest classification model, and constructing the deep forest classification model to classify the image quality, including a multi-granularity scanning forest and a cascade forest; step 5, training the deep forest classification model based on the image quality features and the category labels thereof to obtain the probability thatthe test image belongs to different categories, i.e., statistical information of subjective evaluation results of different people on the image quality; step 6, setting a quality anchor, and fully considering the difference in the subjective evaluation process in combination with the probability that the image belongs to different categories to obtain a final image quality score. According to thenon-reference image quality evaluation method, the difference of different people for image quality cognition is simulated by using the deep forest, so that an image quality evaluation result is given. The method has important theoretical significance and practical value.

Description

technical field [0001] The invention relates to the fields of image processing technology, computer vision and artificial intelligence, in particular to a no-reference image quality evaluation method based on deep forest classification. Background technique [0002] With the popularization of image acquisition equipment, people began to use a large number of pictures to save and collect information in their daily life. However, the process of image acquisition, transmission, recovery and storage will cause different types and degrees of distortion, which seriously affects people's extraction and understanding of image information, and even easily generates wrong information, misleading people's understanding of images. Therefore, how to judge the quality of the image has become an urgent problem to be solved. [0003] According to different reference information, image quality evaluation methods can be divided into three categories: full reference, semi-reference and no ref...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06T7/00G06K9/62
CPCG06T7/0002G06T2207/10024G06T2207/20081G06T2207/30168G06F18/24323G06F18/2415
Inventor 李策刘昊张栋朱子重李兰高伟哲许大有靳山岗贾盛泽
Owner LANZHOU UNIVERSITY OF TECHNOLOGY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products