Patents
Literature
Patsnap Copilot is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Patsnap Copilot

55 results about "Perceptual coding" patented technology

In digital audio perceptual coding is a coding method used to reduce the amount of data needed to produce high-quality sound. Perceptual coding takes advantage of the human ear, screening out a certain amount of sound that is perceived as noise.

Perceptual coding of image signals using separated irrelevancy reduction and redundancy reduction

A perceptual coder is disclosed for encoding image signals, such as speech or music, with different spectral and temporal resolutions for redundancy reduction and irrelevancy reduction. The image signal is initially spectrally shaped using a prefilter. The prefilter output samples are thereafter quantized and coded to minimize the mean square error (MSE) across the spectrum. The disclosed perceptual image coder can use fixed quantizer step-sizes, since spectral shaping is performed by the pre-filter prior to quantization and coding. The disclosed pre-filter and post-filter support the appropriate frequency dependent temporal and spectral resolution for irrelevancy reduction. A filter structure based on a frequency-warping technique is used that allows filter design based on a non-linear frequency scale. The characteristics of the pre-filter may be adapted to the masked thresholds, using techniques known from speech coding, where linear-predictive coefficient (LPC) filter parameters are used to model the spectral envelope of the speech signal. Likewise, the filter coefficients may be efficiently transmitted to the decoder for use by the post-filter using well-established techniques from speech coding, such as an LSP (line spectral pairs) representation, temporal interpolation, or vector quantization.
Owner:AGERE SYST INC

Audio encoding method and system for generating a unified bitstream decodable by decoders implementing different decoding protocols

In a class of embodiments, an audio encoding system (typically, a perceptual encoding system that is configured to generate a single (“unified”) bitstream that is compatible with (i.e., decodable by) a first decoder configured to decode audio data encoded in accordance with a first encoding protocol (e.g., the multichannel Dolby Digital Plus, or DD+, protocol) and a second decoder configured to decode audio data encoded in accordance with a second encoding protocol (e.g., the stereo AAC, HE AAC v1, or HE AAC v2 protocol). The unified bitstream can include both encoded data (e.g., bursts of data) decodable by the first decoder (and ignored by the second decoder) and encoded data (e.g., other bursts of data) decodable by the second decoder (and ignored by the first decoder). In effect, the second encoding format is hidden within the unified bitstream when the bitstream is decoded by the first decoder, and the first encoding format is hidden within the unified bitstream when the bitstream is decoded by the second decoder. The format of the unified bitstream generated in accordance with the invention may eliminate the need for transcoding elements throughout an entire media chain and/or ecosystem. Other aspects of the invention are an encoding method performed by any embodiment of the inventive encoder, a decoding method performed by any embodiment of the inventive decoder, and a computer readable medium (e.g., disc) which stores code for implementing any embodiment of the inventive method.
Owner:DOLBY LAB LICENSING CORP +1

Perceptual coding of audio signals using separated irrelevancy reduction and redundancy reduction

A perceptual audio coder is disclosed for encoding audio signals, such as speech or music, with different spectral and temporal resolutions for redundancy reduction and irrelevancy reduction. The disclosed perceptual audio coder separates the psychoacoustic model (irrelevancy reduction) from the redundancy reduction, to the extent possible. The audio signal is initially spectrally shaped using a prefilter controlled by a psychoacoustic model. The prefilter output samples are thereafter quantized and coded to minimize the mean square error (MSE) across the spectrum. The disclosed perceptual audio coder can use fixed quantizer step-sizes, since spectral shaping is performed by the pre-filter prior to quantization and coding. The disclosed pre-filter and post-filter support the appropriate frequency dependent temporal and spectral resolution for irrelevancy reduction. A filter structure based on a frequency-warping technique is used that allows filter design based on a non-linear frequency scale. The characteristics of the pre-filter may be adapted to the masked thresholds (as generated by the psychoacoustic model), using techniques known from speech coding, where linear-predictive coefficient (LPC) filter parameters are used to model the spectral envelope of the speech signal. Likewise, the filter coefficients may be efficiently transmitted to the decoder for use by the post-filter using well-established techniques from speech coding, such as an LSP (line spectral pairs) representation, temporal interpolation, or vector quantization.
Owner:LUCENT TECH INC

Visual perceptual coding method based on multi-domain JND (Just Noticeable Difference) model

The invention discloses a visual perceptual coding method based on a multi-domain JND (Just Noticeable Difference) model and relates to video information processing. The method comprises the following steps: respectively calculating a space-domain basic JND threshold value, a luminance masking modulation factor, a contrast masking modulation factor and a time domain masking modulation factor of each transformation coefficient in a DCT (Discrete Cosine Transform) block by utilizing a time-space-domain multi-domain JND model so as to obtain the time-space-domain multi-domain JND threshold value of each transformation coefficient; introducing a block perception-based distortion probability evaluative criteria in the transform coding process, and searching a correction factor of each coefficient relative to the JND threshold value through an adaptive searching algorithm so as to obtain a transformation coefficient suppression value; and finally, subtracting the original transformation coefficient from the most appropriate suppression value obtained through corresponding calculation, and taking the coefficient as a novel coefficient to be put at an entropy coding stage. According to the coded suppression strategy of the multi-domain JND model and the block perception-based distortion probability, the coding rate can be effectively reduced on the premise of guaranteeing certain subjective quality, and the compression ratio of the current coding standard is further improved.
Owner:XIAMEN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products