Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Real-time subtitle translation and system implementation method for live broadcast scene

A technology for system implementation and subtitles, applied in the field of deep learning, can solve problems such as limited expressive ability of convolution layer, inability to see long enough voice context information, high time delay, etc., achieve excellent user experience, solve real-time problems, and improve efficiency effect

Active Publication Date: 2021-03-16
TRANSN IOL TECH CO LTD
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

Specifically, although the two-way LSTM technology can model the long-term correlation of speech, it is difficult to apply to the real-time translation of the live broadcast platform because it requires a complete sequence for speech recognition and has a high delay.
CNN technology has been applied to speech recognition around 2012, but because it uses fixed-length frame splicing as input, it cannot see enough speech context information, and CNN is used as a feature extractor, the convolution used There are few layers and limited expression ability, so it is difficult to apply to real-time translation of live broadcast platforms

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Real-time subtitle translation and system implementation method for live broadcast scene
  • Real-time subtitle translation and system implementation method for live broadcast scene
  • Real-time subtitle translation and system implementation method for live broadcast scene

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0033] Such as figure 1 As shown, the present invention discloses a real-time subtitle translation and system implementation method for live broadcast scenes, which is characterized in that it includes the following steps:

[0034] S1. Use the training data set to train the deep convolutional neural network.

[0035] Specifically, S1 specifically includes the following steps:

[0036] S11. Using the training data set to train the deep convolutional neural network to obtain the trained deep convolutional neural network;

[0037] S12, using the gradient descent method to optimize all parameters to reduce the cost function;

[0038] S13. Use the gradient descent method for training, and update all weights of all layers of the network.

[0039] The training data set includes spectrograms of various speech signals and text data corresponding to the speech signals.

[0040] S2. Perform Fourier transform on each frame of input speech, take time and frequency as two dimensions of ...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses an implementation method of a real-time subtitle translation and a system for a live scene. The method comprises the following steps of: S1, training a deep convolutional neural network by using a training data set; S2, performing Fourier transform on the input voice of each frame, and taking time and frequency as the two dimensions of the image, to obtain a spectral map ofthe input voice signal; S3, inputting the obtained spectral map into the trained deep convolutional neural network, to obtain text data corresponding to the voice signal; and S4, displaying the textdata on a screen in real-time. According to the implementation method of the real-time subtitle translation and the system for the live scene, the speech of each person can be translated in real-timeand displayed on the screen in time, so that the writing efficiency is effectively improved, and a more excellent user experience is brought to the audience.

Description

technical field [0001] The present invention relates to a real-time translation method based on computer technology, in particular to a real-time subtitle translation for live broadcast scenes and a system implementation method, which belongs to the field of deep learning technology. Background technique [0002] In recent years, the webcasting industry and its related technologies have developed extremely rapidly. Various livestreaming platforms have emerged in an endless stream, and the content of livestreaming has also become increasingly rich. Not only are there individual livestreaming sessions on various topics, but also conferences and online courses in the form of livestreaming. . It is precisely based on this industry development trend that nowadays, in various live broadcast platforms, the demand for real-time translation and display of the live broadcaster's words is becoming stronger and stronger. [0003] At present, there are also some technical solutions for ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): H04N5/278G10L17/18G10L17/04G10L15/26
CPCG10L15/26G10L17/04G10L17/18H04N5/278
Inventor 张晖丁一全
Owner TRANSN IOL TECH CO LTD
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products