Visualized pronunciation teaching method and apparatus

A technology of pronunciation process and pronunciation organs, which is applied in the field of visual pronunciation teaching methods and devices, which can solve the problems of not being able to dynamically display the position changes of organs in the whole process of pronunciation, and achieve the effect of improving scientificity

Inactive Publication Date: 2008-10-22
TIANJIN XUNFEI INFORMATION TECH
0 Cites 9 Cited by

AI-Extracted Technical Summary

Problems solved by technology

[0004] The present invention provides a visual pronunciation teaching method to solve the problem in the prior art that the whole process of pro...
the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Method used

The embodiment of the present invention proposes, the standard pronunciation audio file corresponding to each pronunciation basic unit information and the correct pronunciation animation material file are played synchronously, so that the pronunciation learner can not only hear the correct pronunci...
the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Abstract

The invention discloses a teaching method for visual pronunciation. The method comprises the following steps of: receiving pronunciation basic unit information which is input by a user; searching audio pronunciation information corresponding to the received pronunciation basic unit information according to mapping relation of each pronunciation basic unit information and the corresponding audio pronunciation information; searching a first video pronunciation information corresponding to the received pronunciation basic unit information according to the mapping relation of each pronunciation basic unit information and the corresponding first video pronunciation information; and synchronously playing the searched audio pronunciation information and the first video pronunciation information. The invention discloses a visual pronunciation teaching device at the same time. The proposal of the invention is adopted to make learners hear correct pronunciation of each pronunciation basic unit and intuitively notice dynamic change of each pronunciation organ and strength change of air current of each pronunciation basic unit in the process of correct pronunciation, thereby effectively improving scientificity, intuition and interest of the pronunciation teaching.

Application Domain

Technology Topic

Air currentSpeech recognition +2

Image

  • Visualized pronunciation teaching method and apparatus
  • Visualized pronunciation teaching method and apparatus
  • Visualized pronunciation teaching method and apparatus

Examples

  • Experimental program(1)

Example Embodiment

[0013] The embodiment of the present invention proposes that the standard pronunciation audio file and the correct pronunciation animation material file corresponding to each basic unit of pronunciation are played synchronously, so that the learner of pronunciation can not only hear the correct pronunciation of each basic unit of pronunciation, but also intuitively watch it. In the process of correct pronunciation of each basic unit of pronunciation, the dynamic changes of the various pronunciation organs and the changes in the strength of the airflow have improved the scientific, intuitive and interesting pronunciation of teaching.
[0014] The embodiments of the present invention will be described in detail below in conjunction with the accompanying drawings of the specification.
[0015] Such as figure 1 As shown, it is a flowchart of a visual pronunciation teaching method in an embodiment of the present invention, which specifically includes the following steps:
[0016] Step 101: Establish a pronunciation basic unit information database.
[0017] The basic unit information of pronunciation includes the basic unit of pronunciation and the age and gender information of the person who pronounced the basic unit of pronunciation, and a unique number is assigned to each basic unit of pronunciation.
[0018] In the teaching of Chinese Pinyin, the basic units of pronunciation include 22 initials, 39 finals, 416 atonal syllables, and 1333 toned syllables. The person who pronounces each basic unit of pronunciation is divided according to age groups, which can be divided into There are three age groups: 7 years old, 18 years old and 35 years old. Among them, the 7-year-old age group represents 3 to 12 years old, the 18-year-old age group represents 13 to 24 years old, and the 35-year-old age group represents 25 to 60 years old.
[0019] Taking the basic unit of pronunciation "a" as an example, the information of each basic unit of pronunciation including the basic unit of pronunciation "a" is shown in the following table:
[0020] Table 1:
[0021] serial number
[0022] Step 102: Establish an audio pronunciation information database.
[0023] The audio pronunciation information database is composed of audio pronunciation information corresponding to each basic unit of pronunciation, among which the audio pronunciation information corresponding to the basic unit of pronunciation, each basic unit of pronunciation corresponds to an audio pronunciation information file, and the number of each basic unit of pronunciation is corresponding to it The audio pronunciation information files have the same number. The pronunciation of each basic unit of pronunciation is recorded according to age and gender, and the corresponding recording formants are statistically analyzed. Formant analysis uses non-overlapping rectangular windows with a length of N = 960 (0.12 seconds) to divide the recorded voice signals into K segment, each segment of speech signal is expressed as s(n) (0≤n
[0024] S FFT ( n ) = s ( n ) , 0 ≤ n N 0 N ≤ n ≤ N FFT N FFT = 1024
[0025] Suppose the short-time frequency spectrum of the input digital speech signal frame is |X(ω)|, ω L ≤ω≤ω H , Where ω L And ω H Is the boundary frequency of the input digital speech signal frame, for S FFT (n) Perform FFT transformation to obtain:
[0026] X(ω)=FFT{S FFT (n))
[0027] Divide |X(ω)| into M subbands according to frequency ω|X 1 (ω)|(ω 1 L ≤ω≤ω 1 H ), |X 2 (ω)|(ω 2 L ≤ω≤ω 2 H ),..., |X M (ω)|(ω M L ≤ω≤ω M H ), where ω i L And ω i H Is the boundary frequency of the i-th subband, ω L ≤ ω i L ≤ ω i H ≤ ω H ( 1 ≤ i ≤ M )
[0028] E j i Represents the energy of the i-th subband of the j-th segment signal, namely E j i = | X i j ( ω ) | 2 , Then the energy of each subband is expressed as:
[0029] E i = 1 K Σ j = 1 K E j i ( ω ) , 1 ≤ i ≤ M
[0030] Select the first three sets of records with the largest sub-band energy and record them as the first, second, and third formants, and find the first, second, and third records of all recordings of the same basic unit, the same gender, and the same age After the mean value of the formant, the recording with the formant closest to the mean value is selected as the audio pronunciation information file of the basic unit of the pronunciation, the gender, and the age.
[0031] Each audio pronunciation information file corresponding to each basic unit of pronunciation in Table 1 is shown in Table 2 below:
[0032] Table 2:
[0033] serial number
[0034] Among them, each audio pronunciation information file corresponding to each basic unit information of the pronunciation can be, but not limited to, saved in a wav format.
[0035] Step 103: Establish a first video pronunciation information database.
[0036] The first video pronunciation information database is composed of the correct pronunciation animation material files formed by the correct pronunciation of the basic unit of pronunciation contained in the information of each basic unit of pronunciation. The dynamic changes of the various pronunciation organs and the changes in the strength of the airflow form the correct pronunciation animation material files. The unit information corresponds to a first video pronunciation information (that is, corresponding to a correct pronunciation animation material file), and the number of each pronunciation basic unit information and the number of the corresponding first video pronunciation information (that is, the corresponding correct pronunciation animation material file) the same.
[0037] The speech signal during the pronunciation process is stable for a short period of time. Due to the physical characteristics of the pronunciation organ, its position is stable and unchanging within tens of milliseconds. In order to more accurately show the continuous change of the position of each basic unit of pronunciation during the correct pronunciation process, the position of each basic unit of pronunciation is drawn every 40 milliseconds, including the human nasal cavity, upper lip, lower lip, Upper gingival, lower gingival, anterior hard palate, medium hard palate, posterior hard palate, soft palate, uvula, tip of tongue, middle tip of tongue, behind tip of tongue, front of tongue, behind tongue surface, changes in the position of vocal cords and changes in airflow strength The correct pronunciation animation material file corresponding to the basic unit information of pronunciation.
[0038]Taking the basic unit "O" as an example, the initial state of the dynamic information of the vocal organs during the correct pronunciation of "O" is that the mouth is half open, the upper lip is slightly lifted, the front teeth are slightly exposed, and the lower jaw is basically still. The dynamic information of the organs depicts the gradual change trajectory from the back of the tongue surface to the soft palate, the gradual change trajectory of the tongue retracting, and the gradual change trajectory of the tongue rising to half height. At the same time, the dynamic information of the vocal organs shows that the lips start slowly from spreading lips The trajectory of the round, the glottis from closed to wide open, the vocal cords from static to vibrating, and the airflow from the lungs rushing out of the mouth.
[0039] The correct pronunciation animation material file corresponding to each basic unit of pronunciation information is saved corresponding to the audio pronunciation information file corresponding to the basic unit of pronunciation. That is, the basic unit of pronunciation contained in each basic unit of pronunciation is in the process of correct pronunciation. The change over time of the correct pronunciation animation material file corresponding to the pronunciation basic unit information corresponds to the change over time of the audio pronunciation information file corresponding to the pronunciation basic unit information.
[0040] The specific saving method of the correct pronunciation animation material file is: select the start and end tongue positions of each correct pronunciation animation material file to correspond to the start and end time points of the corresponding audio pronunciation information file, and then select the above correct The climax point time of the pronunciation animation material file corresponds to the climax point time of the above-mentioned audio pronunciation information file. For the same basic unit of pronunciation, multiple climax points can be selected, and the dynamic changes of the pronunciation organs are inserted between the established time corresponding points Display frames, and on this basis, extend the internal action time of all oral cavity by two to three times. The frontal mouth shape is still fully matched with the time of the audio pronunciation information file, and the time is not extended.
[0041] Each correct pronunciation animation material file corresponding to each pronunciation basic unit information in Table 1 above is shown in Table 3 below:
[0042] table 3:
[0043] serial number
[0044] Among them, each correct pronunciation animation material file corresponding to each pronunciation basic unit information can be, but not limited to, saved in swf format.
[0045] Step 104: Establish a second video pronunciation information database.
[0046] According to Chinese education in ethnic minority areas, Mandarin Chinese education in dialect areas, and Chinese education for foreign students, the statistical information of mispronunciation caused by the influence of native language and local accent in the process of Chinese education for foreign students can also establish a second video pronunciation information database. The basic unit of pronunciation contained in the information of each basic unit of pronunciation. In the process of wrong pronunciation, the dynamic changes of each vocal organ and the change of air flow are composed of wrong pronunciation animation material files. Each basic unit of pronunciation corresponds to a second video. The pronunciation information (that is, corresponding to a wrong pronunciation animation material file), and the number of each basic unit of pronunciation information is the same as the number of the corresponding second video pronunciation information (that is, the corresponding wrong pronunciation animation material file).
[0047] The specific method for establishing the second video information database is similar to the above-mentioned method for establishing the first video pronunciation information database, and will not be repeated here.
[0048] Wherein, the change over time of the wrong pronunciation animation material file corresponding to the basic unit information of pronunciation corresponds to the change over time of the correct pronunciation animation material file corresponding to the basic unit information of pronunciation.
[0049] Each wrong pronunciation animation material file corresponding to each pronunciation basic unit information in Table 1 above is shown in Table 4 below:
[0050] Table 4:
[0051] serial number
[0052] Among them, each wrong pronunciation animation material file corresponding to each pronunciation basic unit information can be saved in the tsh format, but is not limited to.
[0053] Step 105: Search for the number of the pronunciation basic unit information corresponding to the pronunciation basic unit information input by the user in the established pronunciation basic unit information database.
[0054] For example, the basic pronunciation unit information input by the user is: the basic pronunciation unit is "a", the age is 18, and the gender is female, then the number 3 corresponding to the above basic pronunciation unit information is found in Table 1.
[0055] Step 106: Search for audio pronunciation information files with the same number as the audio pronunciation information found in step 105 in the established audio pronunciation information database.
[0056] According to the number 3 found in step 105, the audio pronunciation information file number 3 is found in the above table 2 as a-18-女.wav.
[0057] Step 107: Search for a first video pronunciation information file with the same number as the first video pronunciation information found in step 105 in the established first video pronunciation information database.
[0058] According to the number 3 found in step 105, the first video pronunciation information file number 3 is found in the above table 3 as a-18-女.swf.
[0059] Step 108: Search for a second video pronunciation information file with the same number as that found in step 105 in the established second video pronunciation information database.
[0060] According to the number 3 found in step 105, the second video pronunciation information file number 3 is found in the above table 4 as a-18-女.tsh.
[0061] Step 109: Synchronously play the found audio pronunciation information file and the first video pronunciation information file.
[0062] After the audio pronunciation information file found above and the first video pronunciation information file are played synchronously, the pronunciation basic unit contained in this pronunciation basic unit information has an accurate pronunciation organ location map at any point in the pronunciation display process. , Accurately display the person’s nasal cavity, upper lip, lower lip, upper gingival, lower gingival, anterior hard palate, medium hard palate, posterior hard palate, soft palate, uvula, tip of tongue, middle tip of tongue, behind tip of tongue, in front of tongue, behind tongue , The position of the vocal cords and the strength of the airflow.
[0063] Step 110: Synchronously play the found audio pronunciation information file and the second video pronunciation information file.
[0064] After the audio pronunciation information file and the second video pronunciation information file found above are played synchronously, the learner can not only learn the correct pronunciation process of each basic unit of pronunciation, but also each basic unit of pronunciation presented according to the second video pronunciation information. The error-prone pronunciation process deepens the understanding and memory of the correct pronunciation process through comparative learning.
[0065] Wherein, the above-mentioned pronunciation basic unit information, audio pronunciation information, first video pronunciation information, and second video pronunciation information may be stored in the form of a database, but is not limited to.
[0066] According to the above-mentioned process, it can be seen that when pronunciation teaching is carried out by adopting the scheme of the present invention, pronunciation learners can not only hear the correct pronunciation of each basic unit of pronunciation, but also can intuitively see that each basic unit of pronunciation is in the process of correct pronunciation. The dynamic changes of and the changes in the strength of the airflow improve the scientific, intuitive and interesting pronunciation teaching.
[0067] Correspondingly, the present invention also provides a visual pronunciation teaching device.
[0068] Such as figure 2 As shown, the visual pronunciation teaching device includes:
[0069] The receiving unit 201 is configured to receive basic unit information of pronunciation input by the user.
[0070] The storage unit 202 stores the mapping relationship between the basic unit information of each pronunciation and the corresponding audio pronunciation information, the mapping relationship between each basic unit information of the pronunciation and the corresponding first video pronunciation information, and the information of each basic unit of pronunciation and the corresponding second video pronunciation The mapping relationship of information.
[0071] The first searching unit 203 is configured to search for the audio pronunciation information corresponding to the pronunciation basic unit information received by the receiving unit 201 in the mapping relationship between each pronunciation basic unit information stored in the storage unit 202 and the corresponding audio pronunciation information.
[0072] The second searching unit 204 is configured to search for the first video pronunciation corresponding to the pronunciation basic unit information received by the receiving unit 201 in the mapping relationship between each basic unit of pronunciation information stored in the storage unit 202 and the corresponding first video pronunciation information information.
[0073] The third searching unit 205 is configured to search for the second video pronunciation corresponding to the pronunciation basic unit information received by the receiving unit 201 in the mapping relationship between each basic unit information of pronunciation stored in the storage unit 202 and the corresponding second video pronunciation information information.
[0074] The first playing unit 206 is configured to synchronously play the audio pronunciation information found by the first search unit 203 and the first video pronunciation information found by the second search unit 204.
[0075] The second playing unit 207 is configured to synchronously play the second video pronunciation information found by the third finding unit 205 and the audio pronunciation information found by the first finding unit 203.
[0076] Among them, the first video pronunciation information is used to present the pronunciation basic unit contained in the basic pronunciation unit information in the process of correct pronunciation of each vocal organ dynamic change process, the second video pronunciation information is used to present the basic pronunciation unit information contained in the pronunciation basic The dynamic change process of each vocal organ in the process of incorrect pronunciation of the unit
[0077] During the correct pronunciation of the basic unit of pronunciation, the change over time of the audio pronunciation information searched by the first search unit 203 corresponds to the change over time of the first video pronunciation information searched by the second search unit 204, and The change over time of the audio pronunciation information searched by the first search unit 203 is also corresponding to the change over time of the second video pronunciation information searched by the third search unit 205;
[0078] The pronunciation basic unit information stored in the storage unit 202 includes the pronunciation basic unit and the age and gender information of the person who utters the pronunciation basic unit, and the first video pronunciation information stored in the storage unit 202 is the pronunciation contained in the pronunciation basic unit information. Basic unit In the process of correct pronunciation, the animation information formed by the position changes of the nasal cavity, lips, gums, jaw, tongue and vocal cords and the changes in the strength of the airflow. The stored second video pronunciation information is the basic unit The basic unit of pronunciation contained in the information is the animation information formed by the changes in the position of the nasal cavity, lips, gums, jaws, tongue, and vocal cords and the changes in the strength of the airflow during the process of incorrect pronunciation.
[0079] In addition, the visual pronunciation teaching method and device provided by the present invention are not only suitable for teaching Chinese Pinyin, but also suitable for teaching pronunciation in other languages.
[0080] Obviously, those skilled in the art can make various changes and modifications to the present invention without departing from the spirit and scope of the present invention. In this way, if these modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalent technologies, the present invention is also intended to include these modifications and variations.
the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

PUM

no PUM

Description & Claims & Application Information

We can also present the details of the Description, Claims and Application information to help users get a comprehensive understanding of the technical details of the patent, such as background art, summary of invention, brief description of drawings, description of embodiments, and other original content. On the other hand, users can also determine the specific scope of protection of the technology through the list of claims; as well as understand the changes in the life cycle of the technology with the presentation of the patent timeline. Login to view more.
the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to view more

Similar technology patents

Dual medium numerical reservoir simulation parameter adjusting method

InactiveCN108932397AImprove scienceImprove accuracy and productivityFluid removalDesign optimisation/simulationNumerical modelingReservoir simulation
Owner:PETROCHINA CO LTD

Method and system for evaluating on-off performance of high-voltage circuit breaker, and medium

Owner:ELECTRIC POWER RESEARCH INSTITUTE OF STATE GRID NINGXIA ELECTRIC POWER COMPANY +2

Student physical health monitoring method and system

Owner:SHENZHEN RUNAN SCI & TECH DEVCO

Classification and recommendation of technical efficacy words

Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products