Video conversation recording method and device
A technology of video call and recording method, which is applied in the direction of TV, image communication, color TV, etc., and can solve the problem that the user cannot view the content of the call
Inactive Publication Date: 2017-03-15
BEIJING XIAOMI MOBILE SOFTWARE CO LTD
5 Cites 3 Cited by
AI-Extracted Technical Summary
Problems solved by technology
When using text chat, you can record all the text content and/or picture content through the chat interface. After the communication is ...
Method used
[0101] In this embodiment, the superimposed audio data is encoded to obtain encoded audio data, which can reduce redundancy in the audio. For example, Pulse Code Modulation (English: Pulse Code Modulation, PCM for short) may be used to encode the superimposed audio data. Which audio encoding method is specifically adopted to encode the audio data is not limited in this embodiment.
[0102] The superimposed video data is encoded to obtain encoded video data, which can reduce redundancy in the video. For example, Mpeg4 may be used to encode the superimposed video data. It is not limited in this embodiment to specifically adopt which video encoding manner to encode the audio data.
[0103] In this embodiment, the encoded audio data and the encoded video data are superimposed into a multimedia file to reduce redundant information of the multimedia file and reduce the size of the multimedia file.
[0109] In this embodiment, during a video call, the audio stream and the video stream sent by the opposite end of the call are received, and audio decoding is performed on the received audio stream to obtain decoded audio data, which is referred to as first audio data herein. Video decoding is performed on the received video stream to obtain decoded video data, which is referred to as first video data herein. Since the received audio stream and the video stream are ...
Abstract
The invention relates to a video conversation recording method and device. The method comprises the steps that first audio data and first video data transmitted by an opposite conversation side are received and second audio data and second video data of a local conversation side are acquired in a video conversation; the first audio data, the second audio data, the first video data and the second video data are superposed so that multimedia files recording to the video conversation content are acquired; and the content of the video conversation process is reproduced after the end of the video conversation through the multimedia files.
Application Domain
Television system detailsColor television details +1
Technology Topic
Audio frequencyMultimedia
Image
Examples
- Experimental program(1)
Example Embodiment
[0072] The exemplary embodiments will be described in detail here, and examples thereof are shown in the accompanying drawings. When the following description refers to the accompanying drawings, unless otherwise indicated, the same numbers in different drawings represent the same or similar elements. The implementation manners described in the following exemplary embodiments do not represent all implementation manners consistent with the present disclosure. Rather, they are merely examples of devices and methods consistent with some aspects of the present disclosure as detailed in the appended claims.
[0073] figure 1 Is a flowchart of a method for recording a video call according to an exemplary embodiment, such as figure 1 As shown, the video call recording method used in the terminal includes the following steps.
[0074] In step S11, in the video call, the first audio data and the first video data sent by the opposite end of the call are received, and the second audio data and the second video data of the call end are collected.
[0075] In this embodiment, during a video call between the opposite end of the call and the local end of the call, the method of this embodiment is executed on the local end of the call. This embodiment can receive the audio data sent by the opposite end of the call in real time (here referred to as The first audio data) and video data (herein referred to as the first video data), as well as the audio data (herein referred to as the second audio data) and second video data (herein referred to as the second video data). Wherein, the first audio data is used to play the voice collected by the call peer through the local end of the call during the video call, and the first video data is used to display the picture collected by the call peer during the video call.
[0076] Optionally, this embodiment can collect audio data of the local end of the call through the microphone of the terminal in real time, and this embodiment can collect video data of the local end of the call through the camera of the terminal in real time.
[0077] In step S12, the first audio data, the second audio data, the first video data, and the second video data are superimposed and processed to obtain a multimedia file recording video call content.
[0078] In this embodiment, the first audio data, second audio data, second video data, and second video data obtained above are superimposed and processed to obtain a multimedia file. Since the multimedia file includes the audio data and video data of the call peer, And the audio data and video data of the local end of the call, so the multimedia file records the content of the video call. Therefore, when the multimedia file is subsequently played, the content during the video call can be reproduced.
[0079] In summary, the video call recording method provided by this embodiment receives the first audio data and first video data sent by the opposite end of the call during the video call, and collects the second audio data and second video of the local end of the call. Data; the first audio data, the second audio data, the first video data, and the second video data are superimposed and processed to obtain a multimedia file recording the content of the video call; the video call is realized through the multimedia file After the end, reproduce the content of the video call.
[0080] figure 2 Is a flowchart of a method for recording a video call according to another exemplary embodiment, such as figure 2 As shown, the video call recording method used in the terminal includes the following steps.
[0081] In step S21, in the video call, the first audio data and the first video data sent by the opposite end of the call are received, and the second audio data and the second video data of the local end of the call are collected.
[0082] In this embodiment, the specific implementation process of step S21 can be referred to figure 1 The relevant description of the illustrated embodiment will not be repeated here.
[0083] In this embodiment, the first audio data, the second audio data, the first video data, and the second video data are superimposed and processed to obtain a multimedia file recording video call content. Step S22-Step S24.
[0084] In step S22, the first audio data and the second audio data are superimposed to obtain superimposed audio data.
[0085] In this embodiment, after acquiring the first audio data and the second audio data, the first audio data and the second audio data are superimposed to obtain the superimposed audio data. For example: this superposition process is to superimpose the audio frames corresponding to the number of frames in the first audio data and the second audio data, such as: the first audio data in the first audio data and the first second audio data in the audio frame The audio frame of the second audio data is superimposed, and the audio frame in the second first audio data is superimposed with the audio frame in the second second audio data; and so on.
[0086] In step S23, the first video data and the second video data are superimposed to obtain superimposed video data.
[0087] In this embodiment, after acquiring the first video data and the second video data, the first video data and the second video data are superimposed to obtain superimposed video data. For example: this superimposition process is to superimpose the corresponding number of video frames in the first video data and the second video data, such as: the audio frame in the first first video data and the first second video data The video frames in the second video data are superimposed, and the video frames in the second first video data are superimposed on the video frames in the second second video data; and so on.
[0088] Optionally, a feasible implementation of step S23 may be: superimposing the first video data and the second video data according to a preset superimposition rule to obtain superimposed video data. The preset superimposition rule is used to indicate at least one of the following: the positional relationship between the first video frame and the second video frame in each video frame of the superimposed video data, and the position relationship between each video frame of the superimposed video data; The size relationship between the first video frame and the second video frame; the first video frame is a video frame of the first video data, and the second video frame is a video frame of the second video data.
[0089] Wherein, each video frame in the superimposed video data includes a first video frame and a second video frame. For example, the first video frame in the superimposed video data includes the first first video frame and the first video frame. The second video frame, the second video frame in the superimposed video data includes the second first video frame and the second second video frame, and so on; only the first video frame in each video frame The positional relationship and/or size relationship with the second video frame is determined by a preset superimposition rule.
[0090] Wherein, the positional relationship between the first video frame and the second video frame in each video frame of the superimposed video data is used to indicate: the first video frame is located above or below the second video frame, or the first video frame is located Left or right of the second video frame, or the first video frame is embedded in the upper left or lower left or upper right or lower right of the second video frame, or the second video frame is embedded in the upper left of the first video frame Square or bottom left or top right or bottom right. However, this embodiment is not limited to this.
[0091] Wherein, the size relationship between the first video frame and the second video frame in each video frame of the superimposed video data is used to indicate: the size of the first video frame and the second video frame are the same, or the size of the first video frame It is one-Nth of the size of the second video frame, or the size of the second video frame is one-Nth of the size of the first video frame, and N is an integer greater than or equal to 2. However, this embodiment is not limited to this.
[0092] It should be noted that the execution process of step S22 and step S23 is in no particular order.
[0093] In step S24, the superimposed audio data and the superimposed video data are superimposed into the multimedia file.
[0094] In this embodiment, the superimposed audio data obtained in step S22 and the superimposed video data obtained in step S23 are superimposed to obtain a multimedia file for recording video call content.
[0095] Taking the frame rate of the video frame in the video data of 20 frames per second and the frame rate of the audio frame in the audio data of 60 frames per second as an example, in the process of superimposing audio data and video data, you can Based on the video frame, superimpose 3 audio frames, for example: superimpose the video frame in the first video data with the audio frame in the 1-3 audio data, and combine the video frame in the second video data with The audio frames in the 4-6th audio data are superimposed; and so on.
[0096] For example, the format of the multimedia file is, for example, the MP4 format, and this embodiment is not limited to this. According to the format of the multimedia file, the superposition processing principle corresponds to the format of the multimedia file.
[0097] Optionally, a feasible implementation of step S24 in this embodiment may include steps S241-S243:
[0098] In step S241, audio encoding is performed on the superimposed audio data to obtain encoded audio data.
[0099] In step S242, video encoding processing is performed on the superimposed video data to obtain encoded video data.
[0100] In step S243, the encoded audio data and the encoded video data are superimposed into the multimedia file.
[0101] In this embodiment, the superimposed audio data is encoded to obtain encoded audio data. The encoded audio data can reduce the redundancy in the audio. For example, pulse code modulation coding (English: Pulse Code Modulation, PCM for short) can be used to encode the superimposed audio data. The specific audio encoding method used to encode audio data is not limited in this embodiment.
[0102] The superimposed video data is encoded to obtain encoded video data, and the encoded video data can reduce redundancy in the video. For example, Mpeg4 can be used to encode the superimposed video data. The specific video encoding method used to encode audio data is not limited in this embodiment.
[0103] In this embodiment, the encoded audio data and the encoded video data are superimposed into a multimedia file to reduce redundant information of the multimedia file and reduce the size of the multimedia file.
[0104] In summary, the video call recording method provided in this embodiment can obtain a multimedia file for recording the content of the video call through the above-mentioned solution; the multimedia file is used to reproduce the content during the video call after the end of the video call.
[0105] image 3 Is a flowchart of a method for recording a video call according to another exemplary embodiment, such as image 3 As shown, the video call recording method used in the terminal includes the following steps.
[0106] In step S31, in the video call, the audio stream and the video stream sent by the call peer are received.
[0107] In step S32, the audio stream is audio-decoded to obtain first audio data; and the video stream is video-decoded to obtain the first video data.
[0108] In this embodiment, a possible implementation manner of receiving the first audio data and the first video data sent by the call peer may include step S31 and step S32.
[0109] In this embodiment, in a video call, the audio stream and the video stream sent by the call peer are received, and the received audio stream is audio-decoded to obtain decoded audio data, which is referred to herein as first audio data. Perform video decoding on the received video stream to obtain decoded video data, which is referred to herein as first video data. Since the received audio stream is encoded and the video stream is encoded, the transmission bandwidth can be reduced and resources can be saved in the process of transmitting the audio stream and the video stream.
[0110] In step S33, the second audio data and second video data of the local end of the call are collected.
[0111] In this embodiment, the specific implementation process of step S33 can be referred to figure 1 The related description in the illustrated embodiment will not be repeated here.
[0112] In step S34, the first audio data, the second audio data, the first video data, and the second video data are superimposed and processed to obtain a multimedia file recording video call content.
[0113] In this embodiment, the specific implementation process of step S34 can be referred to figure 1 or figure 2 The related description in the illustrated embodiment will not be repeated here.
[0114] Optionally, the method of this embodiment may further include step S35.
[0115] In step S35, the multimedia file is stored.
[0116] In this embodiment, after obtaining the multimedia file, in order to facilitate the user to view the multimedia file at any time, this embodiment may also store the multimedia file. In a feasible implementation manner, this embodiment can store the multimedia file in the local end of the call, that is, in the terminal that executes this embodiment. In a feasible implementation manner, this embodiment can store the multimedia file in the server, which can avoid occupying the storage space of the terminal.
[0117] In summary, the video call recording method provided in this embodiment can obtain a multimedia file for recording the content of the video call through the above-mentioned solution; the multimedia file is used to reproduce the content during the video call after the end of the video call.
[0118] Figure 4 Is a flowchart of a method for recording a video call according to another exemplary embodiment, such as Figure 4 As shown, the video call recording method used in the terminal includes the following steps.
[0119] In step S41, in the video call, an identifier for triggering the video call recording is displayed.
[0120] In step S42, based on the identifier, an operation instruction input by the user to start video call recording is obtained.
[0121] During a video call, this embodiment displays an identifier for triggering video call recording on the display interface. When the user needs to record a video call, the user will click on the identifier to select the identifier. This embodiment can be used for A contact operation or a sliding operation is detected in the area where the indicator that triggers the video call recording is located, an operation instruction to start the video call recording is obtained according to the contact operation or the sliding operation, and then the following steps S43 and S44 are executed.
[0122] In step S43, the first audio data and the first video data sent by the opposite end of the call are received, and the second audio data and second video data of the local end of the call are collected.
[0123] In step S44, the first audio data, the second audio data, the first video data, and the second video data are superimposed and processed to obtain a multimedia file recording video call content.
[0124] In this embodiment, the specific implementation process of step S43 and step S44 can be found in Figure 1-Figure 3 The relevant description in any of the shown method embodiments will not be repeated here.
[0125] In step S45, an indicator for ending the video call recording is displayed.
[0126] In step S46, based on the identifier, an operation instruction to end the video call record input by the user is obtained.
[0127] In this embodiment, during the video call recording, this embodiment displays an identifier for ending the video call recording on the display interface. When the user needs to end the video call recording, the user will click the identifier to select the identifier. For example, a contact operation or a sliding operation can be detected in the area where the mark for ending the video call record is located, and an operation instruction to end the video call record can be obtained according to the contact operation or the sliding operation, and then the end.
[0128] Optionally, after performing step S45 and step S46, step S41 may be performed.
[0129] It should be noted that step S45 can be executed after step S42.
[0130] In summary, the video call recording method provided in this embodiment can obtain a multimedia file for recording video call content through the above solution, and the video call content can be determined by the user to record important content; this is achieved through the multimedia file After the video call ends, reproduce the content of the video call.
[0131] In another alternative solution, when the video call starts, this embodiment starts to perform steps S45 and S46 as described above, and when the video call ends, the method of this embodiment also ends immediately.
[0132] The following are device embodiments of the present disclosure, which can be used to implement the method embodiments of the present disclosure. For details that are not disclosed in the device embodiments of the present disclosure, please refer to the method embodiments of the present disclosure.
[0133] Figure 5 It is a block diagram showing a video call recording device according to an exemplary embodiment. The video call recording device can be implemented as part or all of an electronic device including a touch screen display through software, hardware or a combination of the two. Reference Figure 5 , The device includes a receiving module 100, an acquisition module 200, and an overlay module 300.
[0134] The receiving module 100 is configured to receive first audio data and first video data sent by the opposite end of the call during a video call.
[0135] The collection module 200 is configured to collect the second audio data and the second video data of the local end of the call.
[0136] The superimposing module 300 is configured to superimpose and process the first audio data, the second audio data, the first video data, and the second video data to obtain a multimedia file recording video call content.
[0137] Regarding the device in the foregoing embodiment, the specific manner and technical effect of each module performing operation have been described in detail in the embodiment related to the method, and detailed description will not be given here.
[0138] Image 6 It is a block diagram showing a video call recording device according to another exemplary embodiment. The video call recording device can be implemented as part or all of an electronic device including a touch screen display through software, hardware or a combination of the two. Reference Image 6 , The device is in Figure 5 Based on the device embodiment shown, the superimposition module 300 includes: a first superimposition sub-module 310, a second superimposition sub-module 320, and a third superimposition sub-module 330.
[0139] The first superimposing sub-module 310 is configured to superimpose the first audio data and the second audio data to obtain superimposed audio data.
[0140] The second superimposition submodule 320 is configured to superimpose the first video data and the second video data to obtain superimposed video data.
[0141] The third superimposition submodule 330 is configured to superimpose the superimposed audio data and the superimposed video data into the multimedia file.
[0142] Optionally, the third superimposition sub-module 330 includes: a first coding sub-module 331, a second coding sub-module 332, and a superimposing sub-module 333.
[0143] The first encoding sub-module 331 is configured to perform audio encoding processing on the superimposed audio data to obtain encoded audio data.
[0144] The second encoding submodule 332 is configured to perform video encoding processing on the superimposed video data to obtain encoded video data.
[0145] The superimposing sub-module 333 is configured to superimpose the encoded audio data and the encoded video data into the multimedia file.
[0146] Optionally, the second superimposition submodule 320 is configured to superimpose the first video data and the second video data according to a preset superimposition rule to obtain superimposed video data.
[0147] The preset superimposition rule is used to indicate at least one of the following: the positional relationship between the first video frame and the second video frame in each video frame of the superimposed video data, and the position relationship between each video frame of the superimposed video data; The size relationship between the first video frame and the second video frame; the first video frame is a video frame of the first video data, and the second video frame is a video frame of the second video data.
[0148] Optionally, the receiving module 100 includes: a receiving submodule 110, a first decoding submodule 120, and a second decoding submodule 130.
[0149] The receiving submodule 110 is configured to receive audio streams and video streams sent by the call peer.
[0150] The first decoding submodule 120 is configured to audio-decode the audio stream to obtain the first audio data.
[0151] The second decoding submodule 130 is configured to decode the video of the video stream to obtain the first video data.
[0152] Regarding the device in the foregoing embodiment, the specific manner and technical effect of each module performing operation have been described in detail in the embodiment related to the method, and detailed description will not be given here.
[0153] Figure 7 It is a block diagram showing a video call recording device according to another exemplary embodiment. The video call recording device can be implemented as part or all of an electronic device including a touch screen display through software, hardware or a combination of the two. Reference Figure 7 , The device is in Figure 5 or Image 6 On the basis of the device embodiment shown, the device of this embodiment further includes a storage module 400.
[0154] The storage module 400 is configured to store the multimedia file.
[0155] Optionally, the device of this embodiment further includes: a display module 500 and an acquisition module 600.
[0156] In a feasible implementation manner, the display module 500 is configured to display an identifier for triggering video call recording.
[0157] The obtaining module 600 is configured to obtain an operation instruction input by the user to start video call recording based on the identifier.
[0158] In a feasible implementation manner, the display module 500 is configured to display an identifier for ending the video call recording.
[0159] The obtaining module 600 is configured to obtain an operation instruction input by the user to end the video call record based on the identifier.
[0160] Regarding the device in the foregoing embodiment, the specific manner and technical effect of each module performing operation have been described in detail in the embodiment related to the method, and detailed description will not be given here.
[0161] Figure 8 It is a block diagram showing a video call recording device 800 according to an exemplary embodiment. For example, the device 800 may be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, etc.
[0162] Reference Figure 8 , The device 800 may include one or more of the following components: a processing component 802, a memory 804, a power component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812, a sensor component 814, and a communication component 816.
[0163] The processing component 802 generally controls the overall operations of the device 800, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the foregoing method. In addition, the processing component 802 may include one or more modules to facilitate the interaction between the processing component 802 and other components. For example, the processing component 802 may include a multimedia module to facilitate the interaction between the multimedia component 808 and the processing component 802.
[0164] The memory 804 is configured to store various types of data to support operations in the device 800. Examples of such data include instructions for any application or method operating on the device 800, contact data, phone book data, messages, pictures, videos, etc. The memory 804 can be implemented by any type of volatile or non-volatile storage devices or their combination, such as static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable and Programmable read only memory (EPROM), programmable read only memory (PROM), read only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
[0165] The power component 806 provides power to various components of the device 800. The power component 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the device 800.
[0166] The multimedia component 808 includes a screen that provides an output interface between the device 800 and the user. In some embodiments, the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user. The touch panel includes one or more touch sensors to sense touch, sliding, and gestures on the touch panel. The touch sensor may not only sense the boundary of the touch or slide action, but also detect the duration and pressure related to the touch or slide operation. In some embodiments, the multimedia component 808 includes a front camera and/or a rear camera. When the device 800 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
[0167] The audio component 810 is configured to output and/or input audio signals. For example, the audio component 810 includes a microphone (MIC), and when the device 800 is in an operation mode, such as a call mode, a recording mode, and a voice recognition mode, the microphone is configured to receive external audio signals. The received audio signal may be further stored in the memory 804 or transmitted via the communication component 816. In some embodiments, the audio component 810 further includes a speaker for outputting audio signals.
[0168] The I/O interface 812 provides an interface between the processing component 802 and a peripheral interface module. The peripheral interface module may be a keyboard, a click wheel, a button, and the like. These buttons may include but are not limited to: home button, volume button, start button, and lock button.
[0169] The sensor component 814 includes one or more sensors for providing the device 800 with various aspects of status assessment. For example, the sensor component 814 can detect the on/off status of the device 800 and the relative positioning of the components. For example, the component is the display and the keypad of the device 800. The sensor component 814 can also detect the position change of the device 800 or a component of the device 800. , The presence or absence of contact between the user and the device 800, the orientation or acceleration/deceleration of the device 800, and the temperature change of the device 800. The sensor component 814 may include a proximity sensor configured to detect the presence of nearby objects when there is no physical contact. The sensor component 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor or a temperature sensor.
[0170] The communication component 816 is configured to facilitate wired or wireless communication between the device 800 and other devices. The device 800 can access a wireless network based on a communication standard, such as WiFi, 2G or 3G, or a combination thereof. In an exemplary embodiment, the communication component 816 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel. In an exemplary embodiment, the communication component 816 further includes a near field communication (NFC) module to facilitate short-range communication. For example, the NFC module can be implemented based on radio frequency identification (RFID) technology, infrared data association (IrDA) technology, ultra-wideband (UWB) technology, Bluetooth (BT) technology and other technologies.
[0171] In an exemplary embodiment, the apparatus 800 may be implemented by one or more application specific integrated circuits (ASIC), digital signal processors (DSP), digital signal processing equipment (DSPD), programmable logic devices (PLD), field programmable A gate array (FPGA), controller, microcontroller, microprocessor, or other electronic components are implemented to implement the above methods.
[0172] In an exemplary embodiment, there is also provided a non-transitory computer-readable storage medium including instructions, such as the memory 804 including instructions, which can be executed by the processor 820 of the device 800 to complete the foregoing method. For example, the non-transitory computer-readable storage medium may be ROM, random access memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
[0173] A non-transitory computer-readable storage medium, when the instructions in the storage medium are executed by the processor of the device 800, the device 800 can execute the above video call recording method.
[0174] Those skilled in the art will easily think of other embodiments of the present disclosure after considering the description and practicing the invention disclosed herein. This application is intended to cover any variations, uses, or adaptive changes of the present disclosure. These variations, uses, or adaptive changes follow the general principles of the present disclosure and include common knowledge or conventional technical means in the technical field not disclosed in the present disclosure. . The description and the embodiments are only regarded as exemplary, and the true scope and spirit of the present disclosure are pointed out by the following claims.
[0175] It should be understood that the present disclosure is not limited to the precise structure that has been described above and shown in the drawings, and various modifications and changes can be made without departing from its scope. The scope of the present disclosure is only limited by the appended claims.
PUM


Description & Claims & Application Information
We can also present the details of the Description, Claims and Application information to help users get a comprehensive understanding of the technical details of the patent, such as background art, summary of invention, brief description of drawings, description of embodiments, and other original content. On the other hand, users can also determine the specific scope of protection of the technology through the list of claims; as well as understand the changes in the life cycle of the technology with the presentation of the patent timeline. Login to view more.