Disclosure of Invention
The embodiment of the application provides a live broadcast teaching method, a live broadcast teaching device, a live broadcast teaching terminal and a storage medium, and can solve the problem of low learning efficiency of students caused by boring teaching pictures in the related technology. The technical scheme is as follows:
in a first aspect, a method for live teaching is provided, where the method includes:
displaying a three-dimensional scene animation corresponding to a first course on a first display layer in a live broadcast teaching interface of the first course;
acquiring a live teacher video, and displaying the live teacher video on a second display layer, wherein the second display layer is positioned at the lower layer of the first display layer;
the method comprises the steps of obtaining live video of a student, and displaying the live video of the student on a third display layer, wherein the third display layer is located on the upper layer of the first display layer.
In a possible implementation manner, the displaying, by the first display layer, the three-dimensional scene animation corresponding to the first course is performed by the first terminal, where the displaying includes:
displaying the three-dimensional scene animation corresponding to the first course in a non-full screen mode on a first display layer;
the obtaining of the live teacher video and the displaying of the live teacher video on the second display layer include:
acquiring a live teacher video sent by a second terminal through a live broadcast server, and displaying the live teacher video in a full screen mode on a second display layer, wherein the second terminal is a teacher terminal;
the acquiring of the live video of the student and the displaying of the live video of the student on the third display layer comprise:
and acquiring a live video of the student through a shooting device, and displaying the live video of the student on a third display layer in a non-full screen mode.
In a possible implementation manner, the displaying, by the first display layer, the three-dimensional scene animation corresponding to the first course is performed by using a first terminal as a teacher terminal, and includes:
displaying the three-dimensional scene animation corresponding to the first course in a non-full screen mode on a first display layer;
the obtaining of the live teacher video and the displaying of the live teacher video on the second display layer include:
acquiring a live teacher video through a shooting device, and displaying the live teacher video on a second display layer in a non-full screen mode, wherein the display area of the three-dimensional scene animation is smaller than the display area of the live teacher video;
the acquiring of the live video of the student and the displaying of the live video of the student on the third display layer comprise:
and acquiring a live student video sent by the second terminal through the live broadcast server, and displaying the live student video on a third display layer in a non-full screen mode.
In one possible implementation, the method further includes:
when the selection operation of the target prop in the three-dimensional scene animation is detected, the target prop is controlled to execute a preset action corresponding to the selection operation, and/or a preset audio corresponding to the selection operation is played.
In a possible implementation manner, when a selection operation on a target prop in the three-dimensional scene animation is detected, controlling the target prop to execute a preset action corresponding to the selection operation includes:
when the selection operation of the target prop in the three-dimensional scene animation is detected, sending a target prop control message to a double-end synchronous server, so that the double-end synchronous server forwards the control message to the first terminal and the second terminal simultaneously;
and receiving the target prop control message, and controlling the target prop to execute a preset action corresponding to the first selection operation.
In one possible implementation manner, the first terminal is a student terminal, and the method further includes:
receiving a reading instruction of a first audio, and playing the first audio;
and receiving the follow-up reading audio of the first student, and sending the follow-up reading audio of the first student to the second terminal.
In one possible implementation manner, the first terminal is a student terminal, and the method further includes:
displaying a student self-learning interface of the second course;
displaying the three-dimensional scene animation corresponding to the second course in a full screen mode in the student self-learning interface;
receiving a reading instruction of a second audio frequency, and playing the second audio frequency;
receiving a second student follow-up reading audio, inputting the second audio and the second student follow-up reading audio into a pre-trained voice recognition model, and determining the similarity of the second audio and the second student follow-up reading audio;
and if the similarity is larger than a preset threshold value, displaying a virtual reward item on the student self-learning interface.
In one possible implementation manner, the first terminal is a teacher terminal, and the method further includes:
when a trigger event of adding a target magic expression is detected, acquiring the target magic expression;
carrying out face recognition on each frame of image in the live video of the teacher to determine a face area in each frame of image;
and adding the target magic expression in the face area of each frame of image.
In a possible implementation manner, the obtaining the target magic expression when the trigger event for adding the target magic expression is detected includes:
and when a selection instruction of the target magic expression is received, acquiring the target magic expression.
In a second aspect, an apparatus for live broadcast teaching is provided, where the apparatus is applied to a first terminal, and the apparatus includes:
the first display module is used for displaying the three-dimensional scene animation corresponding to the first course on a first display layer in a live broadcast teaching interface of the first course;
the second display block is used for acquiring a live teacher video and displaying the live teacher video on a second display layer, wherein the second display layer is positioned at the lower layer of the first display layer;
and the third display module is used for acquiring the live video of the student and displaying the live video of the student on a third display layer, wherein the third display layer is positioned on the upper layer of the first display layer.
In a possible implementation manner, the first terminal is a student terminal, and the first display module is configured to:
displaying the three-dimensional scene animation corresponding to the first course in a non-full screen mode on a first display layer;
the second display module is configured to:
acquiring a live teacher video sent by a second terminal through a live broadcast server, and displaying the live teacher video in a full screen mode on a second display layer, wherein the second terminal is a teacher terminal;
the third display module is configured to:
and acquiring a live video of the student through a shooting device, and displaying the live video of the student on a third display layer in a non-full screen mode.
In a possible implementation manner, the first terminal is a teacher terminal, and the first display module is configured to:
displaying the three-dimensional scene animation corresponding to the first course in a non-full screen mode on a first display layer;
the second display module is configured to:
acquiring a live teacher video through a shooting device, and displaying the live teacher video on a second display layer in a non-full screen mode, wherein the display area of the three-dimensional scene animation is smaller than the display area of the live teacher video;
the third display module is configured to:
and acquiring a live student video sent by the second terminal through the live broadcast server, and displaying the live student video on a third display layer in a non-full screen mode.
In one possible implementation, the apparatus further includes:
and the control module is used for controlling the target prop to execute a preset action corresponding to the selection operation and/or playing a preset audio corresponding to the selection operation when the selection operation of the target prop in the three-dimensional scene animation is detected.
In one possible implementation, the control module is configured to:
when the selection operation of the target prop in the three-dimensional scene animation is detected, sending a target prop control message to a double-end synchronous server, so that the double-end synchronous server forwards the control message to the first terminal and the second terminal simultaneously;
and receiving the target prop control message, and controlling the target prop to execute a preset action corresponding to the first selection operation.
In one possible implementation manner, the first terminal is a student terminal, and the apparatus further includes:
the playing module is used for receiving a playing instruction of a first audio and playing the first audio;
and the sending module is used for receiving the audio read-after of the first student and sending the audio read-after of the first student to the second terminal.
In one possible implementation manner, the first terminal is a student terminal, and the apparatus further includes:
the fourth display module is used for displaying a student self-learning interface of a second course, and in the student self-learning interface, displaying the three-dimensional scene animation corresponding to the second course in a full screen mode;
the receiving module is used for receiving a playing instruction of a second audio and playing the second audio;
the determining module is used for receiving a second student follow-up reading audio, inputting the second audio and the second student follow-up reading audio into a pre-trained voice recognition model, and determining the similarity of the second audio and the second student follow-up reading audio;
and the reward module is used for displaying a virtual reward item on the student self-learning interface if the similarity is greater than a preset threshold value.
In one possible implementation manner, the first terminal is a teacher terminal, and the apparatus further includes:
the detection module is used for acquiring a target virtual expression when a trigger event for adding the target virtual expression is detected;
the determining module is used for carrying out face recognition on each frame of image in the live video of the teacher and determining a face area in each frame of image;
and adding the target virtual expression in the face area of each frame of image.
In a possible implementation manner, the detection module is configured to:
and when a selection instruction of the target virtual expression is received, acquiring the target virtual expression.
In a third aspect, there is provided a first terminal, the first terminal comprising a processor and a memory, the memory having stored therein at least one instruction, the at least one instruction being loaded and executed by the processor to implement the method of live education as described in the first aspect above.
In a fourth aspect, there is provided a computer readable storage medium having stored therein at least one instruction that is loaded and executed by the processor to implement the method of live tutoring as described in the first aspect above.
The technical scheme provided by the embodiment of the application has the following beneficial effects:
the terminal can display the three-dimensional scene animation corresponding to the first course on the first display layer in a non-full screen mode in the live broadcast teaching interface of the first course. And the live video of the teacher can be acquired and displayed on a second display layer below the first display layer. In addition, live videos of students can be acquired and displayed on a third display layer on the upper layer of the first display layer. Namely, a teacher live video, a student live video and three-dimensional scene animations corresponding to courses are fused and displayed on different display layers of the same interface. When a student is in class, the student can see the live video of the teacher and also can see the three-dimensional scene animation corresponding to the course, and the three-dimensional scene animation has greater attraction for the infant, so that the attention of the student can be more concentrated, and further, the learning efficiency of the student is higher.
Detailed Description
To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
Fig. 1 is a schematic diagram of an implementation environment provided by an embodiment of the present application. Referring to fig. 1, a first terminal, a second terminal, and a live server may be included in the implementation environment. One of the first terminal and the second terminal is a student terminal, and the other terminal is a teacher terminal. And the first terminal and the second terminal send live broadcast videos through the live broadcast server. That is, the first terminal may send the live video to the live server, and the live server sends the live video to the second terminal, so that the second terminal receives and displays the live video of the first terminal. Similarly, the second terminal can also send the live video to the live server, and the live server sends the live video to the first terminal, so that the second terminal receives and displays the live video of the first terminal.
Fig. 2 is a flowchart of a live broadcast teaching method provided in an embodiment of the present application, where the method may be implemented by a first terminal. Referring to fig. 2, the process flow of the method may include the following steps:
step 201, in a live-broadcast teaching interface of a first course, displaying a three-dimensional scene animation corresponding to the first course on a first display layer.
Different pictures can be displayed on different display layers in the live broadcast teaching interface.
In implementation, the technician animates the three-dimensional scene in advance corresponding to each lesson. The content of the three-dimensional scene animation is related to the corresponding course content, so that the purpose of assisting teaching is achieved. For example, if the lesson is learning the english word "car" and some content related to the car, the three-dimensional scene animation corresponding to the lesson may include a three-dimensional model of a moving car, a three-dimensional model of a friend who communicates with each other, and the like.
A technician may animate a three-dimensional scene using a unity tool, also known as unity3D tool, which is a comprehensive game development tool that allows users to create types of interactive content such as three-dimensional video games, building visualizations, real-time three-dimensional animations, and the like. The following is a brief description of the animation of a three-dimensional scene using a unity tool.
Technicians start a unity tool to create a three-dimensional scene, place required props in the created scene, put the props in the position, place a camera in the scene, and render the scene through the camera to obtain the three-dimensional scene animation. Finally, export is as xxx.unity file, where XXX is the technician's name for this unity file and can be set by itself.
The first terminal can be provided with a live teaching application program, and after the live teaching application program is started, a user can select any course to learn. When the user enters a live broadcast teaching interface of any course, the terminal can run a unity tool to load the XXX. unity file, so that the three-dimensional scene animation can be displayed on a first display layer in the live broadcast interface.
In addition, the display area of the three-dimensional scene animation in different live teaching scenes can be different. For example, if the first terminal is a student side terminal, correspondingly, the live teaching application program is a student live teaching application program, and the three-dimensional scene animation is displayed in a corresponding area in a live teaching interface of the student live teaching application program shown in fig. 3. For another example, if the first terminal is a teacher-side terminal, the live teaching application is a teacher live teaching application, and the three-dimensional scene animation is displayed in a corresponding area in a live teaching interface of the teacher live teaching application as shown in fig. 4.
Step 202, acquiring a live teacher video, and displaying the live teacher video on a second display layer, wherein the second display layer is positioned at the lower layer of the first display layer.
In implementation, a Raw Image (original picture) is placed on the bottom layer of a three-dimensional scene animation of a live broadcast teaching interface (i.e., the second display layer in the embodiment of the application), and after a live broadcast video of a teacher is acquired, each frame of the live broadcast video of the teacher is rendered on the Raw Image for display, so that the effect of displaying the live broadcast video of the teacher on the bottom layer of the three-dimensional scene animation can be presented.
In addition, the manner of acquiring live video of a teacher is different for the first terminal being a teacher terminal or a student terminal, and the following description will be made for the case where the first terminal is a teacher terminal and the first terminal is a student terminal.
The first terminal is a student terminal, and the first terminal can acquire a live teacher video sent by the second terminal through the live broadcast server, and the second terminal is a teacher terminal. That is, the teacher logs in the teacher live teaching application program at the teacher terminal and enters the live teaching interface of any course, and the students can also log in the corresponding student live teaching application program at the student terminal and enter the live teaching interface of the same course as that in the teacher live teaching application program. The teacher terminal sends the teacher live video to the student terminals through the live broadcasting server, and the student terminals display the teacher live video on a second display layer of the live broadcasting teaching interface. In addition, as shown in fig. 3, the live teacher video may be displayed in a full screen in the second display layer of the student terminal, and the second display layer is located below the first display layer.
The first terminal is a teacher terminal, and the first terminal can directly acquire a live video of the teacher through the camera device and display the live video on the second display layer in the live teaching interface. The camera device may be a camera component built in the first terminal, or may also be a camera externally connected to the first terminal through a signal line, which is not limited in the embodiment of the present application. In addition, as shown in fig. 4, the live teacher video may be displayed in a non-full screen in the second display layer of the student terminal, and similarly, the second display layer is located below the first display layer.
And 203, acquiring a live student video, and displaying the live student video on a third display layer in a non-full screen mode, wherein the third display layer is positioned on the upper layer of the first display layer.
In implementation, the first terminal is a teacher terminal or a student terminal, and the manner of acquiring live videos of students is different, and the following description is given for the case where the first terminal is a teacher terminal and the first terminal is a student terminal.
The first terminal is a teacher terminal, and then the first terminal can acquire the live broadcast video of the students sent by the second terminal through the live broadcast server, and the second terminal is a student terminal. After the student terminals enter a live broadcast teaching interface of a live broadcast application program of students, the student terminals send live broadcast videos of the students to the teacher terminal through the live broadcast server, and the teacher terminal displays the live broadcast videos of the students on a third display layer of the live broadcast teaching interface. In addition, as shown in fig. 4, the live student video may be displayed in a non-full screen in the third display layer of the teacher terminal, and the third display layer is located on an upper layer of the first display layer.
The first terminal is a student terminal, and the first terminal can directly acquire live video of students through the camera device and display the live video on the second display layer in the live teaching interface. In addition, as shown in fig. 3, the live teacher video may be displayed in a non-full screen mode in a third display layer of the student terminal, and similarly, the third display layer is located above the first display layer.
In addition, it should be further described that fig. 3 and fig. 4 are only one possible live teaching interface provided in the embodiment of the present application. The method and the device are not limited by the layout of the teacher live video, the student live video and the three-dimensional scene animation in the live teaching interface.
In a possible implementation manner, in order to increase the interest of live teaching and make students better put into learning, for the props in the three-dimensional scene animation of the live teaching interface, the user can control the props to execute some actions, and correspondingly, the processing can be as follows: when the selection operation of the target prop in the three-dimensional scene animation is detected, the target prop is controlled to execute a preset action corresponding to the selection operation, and/or a preset audio corresponding to the selection operation is played.
In implementation, a user can click a prop of a three-dimensional scene animation in a live broadcast teaching interface to control the prop to execute a preset action. When the effect is achieved, the first terminal may detect the selection operation of the user on the property by adopting the following method.
When a user selects an operation, such as a click operation, for a prop in the three-dimensional scene animation, a camera in the three-dimensional scene emits a ray to the three-dimensional scene at the click position of the user. Additionally, the props in the three-dimensional scene that need to interact may include boxcar collider components. Then, when the ray touches a prop that includes a boxtool component, the EasyTouch plugin will return the identity of the targeted prop touched.
And the first terminal determines a preset action corresponding to the identifier of the target prop in the stored corresponding relation between the identifier of the target prop and the preset action according to the identifier of the target prop, and controls the target prop to execute the preset action. For example, for the target property being a car, the preset action may be to rotate the car, control the car to go forward, stop moving, and so on.
Of course, the corresponding relationship is not limited to the preset action, and there may be a preset audio, that is, after the preset audio corresponding to the identifier of the target property is determined, the preset audio is played through the audio plug-in. The audio plug-in unit can be an FMOD audio plug-in unit. For example, for the target property being a car, the preset audio may be a car whistling sound.
In one possible implementation manner, in order to synchronize the three-dimensional scene animation displayed in the teacher terminal and the three-dimensional scene animation displayed in the student terminal, the processing may be as follows: when the selection operation of the target prop in the three-dimensional scene animation is detected, sending a target prop control message to a double-end synchronous server, so that the double-end synchronous server forwards the control message to the first terminal and the second terminal simultaneously; and receiving the target prop control message, and controlling the target prop to execute a preset action corresponding to the first selection operation.
In implementation, after the first terminal detects that the user performs the selection operation on the target item through the method, the selection operation may not be performed immediately. Instead, a target prop control message is sent to the dual-end synchronization server, and the target prop control message may carry an identifier of the target prop. And after receiving the target prop control message, the double-end synchronous server simultaneously forwards the target prop control message to the first terminal and the second terminal. After receiving the target prop control message, the first terminal and the second terminal can acquire the target prop identifier therein to determine the corresponding preset action and control the target prop to execute the preset action.
In a possible implementation manner, in the embodiment of the present application, besides reading with a teacher, a student may also provide a reading mode with the teacher, and accordingly, the processing may be as follows: receiving a reading instruction of the first audio, playing the first audio, receiving the follow-up reading audio of the first student, and sending the follow-up reading audio of the first student to the second terminal.
In an implementation, the first terminal may be a student terminal and, correspondingly, the second terminal may be a teacher terminal. And after the teacher selects the follow-up reading function option, the teacher terminal enters the follow-up reading interface in the live broadcasting teaching interface. Meanwhile, the teacher terminal can send follow-up reading information to the student terminals through the double-end synchronous server, and the student terminals can make the live-broadcast teaching interface enter the follow-up reading interface after receiving the follow-up reading information. The teacher can select an option for starting the reading in a reading following interface in the teacher terminal, and then the teacher terminal can send a reading instruction to the student terminal through the double-end synchronous server. After receiving the reading instruction, the student terminal can start to play preset audio which is the audio read by a group of children made in advance, so that the student terminal has better attraction to the student and can better arouse the enthusiasm of the student in following reading learning. The microphones of the student terminals collect the follow-up reading audio of the students and send the follow-up reading audio to the teacher terminal, the teacher terminal plays the follow-up reading audio of the students, and the teacher can send out virtual reward articles for the students according to the quality of the follow-up reading of the students. For example, if the student follows the reading quality is high, the student can be given one star, and then one star can be added above the following reading interface of the student terminal.
In a possible implementation manner, a student can also be provided with a self-learning follow-up function, and accordingly, the processing can be as follows: and displaying a student self-learning interface of the second course, displaying the three-dimensional scene animation corresponding to the second course in a full screen mode in the student self-learning interface, receiving a reading instruction of the second audio, and playing the second audio. Receiving a second student follow-up reading audio, inputting the second audio and the second student follow-up reading audio into a pre-trained voice recognition model, determining the similarity between the second audio and the second student follow-up reading audio, and displaying a virtual bonus item on a student self-learning interface if the similarity is greater than a preset threshold value.
In implementation, in addition to live teaching, a student can also enter a student self-learning interface of a live teaching application program, and three-dimensional scene animation can be displayed in the interface in a full screen mode. The student can select a reading function option in the student self-learning interface, the first terminal (student terminal) starts to play corresponding second audio, the student can read aloud along with the second audio, the first terminal can acquire the second reading audio of the student through the microphone, the second reading audio and the second audio are input into a pre-trained voice recognition model, the similarity between the second reading audio and the second audio is determined, if the similarity is larger than a preset threshold value, the student reading quality is high, and a preset virtual reward item, such as a star, can be displayed on the student self-learning interface.
In a possible implementation manner, in the live broadcast teaching process, in order to attract the attention of students and make the students focus more on watching the teacher live broadcast, the following processing can be performed: when a trigger event of adding the target magic expression is detected, the target magic expression is obtained, face recognition is carried out on each frame of image in the live video of the teacher, a face area in each frame of image is determined, and the target magic expression is added to the face area in each frame of image.
In implementation, the first terminal is a teacher terminal, and in a live-broadcast teaching interface of the teacher terminal shown in fig. 4, a magic expression style option and an expression option may be displayed, and the teacher may select a magic expression style according to real-time performance of the student and select an expression for the magic expression style. And when the teacher terminal acquires the live teacher video, performing face recognition on each frame of the live teacher video, and adding the selected magic expression in the recognized face area. For example, if the teacher thinks the student is not performing well, the teacher may select a magic expression pattern and select an angry expression to indicate dissatisfaction with the student's performance. Correspondingly, the teacher terminal can send the teacher live broadcast video with the magic expression added to the student terminals when sending the teacher live broadcast video to the student terminals through the live broadcast server. The student terminal can display the live video of the teacher added with the magic expression in the live teaching interface.
In addition, the teacher can start a self-adaptive magic expression adding function at the teacher terminal, when the teacher judges that the follow-up reading quality of students is good or the classroom performance is good, the teacher can send out virtual reward articles to the students, when the teacher detects that the students send out the virtual reward articles, the teacher can obtain the smiling magic expression, and the smiling magic expression is added in each frame of live video of the teacher. Similarly, the teacher terminal can send the teacher live broadcast video with the magic expression added to the student terminals when sending the teacher live broadcast video to the student terminals through the live broadcast server. The student terminal can display the live video of the teacher added with the magic expression in the live teaching interface.
In the embodiment of the application, the terminal can display the three-dimensional scene animation corresponding to the first course on the first display layer in a non-full screen mode in the live-broadcast teaching interface of the first course. And moreover, a live video of the teacher can be acquired and displayed in a full screen mode on a second display layer below the first display layer. In addition, a live video of the student can be acquired and displayed on a third display layer on the upper layer of the first display layer in a non-full screen mode. Namely, a teacher live video, a student live video and three-dimensional scene animations corresponding to courses are fused and displayed on different display layers of the same interface. When a student is in class, the student can see the live video of the teacher and also can see the three-dimensional scene animation corresponding to the course, and the three-dimensional scene animation has greater attraction for the infant, so that the attention of the student can be more concentrated, and further, the learning efficiency of the student is higher.
Based on the same technical concept, an embodiment of the present application further provides a device for live broadcast teaching, as shown in fig. 5, the device includes:
the first display module 510 is configured to display, in a live-broadcast teaching interface of a first course, a three-dimensional scene animation corresponding to the first course on a first display layer;
the second display block 520 is configured to acquire a live teacher video and display the live teacher video on a second display layer, where the second display layer is located at a lower layer of the first display layer;
and a third display module 530, configured to acquire a live student video, and display the live student video on a third display layer, where the third display layer is located on an upper layer of the first display layer.
In a possible implementation manner, the first terminal is a student terminal, and the first display module 510 is configured to:
displaying the three-dimensional scene animation corresponding to the first course in a non-full screen mode on a first display layer;
the second display module 520 is configured to:
acquiring a live teacher video sent by a second terminal through a live broadcast server, and displaying the live teacher video in a full screen mode on a second display layer, wherein the second terminal is a teacher terminal;
the third display module 530 is configured to:
and acquiring a live video of the student through a shooting device, and displaying the live video of the student on a third display layer in a non-full screen mode.
In a possible implementation manner, the first terminal is a teacher terminal, and the first display module 510 is configured to:
displaying the three-dimensional scene animation corresponding to the first course in a non-full screen mode on a first display layer;
the second display module 520 is configured to:
acquiring a live teacher video through a shooting device, and displaying the live teacher video on a second display layer in a non-full screen mode, wherein the display area of the three-dimensional scene animation is smaller than the display area of the live teacher video;
the third display module 530 is configured to:
and acquiring a live student video sent by the second terminal through the live broadcast server, and displaying the live student video on a third display layer in a non-full screen mode.
In one possible implementation, the apparatus further includes:
and the control module is used for controlling the target prop to execute a preset action corresponding to the selection operation and/or playing a preset audio corresponding to the selection operation when the selection operation of the target prop in the three-dimensional scene animation is detected.
In one possible implementation, the control module is configured to:
when the selection operation of the target prop in the three-dimensional scene animation is detected, sending a target prop control message to a double-end synchronous server, so that the double-end synchronous server forwards the control message to the first terminal and the second terminal simultaneously;
and receiving the target prop control message, and controlling the target prop to execute a preset action corresponding to the first selection operation.
In one possible implementation manner, the first terminal is a student terminal, and the apparatus further includes:
the playing module is used for receiving a playing instruction of a first audio and playing the first audio;
and the sending module is used for receiving the audio read-after of the first student and sending the audio read-after of the first student to the second terminal.
In one possible implementation manner, the first terminal is a student terminal, and the apparatus further includes:
the fourth display module is used for displaying a student self-learning interface of a second course, and in the student self-learning interface, displaying the three-dimensional scene animation corresponding to the second course in a full screen mode;
the receiving module is used for receiving a playing instruction of a second audio and playing the second audio;
the determining module is used for receiving a second student follow-up reading audio, inputting the second audio and the second student follow-up reading audio into a pre-trained voice recognition model, and determining the similarity of the second audio and the second student follow-up reading audio;
and the reward module is used for displaying a virtual reward item on the student self-learning interface if the similarity is greater than a preset threshold value.
In one possible implementation manner, the first terminal is a teacher terminal, and the apparatus further includes:
the detection module is used for acquiring a target virtual expression when a trigger event for adding the target virtual expression is detected;
the determining module is used for carrying out face recognition on each frame of image in the live video of the teacher and determining a face area in each frame of image;
and adding the target virtual expression in the face area of each frame of image.
In a possible implementation manner, the detection module is configured to:
and when a selection instruction of the target virtual expression is received, acquiring the target virtual expression.
It should be noted that: the device for live broadcasting teaching provided by the above embodiment is exemplified by only the division of the above functional modules when live broadcasting teaching is performed, and in practical application, the above function distribution can be completed by different functional modules as needed, that is, the internal structure of the first terminal is divided into different functional modules, so as to complete all or part of the above described functions. In addition, the device for live broadcast teaching provided by the above embodiment and the method embodiment for live broadcast teaching belong to the same concept, and specific implementation processes thereof are detailed in the method embodiment and are not described herein again.
Fig. 6 shows a block diagram of a terminal 600 according to an exemplary embodiment of the present application. The terminal 600 may be: a smart phone, a tablet computer, an MP3 player (Moving Picture Experts Group Audio Layer III, motion video Experts compression standard Audio Layer 3), an MP4 player (Moving Picture Experts Group Audio Layer iv, motion video Experts compression standard Audio Layer 4), a notebook computer, or a desktop computer. The terminal 600 may also be referred to by other names such as user equipment, portable terminal, laptop terminal, desktop terminal, etc.
In general, the terminal 600 includes: a processor 601 and a memory 602.
The processor 601 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so on. The processor 601 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 601 may also include a main processor and a coprocessor, where the main processor is a processor for processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 601 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed on the display screen. In some embodiments, processor 601 may also include an AI (Artificial Intelligence) processor for processing computational operations related to machine learning.
The memory 602 may include one or more computer-readable storage media, which may be non-transitory. The memory 602 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in memory 602 is used to store at least one instruction for execution by processor 601 to implement the method of live tutoring provided by method embodiments herein.
In some embodiments, the terminal 600 may further optionally include: a peripheral interface 603 and at least one peripheral. The processor 601, memory 602, and peripheral interface 603 may be connected by buses or signal lines. Various peripheral devices may be connected to the peripheral interface 603 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of a radio frequency circuit 604, a touch screen display 605, a camera 606, an audio circuit 607, a positioning component 608, and a power supply 609.
The peripheral interface 603 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 601 and the memory 602. In some embodiments, the processor 601, memory 602, and peripheral interface 603 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 601, the memory 602, and the peripheral interface 603 may be implemented on a separate chip or circuit board, which is not limited in this embodiment.
The Radio Frequency circuit 604 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 604 communicates with communication networks and other communication devices via electromagnetic signals. The rf circuit 604 converts an electrical signal into an electromagnetic signal to transmit, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 604 comprises: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuitry 604 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: metropolitan area networks, various generation mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the rf circuit 604 may further include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display 605 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 605 is a touch display screen, the display screen 605 also has the ability to capture touch signals on or over the surface of the display screen 605. The touch signal may be input to the processor 601 as a control signal for processing. At this point, the display 605 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 605 may be one, providing the front panel of the terminal 600; in other embodiments, the display 605 may be at least two, respectively disposed on different surfaces of the terminal 600 or in a folded design; in still other embodiments, the display 605 may be a flexible display disposed on a curved surface or on a folded surface of the terminal 600. Even more, the display 605 may be arranged in a non-rectangular irregular pattern, i.e., a shaped screen. The Display 605 may be made of LCD (liquid crystal Display), OLED (Organic Light-Emitting Diode), and the like.
The camera assembly 606 is used to capture images or video. Optionally, camera assembly 606 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 606 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
Audio circuitry 607 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 601 for processing or inputting the electric signals to the radio frequency circuit 604 to realize voice communication. For the purpose of stereo sound collection or noise reduction, a plurality of microphones may be provided at different portions of the terminal 600. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 601 or the radio frequency circuit 604 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, audio circuitry 607 may also include a headphone jack.
The positioning component 608 is used to locate the current geographic location of the terminal 600 to implement navigation or LBS (location based Service). The positioning component 608 can be a positioning component based on the GPS (global positioning System) in the united states, the beidou System in china, the graves System in russia, or the galileo System in the european union.
Power supply 609 is used to provide power to the various components in terminal 600. The power supply 609 may be ac, dc, disposable or rechargeable. When the power supply 609 includes a rechargeable battery, the rechargeable battery may support wired or wireless charging. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, the terminal 600 also includes one or more sensors 610. The one or more sensors 610 include, but are not limited to: acceleration sensor 611, gyro sensor 612, pressure sensor 613, fingerprint sensor 614, optical sensor 615, and proximity sensor 616.
The acceleration sensor 611 may detect the magnitude of acceleration in three coordinate axes of the coordinate system established with the terminal 600. For example, the acceleration sensor 611 may be used to detect components of the gravitational acceleration in three coordinate axes. The processor 601 may control the touch screen display 605 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 611. The acceleration sensor 611 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 612 may detect a body direction and a rotation angle of the terminal 600, and the gyro sensor 612 and the acceleration sensor 611 may cooperate to acquire a 3D motion of the user on the terminal 600. The processor 601 may implement the following functions according to the data collected by the gyro sensor 612: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
The pressure sensor 613 may be disposed on a side frame of the terminal 600 and/or on a lower layer of the touch display screen 605. When the pressure sensor 613 is disposed on the side frame of the terminal 600, a user's holding signal of the terminal 600 can be detected, and the processor 601 performs left-right hand recognition or shortcut operation according to the holding signal collected by the pressure sensor 613. When the pressure sensor 613 is disposed at the lower layer of the touch display screen 605, the processor 601 controls the operability control on the UI interface according to the pressure operation of the user on the touch display screen 605. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The fingerprint sensor 614 is used for collecting a fingerprint of a user, and the processor 601 identifies the identity of the user according to the fingerprint collected by the fingerprint sensor 614, or the fingerprint sensor 614 identifies the identity of the user according to the collected fingerprint. Upon identifying that the user's identity is a trusted identity, the processor 601 authorizes the user to perform relevant sensitive operations including unlocking the screen, viewing encrypted information, downloading software, paying, and changing settings, etc. The fingerprint sensor 614 may be disposed on the front, back, or side of the terminal 600. When a physical button or vendor Logo is provided on the terminal 600, the fingerprint sensor 614 may be integrated with the physical button or vendor Logo.
The optical sensor 615 is used to collect the ambient light intensity. In one embodiment, processor 601 may control the display brightness of touch display 605 based on the ambient light intensity collected by optical sensor 615. Specifically, when the ambient light intensity is high, the display brightness of the touch display screen 605 is increased; when the ambient light intensity is low, the display brightness of the touch display screen 605 is turned down. In another embodiment, the processor 601 may also dynamically adjust the shooting parameters of the camera assembly 606 according to the ambient light intensity collected by the optical sensor 615.
A proximity sensor 616, also known as a distance sensor, is typically disposed on the front panel of the terminal 600. The proximity sensor 616 is used to collect the distance between the user and the front surface of the terminal 600. In one embodiment, when the proximity sensor 616 detects that the distance between the user and the front surface of the terminal 600 gradually decreases, the processor 601 controls the touch display 605 to switch from the bright screen state to the dark screen state; when the proximity sensor 616 detects that the distance between the user and the front surface of the terminal 600 gradually becomes larger, the processor 601 controls the touch display 605 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 6 is not intended to be limiting of terminal 600 and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be used.
In an exemplary embodiment, a computer-readable storage medium, such as a memory, including instructions executable by a processor in a terminal to perform the method of live teaching in the embodiments described below, is also provided. For example, the computer readable storage medium may be a ROM, a Random Access Memory (RAM), a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and the like.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only exemplary of the present application and should not be taken as limiting, as any modification, equivalent replacement, or improvement made within the spirit and principle of the present application should be included in the protection scope of the present application.