Detailed Description
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present application. The present application may be embodied in many other forms than those herein described, and those skilled in the art will readily appreciate that the present application may be similarly embodied without departing from the spirit or essential characteristics thereof, and therefore the present application is not limited to the specific embodiments disclosed below.
The terminology used in the one or more embodiments of the application is for the purpose of describing particular embodiments only and is not intended to be limiting of the one or more embodiments of the application. As used in one or more embodiments of the application and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that the term "and/or" as used in one or more embodiments of the present application refers to and encompasses any or all possible combinations of one or more of the associated listed items.
It should be understood that, although the terms first, second, etc. may be used in one or more embodiments of the application to describe various information, these information should not be limited by these terms. These terms are only used to distinguish one type of information from another. For example, a first may also be referred to as a second, and similarly, a second may also be referred to as a first, without departing from the scope of one or more embodiments of the application. The term "if" as used herein may be interpreted as "at..once" or "when..once" or "in response to a determination", depending on the context.
First, terms related to one or more embodiments of the present application will be explained.
Dynamic capture, recording the limb movements of an observed object (human or animal) by various technical means.
Facial capture is to record the facial actions of an observed object (human or animal) through various technical means.
At present, live broadcasting is mainly performed in a virtual scene by utilizing a dynamic capturing/surface capturing technology, and a diversified virtual image driving mode is lacked.
According to the scheme, more diversified virtual image driving and scene interaction are realized through automatic switching and flexible configuration of the dynamic capturing and surface capturing modes and the operation of control equipment such as a keyboard, a mouse and the like.
In the present application, a live view display method is provided, and the present application relates to a live view display apparatus, a computing device, and a computer-readable storage medium, which are described in detail in the following embodiments one by one.
Fig. 1 shows a scene schematic diagram of a live broadcast picture display method according to an embodiment of the present application, which is specifically as follows:
It should be noted that, the live broadcast picture display method of the present application can be applied to a multicast live broadcast scene, and each anchor can control virtual roles in a virtual scene through an anchor client. The multicast live broadcast scene refers to a scene in which a plurality of anchor can interact in the same virtual scene.
The target live broadcasting room is started, wherein the target live broadcasting room can display virtual roles of the target live broadcasting room, and the target live broadcasting room can contain control modes of three virtual roles:
The control mode 1 is a device-only control mode, if the target anchor uses a computer terminal to carry out live broadcast, the virtual role can be controlled based on control devices such as a mouse and a keyboard, for example, movement of the virtual role is realized based on W, A, S, D keys in the keyboard, mouse sliding changes the visual angle of the virtual role in a virtual scene, space taking off, shift fast running, C key squatting, F object picking interaction and the like, if the target anchor uses a mobile phone terminal to carry out live broadcast, the virtual role can be controlled based on virtual devices such as a virtual rocker, a virtual auxiliary rocker and an icon button, the virtual role control can comprise control of the virtual role, such as forward movement, jumping and the like of the virtual role, and control of interaction of the virtual role with virtual objects in the virtual scene, such as interaction with other virtual roles in the virtual scene, interaction with props in the virtual scene and the like.
The control mode 2 is an action-only control mode, and can be started if the target anchor uses a client comprising the image acquisition equipment, in the mode, the image acquisition equipment acquires an anchor image of the target anchor, analyzes the anchor image to obtain virtual driving data, and controls the virtual role based on the virtual driving data.
The control mode 3 is a semi-automatic control mode, if the default mode of the target live broadcasting room of the target live broadcasting is a device control mode and the action control mode is started, the virtual roles can be switched and controlled between the device control mode and the action control mode based on the requirement of the target live broadcasting in the live broadcasting, so that the control modes of the virtual roles in a live broadcasting scene are enriched, namely, part of the virtual roles can be driven by dynamic capture data without controlling the operation of the device, a mode switching button is included in a client interface of a host broadcasting, and the action control mode can be started or closed through the touch control mode switching button.
Specifically, if the current target anchor operates the virtual character based on the control device, the dynamic capture data of the target anchor can be automatically collected for controlling the virtual character under the condition that the target anchor stops operating, if the target anchor can directly operate the virtual character based on the control device under the condition that the virtual character is controlled based on the action control mode, the virtual character can be directly switched into the device control mode by the action control mode, further, the control can be performed by adopting different control modes according to different virtual character parts, for example, the face of the virtual character is controlled to move by adopting a face capture control mode, the upper body of the virtual character is controlled to move by adopting a device control mode, and the like.
According to the method and the device for displaying the virtual character, under the condition that the initial virtual picture generated based on the equipment control mode is displayed in the target live broadcasting room, the target virtual picture can be displayed based on the control model switching request, so that the control modes of the virtual character are enriched, the switching of different control modes of the virtual character is achieved, the driving data of the virtual character is acquired according to the character part type information, the control modes of different parts of the virtual character are increased, and the control modes of the virtual character are further enriched.
Fig. 2a shows a flowchart of a live broadcast picture display method according to an embodiment of the present application, which specifically includes the following steps:
And 202, displaying an initial virtual picture generated by the target virtual character based on the first control mode in the target live broadcasting room.
The target live broadcasting room is a live broadcasting room created by a live broadcasting platform based on a live broadcasting request of a target live broadcasting, the target virtual role is a virtual role corresponding to the target live broadcasting which can be displayed in the target live broadcasting room, the first control mode is one of virtual role control modes which can be adopted by the target live broadcasting room, in practical application, the first control mode can be a device control mode, an action control mode and a semi-automatic control mode, the device control mode is a mode for controlling the virtual role based on device control data triggered by the target live broadcasting, in the control mode, the target live broadcasting can operate the target virtual role based on control devices, the control devices can be entity devices such as a mouse, a keyboard and a microphone, and can also be virtual devices such as virtual rockers and virtual buttons, and the application is not limited in detail, the action control mode is a mode for controlling the virtual role based on the action data of the target live broadcasting, the semi-automatic control mode is a mode for controlling the virtual role based on the action data and/or the device control data corresponding to the target live broadcasting, and the initial virtual picture is a picture generated by the target live broadcasting through controlling the virtual role.
Specifically, a target virtual role corresponding to a target anchor in a target living room is displayed in the target living room, and the target virtual role can be controlled based on a first control mode to generate an initial virtual picture.
And 204, receiving a control mode switching request generated based on a preset triggering condition, wherein the control mode switching request comprises switching information.
The preset triggering condition refers to a condition capable of triggering a control mode of a target live broadcasting room to switch, specifically, the preset triggering condition can refer to a host broadcasting triggering condition, a control device triggering condition or a role part triggering condition, the host broadcasting triggering condition refers to a condition that a target host broadcasting in the target live broadcasting room can generate a mode switching request through triggering a mode switching control of the target live broadcasting room, the control device triggering condition refers to a control mode corresponding to a virtual role, under the condition of non-device control mode, the target host broadcasting generates the mode switching request through triggering the control device, and the role part triggering condition refers to a condition that the mode switching request is triggered according to role part type information of the virtual role.
The control mode switching request refers to a request for switching the control mode of the target virtual character, the control mode switching request can be generated by a target anchor to switch the mode by itself or can be generated under the condition that control information of other control modes is received, and the switching information refers to information which can uniquely represent the second control mode, such as a mode identifier, a mode name and the like of which the switching information is an action control mode.
Specifically, when the target live broadcasting room displays the initial virtual screen, a control mode switching request generated according to a preset trigger condition may be received, for example, a control mode switching request for switching the control mode from the device control mode to the action control mode is generated according to a condition that the anchor starts the action control mode of the target live broadcasting room.
And step 206, switching the first control mode into a second control mode based on the switching information, wherein the first control mode and the second control mode comprise a device control mode or an action control mode or a semi-automatic control mode.
The second control mode refers to a virtual character control mode which can be adopted by the target live broadcasting room and is different from the first control mode, and the second control mode can be a device control mode, an action control mode and a semi-automatic control mode, for example, the first control mode is determined to be the device control mode, the second control mode is determined to be the action control mode and the like.
The method comprises the steps of analyzing a control mode switching request, determining switching information, determining a second control mode according to the switching information, and switching the control mode of a target live broadcasting room from a first control mode to the second control mode.
And step 208, acquiring role data according to the second control mode, and mapping the role data to the target virtual role to obtain a target virtual picture.
The character data is data for driving a target virtual character, for example, in a device control mode, character control information of the target virtual character is used for driving the target virtual character to generate a corresponding virtual picture, in an action control mode, character driving information of the target virtual character is used for driving the target virtual character to generate a corresponding virtual picture, and the target virtual picture is a virtual picture obtained based on the character data and the target virtual character.
Specifically, the role data corresponding to the mode is obtained based on the second control mode, and the role data is mapped to the target virtual role, so that a target virtual picture in the second control mode is obtained.
And 210, displaying the target virtual picture in the target live broadcasting room.
Specifically, after the target virtual picture corresponding to the preset virtual picture or the virtual character driving data is obtained, the preset virtual picture or the target virtual picture is directly displayed, and in practical application, the control modes corresponding to different character parts can be different, so that if the preset virtual picture and the target virtual picture are received, fusion display is required for the preset virtual picture or the target virtual picture.
Further, when the first control mode is a device control mode and the second control mode is an operation control mode, the control mode is switched, and the step of generating the corresponding virtual screen is as follows:
displaying an initial virtual picture generated by a target virtual character based on a device control mode in a target live broadcasting room;
Receiving a control mode switching request, wherein the control mode switching request carries character part type information;
collecting virtual character driving data corresponding to a target anchor according to the character part type information;
Mapping the virtual character driving data to the target virtual character according to the character part type information to obtain a target virtual picture;
And displaying the target virtual picture in the target live broadcasting room.
The target host can set a control mode of the virtual role based on the requirement in the target live broadcasting room, and when the control mode of the target live broadcasting room is set as a device control mode by target host equipment or a default control mode of the target live broadcasting room is set as a device control mode, an initial virtual picture generated based on the device control mode can be displayed in the target live broadcasting room.
In a specific embodiment of the application, a teaching host player starts a virtual classroom live broadcast room and sets a control mode corresponding to a virtual teacher as an equipment control mode, and the teaching host player can control the movement of the virtual teacher in the virtual classroom live broadcast room by moving a mouse, so that the teaching teacher can display virtual animation generated by the virtual teacher through the mouse control in the virtual classroom live broadcast room.
Further, before the initial virtual picture is displayed in the target live broadcasting room, the initial virtual picture needs to be generated based on the role information corresponding to the target virtual role.
Specifically, before the target live room displays the initial virtual picture generated by the target virtual character based on the device control mode, the method may further include:
receiving a role control request aiming at a target virtual role, and determining equipment control information and control part type information based on the role control request;
And mapping the equipment control information to the target virtual role based on the control part type information to obtain an initial virtual picture.
The character control request refers to a request generated by a target anchor for controlling the virtual character through control equipment, the character control request comprises equipment control information and control part type information, the equipment control information refers to control information of the target virtual character, for example, the equipment control information is used for enabling a virtual teacher to lift the right hand, and the control part type information refers to part types of the virtual character which needs to be controlled through the character control request, for example, the upper body, the lower body, the head and the like of the virtual teacher are controlled.
Specifically, after the anchor client receives the role control request for the target virtual role, the role control request is analyzed, the equipment control information and the control part type information are determined, which part of the target virtual role needs to be adjusted based on the control part type information, and the part adjustment form of the target virtual role can be determined based on the equipment control information.
In one embodiment of the application, a host client used by a teaching host receives a virtual character control request for a virtual teacher, analyzes the virtual character control request to obtain equipment control information as a right arm to be lifted by the virtual teacher, obtains control part type information as an arm part, and maps the equipment control information to the arm part of the virtual teacher to obtain an initial virtual picture.
An initial virtual picture is generated based on the device control information and the control part type information, so that a picture for controlling the virtual character based on the control device is displayed in a target anchor in a target living room.
Further, the character part type information refers to character part type information of the target virtual character, for example, head information, hand information, upper body information, and the like.
In practical application, the control mode switching request can be generated by a target anchor through triggering a mode button in a client interface, for example, the target anchor clicks a semi-automatic control mode starting button of the anchor client interface to generate the control mode switching request, the control mode switching request can be switched by itself under the condition that control information of other modes is received, for example, the target anchor receives anchor face driving data acquired by a camera under the condition that the virtual role is controlled based on control equipment, and the control mode switching request can be generated based on the anchor face driving data, so that the current virtual role animation acquired through face capturing in a target living broadcast can be replaced by virtual role face animation generated based on the control information of the control equipment conveniently.
In a specific embodiment of the application, a lecturer anchor starts a semi-automatic control mode in a virtual classroom live broadcasting room, and when a virtual teacher is controlled based on control equipment, if dynamic capture data of the lecturer teacher is received, a control mode switching request is generated based on the dynamic capture data, so that the current equipment control is adjusted to be action control based on the control mode switching request.
The virtual animation is then adjusted based on the control mode switch request by receiving the control mode switch request.
Further, after receiving the control mode switching request, character part type information may be determined based on the control mode switching request, where the character part type information refers to character part correspondence information of the target virtual character, for example, the character part type information is header information, hand information, etc., and data for driving the target virtual character in a new control mode may be collected based on the character part type information.
The virtual character driving data refers to data for driving a target virtual character to generate a virtual picture, for example, the virtual character driving data is pickup item data, blink data, or the like.
The method comprises the steps of analyzing and receiving a control mode switching request, obtaining character part type information, namely determining which parts of a target virtual character need to be subjected to control mode switching, and further collecting virtual character driving data corresponding to a target anchor based on the character part type information, wherein the virtual character driving data are used for driving character parts of the target virtual character.
In an actual application, the method for collecting virtual character driving data corresponding to the target anchor according to the character part type information may include:
Starting target acquisition equipment and acquiring a host image of a target host based on the target acquisition equipment;
And acquiring virtual character driving data in the anchor image based on the character part type information.
The target acquisition device is a device capable of acquiring an image of a target anchor, for example, the target acquisition device is an image acquisition device configured on an anchor client or an image acquisition device in transmission connection with the anchor client, and the anchor image is an image of the target anchor acquired by the target acquisition device, such as an anchor video, an anchor photo, and the like.
Further, the method for acquiring virtual character driving data in the anchor image based on the character part type information may include:
Analyzing the anchor image to obtain anchor action information;
and extracting virtual character driving data from the anchor action information according to the character part type information.
The anchor action information refers to action information of a target anchor analyzed in an anchor image, for example, action information of anchor hands, fist making and the like in anchor video.
The method comprises the steps of acquiring a host image acquired by target acquisition equipment, inputting the host image into an action information acquisition model, acquiring host action information output by the action information acquisition model, determining that a target virtual character is at a target character part based on character part type information, and acquiring host action information corresponding to the target character part in the host action information as virtual character driving data.
In a specific embodiment of the application, an image acquisition device connected with a data transmission of an anchor client is started in response to a control mode switching request, an anchor video of a target anchor acquired by the image acquisition device is acquired, anchor action information in the anchor video is extracted, a target role position of a virtual teacher is determined according to role position type information, and anchor action information corresponding to the target role position is screened in the anchor action information to be used as virtual role driving data.
In an actual application, before the initial virtual picture generated by the target virtual character based on the first control mode is displayed in the target live broadcasting room, the method further comprises the following steps:
detecting the corresponding shooting equipment of the target live broadcasting room and the corresponding starting state of the shooting equipment;
in the case where it is determined that the target image capturing apparatus is in the on state, it is determined that the first control mode is an action control mode.
As shown in fig. 2b, fig. 2b is a schematic flow chart of detecting the image capturing apparatus according to an embodiment of the present application, and the specific detecting method includes steps s1 to s7:
step s1, the target anchor enters the target living room.
And step S2, loading virtual scene information and virtual role information of the target live broadcasting room.
Step s3, detecting whether available image capturing devices and the corresponding quantity of the image capturing devices exist at the anchor client where the target anchor exists, if so, executing step s4, if only one image capturing device exists, executing step s5, and if not, executing step s6.
Step s4, selecting a default image capturing apparatus from the plurality of image capturing apparatuses, and continuing to execute step s5.
And step 5, automatically starting an action control mode in the target live broadcasting room.
Step S6, the action control mode is in a closed state, and the failure of opening the action control mode is prompted.
And step S7, in the action control mode, the virtual roles are controlled based on virtual role driving data acquired by a default camera, so that the target anchor controls the target virtual roles to interact in the virtual scene.
And determining virtual character driving data in the anchor image based on the character part type information so as to drive the target virtual character based on the virtual character driving data.
Further, after the virtual character driving data is determined, the character part of the target virtual character may be driven according to the virtual character driving data, thereby obtaining the target virtual picture.
Wherein, the target virtual picture refers to the virtual picture obtained based on the virtual character data and the target virtual character.
Specifically, after the virtual character driving data corresponding to the character part type information is obtained, the virtual character driving data can be mapped to the virtual character part corresponding to the character part type information based on the character part type information, that is, the virtual character driving data is mapped to the target virtual character, so that the target virtual picture for performing virtual character control based on the virtual character driving data is obtained.
In an actual application, the method for mapping the virtual character driving data to the target virtual character according to the character part type information to obtain the target virtual picture may include:
Determining a target role part of the target virtual role according to the role part type information;
determining character part attribute information corresponding to the target character part;
And adjusting the character part attribute information based on the virtual character driving data to obtain a target virtual picture.
The character part attribute information refers to attribute information corresponding to each target character part, for example, the attribute information corresponding to the arm part is side lifting, downward 90-degree bending and the like.
Specifically, according to the character part type information in the control mode switching request, a target character part, such as a head, a hand and the like, is determined in a target virtual character, a character database corresponding to the target character part is obtained, character part attribute information corresponding to the target character part is screened in the character database, character attribute values, such as a part height value, a part angle value and the like, are contained in the character part attribute information, character driving values contained in virtual character driving data are determined, and the character attribute values are replaced based on the character driving values, so that a target virtual picture is obtained.
In a specific embodiment of the application, a character hand and a character head are determined in a virtual part corresponding to a virtual teacher according to character part type information of the hand and the head, hand attribute values corresponding to the character hand and head attribute values corresponding to the character head are collected in a database corresponding to the virtual teacher, the hand attribute values are replaced according to driving hand attribute values in virtual character data, and head values are replaced according to driving head attribute values, so that a target virtual picture is obtained.
In practical application, each virtual character can be controlled by corresponding character states, such as dancing state, walking state, picking up object state and the like, in the virtual scene, and under the condition of different character states, different character parts can be controlled in different control modes, so that the stability of the virtual picture is ensured.
Specifically, before the virtual character driving data corresponding to the target anchor is collected according to the character part type information, the method may further include:
Acquiring a current role state of the target virtual role;
And determining static part information and dynamic part information in the character part type information based on the current character state.
The method comprises the steps of determining a target virtual character in a target living room, wherein the current character state refers to a character state of the target virtual character in the target living room, such as a dancing state, a flying state and the like, dividing character part type information into static part information and dynamic part information based on the current character state, wherein the static part information refers to a character part of the target virtual character which can be static in the current character state, the dynamic part information refers to a character part of the target virtual character which can be dynamic in the current character state, for example, when the virtual character is in a sitting character state, a lower part of the target virtual character can be a static part, an upper part of the target virtual character can be a dynamic part, a face can be a static part, and an upper part of the target virtual character can be a dynamic part and the like.
In practical application, if the virtual character is in a dynamic condition, when the collected action data or face data of the host is mapped to the virtual character, the stability of the virtual picture is poor, and the watching experience of a user is affected, so that the character part type information can be classified according to the current character state information, and the static part information and the dynamic part information can be determined.
In order to solve the problem of poor stability of the virtual picture, the method and the device respectively process the part corresponding to the static part information and the part corresponding to the dynamic part information, and specifically can drive the part corresponding to the static part information by using action data or facial data of a host, and display the part corresponding to the dynamic part information based on a preset virtual picture.
Specifically, in the case that the character part information includes static part information, the method for collecting virtual character driving data corresponding to the target anchor according to the character part type information may include:
Determining the static part of the target virtual character according to the static part information;
And acquiring virtual character driving data corresponding to the target anchor based on the static part.
The static part refers to a part where the target virtual character can be in a static state in the current character state.
In practical applications, the character part type information may include only static part information, only dynamic part information, or both static part information and dynamic part information, for example, when the current state of the virtual character is a standing state, the character part type information may include only static part information, and when the current state of the virtual character is a dancing state, the character part type information may include only dynamic part information.
In a specific embodiment of the application, the current character state of the virtual teacher is obtained as a standing state, the target character part type information comprises head part information and upper body part information, the head part information and the upper body part information are both determined to be static part information, further, target anchor corresponding to the virtual teacher, namely surface capturing data and dynamic capturing data of the teaching teacher are collected according to each static part information, and virtual character driving data is determined based on the surface capturing data and the dynamic capturing data.
Specifically, in the case that the character part information includes dynamic part information, the method for acquiring virtual character driving data corresponding to the target anchor according to the character part type information may include:
Determining the dynamic part of the target virtual character according to the dynamic part information;
And acquiring a preset virtual character animation corresponding to the target anchor based on the dynamic part.
The dynamic part is a part where the target virtual character is in a dynamic state under the current character state, the preset virtual character animation is a virtual character dynamic corresponding to the dynamic part, and is a virtual character animation generated for the dynamic part in advance, and when the preset virtual character animation is displayed, a target anchor is not required to control.
In a specific embodiment of the application, the current character state of the virtual teacher is obtained as a dancing state, the target character part type information comprises upper body part information and lower body part information, the upper body part information and the lower body part information are both determined to be dynamic part information according to the dancing state, and further, a preset virtual character animation corresponding to each dynamic part information is obtained.
In a preferred embodiment of the present application, it may be determined, based on the preset character status list, that preset virtual character animation and/or virtual character driving data corresponding to character part type information need to be acquired, as shown in the following table 1:
TABLE 1
Determining face information, head information and upper body information as static part information when the virtual character is in a standing state, namely mapping dynamic capture data corresponding to an acquisition target anchor to the virtual character, and determining lower body information as dynamic part information, namely acquiring a preset virtual picture corresponding to lower body information;
determining face information and head information as static part information and upper body information and lower body information as dynamic part information in a walking state of the virtual character
Determining face information, head information and upper body information as static part information and lower body information as dynamic part information when the virtual character is in a sitting state;
when the virtual character is in the object picking state, determining that the face information is static position information, and determining that the head information, the upper body information and the lower body information are dynamic position information, wherein after the object picking state is finished, the object can be held for capturing, namely, the object holding standing state is achieved;
Determining face information as static part information and lower body information, head information and upper body information as dynamic part information when the virtual character is in a hit state, such as displaying facial expression of a host player obtained by face capturing in the flying process;
determining face information as static part information and lower body information, head information and upper body information as dynamic part information when the virtual character is in a dancing state, such as displaying and displaying the facial expression of a host player obtained by face capturing in the dancing process;
In this embodiment, since the target anchor in the living room is normally in the sitting position to perform living broadcast, the motion of the lower body is a preset animation display, and if the lower body needs to be captured in practical application, the table 1 can be adjusted based on the requirement.
It should be noted that after the virtual character driving data corresponding to the static part information is obtained, the virtual character driving data is mapped to the target virtual character according to the static part information to obtain the target virtual picture, and after the preset virtual picture corresponding to the dynamic part information is obtained, the preset virtual picture can be directly displayed in the target live broadcasting room.
Further, in addition to determining whether the motion control mode is used for the diagonal portion based on the current character state, a portion mode switching control, such as a manual capturing control, a head capturing control and the like, can be configured on the anchor client interface, and the target anchor can start or close the motion control mode of the corresponding portion based on the portion mode switching control, such as starting the hand capturing, closing the head capturing and displaying the preset virtual animation corresponding to the head.
The above scheme describes the step of switching from the device control mode to the action control mode to control the virtual character, in practical application, the virtual character can also be controlled by switching from the action control mode to the device control mode, so that the control method of the virtual character is enriched, and the target anchor can conveniently control the virtual character.
Specifically, the method further comprises the following steps:
displaying a first virtual picture generated by a target virtual character based on an action control mode in the target live broadcasting room;
Receiving a control mode switching request, wherein the control mode switching request carries character part type information and character control information;
mapping the role control information to the target virtual role based on the role part type information to obtain a second virtual picture;
and displaying the second virtual picture in the target live broadcasting room.
The first virtual picture is a picture generated by operating the virtual character through dynamic data capture, the character control information is control information sent by the control equipment, for example, the character control information is to move the target virtual character forward by two steps, and the second virtual picture is a picture generated by operating the virtual character based on the character part type information and the character control information.
In practical application, the control mode switching request can be triggered by a target anchor or generated after role information is received, namely, under the condition that the control mode is in a semi-automatic mode, the control mode can be switched from an action control mode to a device control mode through the received role information, or the control mode can be switched from the device control mode to the action control mode through the received action capturing data, or different parts of a target virtual role are controlled by the action control mode and the device control mode at the same time, so that virtual role control modes are enriched, and the use experience of a user is improved.
After the anchor client generates the target virtual picture or the second virtual picture, the target virtual picture or the second virtual picture can be pushed to the server in the form of live broadcast stream, and the audience client entering the target live broadcast room pulls the direct broadcast stream in the server, so that the watching of the content in the target live broadcast room is realized.
When the control mode of the target virtual character controlled by the target live broadcasting room is an action control mode, the control can be further performed by switching the action control mode to a semiautomatic control mode, specifically, when the first control mode is the action control mode and the second control mode is the semiautomatic control mode, an initial virtual picture generated by the target virtual character based on the first control mode is displayed in the target live broadcasting room, and the method comprises the following steps:
displaying a first virtual picture generated by a target virtual character based on an action control mode in the target live broadcasting room;
Receiving a control mode switching request, wherein the control mode switching request carries target role part type information and target role control information;
determining other role part type information corresponding to the target virtual role based on the target role part type information;
Collecting target virtual character driving data corresponding to a target anchor according to the other character part type information;
Mapping the role control information to a target role part of the target virtual role based on the target role part type information, and mapping the target virtual role driving data to other role parts of the target virtual role based on the other role part type information, so as to obtain a second virtual picture;
and displaying the second virtual picture in the target live broadcasting room.
Specifically, when the control mode is in the motion control mode, the received control mode switching request is analyzed, and target character position type information is determined, wherein the target character position type information refers to type information of a target character position of a target virtual character, for example, when a head of the target virtual character is determined to be the target position, the type information corresponding to the head is determined to be the target character position type information, the target character position corresponding to the target character position type information is determined to be the type information of other character positions of the target virtual character to be controlled by the control device, and other character position type information of other character positions of the target virtual character is determined according to the target character position type information, wherein the other character position type information refers to type information corresponding to other character positions of the target virtual character except the target character position, for example, when the target character position of the target virtual character is determined to be the head, the other character positions are determined to be the upper body and the lower body, namely, the type information corresponding to the upper body and the lower body are the type information of other character position types.
After the target role part and other role parts are determined, role control information is mapped to the target role part of the target virtual role according to the target role part type information, target virtual role driving data is acquired according to the other role part type information, and the target virtual role driving data is mapped to the other role parts of the target virtual role to generate a second virtual picture, so that the target virtual role can be controlled simultaneously by the action control mode and the equipment control mode in the semi-automatic control mode.
In a specific embodiment of the application, a live broadcast picture for controlling a virtual teacher based on an action control mode is displayed in a teaching live broadcast room, a control mode switching request for the teaching live broadcast room is received, the control mode switching request comprises upper body role type information and corresponding role control information, other head type information of the virtual teacher is determined according to the upper body role part type information, target virtual role driving data of a target anchor, namely, capturing data of the target anchor, is acquired according to the head type information, the role control data is mapped to the upper body part of the virtual teacher, the capturing data is mapped to the face part of the virtual teacher, and the lower body part of the virtual teacher is displayed by a preset virtual animation, so that the control mode of the virtual teacher is switched from the action-only control mode to the semiautomatic control mode.
The live broadcast picture display method includes the steps of displaying an initial virtual picture generated by a target virtual role based on a first control mode in a target live broadcast room, receiving a control mode switching request generated based on a preset triggering condition, wherein the control mode switching request comprises switching information, switching the first control mode into a second control mode based on the switching information, wherein the first control mode and the second control mode comprise a device control mode or an action control mode or a semi-automatic control mode, acquiring role data according to the second control mode, mapping the role data to the target virtual role to obtain the target virtual picture, and displaying the target virtual picture in the target live broadcast room.
According to the embodiment of the application, under the condition that the initial virtual picture generated based on the first control mode is displayed in the target live broadcasting room, the target virtual picture can be switched to the second control mode based on the control mode switching request, and the target virtual picture in the second control mode is displayed, so that the control mode of the virtual character is enriched.
The live broadcast picture display method provided by the application is further described below by taking the application of the live broadcast picture display method in a game scene as an example in combination with fig. 3. Fig. 3 shows a process flow chart of a live broadcast picture display method applied to a game scene according to an embodiment of the present application, which specifically includes the following steps:
step 302, determining a target virtual role corresponding to a target anchor in a target game live room.
Step 304, receiving a role control request for the target virtual role, and determining device control information and control part type information based on the role control request.
And 306, mapping the equipment control information to the target virtual roles based on the control part type information to obtain an initial virtual picture.
And 308, displaying an initial virtual picture of the target virtual character in the target live broadcasting room.
Step 310, a first control mode switching request is received, wherein the first control mode switching request carries character part type information.
Step 312, responding to the first control mode switching request, starting the target acquisition equipment, and acquiring the anchor image of the target anchor based on the target acquisition equipment.
Step 314, obtaining the anchor action information in the anchor image, and extracting the virtual character driving data from the anchor action information according to the character part type information.
Step 316, obtaining the current character state of the target virtual character, and determining the static part information and the dynamic part information in the character part type information based on the current character state.
Step 318, determining virtual character driving data corresponding to the target anchor according to the static position information, and determining preset virtual character animation corresponding to the target anchor according to the dynamic position information.
And 320, determining the static part of the target virtual character according to the static part information, and determining the part attribute information corresponding to the static part.
And 322, adjusting character part attribute information based on the virtual character driving data to obtain a virtual picture.
And 324, fusing the virtual picture and the preset virtual character animation to obtain a target virtual picture, and displaying the target virtual picture in the target game live room.
And 326, receiving a second control mode switching request, wherein the second control mode switching request carries the character part type information and the character control information.
Step 328, mapping the character control information to the target virtual character based on the character part type information, and obtaining a second virtual picture.
And 330, displaying a second virtual picture in the target game live room.
The live broadcast picture display method includes the steps of displaying an initial virtual picture generated by a target virtual role based on a device control mode in a target live broadcast room, receiving a control mode switching request, wherein the control mode switching request carries role part type information, acquiring virtual role driving data corresponding to a target host according to the role part type information, mapping the virtual role driving data to the target virtual role according to the role part type information to obtain the target virtual picture, and displaying the target virtual picture in the target live broadcast room.
According to the method and the device for displaying the virtual character, under the condition that the initial virtual picture generated based on the equipment control mode is displayed in the target live broadcasting room, the target virtual picture can be displayed based on the control model switching request, so that the control modes of the virtual character are enriched, the switching of different control modes of the virtual character is achieved, the driving data of the virtual character is acquired according to the character part type information, the control modes of different parts of the virtual character are increased, and the control modes of the virtual character are further enriched.
Corresponding to the method embodiment, the application also provides an embodiment of a live broadcast picture display device, and fig. 4 shows a schematic structural diagram of the live broadcast picture display device according to an embodiment of the application. As shown in fig. 4, the apparatus includes:
a first presentation module 402 configured to present, at a target living room, an initial virtual picture generated by a target virtual character based on a first control mode;
A receiving module 404, configured to receive a control mode switching request generated based on a preset trigger condition, where the control mode switching request includes switching information;
An acquisition module 406 configured to switch the first control mode to a second control mode based on the switching information, wherein the first control mode and the second control mode include a device control mode or an action control mode or a semiautomatic control mode;
a mapping module 408 configured to obtain character data according to the second control mode, and map the character data to the target virtual character to obtain a target virtual picture;
A second presentation module 410 configured to present the target virtual picture in the target live room.
Optionally, the preset trigger conditions comprise a main broadcasting trigger condition, a control device trigger condition and a role part trigger condition.
Optionally, the apparatus further comprises an action control sub-module configured to:
displaying an initial virtual picture generated by a target virtual character based on a device control mode in a target live broadcasting room;
Receiving a control mode switching request, wherein the control mode switching request carries character part type information;
collecting virtual character driving data corresponding to a target anchor according to the character part type information;
Mapping the virtual character driving data to the target virtual character according to the character part type information to obtain a target virtual picture;
And displaying the target virtual picture in the target live broadcasting room.
Optionally, the apparatus further comprises a device control sub-module configured to:
displaying a first virtual picture generated by a target virtual character based on an action control mode in the target live broadcasting room;
Receiving a control mode switching request, wherein the control mode switching request carries character part type information and character control information;
mapping the role control information to the target virtual role based on the role part type information to obtain a second virtual picture;
and displaying the second virtual picture in the target live broadcasting room.
Optionally, the apparatus further comprises a semi-automatic control sub-module configured to:
displaying a first virtual picture generated by a target virtual character based on an action control mode in the target live broadcasting room;
Receiving a control mode switching request, wherein the control mode switching request carries target role part type information and target role control information;
determining other role part type information corresponding to the target virtual role based on the target role part type information;
Collecting target virtual character driving data corresponding to a target anchor according to the other character part type information;
Mapping the role control information to a target role part of the target virtual role based on the target role part type information, and mapping the target virtual role driving data to other role parts of the target virtual role based on the other role part type information, so as to obtain a second virtual picture;
and displaying the second virtual picture in the target live broadcasting room.
Optionally, the apparatus further comprises a mapping sub-module configured to:
receiving a role control request aiming at a target virtual role, and determining equipment control information and control part type information based on the role control request;
And mapping the equipment control information to the target virtual role based on the control part type information to obtain an initial virtual picture.
Optionally, the acquisition module 406 is further configured to:
Starting target acquisition equipment and acquiring a host image of a target host based on the target acquisition equipment;
And acquiring virtual character driving data in the anchor image based on the character part type information.
Optionally, the acquisition module 406 is further configured to:
Analyzing the anchor image to obtain anchor action information;
and extracting virtual character driving data from the anchor action information according to the character part type information.
Optionally, the mapping module 408 is further configured to:
Determining a target role part of the target virtual role according to the role part type information;
determining character part attribute information corresponding to the target character part;
And adjusting the character part attribute information based on the virtual character driving data to obtain a target virtual picture.
Optionally, the apparatus further comprises an acquisition sub-module configured to:
Acquiring a current role state of the target virtual role;
And determining static part information and dynamic part information in the character part type information based on the current character state.
Optionally, the acquisition module 406 is further configured to:
Determining the static part of the target virtual character according to the static part information;
And acquiring virtual character driving data corresponding to the target anchor based on the static part.
Optionally, the acquisition module 406 is further configured to:
Determining the dynamic part of the target virtual character according to the dynamic part information;
And acquiring a preset virtual character animation corresponding to the target anchor based on the dynamic part.
Optionally, the apparatus further comprises a detection module configured to:
detecting the corresponding shooting equipment of the target live broadcasting room and the corresponding starting state of the shooting equipment;
in the case where it is determined that the target image capturing apparatus is in the on state, it is determined that the first control mode is an action control mode.
The live broadcast picture display method includes the steps of displaying an initial virtual picture generated by a target virtual role based on a first control mode in a target live broadcast room, receiving a control mode switching request generated based on a preset triggering condition, wherein the control mode switching request comprises switching information, switching the first control mode into a second control mode based on the switching information, wherein the first control mode and the second control mode comprise a device control mode or an action control mode or a semi-automatic control mode, acquiring role data according to the second control mode, mapping the role data to the target virtual role to obtain the target virtual picture, and displaying the target virtual picture in the target live broadcast room.
According to the embodiment of the application, under the condition that the initial virtual picture generated based on the first control mode is displayed in the target live broadcasting room, the target virtual picture can be switched to the second control mode based on the control mode switching request, and the target virtual picture in the second control mode is displayed, so that the control mode of the virtual character is enriched.
The above is a schematic solution of a live broadcast picture display device of this embodiment. It should be noted that, the technical solution of the live broadcast picture display device and the technical solution of the live broadcast picture display method belong to the same concept, and details of the technical solution of the live broadcast picture display device which are not described in detail can be referred to the description of the technical solution of the live broadcast picture display method.
Fig. 5 illustrates a block diagram of a computing device 500, provided in accordance with an embodiment of the present application. The components of the computing device 500 include, but are not limited to, a memory 510 and a processor 520. Processor 520 is coupled to memory 510 via bus 530 and database 550 is used to hold data.
Computing device 500 also includes access device 540, access device 540 enabling computing device 500 to communicate via one or more networks 560. Examples of such networks include public switched telephone networks (PSTN, public Switched Telephone Network), local area networks (LAN, local Area Network), wide area networks (WAN, wide Area Network), personal area networks (PAN, personal Area Network), or combinations of communication networks such as the internet. The access device 540 may include one or more of any type of network interface, wired or wireless, such as a network interface card (NIC, network interface controller), such as an IEEE802.11 wireless local area network (WLAN, wireless Local Area Network) wireless interface, a worldwide interoperability for microwave access (Wi-MAX, worldwide Interoperability for Microwave Access) interface, an ethernet interface, a universal serial bus (USB, universal Serial Bus) interface, a cellular network interface, a bluetooth interface, a Near Field Communication (NFC) interface, and so forth.
In one embodiment of the application, the above-described components of computing device 500, as well as other components not shown in FIG. 5, may also be connected to each other, such as by a bus. It should be understood that the block diagram of the computing device illustrated in FIG. 5 is for exemplary purposes only and is not intended to limit the scope of the present application. Those skilled in the art may add or replace other components as desired.
Computing device 500 may be any type of stationary or mobile computing device, including a mobile computer or mobile computing device (e.g., tablet, personal digital assistant, laptop, notebook, netbook, etc.), mobile phone (e.g., smart phone), wearable computing device (e.g., smart watch, smart glasses, etc.), or other type of mobile device, or a stationary computing device such as a desktop computer or personal computer (PC, personal Computer). Computing device 500 may also be a mobile or stationary server.
Wherein, the processor 520 implements the steps of the live view display method when executing the computer instructions.
The foregoing is a schematic illustration of a computing device of this embodiment. It should be noted that, the technical solution of the computing device and the technical solution of the live broadcast picture display method belong to the same concept, and details of the technical solution of the computing device, which are not described in detail, can be referred to the description of the technical solution of the live broadcast picture display method.
An embodiment of the present application also provides a computer readable storage medium storing computer instructions that, when executed by a processor, implement the steps of the live view presentation method as described above.
The above is an exemplary version of a computer-readable storage medium of the present embodiment. It should be noted that, the technical solution of the storage medium and the technical solution of the live broadcast picture display method belong to the same concept, and details of the technical solution of the storage medium which are not described in detail can be referred to the description of the technical solution of the live broadcast picture display method.
The foregoing describes certain embodiments of the present application. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the claims can be performed in a different order than in the embodiments and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.
The computer instructions include computer program code that may be in source code form, object code form, executable file or some intermediate form, etc. The computer readable medium may include any entity or device capable of carrying the computer program code, a recording medium, a U disk, a removable hard disk, a magnetic disk, an optical disk, a computer Memory, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), an electrical carrier signal, a telecommunications signal, a software distribution medium, and so forth. It should be noted that the computer readable medium contains content that can be appropriately scaled according to the requirements of jurisdictions in which such content is subject to legislation and patent practice, such as in certain jurisdictions in which such content is subject to legislation and patent practice, the computer readable medium does not include electrical carrier signals and telecommunication signals.
It should be noted that, for the sake of simplicity of description, the foregoing method embodiments are all expressed as a series of combinations of actions, but it should be understood by those skilled in the art that the present application is not limited by the order of actions described, as some steps may be performed in other order or simultaneously in accordance with the present application. Further, those skilled in the art will appreciate that the embodiments described in the specification are all preferred embodiments, and that the acts and modules referred to are not necessarily all required for the present application.
In the foregoing embodiments, the descriptions of the embodiments are emphasized, and for parts of one embodiment that are not described in detail, reference may be made to the related descriptions of other embodiments.
The preferred embodiments of the application disclosed above are intended only to assist in the explanation of the application. Alternative embodiments are not intended to be exhaustive or to limit the application to the precise form disclosed. Obviously, many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the application and the practical application, to thereby enable others skilled in the art to best understand and utilize the application. The application is limited only by the claims and the full scope and equivalents thereof.