Detailed Description
The spirit of the present disclosure will be apparent from the accompanying drawings and detailed description, and any person skilled in the art, who can understand the embodiments of the present disclosure, can make changes and modifications from the technology taught by the present disclosure without departing from the spirit and scope of the present disclosure.
As used herein, "coupled" may mean that two or more elements are in direct physical or electrical contact with each other, or in indirect physical or electrical contact with each other, and "coupled" may mean that two or more elements are in mutual operation or action.
As used herein, the terms "first," "second," …, etc. do not denote any order or order, nor are they used to limit the invention, but rather are used to distinguish one element from another element or operation described by the same technical terms.
As used herein, the terms "comprising," "including," "having," "containing," and the like are open-ended terms that mean including, but not limited to.
As used herein, "and/or" includes any and all combinations of the described items.
With respect to the terms (terms) used herein, it is common, unless otherwise noted, to have the ordinary meaning of each term used in the art, in the disclosure herein, and in the specific context. Certain terms used to describe the present disclosure will be discussed below or elsewhere in this specification to provide additional guidance to those skilled in the art in describing the present disclosure.
As used herein, the terms "substantially," "about," and the like are generally used to refer to any value or range that is close to the recited value or range, which may vary according to the particular art involved, and which is to be accorded the widest interpretation so as to encompass all such modifications and similar structures as would be appreciated by those skilled in the art. In some embodiments, the range of slight variations or errors that such terms modify is 20%, in some preferred embodiments 10%, and in some more preferred embodiments 5%. In addition, all numerical values mentioned herein mean approximate values, and in the case where no additional description is made, the words "substantially" and "about" are intended to be implied.
An example of the present disclosure relates to an electronic device. For clarity of explanation, in the following paragraphs, details of the electronic device will be described by taking a live box as an example. However, other electronic devices, such as tablet computers, smart phones, desktop computers, etc., are also within the scope of the present disclosure.
Fig. 1 is a schematic diagram of an electronic device 100 according to an embodiment of the disclosure. In the present embodiment, the electronic device 100 is electrically connected to the eye tracker 20 and the one or more photographing devices 30. In the present embodiment, the electronic device 100 is configured to receive the eyeball tracking data from the eyeball tracker 20. In the present embodiment, the electronic device 100 is configured to receive one or more pictures from one or more cameras 30.
In the present embodiment, the electronic device 100 includes one or more processing elements 110 and a memory 120. In the present embodiment, the one or more processing elements 110 are electrically connected to the memory 120. In fig. 1 and the following description, only one processing element 110 is taken as an example, and the disclosure is not limited thereto.
In one embodiment, the processing element 110 can be implemented by, but is not limited to, a central processing unit and/or a processor such as a microprocessor. In one embodiment, the memory 120 may include one or more memory devices, where each memory device or set of memory devices includes a computer-readable recording medium. The memory 120 may include a read-only memory, a flash memory, a floppy disk, a hard disk, a compact disk, a flash disk, a magnetic tape, a database accessible via a network, or a computer-readable recording medium with the same functions as those easily recognized by those skilled in the art.
In one embodiment, the processing element 110 may execute or execute various software programs and/or sets of instructions stored in the memory 120 to perform various functions of the electronic device 100.
It should be noted that the implementation manner of the components in the electronic device 100 is not limited to the embodiments disclosed above, and the connection relationship is not limited to the embodiments, and all the connection manners and implementation manners that are sufficient for the electronic device 100 to implement the following technical contents can be applied to the present invention.
In one embodiment, the processing element 110 can obtain one or more photographic frames through one or more photographic devices 30. The processing element 110 can map the photographed image to a system coordinate system and obtain the image boundary of the photographed image in the system coordinate system.
In addition, the processing element 110 can obtain a gaze coordinate in the system coordinate system according to the eye tracking data from the eye tracker 20, wherein the gaze coordinate corresponds to the gazing position of the user. In one embodiment, the processing element 110 is capable of analyzing and calculating eye tracking data to obtain eye gaze coordinates.
In the case that the sight line coordinate is located within the frame boundary of the one or more frames in the system coordinate system, the processing element 110 may determine a local output frame corresponding to the user's gaze location in the one or more frames. In an embodiment, the processing element 110 may output the local output frame to an output object device (e.g., the output object device 600 in fig. 6), such as a live broadcast server, an image capture box, a personal computer, but not limited thereto.
By such operation, the live broadcast person can change the live broadcast picture seen by the viewer by using the sight line, and further the interaction between the live broadcast person and the viewer can be increased.
Wherein the system coordinate system may be a desktop coordinate system or other similar coordinate system. The system coordinate system is, for example, a coordinate system based on a desktop on which the target OBJ and the reference points NA, NB, and NC shown in fig. 3 are located. For example, a first direction edge of the table top may be aligned with an x-axis of a system coordinate system and a second direction edge of the table top may be aligned with a y-axis of the system coordinate system. It should be noted that the above-mentioned embodiments are only examples, and the disclosure is not limited thereto.
The output frame determining method shown in fig. 2 is combined to provide more detailed details of the present disclosure, but the present disclosure is not limited to the following embodiments.
It should be noted that the output frame determining method can be applied to an electronic device having the same or similar structure as that shown in FIG. 1. For simplicity, the method for determining the output screen will be described below by taking the electronic device 100 in fig. 1 as an example according to an embodiment of the present invention, but the present invention is not limited to this application.
In addition, the output image determining method can also be implemented as a computer program and stored in a non-transitory computer readable recording medium, so that the computer or the electronic device executes the virtual reality method after reading the recording medium. The non-transitory computer readable recording medium can be a read-only memory, a flash memory, a floppy disk, a hard disk, a compact disk, a portable disk, a magnetic tape, a database accessible through a network, or a non-transitory computer readable recording medium with the same functions as those easily recognized by those skilled in the art.
In addition, it should be understood that the operations of the output frame determining method in the present embodiment, except for the specific description of the sequence, can be performed simultaneously or partially simultaneously by adjusting the sequence according to the actual requirement.
Moreover, such operations may be adaptively added, replaced, and/or omitted in various embodiments.
Referring to fig. 1 and 2, the output frame determining method 200 includes the following operations.
In operation S1, the processing element 110 obtains one or more photographic frames. In one embodiment, the one or more pictures are from one or more cameras 30, but the disclosure is not limited thereto.
In operation S2, the processing element 110 obtains frame coordinates of the one or more frames in a system coordinate system. In an embodiment, the origin and the coordinate axis of the system coordinate system may be a predetermined origin and coordinate axis, but not limited thereto. In one embodiment, the frame coordinates may correspond to a vertex, a boundary, a center, and/or other reference positions of the one or more frames, but not limited thereto. In one embodiment, the processing element 110 utilizes a plurality of reference points to obtain the frame coordinates.
For example, referring to fig. 3 and 4, in the present example, reference points NA, NB, and NC exist in the photographing screen CMI. As shown in fig. 4, the reference points NA, NB, and NC have their coordinates (hereinafter referred to as reference coordinates) in the system coordinate system SYC. In some embodiments, the reference coordinate may be a preset value. In some embodiments, the reference coordinates may be pre-stored in the electronic device 100, and the user may set the reference points NA, NB, and NC according to a preset instruction when the environment is initialized. It should be noted that the above description is only exemplary and the disclosure is not limited to the above embodiments.
Further, it is assumed that coordinates of the reference points NA, NB, and NC in the photographing screen CMI are (1, 3), (2, 4), and (3, 6), and coordinates of the reference points NA, NB, and NC in the system coordinate system SYC may be (-1, -3), (10, 20), (30, 40). Further, let the coordinates of the vertices of the photographic frame CMI in the photographic frame CMI be (0, 0), (50, 0), (0, 50), (50, 50), and the coordinates of these vertices in the system coordinate system SYC may be (-15 ), (-15, 100), (100, -15), (100 ).
In the present embodiment, the processing element 110 obtains the reference coordinates of the reference points NA, NB, NC in the system coordinate system SYC (operation S21 in fig. 7). On the other hand, the processing element 110 obtains respective positions of the reference points NA, NB, and NC in the photographing screen CMI (see fig. 3) (operation S22 in fig. 7). Then, the processing element 110 establishes a correspondence relationship (hereinafter referred to as a second correspondence relationship) between the photographing screen CMI and the system coordinate system SYC based on the reference coordinates of the reference points NA, NB, and NC in the system coordinate system SYC and the positions of the reference points NA, NB, and NC in the photographing screen CMI (fig. 7 operation S23). According to the second corresponding relationship, the processing element 110 may generate frame coordinates of the vertex, the boundary, the center, and/or other reference points of the photographed frame CMI in the system coordinate system SYC (fig. 7 operation S24). In an embodiment, the second correspondence relationship may be implemented by a transformation matrix, a look-up table, a mathematical function, or any other feasible manner, which is not limited by the disclosure.
For example, in other embodiments, there may be other cameras (i.e., there are a plurality of cameras) besides the camera CMI, the processing element 110 may establish a corresponding relationship (all referred to as a second corresponding relationship) between each camera and the system coordinate system SYC, and generate the frame coordinates of the vertex, the boundary, the center, and/or other reference points of each camera in the system coordinate system SYC according to the second corresponding relationship.
In this way, the processing element 110 may also obtain the frame coordinates of other photographed frames (if any) in the system coordinate system SYC.
In operation S3, the processing element 110 determines one or more frame boundaries CNT of the one or more frames in the system coordinate system SYC according to the frame coordinates of the one or more frames.
For example, referring to fig. 4, the processing element 110 may determine the frame boundary CNT of the photographic frame CMI in the system coordinate system SYC according to the frame coordinates (in the embodiment, the system coordinate system may be a coordinate system corresponding to the desktop in fig. 4, for example). Several embodiments are described herein, but the disclosure is not limited thereto. In one embodiment, the processing element 110 determines the frame boundary CNT in the system coordinate system SYC according to frame coordinates corresponding to at least some of the vertices (e.g., two diagonal vertices, a center point, and at least one vertex) of the captured frame CMI. In another embodiment, the processing element 110 may determine the frame boundary CNT in the system coordinate system SYC according to the frame coordinates corresponding to the boundary of the photographed frame CMI. In another embodiment, the processing element 110 may estimate the size of the photographic image CMI in the system coordinate system SYC according to the second corresponding relationship between the photographic image CMI and the system coordinate system SYC, and then determine the image boundary CNT in the system coordinate system SYC according to the image coordinates corresponding to the center of the photographic image CMI.
It should be noted that although the frame boundary CNT is shown as a circular dotted line in fig. 4, the frame boundary CNT may have other shapes, such as a fan shape, a parallelogram shape, or an irregular shape, and thus the shape of the frame boundary CNT in the present disclosure is not limited to that shown in fig. 4.
In operation S4, the processing element 110 obtains a gaze coordinate in the system coordinate system through the eye tracker 20. In one embodiment, the gaze coordinate corresponds to a specific location in the camera CMI, and the specific location is the location of the user's gaze in the camera CMI. It should be noted that operation S4 and operations S1, S2, S3 may be performed simultaneously, or in reverse order.
In one embodiment, the processing element 110 may obtain the aforementioned gaze coordinate in the system coordinate system according to a correspondence relationship (hereinafter referred to as a first correspondence relationship) between the system coordinate system and the eye tracking data from the eye tracker 20. In one embodiment, the processing element 110 utilizes a plurality of reference points to obtain a first corresponding relationship between the system coordinate system and the eye tracking data.
For example, referring to fig. 4 and 5, first, the processing element 110 obtains the reference coordinates of the reference points NA, NB, and NC in the system coordinate system SYC (operation S41 in fig. 8). The user can use the controller or other user input interface to transmit the user signal USI to the electronic device 100 while looking at the reference point NA. In one embodiment, the user signal USI is a trigger signal. In one embodiment, a user may press a controller, for example, to generate the user signal USI. The processing element 110 may retrieve the eye tracking data from the eye tracker 20 corresponding to the user signal USI as the reference eye tracking data corresponding to the reference point NA (operation S42 in fig. 8). In a similar manner, the processing element 110 may obtain reference eye tracking data corresponding to the reference points NB, NC. Then, the processing element 110 may obtain a first corresponding relationship between the system coordinate system SYC and the eye tracking data according to the reference coordinates of the reference points NA, NB, and NC in the system coordinate system SYC and the reference eye tracking data corresponding to the reference points NA, NB, and NC (fig. 8 operation S43). It is noted that in different embodiments, the processing element 110 may also establish the first correspondence between the system coordinate system SYC and the eye tracking data using other reference points than the reference points NA, NB, NC.
Then, the processing element 110 may obtain the line-of-sight coordinate of the gazing position of the user in the system coordinate system SYC according to the eyeball tracking data from the eyeball tracker 20 by using the first corresponding relationship between the system coordinate system SYC and the eyeball tracking data (operation S44 in fig. 8).
In operation S5, the processing element 110 determines a partial output frame corresponding to the line-of-sight coordinate in the one or more frames corresponding to the line-of-sight coordinate, wherein the partial output frame is located within a frame boundary corresponding to the one or more frames (e.g., when a target focused by the user is within a range covered by any one of the frames).
For example, referring to fig. 3, 4, and 6, when the user gazes at the target OBJ, the processing element 110 may obtain the line-of-sight coordinates corresponding to the target OBJ in the system coordinate system SYC. In one embodiment, the target OBJ may include, but is not limited to, an article, text, a frame, and the like. The processing element 110 can determine whether the line-of-sight coordinate is located in the frame boundary CNT. In the case that the line-of-sight coordinate is located in the frame boundary CNT, the processing element 110 may obtain the line-of-sight position (e.g., the position of the object OBJ in the photographic frame CMI in fig. 3) corresponding to the line-of-sight coordinate in the photographic frame CMI according to the second corresponding relationship between the photographic frame CMI and the system coordinate system SYC (operation S51 in fig. 9). Then, the processing element 110 may determine a local output image ORI (refer to fig. 6) according to the line-of-sight position corresponding to the line-of-sight coordinate in the photographing image CMI (fig. 9 operation S52).
For example, in other embodiments, besides the camera frame CMI and the frame boundary CNT, there may be other camera frames and frame boundaries (that is, there are a plurality of camera frames and a plurality of frame boundaries of the plurality of camera frames in the system coordinate system SYC), the processing element 110 may obtain the view-line position corresponding to the view-line coordinate in the camera frame according to the frame boundary where the view-line coordinate is located and the corresponding second corresponding relation, and then determine to output the local output frame according to the view-line position corresponding to the view-line coordinate in the camera frame.
In contrast, in an embodiment, in the case that the sight line coordinate is located outside the frame boundary of all the photographing frames in the system coordinate system, the processing element 110 may determine that the output frame is one of the one or more photographing frames (e.g., the output frame is the photographing frame CMI in fig. 3). That is, when there is the photographing screen A, B, C and the user's gaze coordinate is outside the range covered by the photographing screen A, B, C, the processing element 110 can determine that the output screen is a predetermined one of the photographing screens A, B, C (e.g., the photographing screen aimed at the main photographing target). Therefore, when the sight of the user is temporarily removed, the output picture can return to the preset main shooting picture, and the output of meaningless shooting pictures can be avoided.
In one embodiment, in the case that the line-of-sight coordinate is located within only one of the one or more frame boundaries (e.g., the frame CMI in FIG. 3), the local output frame is a portion of the frame corresponding to the one of the one or more frame boundaries (e.g., the portion of the frame CMI in FIG. 3 corresponding to the target OBJ). For example, when there is the photographic image A, B, C (i.e., there are a plurality of photographic images) and the user is only gazing at the target within the range covered by the photographic image a, the local output image only shows a part of the photographic image a, and does not show any part of the photographic image B, C.
In other embodiments, for example, only the captured frame B exists (i.e., only one captured frame exists) and the line-of-sight coordinate is located only in the frame boundary corresponding to the captured frame B, the local output frame is a portion corresponding to only the captured frame (in this example, the local output frame is a portion corresponding to the captured frame B). In one embodiment, in the case that the gaze coordinate is located within more than one of the one or more frame boundaries, the processing element 110 may determine the local output frame to be a part of the captured frame corresponding to one of the one or more frame boundaries according to the actual requirement or the predetermined priority, that is, when the captured frame A, B, C exists and the target watched by the user is within the range covered by the captured frame A, B, C, the processing element 110 may determine which part of the captured frame a, the captured frame B, or the captured frame C is to be displayed by the local output frame according to the actual requirement or the predetermined priority.
It should be noted that although the present disclosure is described with respect to live broadcasting as an example, the application field of the present disclosure is not limited to live broadcasting, and other applications in which a local output screen corresponding to the sight line coordinate can be determined in a photographic screen are also within the scope of the present disclosure.
By applying one of the above embodiments, a local output frame corresponding to the line-of-sight coordinate can be determined in the photographing frame CMI. Therefore, the live broadcast user does not need to manually operate, the live broadcast picture which can be seen by the viewer can be changed, the interaction between the live broadcast user and the viewer can be further increased, and the convenience in use of the live broadcast user can be improved.
While the foregoing is directed to embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow.