WO2018142756A1 - 情報処理装置及び情報処理方法 - Google Patents
情報処理装置及び情報処理方法 Download PDFInfo
- Publication number
- WO2018142756A1 WO2018142756A1 PCT/JP2017/043700 JP2017043700W WO2018142756A1 WO 2018142756 A1 WO2018142756 A1 WO 2018142756A1 JP 2017043700 W JP2017043700 W JP 2017043700W WO 2018142756 A1 WO2018142756 A1 WO 2018142756A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- category
- recognition
- unit
- information processing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/98—Detection or correction of errors, e.g. by rescanning the pattern or by human intervention; Evaluation of the quality of the acquired patterns
- G06V10/987—Detection or correction of errors, e.g. by rescanning the pattern or by human intervention; Evaluation of the quality of the acquired patterns with the intervention of an operator
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
Definitions
- the present invention relates to a technique for recognizing an image.
- category classification is also referred to as general object recognition, and it only recognizes the category or proper noun to which an object in the image belongs, and does not specify where the object is in the image.
- object detection is also referred to as specific object recognition. In addition to specifying the category or proper noun of an object in the image, the object detection also specifies the position in the image (and the contour of the object).
- Object detection is known to be more difficult than categorization because, for example, there are many categories in the object group, the position or orientation of the object is indeterminate, and lighting fluctuations.
- the scale is indefinite or the object is deformed.
- the category recognizing unit includes a first category recognizing unit that recognizes a category to which an object included in the captured image belongs, and a second category recognizing unit that recognizes a category to which the object included in the captured image belongs.
- a second category recognizing unit that recognizes a category different from a category that can be recognized by the recognizing unit, and the recognition control unit is not recognized by the object recognizing unit and the first category recognizing unit.
- the recognition control unit displays an image specifying the position of the object in the captured image in the captured image for an imaging display device that captures an image according to the user's operation and displays the captured image. By instructing the user, the user may be guided to the imaging method.
- the recognition control unit may guide the imaging method to the user by instructing the imaging display device to display an image imitating the object superimposed on the captured image.
- the synthesized image corresponding to the object is synthesized according to the position of the object recognized in the captured image according to the operation of the user.
- An instruction is given to an imaging display device that performs imaging and displays a captured image, and when the category is recognized by the category recognition unit, the imaging display device is configured to synthesize a composite image corresponding to the category in the captured image. You may make it provide the image composition instruction
- the present invention also includes a step of trying to recognize a position of an object included in the captured image, a step of attempting to recognize a category to which the object included in the captured image belongs, and the position of the object is not recognized and the object belongs to
- an information processing method is provided that includes guiding an imaging method to a user and attempting to recognize an object included in a captured image captured according to the guidance.
- the probability that an object can be detected in image recognition can be improved.
- DESCRIPTION OF SYMBOLS 1 ... Information processing system, 10 ... Information processing apparatus, 11 ... Image recognition part, 12 ... 1st synthesized image determination part, 13 ... Image composition instruction
- FIG. 1 is a diagram illustrating the configuration of an information processing system 1 according to an embodiment of the invention.
- the information processing system 1 includes an information processing device 10 that recognizes an image, an imaging display device 20 that captures and displays an image, and a network 90 that connects the information processing device 10 and the imaging display device 20 so that they can communicate with each other.
- the network 90 is, for example, a LAN (Local Area Network), a WAN (Wide Area Network), or a combination thereof, and may include a wired section or a wireless section.
- 1 shows one information processing apparatus 10 and one imaging display apparatus 20, one or more of them may be provided.
- the information processing apparatus 10 acquires a captured image obtained by the imaging display device 20 and performs image recognition on the captured image. Then, the information processing apparatus 10 synthesizes a different synthesized image (referred to as a first synthesized image) with the captured image on the imaging display apparatus 20 according to the image recognition result, and further, a predetermined position with respect to the first synthesized image. And instructing to synthesize a synthesized image (referred to as a second synthesized image) different from the first synthesized image. For example, when a landscape is imaged by the imaging display device 20 (FIG.
- a moving character image imitating a human running is synthesized as the first synthesized image G1 corresponding to the landscape.
- a predetermined position in the first composite image G1, that is, a region G2 corresponding to the human head represented by the first composite image G1 is blank.
- the captured face image is synthesized as the second synthesized image G3 with respect to the region G2 of the first synthesized image G1.
- FIG. 8D a scene in which a human character image having the user's own face is running is synthesized as a moving image with the landscape image displayed on the imaging display device 20, and the user can view these images. You can enjoy watching the fun of the combination and movement.
- the image recognition function of the information processing apparatus 10 includes an object recognition function and a category recognition function. Furthermore, the category recognition function includes two types of functions, a first category recognition function and a second category recognition function.
- the object recognition function is also called object detection or specific object recognition in the field of image recognition for unconstrained real world scene images.
- the object recognition function Is specified (and the contour of the object).
- the object recognition function is realized, for example, by performing local feature extraction of an image and using a statistical learning method for the extraction result. In this embodiment, object recognition is used to recognize what kind of object is in the image and display a composite image corresponding to the object at a position corresponding to the image recognition result.
- the category recognition function is also called category classification or general object recognition in the field of image recognition for an image of an unrestricted real world scene, and recognizes the category or proper noun to which the object in the image belongs (hereinafter simply referred to as category It does not specify where the object is in the image.
- the category recognition function is realized, for example, by extracting a feature of an image to calculate a vector quantization histogram, and applying the calculation result to a category-specific histogram prepared in advance.
- the first category recognition function is used for recognizing a category to which an object in an image belongs, and displaying a composite image corresponding to the image recognition result. Perform category recognition for objects belonging to.
- the recognition algorithm for recognizing the category to which the object in the image belongs is common to the first category recognition function, but the second category recognition function is different from the first category group in a predetermined second category group. Perform category recognition for the object to which it belongs. That is, the category group covered by the first category recognition function is different from the second category group covered by the second category recognition function.
- the first category group is a set of objects for the purpose of category recognition, and it can be arbitrarily determined which object belongs to the first category group, but in this embodiment, for example, a vehicle or a person A collection of objects such as
- the second category group is a set of objects for the purpose of finally performing object recognition through category recognition, and which object belongs to the second category group can be arbitrarily determined.
- an object such as a point card or a company logo issued by a company to a consumer.
- the information processing apparatus 10 displays a composite image corresponding to the image recognition result at a position corresponding to the image recognition result.
- the imaging display device 20 is instructed.
- the information processing apparatus 10 displays a composite image corresponding to the image recognition result.
- the imaging display device 20 is instructed.
- the information processing apparatus 10 An imaging method capable of recognizing an object is guided, and the object recognition function is made to try to recognize the captured image captured according to the guidance again.
- the information processing apparatus 10 displays the composite image corresponding to the image recognition result at a position corresponding to the image recognition result.
- the reason for dividing the first category group covered by the first category recognition function from the second category group covered by the second category recognition function is that the first category group can realize category recognition.
- the purpose of image recognition for each group is different so that the second category group ultimately wants to realize even object recognition. That is, the object belonging to the second category group is originally intended for object recognition, but the object recognition is performed because the position or orientation of the object in the captured image is indefinite or the object is deformed. Some functions may not be recognized correctly.
- the user can recognize the object by adjusting the position, posture, or shape of the object in the captured image according to the guided imaging method. In other words, it can be said that the object group covered by the object recognition function and the object group covered by the second category recognition function at least partially overlap.
- first category group covered by the first category recognition function and the second category group covered by the second category recognition function are integrated into a single category group, the accuracy of category recognition deteriorates. Therefore, there is also a reason that these two groups are separated.
- FIG. 2 is a diagram illustrating a hardware configuration of the imaging display device 20.
- the imaging display device 20 is a device that performs imaging in accordance with a user operation and displays a captured image, and is a communicable computer such as a smartphone, a tablet, or a wearable terminal.
- the imaging display device 20 includes a CPU 201 (Central Processing Unit), a ROM (Read Only Memory) 202, a RAM (Random Access Memory) 203, an auxiliary storage device 204, a communication IF 205, a camera 206, and a UI device 207.
- the CPU 201 is a processor that performs various calculations.
- the ROM 202 is a non-volatile memory that stores programs and data used for starting up the imaging display device 20, for example.
- a RAM 203 is a volatile memory that functions as a work area when the CPU 201 executes a program.
- the auxiliary storage device 204 is a non-volatile storage device such as an HDD (Hard Disk Drive) or an SSD (Solid State Drive), and stores programs and data used in the imaging display device 20. When the CPU 201 executes this program, the imaging display device 20 functions as a computer device, and a function shown in FIG. 5 described later is realized.
- the communication IF 205 is an interface for performing communication via the network 90 in accordance with a predetermined communication standard. This communication standard may be a wireless communication standard or a wired communication standard.
- the communication IF 205 includes an antenna and a communication circuit that operate according to a communication standard such as LTE (Long Term Evolution) or Wi-Fi (registered trademark).
- the camera 206 includes a lens and an optical element, and generates image data indicating a captured image.
- the camera 206 includes an in camera provided on the front side of the imaging display device 20 and an out camera provided on the back side of the imaging display device 20.
- the in-camera is mainly used when imaging the user himself, and the out-camera is mainly used when imaging scenery other than the user himself.
- the UI device 207 includes an operation unit provided with operators such as a key and a touch sensor, a display unit such as a liquid crystal panel and a liquid crystal drive circuit, and an audio output unit such as a speaker or an earphone terminal.
- the imaging display device 20 may include other configurations such as a GPS (Global Positioning System) unit, a gyro sensor, or an acceleration sensor.
- GPS Global Positioning System
- FIG. 3 is a diagram illustrating a hardware configuration of the information processing apparatus 10.
- the information processing apparatus 10 is a computer apparatus having a CPU 101, a ROM 102, a RAM 103, an auxiliary storage device 104, and a communication IF 105.
- the CPU 101 is a processor that performs various calculations.
- the ROM 102 is a non-volatile memory that stores programs and data used for starting up the information processing apparatus 10, for example.
- the RAM 103 is a volatile memory that functions as a work area when the CPU 101 executes a program.
- the auxiliary storage device 104 is a nonvolatile storage device such as an HDD or an SSD, and stores programs and data used in the information processing apparatus 10. When the CPU 101 executes this program, the functions shown in FIG. 5 described later are realized.
- the communication IF 105 is an interface for performing communication via the network 90 in accordance with a predetermined communication standard.
- the information processing apparatus 10 may include other configurations such as a UI device in addition to
- auxiliary storage device 104 in addition to the image recognition algorithm for performing image recognition and the first composite image data group for displaying the first composite image described above on the imaging display device 20, which first composite image is displayed on the imaging display device 20.
- a composite image determination table for determining whether to display an image is stored.
- FIG. 4 is a diagram illustrating a composite image determination table.
- the type of the image recognition function, the image recognition result, the first composite image ID for identifying the first composite image, and the display position of the first composite image are associated with each other.
- the first composite image with the first composite image ID “C001” corresponding to the classified category Is displayed at absolute coordinates X1, Y1 in the image.
- the absolute coordinates here are coordinates according to XY coordinate axes defined in advance in the display area of the imaging display device 20. That is, in this example, the first composite image is a position designated by absolute coordinates in the display area of the imaging display device 20 (for example, the central part of the display area) regardless of the position of the object “bridge” in the image. Is synthesized.
- the object in the image can be recognized, for example, for the category “building”, “clock tower”, proper noun and position in the image
- the object It is defined that the first composite image of the first composite image ID “C020” corresponding to the category or proper noun is displayed at the relative coordinates x1, y1.
- the relative coordinates are coordinates according to an xy coordinate axis (for example, an xy coordinate axis having a unit scale corresponding to the size of the object with the center of gravity of the object as the origin) defined based on the detected object “clock tower”. It is.
- the first composite image is synthesized at a position designated by relative coordinates with respect to the position of the object “clock tower” in the image. Accordingly, the display position of the first composite image is different. For example, as shown in FIG. 8B, the first composite image G1 is always displayed at an upper position in the gravity direction of the roof portion with respect to the object “clock tower”.
- each first composite image data stored in the auxiliary storage device 104 includes data specifying the display position of the second composite image with respect to the first composite image. For example, when a moving character image imitating a human running is synthesized as the first synthesized image G1 (FIG. 8B), the position of the region G2 corresponding to the human head in the first synthesized image G1. Is designated as a predetermined position in the first composite image data.
- FIG. 5 is a block diagram showing a functional configuration of the information processing system 1.
- an imaging unit 21 that captures an image
- a second composite image acquisition unit 22 that acquires a second composite image
- an image captured by the imaging unit 21 are processed.
- An image data generation unit 23 that generates image data obtained by combining the first composite image and the second composite image, and a display unit 24 that displays an image corresponding to the image data generated by the image data generation unit 23 are realized. .
- an image recognition unit 11 that recognizes an image captured by the imaging unit 21, and a first composite image determination unit 12 that determines a different first composite image according to a recognition result by the image recognition unit 11.
- the first composite image determined by the first composite image determination unit 12 is combined with the image captured by the imaging unit 21, and the second composite image is combined at a predetermined position with respect to the first composite image combined with the image.
- the image composition instruction unit 13 that instructs the imaging display device 20 to do so is realized.
- the image recognition unit 11 includes an object recognition unit 111, a category recognition unit 112 including a first category recognition unit 1121 and a second category recognition unit 1122, and a recognition control unit 113.
- the object recognition unit 111 realizes the above-described object recognition function
- the first category recognition unit 1121 realizes the above-described first category recognition function
- the second category recognition unit 1122 realizes the above-described second category recognition function.
- the recognition control unit 113 guides the imaging method to the user and performs the guidance.
- the object recognition unit 111 recognizes the captured image captured according to the above.
- the recognition control unit 113 instructs the imaging display device 20 to display an image designating the category-recognized object position in the captured image. Guide you through the method. More specifically, the recognition control unit 113 instructs the imaging method to the user by instructing the imaging display device to display an image imitating the category-recognized object superimposed on the captured image.
- the image display device 20 When the object recognizing unit 111 recognizes the image composition instruction unit 13, the image display device 20 synthesizes a composite image corresponding to the object according to the position of the object recognized in the captured image. To instruct. In addition, when the first category recognition unit 1121 recognizes the image composition instruction unit 13 without being recognized by the object recognition unit 111, the image composition instruction unit 13 instructs the imaging display device 20 to synthesize a composite image corresponding to the category in the captured image. Instruct.
- the imaging unit 21 and the second composite image acquisition unit 22 are realized by the camera 206 of the imaging display device 20, the image data generation unit 23 is realized by the CPU 201 of the imaging display device 20, and the display unit 24 is a UI device of the imaging display device 20. 207.
- the image recognition unit 11 is realized by the CPU 101 of the information processing device 10
- the first composite image determination unit 12 is realized by the CPU 101 and the auxiliary storage device 104 of the information processing device 10
- the image composition instruction unit 13 is the CPU 101 of the information processing device 10.
- the image data generation unit 23 and the image composition instruction unit 13 implement an image composition unit 30 included in the information processing system 1 according to the present invention.
- the image synthesis unit 30 synthesizes a first synthesized image that differs according to the recognition result by the image recognition unit 11 with the image captured by the imaging unit 21.
- FIG. 6 An example of display control of the information processing system 1 will be described with reference to FIGS.
- the user activates the imaging unit 21 (out camera) of the imaging display device 20 and starts imaging an object desired by the user (step S ⁇ b> 11).
- the imaging result by the imaging unit 21 is displayed in real time in the display area of the imaging display device 20, and when the imaging direction of the imaging unit 21 changes, the image on this display area is also changed.
- the captured image is displayed on the imaging display device 20 as shown in FIG. Displayed in the display area.
- the captured image data indicating the captured image is transmitted from the communication IF 205 to the information processing apparatus 10 (step S12).
- the transmission of the captured image data may be started, for example, when a user operation is performed, or may be automatically started without waiting for the user operation. Further, the transmission of the captured image data may be continuously performed from the start time point, or may be periodically performed from the start time point.
- the imaging display device 20 desirably compresses the captured image data as appropriate to reduce the data size.
- the image recognition unit 11 of the information processing apparatus 10 attempts image recognition on the captured image data (step S13). Details of this image recognition processing will be described later with reference to FIG.
- the first composite image determination unit 12 determines a first composite image to be combined with the captured image in the imaging display device 20 based on the recognition result (step S14). If the image recognition unit 11 succeeds in category recognition (recognition by the first category recognition unit 1121 here), the first composite image determination unit 12 sets the category of the object in the table illustrated in FIG. The corresponding first composite image ID is specified, and the first composite image data of the first composite image ID is read from the auxiliary storage device 104.
- the image composition instruction unit 13 transmits the read first composition image data to the imaging display device 20 as an image composition instruction together with the absolute coordinates corresponding to the category in the table illustrated in FIG. 4 (step S15). At this time, as described above, the first composite image data includes data designating the display position of the second composite image with respect to the first composite image.
- the first composite image determination unit 12 specifies the first composite image ID corresponding to the object in the table illustrated in FIG. Then, the first composite image data of the first composite image ID is read from the auxiliary storage device 104.
- the image composition instruction unit 13 displays the read first composition image data together with the coordinates of the object in the display area of the imaging display device 20 and the relative coordinates corresponding to the object in the table illustrated in FIG. To the imaging display device 20 (step S15). Also at this time, as described above, the first composite image data includes data designating the display position of the second composite image with respect to the first composite image.
- the image data generation unit 23 guides the user to capture the second composite image (step S17).
- This guidance may be realized by, for example, a method of displaying a message “Please take a picture of yourself” and designating the imaging position of the face on the display area of the imaging display device 20.
- the user images his / her face with the imaging unit 21 (in-camera).
- the imaging unit 21 in-camera.
- the upper body including the user's face is displayed in the display area of the imaging display device 20.
- the image data generation unit 23 extracts a partial image corresponding to the user's head from this image using an image recognition technique or the like, and acquires it as a second composite image (step S18).
- the image data generation unit 23 generates image data obtained by combining the first combined image and the second combined image with the captured image (Step S19), and displays an image corresponding to the image data (Step S20).
- the second composite image showing the user's face is arranged at the position of the head in the first composite image imitating a human running, and these are the roof portions of the clock tower. Will be displayed above.
- step S13 in FIG. 6 the image recognition processing (step S13 in FIG. 6) of the information processing apparatus 10 will be described with reference to FIG.
- the object recognizing unit 111 of the image recognizing unit 11 attempts object recognition on the captured image data. If the object recognition is successful (step S111; YES), as described above, the first composite image determination unit 12 reads out the first composite image data of the first composite image ID corresponding to the object from the auxiliary storage device 104. . The image compositing instruction unit 13 then captures the read first composite image data as an image compositing instruction together with the coordinates of the object in the display area of the image capturing display device 20 and the relative coordinates corresponding to the object. (FIG. 6: Steps S14 to S15).
- step S111 if the object recognition by the object recognition unit 111 fails (step S111; NO), the first category recognition unit 1121 of the image recognition unit 11 tries to perform category recognition on the captured image data. If this category recognition is successful (step S112; YES), as described above, the first composite image determination unit 12 stores the first composite image data of the first composite image ID corresponding to the category of the object in the auxiliary storage device 104. Read from. Then, the image composition instruction unit 13 transmits the read first composition image data together with the absolute coordinates corresponding to the category as an image composition instruction to the imaging display device 20 (FIG. 6: steps S14 to S15).
- step S112 If the object recognition by the first category recognition unit 1121 fails (step S112; NO), the second category recognition unit 1122 of the image recognition unit 11 attempts category recognition on the captured image data. If the category recognition is successful (step S113; YES), the recognition control unit 113 of the image recognition unit 11 guides the user to the imaging method (step S115).
- the point card when a point card for giving a privilege to the user from the company A is imaged as an object by the imaging display device 20 (FIG. 9A), the point card is imaged at a position, posture, or shape where the object can be recognized. May not have been. Even in such a case, if the reward card is included in the second category group that can be recognized by the second category recognizing unit 1122, category recognition by the second category recognizing unit 1122 may be possible. In this case, the recognition control unit 113 of the image recognition unit 11 transmits image data simulating an object called a category-recognized point card to the imaging display device 20, and superimposes an image corresponding to the image data on the captured image. To display.
- the image data transmitted to the imaging display device 20 displays a translucent image imitating the point card on the imaging display device 20 so that the position, posture, and shape of the point card can be recognized.
- Data As a result, a translucent image G4 simulating a point card is displayed in the captured image of the imaging display device 20 (FIG. 9B).
- a message such as “Please take pictures according to the image” is also displayed.
- the captured image data is transmitted from the imaging display device 20 to the information processing device 10.
- the object recognition unit 111 of the image recognition unit 11 acquires the captured image data (step S116 in FIG. 7; YES), the object recognition unit 111 tries to perform object recognition on the captured image data.
- the point card in the captured image is captured at a position, posture, and shape that allow object recognition, there is a high possibility of successful object recognition.
- the first composite image determination unit 12 reads the first composite image data of the first composite image ID corresponding to the object from the auxiliary storage device 104, and The image composition instruction unit 13 transmits the read first composite image data to the image capture display device 20 as an image composition instruction together with the coordinates of the object in the display area of the image capture display device 20 and the relative coordinates corresponding to the object. (FIG. 6: Steps S14 to S15).
- first composite image G5 corresponding to an object called a point card
- first composite image ID: C022 in FIG. 4 a moving character image imitating a human figure
- image G7 an animated character image (first composite image ID: C023 in FIG. 4) dancing with the human is synthesized
- first composite image ID: C023 in FIG. 4 an animated character image
- FIG. 9D a region G6 corresponding to a human head is blank.
- step S113 when the category recognition by the second category recognition unit 1122 of the image recognition unit 11 fails (step S113 in FIG. 7; NO), the recognition control unit 113 of the image recognition unit 11 cannot recognize the image.
- Predetermined error processing such as notification to the imaging display device 20 is performed (step S114).
- the object recognition function may not be able to correctly recognize the object because the position or orientation of the object is indeterminate in the captured image or the object is deformed, but even in such a case, the guided imaging method is used.
- the user can recognize the object by adjusting the position, posture, or shape of the object in the captured image.
- This guidance of the imaging method is realized in such a form that a translucent image imitating an object and a message such as “Please take a picture according to the image” are displayed, so that it is easy for the user to understand.
- the first category group covered by the first category recognition function and the second category group covered by the second category recognition function are separated, the accuracy of category recognition for objects belonging to the respective category groups is reduced. There is nothing to do.
- a moving image in which a human character image (first composite image) having a user's own face (second composite image) is running is combined with the captured image displayed on the image display device 20.
- first composite image differs depending on the recognition result of the captured image displayed on the imaging display device 20, what kind of first composite image is synthesized by the user capturing various objects. You can feel a sense of expectation and enjoy the results of that expectation.
- image of the user's own face is used as the second composite image, it is possible to enjoy a feeling as if the user himself / herself existed in the virtual space of the captured image.
- the user since the user captures his / her face as the second composite image after seeing the first composite image combined with the captured image, the user captures the direction or expression of his / her face to be combined.
- the user In order to match the atmosphere of the image and the first composite image, for example, it is possible to turn sideways or make a comical expression or a desperate expression. This is also one of the fun for the user.
- the embodiment described above can be modified as follows. The following modifications may be implemented in combination with each other.
- [Modification 1] The first category recognition unit 1121 and the second category recognition unit 1122 are simply integrated without dividing the first category group covered by the first category recognition function and the second category group covered by the second category recognition function. It may be a single function. That is, the present invention includes an object recognition unit that recognizes the position of an object included in a captured image, a category recognition unit that recognizes a category to which an object included in the captured image belongs, and the category recognition unit that is not recognized by the object recognition unit.
- the information processing apparatus includes a recognition control unit that guides the user to the imaging method and causes the object recognition unit to recognize the captured image captured according to the guidance. May be. In this way, it is possible to save the cost for separately designing the first category group covered by the first category recognition function and the second category group covered by the second category recognition function.
- the recognition control unit 113 instructs the imaging display device 20 to display a semitransparent image imitating an object superimposed on the captured image.
- the guidance of the imaging method is limited to the example of the embodiment.
- the recognition control unit 113 may instruct the imaging display device 20 to display an image specifying the position of the object in the captured image in the captured image and guide the imaging method.
- the image compositing unit 30 is not limited to compositing the first composite image having different appearances according to the recognition result by the image recognizing unit 11.
- the first composite image having at least one of size, position, motion, or number is different. May be synthesized.
- the first composite image is not limited to a moving image and may be a still image.
- the image synthesis unit 30 may synthesize different first synthesized images according to the user attribute or the position or orientation of the imaging display device 20.
- the image composition unit 30 may be configured such that a character image that is a first composite image flies and lands on the roof of the clock tower, a character image that is the first composite image opens a door of the clock tower, You may make it synthesize
- a character image that is a first composite image flies and lands on the roof of the clock tower
- a character image that is the first composite image opens a door of the clock tower
- combine the AR (Augmented Reality) image of a moving image in which the character image which is a synthesized image rewinds the clock hands of the clock tower.
- the image composition unit 30 sets a face image to which a tag corresponding to the recognition result by the image recognition unit 11 is assigned as a second composite image.
- the recognition result by the image recognition unit 11 and the type of user's appearance are associated with each other and stored in the auxiliary storage device 204, and the image composition unit 30 corresponds to the recognition result by the image recognition unit 11.
- the tag to be identified is specified, and the image data of the face to which the tag is attached is read from the auxiliary storage device 204.
- the image composition unit 30 is a second composite image in which at least one of appearance, size, position, movement, or number differs according to the recognition result by the image recognition unit 11. May be synthesized.
- the second composite image is not limited to the captured image, and may be any image such as an avatar or an animation image. As described above, when the various elements of the second composite image are different depending on the image recognition result, it is more interesting for the user.
- the image composition unit 30 may compose a different second composite image according to the first composite image. For example, when a landscape is imaged by the imaging display device 20, the first composite image G1 corresponding to the landscape is synthesized.
- the image composition unit 30 presents, for example, a plurality of second composite image candidates to the user according to the first composite image G1, and when the user selects one of them, the second composite image is composed.
- the first composite image or the second composite image may be used as an advertisement.
- the first composite image an image in which a certain character has the plastic bottle (advertisement object) of the manufacturer, or a picture of the plastic bottle (advertisement object) in the middle of the character's clothes is described.
- An example of an image can be considered.
- the imaging display device 20 is held over a beverage bottle (advertisement object) of a certain beverage manufacturer, the beverage manufacturer's character image is displayed as a first composite image at a certain position of the plastic bottle, and the character image is further displayed.
- the user's face may be arranged as the second composite image.
- the imaging display device 20 when the imaging display device 20 is held over a large-scale intersection in a certain city, a certain first composite image is combined with the area of the screen panel installed at the intersection, and an advertisement object is displayed on the first composite image. You may make it reproduce
- the combination of the first composite image and the second composite image may be changed according to the user attribute or the position of the imaging display device 20, or may be changed according to the unit price paid by the advertiser. If it does in this way, it will become possible to utilize this invention for an advertisement.
- each functional block may be realized by one device physically and / or logically coupled, and two or more devices physically and / or logically separated may be directly and / or indirectly. (For example, wired and / or wireless) and may be realized by these plural devices.
- Each aspect / embodiment described in this specification includes LTE (Long Term Evolution), LTE-A (LTE-Advanced), SUPER 3G, IMT-Advanced, 4G, 5G, FRA (Future Radio Access), W-CDMA.
- LTE Long Term Evolution
- LTE-A Long Term Evolution-Advanced
- SUPER 3G IMT-Advanced
- 4G 5G
- FRA Full Radio Access
- W-CDMA Wideband
- GSM registered trademark
- CDMA2000 Code Division Multiple Access 2000
- UMB User Mobile Broadband
- IEEE 802.11 Wi-Fi
- IEEE 802.16 WiMAX
- IEEE 802.20 UWB (Ultra-WideBand
- the present invention may be applied to a Bluetooth (registered trademark), a system using other appropriate systems, and / or a next generation system extended based on these systems.
- system and “network” used in this specification are used interchangeably.
- determining and “determining” used in this specification may encompass a wide variety of actions. “Judgment” and “determination” are, for example, judgment (judging), calculation (calculating), calculation (processing), processing (deriving), investigating (investigating), searching (looking up) (for example, table , Searching in a database or another data structure), considering ascertaining as “determining”, “deciding”, and the like.
- determination and “determination” are reception (for example, receiving information), transmission (for example, transmitting information), input (input), output (output), and access. (Accessing) (eg, accessing data in a memory) may be considered as “determined” or “determined”.
- determination and “determination” means that “resolving”, “selecting”, “choosing”, “establishing”, and “comparing” are regarded as “determining” and “determining”. May be included. In other words, “determination” and “determination” may include considering some operation as “determination” and “determination”.
- software whether it is called software, firmware, middleware, microcode, hardware description language, or other names, instructions, instruction sets, codes, code segments, program codes, programs, subprograms, software modules , Application, software application, software package, routine, subroutine, object, executable file, execution thread, procedure, function, etc. should be interpreted broadly.
- software, instructions, etc. may be transmitted / received via a transmission medium.
- software may use websites, servers, or other devices utilizing wired technology such as coaxial cable, fiber optic cable, twisted pair and digital subscriber line (DSL) and / or wireless technology such as infrared, wireless and microwave. When transmitted from a remote source, these wired and / or wireless technologies are included within the definition of transmission media.
- connection means any direct or indirect connection or coupling between two or more elements and It can include the presence of one or more intermediate elements between two “connected” or “coupled” elements.
- the coupling or connection between the elements may be physical, logical, or a combination thereof.
- the two elements include radio frequency by utilizing one or more wires, cables, and / or printed electrical connections, as well as some non-limiting and non-inclusive examples.
- electromagnetic energy such as electromagnetic energy having wavelengths in the region, the microwave region and the light (both visible and invisible) region can be considered “connected” or “coupled” to each other.
- the information processing apparatus can be realized by an apparatus that is provided with all of the functions, and the functions of the respective apparatuses are further distributed to a plurality of apparatuses. It may be an implemented system.
- the present invention can also be implemented in the form of an information processing method performed by the information processing apparatus 10 or a program for causing a computer to function as the information processing apparatus 10.
- a program may be provided in a form recorded on a recording medium such as an optical disc, or may be provided in a form such that the program is downloaded to a computer via a network such as the Internet, and the program can be installed and used. Is possible.
- This program whether it is called software, firmware, middleware, microcode, hardware description language or other name, instructions, instruction set, code, code segment, program code, program, subprogram, software It should be interpreted broadly to mean modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Quality & Reliability (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- User Interface Of Digital Computer (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
[構成]
図1は、本発明の一実施形態に係る情報処理システム1の構成を例示する図である。情報処理システム1は、画像の認識を行う情報処理装置10と、画像の撮像及び表示を行う撮像表示装置20と、これら情報処理装置10及び撮像表示装置20を通信可能に接続するネットワーク90とを備えている。ネットワーク90は、例えばLAN(Local Area Network)又はWAN(Wide Area Network)、若しくはこれらの組み合わせであり、有線区間又は無線区間を含んでいてもよい。なお、図1には、情報処理装置10及び撮像表示装置20を1つずつ示しているが、これらはそれぞれ複数であってもよい。
図6~図9を参照して、情報処理システム1の表示制御の例について説明する。図6において、ユーザは撮像表示装置20の撮像部21(アウトカメラ)を起動して、自身が所望する物体の撮像を開始する(ステップS11)。撮像部21による撮像結果は、撮像表示装置20の表示領域にリアルタイムに表示され、撮像部21の撮像方向が変わるとこの表示領域上の画像も変更される。例えばユーザが時計台の風景が撮像されるような向きに撮像表示装置20を構えてユーザが所定の操作を行うと、図8(A)に示すように、その撮像画像が撮像表示装置20の表示領域に表示される。この撮像画像を示す撮像画像データは、通信IF205から情報処理装置10に送信される(ステップS12)。この撮像画像データの送信は、例えば、ユーザの操作があったときに開始されてもよいし、ユーザの操作を待たず自動的に開始されてもよい。また、撮像画像データの送信はその開始時点から連続的に成されてもよいし、又はその開始時点から定期的になされてもよい。なお、撮像表示装置20は撮像画像データを適宜圧縮してデータサイズを小さくすることが望ましい。
例えば撮像画像において物体の位置又は姿勢が不定であるとか物体が変形するなどの理由から、物体認識機能によって正しく認識できない場合があるが、このような場合であっても、案内された撮像方法に従ってユーザが撮像画像における物体の位置又は姿勢或いは形状を調整することで物体認識が可能となる。この撮像方法の案内は、例えば物体を模した半透明の画像と、「画像に合わせて撮ってください」というようなメッセージとを表示するような形態で実現されるから、ユーザにとって分かりやすい。また、第1カテゴリ認識機能がカバーする第1カテゴリグループと、第2カテゴリ認識機能がカバーする第2カテゴリグループとを分けているので、それぞれのカテゴリグループに属する物体についてのカテゴリ認識の精度が低下することがない。
上述した実施形態は次のような変形が可能である。また、以下の変形例を互いに組み合わせて実施してもよい。
[変形例1]
第1カテゴリ認識機能がカバーする第1カテゴリグループと、第2カテゴリ認識機能がカバーする第2カテゴリグループとを分けずに、第1カテゴリ認識部1121及び第2カテゴリ認識部1122を統合して単一の機能としてもよい。つまり、本発明を、撮像画像に含まれる物体の位置を認識する物体認識部と、撮像画像に含まれる物体が属するカテゴリを認識するカテゴリ認識部と、前記物体認識部による認識がされず前記カテゴリ認識部による認識がされた場合には、ユーザに対して撮像の方法を案内し、当該案内に従って撮像された撮像画像について前記物体認識部に認識させる認識制御部とを備える情報処理装置として構成してもよい。このようにすれば、第1カテゴリ認識機能がカバーする第1カテゴリグループと、第2カテゴリ認識機能がカバーする第2カテゴリグループとを分けて設計するためのコストを省くことができる。
実施形態において認識制御部113は、物体を模した半透明の画像を撮像画像に重畳して表示するよう撮像表示装置20に指示していたが、この撮像方法の案内は実施形態の例に限定されない。要するに、認識制御部113は、撮像表示装置20に対して、撮像画像における物体の位置を指定する画像を当該撮像画像中に表示するよう指示して、撮像方法を案内すればよい。
画像合成部30は、画像認識部11による認識結果に応じて外観が異なる第1合成画像を合成することに限らず、例えば大きさ、位置、動き又は数の少なくともいずれかが異なる第1合成画像を合成するようにしてもよい。また、第1合成画像は動画に限らず静止画であってもよい。
また、画像合成部30は、画像認識部11による認識結果に加え、ユーザの属性又は撮像表示装置20の位置又は姿勢に応じて異なる第1合成画像を合成するようにしてもよい。
また、画像合成部30は、画像認識部11による認識結果に応じて異なる第1合成画像を合成すればよく、例えば建物を撮像したときに、その建物が設置された地面の平面を認識し、その平面上に第1合成画像を合成してもよい。また、画像合成部30は、物体の部分(例えば物体が時計台であれば、その時計台の屋根、ドア、時計などが部分に相当する)が画像認識部11により認識される場合には、その部分に対する所定の位置に第1合成画像を合成してもよい。例えば、画像合成部30は、時計台の屋根に対して第1合成画像であるキャラクタ画像が飛行して着地するとか、第1合成画像であるキャラクタ画像が時計台のドアを開けるとか、第1合成画像であるキャラクタ画像が時計台の時計の針を巻き戻すというような動画のAR(Augmented Reality)画像を合成するようにしてもよい。
このように第1合成画像の様々な要素が画像認識結果に応じて異なるほうが、ユーザにとっての面白味が増す。
実施形態では、画像合成部30は、画像認識部11による認識結果に関わらず、ユーザによって撮像された顔の画像を常に第2合成画像としていたが、画像認識部11による認識結果に応じて異なる第2合成画像を合成するようにしてもよい。例えば事前にユーザが、様々に異なる表情又は向きの顔を自分で撮像して撮像表示装置20の補助記憶装置204に記憶させておき、さらに、その顔の表情又は向きから連想するユーザの様子(例えば楽しい、悲しい、笑い顔、泣き顔、知らんぷり、無表情など)を意味するタグを付与しておく。このタグの付与は、ユーザ自身が行ってもよいし、画像認識技術を用いて撮像表示装置20又は情報処理装置10が行ってもよい。画像合成部30は、画像認識部11による認識結果に対応するタグが付与された顔の画像を第2合成画像とする。この場合、画像認識部11による認識結果とユーザの様子を類型化したものとが対応付けられて補助記憶装置204に記憶されており、画像合成部30は、画像認識部11による認識結果に対応するタグを特定し、そのタグが付与された顔の画像データを補助記憶装置204から読み出す 。さらに、このような顔の表情の異同に限らず、画像合成部30は、画像認識部11による認識結果に応じて外観、大きさ、位置、動き又は数の少なくともいずれかが異なる第2合成画像を合成してもよい。なお、第2合成画像は撮像画像に限らず、例えばアバター又はアニメの画像など、どのような画像であってもよい。このように第2合成画像の様々な要素が画像認識結果に応じて異なるほうが、ユーザにとっての面白味が増す。
画像合成部30は、第1合成画像に応じて異なる第2合成画像を合成するようにしてもよい。例えば、撮像表示装置20により風景が撮像された場合、その風景に応じた第1合成画像G1が合成される。画像合成部30は、この第1合成画像G1に応じて例えば複数の第2合成画像の候補をユーザに提示し、ユーザがそのうちのいずれかを選択すると、その第2合成画像を合成する。
また、複数の第2合成画像を選択的に第1合成画像に組み合わせる例として、例えば撮像された建築物を徐々に登っていく動きと位置変化をするような、人間の体を模した動画の第1合成画像に対し、その建築物の低層においては楽しそうな表情の顔の第2合成画像を組み合わせ、高層に進むにつれて苦しそうな表情の顔の第2合成画像を組み合わせるようにしてもよい。つまり、第1合成画像の動き又は位置の変化に応じて、第2合成画像を変化させてもよい。
このように第2合成画像の様々な要素が第1合成画像に応じて異なるほうが、ユーザにとっての面白味が増す。
第1合成画像データを情報処理装置10が記憶するのではなく、撮像表示装置20が記憶しておき、情報処理装置10はどの第1合成画像データを用いるかを第1合成画像IDで撮像表示装置20に指示するだけでもよい。このようにすれば情報処理装置10及び撮像表示装置20間の通信データ量を削減することができる。
第1合成画像及び第2合成画像が合成される画像は、撮像部21によってリアルタイムに撮像された画像に限らず、例えば過去に撮像部21によって撮像された画像が補助記憶装置204に記憶されており、その中から選ばれたものであってもよいし、撮像表示装置20が他の装置から取得したものであってもよい。このようにすれば、ユーザにとって、第1合成画像及び第2合成画像が合成される対象となる画像の選択の余地が広がる。
第1合成画像又は第2合成画像を広告として利用してもよい。
例えば第1合成画像として、或るキャラクタがそのメーカのペットボトル(広告宣伝対象物)を持っている画像とか、そのキャラクタの服の真ん中にそのペットボトル(広告宣伝対象物)の絵が記載されている画像というような例が考えられる。
また、撮像表示装置20を或る飲料メーカのペットボトル(広告宣伝対象物)にかざすと、そのペットボトルの或る位置において第1合成画像として飲料メーカのキャラクタ画像が表示され、さらにそのキャラクタ画像に対して第2合成画像としてユーザの顔が配置されてもよい。
また、例えば或る街の大規模交差点に撮像表示装置20をかざすと、その交差点に設置されたスクリーンパネルの領域に或る第1合成画像を合成し、その第1合成画像上に広告宣伝対象物の動画の第2合成画像として再生するようにしてもよい。このようにすれば、あたかも、その街で広告が再生されているように見える。
これら第1合成画像及び第2合成画像の組み合わせは、ユーザの属性又は撮像表示装置20の位置によって変更してもよいし、広告主の支払う単価などにより変更してもよい。
このようにすれば、本発明を広告に利用することが可能となる。
上記実施の形態の説明に用いたブロック図は、機能単位のブロックを示している。これらの機能ブロック(構成部)は、ハードウェア及び/又はソフトウェアの任意の組み合わせによって実現される。また、各機能ブロックの実現手段は特に限定されない。すなわち、各機能ブロックは、物理的及び/又は論理的に結合した1つの装置により実現されてもよいし、物理的及び/又は論理的に分離した2つ以上の装置を直接的及び/又は間接的に(例えば、有線及び/又は無線)で接続し、これら複数の装置により実現されてもよい。
Claims (6)
- 撮像画像に含まれる物体の位置を認識する物体認識部と、
撮像画像に含まれる物体が属するカテゴリを認識するカテゴリ認識部と、
前記物体認識部による認識がされず前記カテゴリ認識部による認識がされた場合には、ユーザに対して撮像の方法を案内し、当該案内に従って撮像された撮像画像について前記物体認識部に認識させる認識制御部と
を備えることを特徴とする情報処理装置。 - 前記カテゴリ認識部は、
撮像画像に含まれる物体が属するカテゴリを認識する第1カテゴリ認識部と、
撮像画像に含まれる物体が属するカテゴリを認識する第2カテゴリ認識部であって、前記第1カテゴリ認識部が認識可能なカテゴリとは異なるカテゴリについて認識を行う第2カテゴリ認識部とを含み、
前記認識制御部は、前記物体認識部及び前記第1カテゴリ認識部による認識がされず前記第2カテゴリ認識部による認識がされた場合には、ユーザに対して撮像の方法を案内し、当該案内に従って撮像された撮像画像について前記物体認識部に認識させる
ことを特徴とする請求項1記載の情報処理装置。 - 前記認識制御部は、前記ユーザの操作に応じて撮像を行い且つ撮像画像を表示する撮像表示装置に対して、当該撮像画像における前記物体の位置を指定する画像を当該撮像画像中に表示するよう指示することで、前記ユーザに対して撮像の方法を案内する
ことを特徴とする請求項1又は2に記載の情報処理装置。 - 前記認識制御部は、前記物体を模した画像を前記撮像画像に重畳して表示するよう撮像表示装置に指示することで、前記ユーザに対して撮像の方法を案内する
ことを特徴とする請求項3に記載の情報処理装置。 - 前記物体認識部により物体の位置が認識された場合には、当該物体に応じた合成画像を、前記撮像画像において認識された当該物体の位置に応じて合成するよう、前記ユーザの操作に応じて撮像を行い且つ撮像画像を表示する撮像表示装置に指示し、前記カテゴリ認識部によりカテゴリが認識された場合には、前記撮像画像において前記カテゴリに応じた合成画像を合成するよう前記撮像表示装置に指示する画像合成指示部を備える
ことを特徴とする請求項1~4のいずれか1項に記載の情報処理装置。 - 撮像画像に含まれる物体の位置の認識を試みるステップと、
撮像画像に含まれる物体が属するカテゴリの認識を試みるステップと、
前記物体の位置が認識されず、前記物体が属するカテゴリが認識された場合に、ユーザに対して撮像の方法を案内し、当該案内に従って撮像された撮像画像に含まれる物体の認識を試みるステップと
を有する情報処理方法。
Priority Applications (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/477,577 US10977510B2 (en) | 2017-01-31 | 2017-12-05 | Information-processing device and information-processing method |
| CN201780083269.2A CN110192386B (zh) | 2017-01-31 | 2017-12-05 | 信息处理设备和信息处理方法 |
| JP2018565967A JP6750046B2 (ja) | 2017-01-31 | 2017-12-05 | 情報処理装置及び情報処理方法 |
| EP17895113.3A EP3557861A4 (en) | 2017-01-31 | 2017-12-05 | INFORMATION PROCESSING DEVICE AND INFORMATION PROCESSING METHOD |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2017015128 | 2017-01-31 | ||
| JP2017-015128 | 2017-01-31 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2018142756A1 true WO2018142756A1 (ja) | 2018-08-09 |
Family
ID=63039596
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2017/043700 Ceased WO2018142756A1 (ja) | 2017-01-31 | 2017-12-05 | 情報処理装置及び情報処理方法 |
Country Status (5)
| Country | Link |
|---|---|
| US (1) | US10977510B2 (ja) |
| EP (1) | EP3557861A4 (ja) |
| JP (1) | JP6750046B2 (ja) |
| CN (1) | CN110192386B (ja) |
| WO (1) | WO2018142756A1 (ja) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2020110323A1 (ja) * | 2018-11-30 | 2020-06-04 | 株式会社ドワンゴ | 動画合成装置、動画合成方法及び記録媒体 |
| JP2020087277A (ja) * | 2018-11-30 | 2020-06-04 | 株式会社ドワンゴ | 動画合成装置、動画合成方法及び動画合成プログラム |
Families Citing this family (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP7171349B2 (ja) * | 2018-09-28 | 2022-11-15 | 富士フイルム株式会社 | 画像処理装置、画像処理方法、プログラムおよび記録媒体 |
| US11561100B1 (en) * | 2018-10-26 | 2023-01-24 | Allstate Insurance Company | Exit routes |
| CN110941987B (zh) * | 2019-10-10 | 2023-04-07 | 北京百度网讯科技有限公司 | 目标对象识别方法、装置、电子设备及存储介质 |
| CN110944230B (zh) * | 2019-11-21 | 2021-09-10 | 北京达佳互联信息技术有限公司 | 视频特效的添加方法、装置、电子设备及存储介质 |
| CN111464743A (zh) * | 2020-04-09 | 2020-07-28 | 上海城诗信息科技有限公司 | 一种摄影构图匹配方法及系统 |
| JP7360997B2 (ja) * | 2020-06-18 | 2023-10-13 | 京セラ株式会社 | 情報処理システム、情報処理装置、および情報処理方法 |
| JP7746728B2 (ja) * | 2021-08-17 | 2025-10-01 | 富士フイルムビジネスイノベーション株式会社 | 遠隔支援システム、端末装置、遠隔装置、誘導画像表示プログラム、及び遠隔支援プログラム |
| US11792506B2 (en) | 2022-02-09 | 2023-10-17 | Motorola Mobility Llc | Electronic devices and corresponding methods for defining an image orientation of captured images |
| CN114841955A (zh) * | 2022-04-28 | 2022-08-02 | 杭州睿胜软件有限公司 | 一种生物物种的识别方法、装置、设备及存储介质 |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2011081193A1 (ja) * | 2009-12-28 | 2011-07-07 | サイバーアイ・エンタテインメント株式会社 | 画像表示システム |
| WO2011126134A1 (ja) * | 2010-04-09 | 2011-10-13 | サイバーアイ・エンタテインメント株式会社 | リアルタイム動画像収集・認識・分類・処理・配信サーバシステム |
| WO2012176317A1 (ja) * | 2011-06-23 | 2012-12-27 | サイバーアイ・エンタテインメント株式会社 | 画像認識システムを組込んだ関連性検索によるインタレスト・グラフ収集システム |
| JP2013218597A (ja) * | 2012-04-11 | 2013-10-24 | Sony Corp | 情報処理装置、表示制御方法及びプログラム |
| JP2015046071A (ja) | 2013-08-28 | 2015-03-12 | ヤフー株式会社 | 画像認識装置、画像認識方法及び画像認識プログラム |
| JP2016212827A (ja) * | 2015-04-28 | 2016-12-15 | バイドゥ オンライン ネットワーク テクノロジー (ベイジン) カンパニー リミテッド | 情報検索方法及び情報検索装置 |
Family Cites Families (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH10243274A (ja) * | 1997-02-25 | 1998-09-11 | Sanyo Electric Co Ltd | 画像撮影装置 |
| EP2357796A3 (en) * | 2002-05-28 | 2013-05-15 | Casio Computer Co., Ltd. | Composite image output apparatus and composite image delivery apparatus |
| JP2015090593A (ja) | 2013-11-06 | 2015-05-11 | ソニー株式会社 | 情報処理装置、情報処理方法および情報処理システム |
| KR102174470B1 (ko) * | 2014-03-31 | 2020-11-04 | 삼성전자주식회사 | 영상 저장 시스템 및 방법 |
-
2017
- 2017-12-05 EP EP17895113.3A patent/EP3557861A4/en not_active Withdrawn
- 2017-12-05 CN CN201780083269.2A patent/CN110192386B/zh not_active Expired - Fee Related
- 2017-12-05 US US16/477,577 patent/US10977510B2/en active Active
- 2017-12-05 JP JP2018565967A patent/JP6750046B2/ja active Active
- 2017-12-05 WO PCT/JP2017/043700 patent/WO2018142756A1/ja not_active Ceased
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2011081193A1 (ja) * | 2009-12-28 | 2011-07-07 | サイバーアイ・エンタテインメント株式会社 | 画像表示システム |
| WO2011126134A1 (ja) * | 2010-04-09 | 2011-10-13 | サイバーアイ・エンタテインメント株式会社 | リアルタイム動画像収集・認識・分類・処理・配信サーバシステム |
| WO2012176317A1 (ja) * | 2011-06-23 | 2012-12-27 | サイバーアイ・エンタテインメント株式会社 | 画像認識システムを組込んだ関連性検索によるインタレスト・グラフ収集システム |
| JP2013218597A (ja) * | 2012-04-11 | 2013-10-24 | Sony Corp | 情報処理装置、表示制御方法及びプログラム |
| JP2015046071A (ja) | 2013-08-28 | 2015-03-12 | ヤフー株式会社 | 画像認識装置、画像認識方法及び画像認識プログラム |
| JP2016212827A (ja) * | 2015-04-28 | 2016-12-15 | バイドゥ オンライン ネットワーク テクノロジー (ベイジン) カンパニー リミテッド | 情報検索方法及び情報検索装置 |
Non-Patent Citations (1)
| Title |
|---|
| See also references of EP3557861A4 |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2020110323A1 (ja) * | 2018-11-30 | 2020-06-04 | 株式会社ドワンゴ | 動画合成装置、動画合成方法及び記録媒体 |
| JP2020087277A (ja) * | 2018-11-30 | 2020-06-04 | 株式会社ドワンゴ | 動画合成装置、動画合成方法及び動画合成プログラム |
| WO2020110322A1 (ja) * | 2018-11-30 | 2020-06-04 | 株式会社ドワンゴ | 動画合成装置、動画合成方法及び記録媒体 |
| JP2020087278A (ja) * | 2018-11-30 | 2020-06-04 | 株式会社ドワンゴ | 動画合成装置、動画合成方法及び動画合成プログラム |
| US11195302B2 (en) | 2018-11-30 | 2021-12-07 | Dwango Co., Ltd. | Video synthesis device, video synthesis method and recording medium |
| US11367260B2 (en) | 2018-11-30 | 2022-06-21 | Dwango Co., Ltd. | Video synthesis device, video synthesis method and recording medium |
| US11625858B2 (en) | 2018-11-30 | 2023-04-11 | Dwango Co., Ltd. | Video synthesis device, video synthesis method and recording medium |
Also Published As
| Publication number | Publication date |
|---|---|
| CN110192386B (zh) | 2021-06-11 |
| US20190370593A1 (en) | 2019-12-05 |
| EP3557861A4 (en) | 2019-12-25 |
| US10977510B2 (en) | 2021-04-13 |
| EP3557861A1 (en) | 2019-10-23 |
| JPWO2018142756A1 (ja) | 2019-11-07 |
| JP6750046B2 (ja) | 2020-09-02 |
| CN110192386A (zh) | 2019-08-30 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP6750046B2 (ja) | 情報処理装置及び情報処理方法 | |
| KR102503413B1 (ko) | 애니메이션 인터랙션 방법, 장치, 기기 및 저장 매체 | |
| CN109688451B (zh) | 摄像机效应的提供方法及系统 | |
| CN110716645A (zh) | 一种增强现实数据呈现方法、装置、电子设备及存储介质 | |
| CN113721804A (zh) | 一种显示方法、装置、电子设备及计算机可读存储介质 | |
| US12519924B2 (en) | Multi-perspective augmented reality experience | |
| JP6730461B2 (ja) | 情報処理システム及び情報処理装置 | |
| US12417593B2 (en) | Generating immersive augmented reality experiences from existing images and videos | |
| WO2016122973A1 (en) | Real time texture mapping | |
| EP4581468A1 (en) | Multi-dimensional experience presentation using augmented reality | |
| JP2023515988A (ja) | 拡張現実ゲスト認識システム及び方法 | |
| US20240362743A1 (en) | Late warping to minimize latency of moving objects | |
| CN116943191A (zh) | 基于故事场景的人机互动方法、装置、设备及介质 | |
| WO2024130194A1 (en) | Generating augmented reality content including translations | |
| US12067693B2 (en) | Late warping to minimize latency of moving objects | |
| CN116993949A (zh) | 虚拟环境的显示方法、装置、可穿戴电子设备及存储介质 | |
| CN116977157A (zh) | 一种图像处理方法、装置、设备、介质及程序产品 | |
| JP6857537B2 (ja) | 情報処理装置 | |
| KR20230081695A (ko) | 증강현실 제공 장치 | |
| KR20230081696A (ko) | 증강현실 제공 장치 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17895113 Country of ref document: EP Kind code of ref document: A1 |
|
| ENP | Entry into the national phase |
Ref document number: 2018565967 Country of ref document: JP Kind code of ref document: A |
|
| ENP | Entry into the national phase |
Ref document number: 2017895113 Country of ref document: EP Effective date: 20190715 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| WWW | Wipo information: withdrawn in national office |
Ref document number: 2017895113 Country of ref document: EP |