WO2014064325A1 - Système de re-mélange de milieu - Google Patents
Système de re-mélange de milieu Download PDFInfo
- Publication number
- WO2014064325A1 WO2014064325A1 PCT/FI2012/051033 FI2012051033W WO2014064325A1 WO 2014064325 A1 WO2014064325 A1 WO 2014064325A1 FI 2012051033 W FI2012051033 W FI 2012051033W WO 2014064325 A1 WO2014064325 A1 WO 2014064325A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user devices
- event
- cluster
- user device
- sensor data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
- H04S7/303—Tracking of listener position or orientation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R27/00—Public address systems
Definitions
- Multimedia capturing capabilities have become common features in portable devices.
- an event such as a music concert, a sport event or a private event such as a birthday or a wedding, they are attending.
- there are multiple attendants capturing content from an event whereby variations in capturing location, view, equipment, etc. result in a plurality of captured versions of the event with a high amount of variety in both the quality and the content of the captured media.
- Media remixing is an application where multiple media recordings are combined in order to obtain a media mix that contains some segments selected from the plurality of media recordings.
- Video remixing is one of the basic manual video editing applications, for which various software products and services are already available.
- automatic video remixing or editing systems which use multiple instances of user-generated or professional recordings to automatically generate a remix that combines content from the available source content.
- Some automatic video remixing systems depend only on the recorded content, while others are capable of utilizing environmental context data that is recorded together with the video content.
- the context data may be, for example, sensor data received from a compass, an accelerometer, or a gyroscope, or global positioning system (GPS) location data.
- GPS global positioning system
- a generic event such as a music concert, sports event, etc.
- there may be multiple users in the audience capturing media content from the event i.e. recording videos, audio clips and/or taking pictures.
- Some users capturing media content may be close to each other, whereas others may be further away.
- commonality is redundant media content, which unnecessarily consumes network resources when uploaded or up- streamed and data storage resource when stored e.g. in the content management system.
- a method comprising: receiving sensor data from a plurality of user devices attending an event; generating at least a spatial distribution of the user devices on the basis of the sensor data; dividing the user devices into clusters of one or more user devices on the basis of the spatial distribution; and selecting a representative user device for a cluster to represent the audio scene for the area of the cluster to which the representative user device belongs.
- the method further comprises determining a threshold distance for each cluster such that the user devices located within the threshold distance belong to the same cluster.
- parameters affecting on the determination of the threshold distance for the clusters include one or more of the following:
- the method further comprises generating a temporal distribution of the user devices on the basis of the sensor data.
- the sensor data includes at least one of the following:
- the method further comprises generating the one or more distributions of the user devices on the basis of the sensor data received from user devices recording content from the event.
- the method further comprises receiving sensor data updates from the plurality of user devices attending the event; and updating the one or more distributions of the user devices on the basis of the updated sensor data.
- selecting the representative user device for a cluster comprises determining the best quality audio from among the user devices in the cluster; and determining the best audio scene perspective among the user devices in the cluster by comparing location of interesting parts of the event and orientation of the user devices for a majority of interesting parts.
- the method further comprises uploading or up- streaming, within a particular cluster, captured media content from the selected representative user device only.
- the method further comprises synchronizing device clocks of the user devices recording media content from the event to each other.
- audio quality of an audio track captured by a user device is analyzed by choosing, for a given audio track, a random sampling position for analyzing the quality of a small temporal segment of the audio; in response to the small segment being of good quality, choosing a subsequent sampling position for an analysis; and
- choosing the subsequent sampling position is performed by using a half-interval search.
- an apparatus comprising at least one processor, memory including computer program code, the memory and the computer program code configured to, with the at least one processor, cause the apparatus to at least: receiving sensor data from a plurality of user devices attending an event; generating at least a spatial distribution of the user devices on the basis of the sensor data; dividing the user devices into clusters of one or more user devices on the basis of the spatial distribution; and selecting a representative user device for a cluster to represent the audio scene for the area of the cluster to which the representative user device belongs.
- a computer program embodied on a non-transitory computer readable medium, the computer program comprising instructions causing, when executed on at least one processor, at least one apparatus to: receiving sensor data from a plurality of user devices attending an event; generating at least a spatial distribution of the user devices on the basis of the sensor data; dividing the user devices into clusters of one or more user devices on the basis of the spatial distribution; and selecting a representative user device for a cluster to represent the audio scene for the area of the cluster to which the representative user device belongs.
- a computer readable storage medium stored with code thereon for use by an apparatus, which when executed by a processor, causes the apparatus to perform: receiving sensor data from a plurality of user devices attending an event; generating at least a spatial distribution of the user devices on the basis of the sensor data; dividing the user devices into clusters of one or more user devices on the basis of the spatial distribution; and selecting a representative user device for a cluster to represent the audio scene for the area of the cluster to which the representative user device belongs.
- a system comprising at least one processor, memory including computer program code, the memory and the computer program code configured to, with the at least one processor, cause the system to at least: receive sensor data from a plurality of user devices attending an event; generate at least a spatial distribution of the user devices on the basis of the sensor data; divide the user devices into clusters of one or more user devices on the basis of the spatial distribution; and select a representative user device for a cluster to represent the audio scene for the area of the cluster to which the representative user device belongs.
- an apparatus comprising: means for receiving sensor data from a plurality of user devices attending an event; means for generating at least a spatial distribution of the user devices on the basis of the sensor data; means for dividing the user devices into clusters of one or more user devices on the basis of the spatial distribution; and means for selecting a representative user device for a cluster to represent the audio scene for the area of the cluster to which the representative user device belongs.
- Figs. 1 a and 1 b show a system and devices suitable to be used in an automatic media remixing service according to an embodiment
- Fig. 2 shows an exemplified service architecture for creating a media remix
- Fig. 3 shows an exemplified implementation of a method according to some of the embodiments in a media remix application
- Fig. 4 shows, according to an embodiment, a process of clustering user devices in an event
- Fig. 5 shows, according to an embodiment, a method for selective temporal- segment sampling of an audio track performed on a user device to determine an approximate quality of the audio
- Figs. 6a - 6c show an example of the temporal segment sampling process of Figure
- UGC user generated content
- SMP social media portals
- the media content to be used in media remixing services may comprise at least video content including 3D video content, still images (i.e. pictures), and audio content including multi-channel audio content.
- the embodiments disclosed herein are mainly described from the viewpoint of creating an automatic media remix from video and audio content of source videos, but the embodiments are not limited to video and audio content of source videos, but they can be applied generally to any type of media content.
- Figs. 1 a and 1 b show a system and devices suitable to be used in an automatic media remixing service according to an embodiment. In Fig.
- the different devices may be connected via a fixed network 210 such as the Internet or a local area network; or a mobile communication network 220 such as the Global System for Mobile communications (GSM) network, 3rd Generation (3G) network, 3.5th Generation (3.5G) network, 4th Generation (4G) network, Wireless Local Area Network (WLAN), Bluetooth®, or other contemporary and future networks.
- GSM Global System for Mobile communications
- 3G 3rd Generation
- 3.5G 3.5th Generation
- 4G 4th Generation
- WLAN Wireless Local Area Network
- Bluetooth® Wireless Local Area Network
- Different networks are connected to each other by means of a communication interface 280.
- the networks comprise network elements such as routers and switches to handle data (not shown), and communication interfaces such as the base stations 230 and 231 in order for providing access for the different devices to the network, and the base stations 230, 231 are themselves connected to the mobile network 220 via a fixed connection 276 or a wireless connection 277.
- servers 240, 241 and 242 each connected to the mobile network 220, which servers may be arranged to operate as computing nodes (i.e. to form a cluster of computing nodes or a so-called server farm) for the automatic media remixing service.
- Some of the above devices for example the computers 240, 241 , 242 may be such that they are arranged to make up a connection to the Internet with the communication elements residing in the fixed network 210.
- the various devices may be connected to the networks 210 and 220 via communication connections such as a fixed connection 270, 271 , 272 and 280 to the internet, a wireless connection 273 to the internet 210, a fixed connection 275 to the mobile network 220, and a wireless connection 278, 279 and 282 to the mobile network 220.
- Fig. 1 b shows devices for automatic media remixing according to an example embodiment.
- the server 240 contains memory 245, one or more processors 246, 247, and computer program code 248 residing in the memory 245 for implementing, for example, automatic media remixing.
- the different servers 241 , 242, 290 may contain at least these elements for employing functionality relevant to each server.
- the end-user device 251 contains memory 252, at least one processor 253 and 256, and computer program code 254 residing in the memory 252 for implementing, for example, gesture recognition.
- the end-user device may also have one or more cameras 255 and 259 for capturing image data, stereo video, 3D video or alike.
- the end-user device may also contain one, two or more microphones 257 and 258 for capturing sound.
- the end-user device may also contain sensors for generating the depth information using any suitable technology.
- the different end- user devices 250, 260 may contain at least these same elements for employing functionality relevant to each device.
- the depth maps i.e.
- the end-user device may also have a time-of-flight camera, whereby the depth map may be obtained from a time-of-flight camera or from a combination of stereo (or multiple) view depth map and a time-of-flight camera.
- the end-user device may generate depth map for the captured content using any available and suitable mechanism.
- the end user devices may also comprise a screen for viewing single-view, stereoscopic (2-view), or multiview (more-than-2-view) images.
- the end-user devices may also be connected to video glasses 290, e.g., by means of a communication block 293 able to receive and/or transmit information.
- the glasses may contain separate eye elements 291 and 292 for the left and right eye. These eye elements may either show a picture for viewing, or they may comprise a shutter functionality e.g., to block every other picture in an alternating manner to provide the two views of three-dimensional picture to the eyes, or they may comprise an orthogonal polarization filter (compared to each other), which, when connected to similar polarization realized on the screen, provide the separate views to the eyes.
- Stereoscopic or multiview screens may also be autostereoscopic, i.e., the screen may comprise or may be overlaid by an optics arrangement, which results into a different view being perceived by each eye.
- Single-view, stereoscopic, and multiview screens may also be operationally connected to viewer tracking such a manner that the displayed views depend on viewer's position, distance, and/or direction of gaze relative to the screen.
- parallelized processes of the automatic media remixing may be carried out in one or more processing devices; i.e., entirely in one user device like 250, 251 or 260, or in one server device 240, 241 , 242 or 290, or across multiple user devices 250, 251 , 260 or across multiple network devices 240, 241 , 242, 290, or across both user devices 250, 251 , 260 and network devices 240, 241 , 242, 290.
- the elements of the automatic media remixing process may be implemented as a software component residing on one device or distributed across several devices, as mentioned above, for example so that the devices form a so-called cloud.
- One or more of the computers disclosed in Fig. 1 a may be configured to operate a multimedia content remix service, which can be referred to as a media remix service.
- the media remix service is a service infrastructure that is capable of receiving user communication requests for inviting other users.
- the media remix service, together with the computer(s) running the service, further comprise networking capability to receive and process media content and corresponding context data from other data processing devices, such as servers operating social media portals (SMP).
- SMP social media portal
- UGC user generated content
- the UGC media content can be stored in various formats, for example, using the formats described in the Moving Picture Experts Group MPEG-4 standard.
- the context data may be stored in suitable fields in the media data container file formats, or in separate files with database entries or link files associating the media files and their timestamps with sensor information and their timestamps.
- Some examples of popular SMPs are YouTube, Flickr®, and PicasaTM. It is apparent for a skilled person that the media remix service and the social media portals SMP are implemented as network domains, wherein the operation may be distributed among a plurality of servers.
- a media remix can be created according to the preferences of a user.
- the source content refers to all types of media that is captured by users, wherein the source content may involve any associated context data.
- videos, images, audio captured by users may be provided with context data, such as information from various sensors, such as from a compass, an accelerometer, a gyroscope, or information indicating location, altitude, temperature, illumination, pressure, etc.
- a particular sub-type of source content is a source video, which refers to videos captured by the user, possibly provided with the above-mentioned context information.
- a user can request the media remix service an automatically created media remix version from the material available for the service about an event, such as a concert.
- the service may be available to any user or it may be limited to registered users only. It is also possible to create a media remix version from private video material only.
- the service creates an automatic cut of the video clips of the users.
- the service may analyze the sensory data to determine which are interesting points at each point in time during the event, and then makes switches between different source media in the final cut. Audio alignment is used to find a common timeline for all the source videos, and, for example, dedicated sensor data (accelerometer, compass) analysis algorithms are used to detect when several users are pointing to the same location on the stage, most likely indicating an interesting event.
- music content analysis (beats, downbeats), is used to find a temporal grid of potential cut points in the event sound track.
- Fig. 2 shows exemplified service architecture for creating an automatically created media remix.
- the service architecture may include components, known as such from contemporary video editing services, for example an interface 200 for the users contributing their recorded content from the event, which interface may annotate the contributed content for clustering the content related to the same event for generating the media remix, a content management system (CMS; 202) to store/tag/organize the content, and an interface 204 for delivering the media remix and its related source content to the users to consume.
- CMS content management system
- the service architecture of Fig. 2 may further comprise a feedback module (FBM; 206) to capture the content consumption feedback about the content contributed by the users and the media remix versions that have been generated.
- the feedback information may be provided to a synergistic intelligence module (SIM; 208), which contains the required intelligence or the logic required to analyze and create the information about the user contributed source content that is contributed to the service.
- SIM is connected to a user apparatus 214 via a signalling interface 212, which enables the user to request a media remix to be created according to user- defined parameters and also to provide new UGC content to be used in the media remix generation process.
- the SIM may utilize, in addition to the feedback information, also information about the arrival distribution pattern of the source content.
- the SIM may use the UGC contribution data from past events in various locations and use it to generate a probabilistic model to predict user content contribution's arrival time (or upload time) to the service.
- the information provided by the SIM are received in a synergizing engine (SE; 210), which may be implemented as a separate module that interacts with the CMS, the SIM and the FBM to generate the media remix versions that match the criteria signalled by the user requesting a media remix.
- SE synergizing engine
- the information provided by the SIM enables the SE to utilize the previous media remix versions and their consumption feedback as inputs, in addition to the newly provided source content and its consumption feedback, wherein the SE changes the weights of different parameters which are used to combine the multitude of content.
- a generic event such as a music concert, sports event, etc.
- there may be multiple users in the audience capturing media content from the event i.e., recording videos, audio clips and/or taking pictures.
- the presence of multiple users recording at the event implies that there may be some redundancy in the captured content in such a way that there may be multiple users recording at the same time during some time intervals, there may be just one user recording at some time intervals or there may be no users recording at some other time intervals during the event.
- Some users capturing media content may be close to each other, whereas others may be further away. There is a higher likelihood of the users that are close to each other having higher commonality in the audio scene and thus in the captured media than users that are far apart.
- a spatial sampling method is now presented for selecting a subset of all the user devices capturing media content from the event as candidate audio scene representatives based on their spatial distribution.
- the user devices are clustered such that user devices that are closer than a predefined threshold distance to each other are considered to belong to same cluster.
- the predefined threshold distance may be modulated based on the spatial expanse of the event.
- One user device is selected as an audio scene representative from each cluster.
- a method according to some of the embodiments is illustrated in the flow chart of Figure 3, wherein the operation is described from the perspective of a media remix application, typically executed on one or more servers in a network.
- the media remix application receives sensor data from a plurality of user devices attending an event (300).
- the sensor data may include at least one or more of the following:
- the position may be determined, for example, using satellite positioning system such as GPS
- the position may also be determined as relative position to other user devices or a reference point.
- orientation information indicating the orientation of the user device with relation to the magnetic north.
- the orientation information may be determined, for example, using a compass.
- the 3d space position may be determined using a gyroscope.
- the media remix application On the basis of the sensor data collected from the plurality of user devices participating in the event, the media remix application generates at least a spatial and optionally a temporal distribution of the user devices (302). According to an embodiment, only those user devices that are recording the media content from the event are considered, when generating update the spatial and the temporal distribution of the user devices.
- the user devices may continuously update their sensor data and the media remix application may consequently update the spatial and the temporal distribution of the users.
- Information on the spatial and the temporal distribution of the users is used to divide the user devices into clusters of one or more user devices (304).
- the clusters may be formed such that the user devices located within a predefined threshold distance to each other are determined to form a cluster.
- Parameters affecting on the determination of the threshold distance for the clusters may include one or more of the following:
- the size of the event venue The bigger is the size of the venue, the larger threshold would likely be needed for cluster formation.
- a representative user device is selected (306) from all candidate user devices of the cluster to represent the audio scene for the area encompassed by the cluster to which the user device belongs.
- selecting the representative user device may consist of determination of the best quality audio from among the candidates in the cluster and determination of the best audio scene perspective by comparing the location of the interesting parts of the event and the candidates' orientation for the majority of the considered temporal interval.
- the best candidate is chosen to represent the audio scene.
- the captured media content is then preferably uploaded or up-streamed from the selected representative user device only. This results in avoidance of uploading or up-streaming and processing of content from other user devices in the cluster.
- the device clocks of the user devices recording media content from the event are synchronized to each other.
- the sensor data received from the user devices may be timely synchronized, which enables the media remix application to update the spatial and the temporal distribution of the users, the confirmation of the clusters and the representative of the cluster flexibly and in timely manner.
- the process of clustering user devices is illustrated in an example of Figure 4, wherein an audience is gathered in an event, for example a music concert. Within the audience, altogether 16 user devices recording media content from the event are detected. The recording user devices send their sensor data to the server comprising the media remix application. On the basis of the sensor data and the parameters derived therefrom, the media remix application determines the threshold distances for clustering the user devices.
- cluster 1 consisting of user devices 1 , 2, 3 and 4
- C2 consisting of user devices 1 1 and 12, C3 consisting of user devices 5 and 6, and C4 consisting of user devices 7, 8 and 9.
- User devices 10, 13, 14, 15 and 16 are determined as clusters of single user device.
- Figure 4 illustrates well how the threshold distance may vary between the clusters within the event venue. For example, due to parameter changes in audio level of the common ambient audio scene in the event, regional variations of the audio level within the event and/or variation in density of the users, the threshold distance for cluster 1 (C1 ) is larger than the threshold distance of cluster 3 (C3). For the single user clusters, the threshold distance is determined to be so small that no other user devices are located within the threshold distance.
- An important aspect in the process of determining the representative user device with the best quality audio from among the candidate user devices in the cluster is a method for selective temporal-segment sampling of an audio track from the candidate user devices in the cluster is performed locally on each user device to determine an approximate quality of the audio.
- the segments of audio that are of not good enough quality may be determined.
- the method is illustrated in the flow chart of Figure 5. For a given audio track, a random sampling position is chosen for analyzing the quality of a small temporal segment of the audio (500). If the analysis of the small segment results in (502) determining the segment of good quality, the analysis may be continued by choosing a subsequent sampling position (500). For choosing the subsequent sampling position, any well-known pattern, such as a half-interval search may be used.
- the steps (500, 502) are repeated N times such that a counter value is increased by one (504) for every good quality segment, and if the result of the first N evaluations is good quality (506), the audio track may be classified as a good quality audio track (508).
- a binary search is performed (510) before and after the bad quality segment. If the binary search reveals further bad quality segments (512), the bad quality segments may be subjected to a further binary search, until the width of the bad quality audio content in the audio track can be localized. The bad quality audio content in the audio track may then be discarded (514).
- FIGS 6a - 6c An example of the temporal segment sampling process is illustrated in Figures 6a - 6c, wherein the audio quality from user device 1 in cluster C1 is determined. Within a time interval T, initially only a small central temporal segment (S1 ) is chosen for analysis, as shown in Figure 6a. If the segment S1 is determined to be of good quality, the audio for the whole time interval T is selected for further analysis. Next, two other temporal segments (S2, S3) may be chosen for analysis at the quarter and three-quarter intervals, as shown in Figure 6b.
- S1 central temporal segment
- S2 two other temporal segments
- segment S3 turns out to be of bad quality, which results in further analysis in two temporal segments (S4, S5), one on both sides of segment S3, as shown in Figure 6c.
- Segment S4 is determined to be of good quality
- segment S5 is of bad quality.
- an audio segment starting from the beginning of the time interval T up to the segment S4 included is determined to be good quality content
- an audio segment of the rest of the time interval T, starting after segment S4, is determined as content to be skipped.
- an audio segment from another user device in cluster C1 may be selected to represent the audio scene of cluster C1 .
- the number of iteration rounds N may vary depending on the required level of accuracy in granularity. According to an embodiment, the number of iteration rounds N may optimized in terms of the amount of analysis that needs to be performed on the user device and the savings in content upload or up-stream.
- the spatial and temporal sampling of media can be used to generate an audio representation of the ambient audio scene at the event with minimal upload or upstream of audio content.
- the media remix service has been described above as implemented in a client- server-type media remix service.
- the implementation is not limited to client- server-type system, but according to an embodiment, the media remix service may be implemented as peer-to-peer-type system, where the processing steps described above are performed on one or more user devices.
- the system is client-server-type, but at least some of the steps described above are performed on the user device.
- the various embodiments may provide advantages over state of the art. For example, for creating a media remix from an event, audio content upload or up- stream requirements may be minimized while preserving audio scene information from the event. The minimized amount of the uploaded or up-streamed content from the event may enable faster collaborative media remixing. Since only a few user devices upload or up-stream the captured media content to the media remix application, significant power savings on average may be achieved in the user devices and bandwidth consumption savings on various network connections. The power savings may enable longer battery life in the user devices.
- the various embodiments of the invention may be implemented in hardware or special purpose circuits, software, logic or any combination thereof.
- some aspects may be implemented in hardware, while other aspects may be implemented in firmware or software which may be executed by a controller, microprocessor or other computing device, although the invention is not limited thereto.
- firmware or software which may be executed by a controller, microprocessor or other computing device, although the invention is not limited thereto.
- While various aspects of the invention may be illustrated and described as block diagrams, flow charts, or using some other pictorial representation, it is well understood that these blocks, apparatus, systems, techniques or methods described herein may be implemented in, as non-limiting examples, hardware, software, firmware, special purpose circuits or logic, general purpose hardware or controller or other computing devices, or some combination thereof.
- the embodiments of this invention may be implemented by computer software executable by a data processor of the mobile device, such as in the processor entity, or by hardware, or by a combination of software and hardware.
- any blocks of the logic flow as in the Figures may represent program steps, or interconnected logic circuits, blocks and functions, or a combination of program steps and logic circuits, blocks and functions.
- the software may be stored on such physical media as memory chips, or memory blocks implemented within the processor, magnetic media such as hard disk or floppy disks, and optical media such as for example DVD and the data variants thereof, or CD.
- the memory may be of any type suitable to the local technical environment and may be implemented using any suitable data storage technology, such as semiconductor based memory devices, magnetic memory devices and systems, optical memory devices and systems, fixed memory and removable memory.
- the data processors may be of any type suitable to the local technical environment, and may include one or more of general purpose computers, special purpose computers, microprocessors, digital signal processors (DSPs) and processors based on multi core processor architecture, as non-limiting examples.
- Embodiments of the inventions may be practiced in various components such as integrated circuit modules.
- the design of integrated circuits is by and large a highly automated process. Complex and powerful software tools are available for converting a logic level design into a semiconductor circuit design ready to be etched and formed on a semiconductor substrate.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Information Transfer Between Computers (AREA)
Abstract
La présente invention porte sur un procédé comprenant : recevoir des données de capteur provenant d'une pluralité de dispositifs d'utilisateur assistant à un évènement ; générer au moins une distribution spatiale des dispositifs d'utilisateur sur la base des données de capteurs ; diviser les dispositifs d'utilisateur en groupes d'un ou plusieurs dispositifs d'utilisateur sur la base de la distribution spatiale ; et sélectionner un dispositif d'utilisateur représentatif pour un groupe pour représenter la scène audio pour la zone du groupe auquel le dispositif d'utilisateur représentatif appartient.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/FI2012/051033 WO2014064325A1 (fr) | 2012-10-26 | 2012-10-26 | Système de re-mélange de milieu |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/FI2012/051033 WO2014064325A1 (fr) | 2012-10-26 | 2012-10-26 | Système de re-mélange de milieu |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2014064325A1 true WO2014064325A1 (fr) | 2014-05-01 |
Family
ID=50544078
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/FI2012/051033 Ceased WO2014064325A1 (fr) | 2012-10-26 | 2012-10-26 | Système de re-mélange de milieu |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2014064325A1 (fr) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP3198721A4 (fr) * | 2014-09-23 | 2018-05-30 | Denton, Levaughn | Procédé et appareil d'ajustement audio à base de grappe mobile |
| EP3349480B1 (fr) * | 2017-01-16 | 2020-09-02 | Vestel Elektronik Sanayi ve Ticaret A.S. | Afficheur vidéo et procédé de fonctionnement correspondant |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6850496B1 (en) * | 2000-06-09 | 2005-02-01 | Cisco Technology, Inc. | Virtual conference room for voice conferencing |
| US20050193421A1 (en) * | 2004-02-26 | 2005-09-01 | International Business Machines Corporation | Method and apparatus for cooperative recording |
| EP1613124A2 (fr) * | 2004-06-30 | 2006-01-04 | Polycom, Inc. | Traitement des signaux stéréo de microphone pour le système de téléconférences |
| WO2009026347A1 (fr) * | 2007-08-21 | 2009-02-26 | Syracuse University | Système et procédé pour enregistrement audio distribué et mélange collaboratif |
| US20090087161A1 (en) * | 2007-09-28 | 2009-04-02 | Graceenote, Inc. | Synthesizing a presentation of a multimedia event |
| US20100183280A1 (en) * | 2008-12-10 | 2010-07-22 | Muvee Technologies Pte Ltd. | Creating a new video production by intercutting between multiple video clips |
| WO2012028902A1 (fr) * | 2010-08-31 | 2012-03-08 | Nokia Corporation | Appareil de scène audio |
-
2012
- 2012-10-26 WO PCT/FI2012/051033 patent/WO2014064325A1/fr not_active Ceased
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6850496B1 (en) * | 2000-06-09 | 2005-02-01 | Cisco Technology, Inc. | Virtual conference room for voice conferencing |
| US20050193421A1 (en) * | 2004-02-26 | 2005-09-01 | International Business Machines Corporation | Method and apparatus for cooperative recording |
| EP1613124A2 (fr) * | 2004-06-30 | 2006-01-04 | Polycom, Inc. | Traitement des signaux stéréo de microphone pour le système de téléconférences |
| WO2009026347A1 (fr) * | 2007-08-21 | 2009-02-26 | Syracuse University | Système et procédé pour enregistrement audio distribué et mélange collaboratif |
| US20090087161A1 (en) * | 2007-09-28 | 2009-04-02 | Graceenote, Inc. | Synthesizing a presentation of a multimedia event |
| US20100183280A1 (en) * | 2008-12-10 | 2010-07-22 | Muvee Technologies Pte Ltd. | Creating a new video production by intercutting between multiple video clips |
| WO2012028902A1 (fr) * | 2010-08-31 | 2012-03-08 | Nokia Corporation | Appareil de scène audio |
Non-Patent Citations (1)
| Title |
|---|
| HIMAWAN, IVAN ET AL.: "Clustering of ad hoc microphone arrays for robust blind beamforming", PROCEEDINGS OF 2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS SPEECH AND SIGNAL PROCESSING (ICASSP), pages 2814 - 2817 * |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP3198721A4 (fr) * | 2014-09-23 | 2018-05-30 | Denton, Levaughn | Procédé et appareil d'ajustement audio à base de grappe mobile |
| EP4054211A1 (fr) * | 2014-09-23 | 2022-09-07 | Denton, Levaughn | Procédé et appareil d'ajustement audio à base de grappe mobile |
| EP3349480B1 (fr) * | 2017-01-16 | 2020-09-02 | Vestel Elektronik Sanayi ve Ticaret A.S. | Afficheur vidéo et procédé de fonctionnement correspondant |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11381739B2 (en) | Panoramic virtual reality framework providing a dynamic user experience | |
| US9940970B2 (en) | Video remixing system | |
| US10721439B1 (en) | Systems and methods for directing content generation using a first-person point-of-view device | |
| US9570111B2 (en) | Clustering crowdsourced videos by line-of-sight | |
| US20220256231A1 (en) | Systems and methods for synchronizing data streams | |
| JP6654134B2 (ja) | マルチビューオーディオおよびビデオインタラクティブプレイバック | |
| CN104012106B (zh) | 使表示不同视点的视频对准 | |
| US20130259447A1 (en) | Method and apparatus for user directed video editing | |
| US9942533B2 (en) | Method and apparatus for generating multi-channel video | |
| US20130259446A1 (en) | Method and apparatus for user directed video editing | |
| EP2724343B1 (fr) | Système de remixage vidéo | |
| US10574715B2 (en) | Method and system for aggregating content streams based on sensor data | |
| EP2816564B1 (fr) | Procédé et appareil de rendu vidéo intelligent | |
| WO2010096282A2 (fr) | Partage de vidéo | |
| US20150208000A1 (en) | Personalized media remix | |
| WO2014064325A1 (fr) | Système de re-mélange de milieu | |
| EP3350720A1 (fr) | Procédés et appareil de capture et de présentation d'informations | |
| WO2013155708A1 (fr) | Système assurant la fonction de zoomage sélectif et intelligent dans un flux media issu d'une collaboration | |
| US20150074123A1 (en) | Video remixing system | |
| Mate | Automatic Mobile Video Remixing and Collaborative Watching Systems | |
| WO2014037604A1 (fr) | Remélange d'un média multisource | |
| US20130169828A1 (en) | Method and apparatus for tagging location information | |
| HK40044566A (en) | Synchronising data streams | |
| DTO et al. | Deliverable D6. |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12886981 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 12886981 Country of ref document: EP Kind code of ref document: A1 |