WO2014081384A1 - Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application - Google Patents
Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application Download PDFInfo
- Publication number
- WO2014081384A1 WO2014081384A1 PCT/SG2012/000439 SG2012000439W WO2014081384A1 WO 2014081384 A1 WO2014081384 A1 WO 2014081384A1 SG 2012000439 W SG2012000439 W SG 2012000439W WO 2014081384 A1 WO2014081384 A1 WO 2014081384A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- head
- application program
- input indicating
- parameter
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04847—Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0486—Drag-and-drop
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S1/00—Two-channel systems
- H04S1/002—Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
- H04S1/005—For headphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/01—Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]
Definitions
- Various embodiments generally relate to methods for outputting a modified audio signal and to graphical user interfaces produced by an application program.
- a head-reflectance transfer function (or head-related transfer function; HRTF) may be applied to an incoming analog stereo audio signal in order to create the illusion of a multi-channel audio system through typical stereo headphones.
- This HRTF may have to be calibrated to a specific user.
- a method for outputting a modified audio signal may be provided.
- the method may include: receiving from a user an input indicating an angle; determining a parameter for a head-related transfer function based on the received input indicating the angle; modifying an audio signal in accordance with the head-related transfer function based on the determined parameter; and outputting the modified audio signal.
- a graphical user interface produced by an application program may be provided.
- the graphical user interface may include: an application program window generated by the application program, wherein the application program window may include: a visual representation of the user; a visual representation of a speaker on a geometric shape around the user; and an input for inputting an indication of an angle on the geometric shape with respect to the visual representation of the speaker.
- FIG. 1 shows a flow diagram illustrating a method for outputting a modified audio signal in accordance with an embodiment
- FIG. 2 shows an audio output device in accordance with an embodiment
- FIG. 3 shows a graphical user interface in accordance with an embodiment
- FIG. 4 shows a graphical user interface in accordance with an embodiment
- FIG. 5 A shows a diagram of an application program window in accordance with one embodiment
- FIG. 5B shows a plurality of ear shapes
- FIG. 5 shows a screen shot of a graphical user interface for calibrating virtual speaker positions in accordance with an embodiment
- FIG. 6A shows a screen shot of a graphical user interface for calibrating virtual speaker positions in accordance with an embodiment
- FIG. 6B shows a screen shot of a graphical user interface in accordance with an embodiment, wherein a virtual speaker location marker is shown when the virtual speaker location 616 is selected;
- FIG. 7 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein an audio output device may be set;
- FIG. 8 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein general audio output parameters may be set;
- FIG. 9 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein equalizer parameters may be set;
- FIG. 10 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein the position of a virtual speaker may be adjusted;
- FIG. 11 shows a screen shot of a graphical user interface in accordance with an embodiment, wherein a marker indicating an angle for the chosen virtual speaker may be set;
- FIG. 12 shows a screen shot of a graphical user interface in accordance with an embodiment, wherein a marker indicating an angle for the chosen virtual speaker may be set;
- FIG. 13 shows a screen shot of a graphical user interface or application program window showing alternative representation of the speakers
- FIG. 14 shows an application window according to an embodiment
- the audio output device may include a memory which is for example used in the processing carried out by the audio output device.
- a memory used in the embodiments may be a volatile memory, for example a DRAM (Dynamic Random Access Memory) or a non- volatile memory, for example a PROM (Programmable Read Only Memory), an EPROM (Erasable PROM), EEPROM (Electrically Erasable PROM), or a flash memory ⁇ e.g., a floating gate memory, a charge trapping memory, an MRAM (Magnetoresistive Random Access Memory) or a PCRAM (Phase Change Random Access Memory).
- DRAM Dynamic Random Access Memory
- PROM Programmable Read Only Memory
- EPROM Erasable PROM
- EEPROM Electrical Erasable PROM
- flash memory e.g., a floating gate memory, a charge trapping memory, an MRAM (Magnetoresistive Random Access Memory) or a PCRAM (Phase Change Random Access Memory).
- a “circuit” may be understood as any kind of a logic implementing entity, which may be special purpose circuitry or a processor executing software stored in a memory, firmware, or any combination thereof.
- a “circuit” may be a hard-wired logic circuit or a programmable logic circuit such as a programmable processor, e.g. a microprocessor (e.g. a Complex Instruction Set Computer (CISC) processor or a Reduced Instruction Set Computer (RISC) processor).
- a “circuit” may also be a processor executing software, e.g. any kind of computer program, e.g. a computer program using a virtual machine code such as e.g. Java.
- circuit any other kind of implementation of the respective functions which will be described in more detail below may also be understood as a “circuit” in accordance with an alternative embodiment. It will be understood that what is described herein as circuits with different names (for example “circuit A” and “circuit B”) may also be provided in one physical circuit like described above.
- a geometric shape may be or may include a conic section, for example a circle, an ellipse, a parabola or a hyperbola, or may include or may be a polygon, or may include or may be any other kind of geometric shape.
- FIG. 1 shows a flow diagram 100 illustrating a method for outputting a modified audio signal in accordance with an embodiment.
- an input indicating an angle may be received from a user.
- a parameter (or a plurality of parameters) for a head-related transfer function may be determined based on the received input indicating the angle.
- an audio signal may be modified in accordance with the head-related transfer function based on the determined parameter (or the plurality of determined parameters).
- the modified audio signal may be output.
- the input indicating the angle may be a graphical input indicating the angle by a point on a geometric shape.
- the input indicating the angle may be a graphical input indicating the angle by a direction from a center of a geometric shape.
- the input indicating the angle may be a real number indicating the angle.
- the method may further include displaying a presently set angle.
- the receiving the input indicating the angle from the user may include receiving an indication for increasing or decreasing the angle in response to the displaying.
- the method may further include setting the angle based on the indication.
- the method may further include receiving from the user an input indicating a head size of the user.
- the parameter (or the plurality of parameters) for the head-related transfer function may be determined further based on the received input indicating the head size of the user.
- the method may further include receiving from the user an input indicating a head shape of the user.
- the parameter (or the plurality of parameters) for the head-related transfer function may be determined further based on the received input indicating the head shape of the user.
- the method may further include receiving from the user an input indicating an ear size of the user.
- the parameter (or the plurality of parameters) for the head- related transfer function may be determined further based on the received input indicating the ear size of the user.
- the method may further include receiving from the user an input indicating an ear shape of the user.
- the parameter (or the plurality of parameters) for the head-related transfer function may be determined further based on the received input indicating the ear shape of the user.
- the receiving and the determining may be performed for a plurality of virtual speaker positions.
- the method may further include sending the determined parameter (or plurality of determined parameters) to a se er in a cloud.
- the method may further include: receiving a parameter (or a plurality of parameters) for the head-related transfer function from a server in a cloud; modifying the audio signal in accordance with the head-related transfer function based on the received parameter (or the plurality of received parameters); and outputting the modified audio signal.
- the parameter (or the plurality of parameters) for the head-related transfer function may be determined based on the received input indicating the angle using a lookup table storing a relation between angles and parameters (or the plurality of parameters).
- FIG. 2 shows an audio output device 200 in accordance with an embodiment.
- the audio output device 200 may include an input circuit 202 configured to receive from a user an input indicating an angle.
- the audio output device 200 may further include a determination circuit 204 configured to determine a parameter (or a plurality of parameters) for a head-related transfer function based on the received input indicating the angle.
- the audio output device 200 may further include a modification circuit 206 , configured to modify an audio signal in accordance with the head-related transfer function based on the determined parameter (or the plurality of determined parameters).
- the audio output device 200 may further include an output circuit 208 configured to output the modified audio signal.
- the input circuit 202, the determination circuit 204, the modification circuit 206, and the output circuit 208 may be connected via a connection 210 (or a plurality of separate connections), for example an electrical or optical connection, for example any kind of cable or bus.
- the input circuit 202 may be configured to receive a graphical input indicating the angle by a point on a geometric shape.
- the input circuit 202 may be configured to receive a graphical input indicating the angle by a direction from a center of a geometric shape.
- the input circuit 202 may be configured to receive a real number indicating the angle.
- the audio output device 200 may further include a display circuit (not shown) configured to display a presently set angle.
- the input circuit 202 may be configured to receive from the user an indication for increasing or decreasing the angle in response to the displayed presently set angle.
- the audio output device 200 may further include a setting circuit (not shown) configured to set the angle based on the indication.
- the input circuit 202 may further be configured to receive from the user an input indicating a head size of the user.
- the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the head size of the user.
- the input circuit 202 may further be configured to receive from the user an input indicating a head shape of the user.
- the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the head shape of the user.
- the input circuit 202 may further be configured to receive from the user an input indicating an ear size of the user.
- the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the ear size of the user.
- the input circuit 202 may further be configured to receive from the user an input indicating an ear shape of the user.
- the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the ear shape of the user.
- the input circuit 202 and the determination circuit 204 may be configured to perform the receiving and the determining for a plurality of virtual speaker positions.
- the audio output device 200 may further include a sending circuit (not shown) configured to send the determined parameter (or the plurality of determined parameters) to a server in a cloud.
- a sending circuit (not shown) configured to send the determined parameter (or the plurality of determined parameters) to a server in a cloud.
- the audio output device 200 may further include a receiving circuit (not shown) configured to receive a parameter (or a plurality of parameters) for the head-related transfer function from a server in a cloud.
- the modification circuit 206 may be configured to modify the audio signal in accordance with the head-related transfer function based on the received parameter (or the plurality of received parameters).
- the output circuit 208 may be configured to output the modified audio signal.
- the determination circuit 204 may be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function based on the received input indicating the angle using a lookup table storing a relation between angles and parameters (or the plurality of parameters).
- FIG. 3 shows a graphical user interface 300 in accordance with an embodiment.
- the graphical user interface 300 may for example be displayed on a computer screen 302.
- the graphical user interface 300 may include an application program window 304 generated by the application program.
- the application program window 304 may include a visual representation 306 of the user (here shown as a geometric shape).
- the application program window 304 may further include a visual representation 308 of a speaker (here shown as a geometric shape) on a geometric shape 310 around the user.
- the geometric shape 310 in the graphical user interface may be displayed as any kind of geometrical form, for example an ellipse, wherein the user 306 is located at the center of the geometric shape, for example the ellipse 310.
- the speaker 308 is movable along the geometric shape 310, wherein the positioning of the speaker 308 translates to inputting an indication of an angle on the geometric shape 310 with respect to the user 306.
- the application program window 304 may further include an input 312 for inputting an indication of an angle on the geometric shape with respect to the visual representation of the speaker (here shown as a geometric shape that may be moved on the geometric shape 310), wherein the input 312 is associated to the speaker 308. The moving of the speaker 308 adjusts the angle.
- FIG. 4 shows an alternative implementation 400.
- a movable input 312 may be used to adjust the angle.
- the visual representation 308 of the speaker may include an image of a speaker or other images of output speaker device.
- the visual representation 308 of the speaker may include a ball.
- the input 3 12 may include a marker configured to be moved on the geometric shape 3 10.
- the input 312 may include the visual representation of the speaker configured to be moved on the geometric shape.
- the visual representation 308 of the speaker or the input 312 may include a needle of a compass configured to be moved around the geometric shape 310 with respect to the user 306.
- the graphical user interface 300 may be configured to send the input indication of the angle to the application program.
- the application program may be configured to: determine a parameter (or a plurality of parameters) for a head-related transfer function based on the received input indicating the angle, modify an audio signal in accordance with the head-related transfer function based on the determined parameter (or the plurality of determined parameters); and output the modified audio signal.
- the application program window 304 may further be configured to receive from the user an input indicating a head size of the user.
- the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the head size of the user.
- the application program window 304 may further be configured to receive from the user an input indicating a head shape of the user.
- the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the head shape of the user.
- the application program window 304 may further be configured to receive from the user an input indicating an ear size of the user.
- the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the ear size of the user.
- the application program window 304 may further be configured to receive from the user an input indicating an ear shape of the user.
- the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the ear shape of the user.
- the application program window 304 may include visual representations of a plurality of speakers and the input 312 may be for inputting an angle for each of the visual representation of each speaker.
- the application program window 304 may further include a sender input for receiving an input for instructing the application program to send a parameter (or a plurality of parameters) determined based on the input angle to a server in a cloud.
- the application program window 304 may further include a receiver for receiving a parameter (or a plurality of parameters) for a head-related transfer function from a server in a cloud.
- the application program may be configured to modify the audio signal in accordance with the head-related transfer function based on the received parameter (or the plurality of received parameters) and to output the modified audio signal.
- the application program may be configured to determine a parameter (or a plurality of parameters) for a head-related transfer function based on the input indication of the angle using a lookup table storing a relation between angles and parameters (or the plurality of parameters) for the head- related transfer function.
- the seven speakers may represent a 7.1 sound system.
- Typical sound systems may be 5.1 (e.g. cinema theatre).
- the UI may be provided to (i) allow user to calibrate the sound setting of a 7.1 audio headset and/ or (ii) perform virtualization so that a 2.1 or 5.1 headset sounds like a 7.1 system to the user.
- a method of HRTF calibration may be provided.
- a head-reflectance transfer function (or head-related transfer function; HRTF) may be applied to an incoming analog stereo audio signal in order to create the illusion of a multi-channel audio system through typical stereo headphones.
- HRTF head-related transfer function
- a method of calibrating an HRTF system using a graphical user interface to position virtual speaker positions may be provided, in a way which is easy to understand and manipulate by a novice user with no prior experience in tuning an HRTF.
- further the association of the HRTF calibration parameters determined by the user with a unique cloud identifier for that user, storing these settings for use across any device connecting to the cloud service may be provided. Cloud identification may enable not only the storage of an HRTF calibration profile for a particular user, but also the machine and devices in the audio reproduction environment such as the digital-to-analog converter (DAC), headphone amplifier, and the headphones or headset used to reproduce sound.
- DAC digital-to-analog converter
- a head-related transfer function may be a response that may describe how an ear receives a sound from a point in space; a pair of HRTFs for two ears may be used to synthesize a binaural sound that seems to come from a particular point in space. It may be a transfer function, describing how a sound from a specific point will arrive at the ear (generally at the outer end of the auditory canal). Some consumer home entertainment products designed to reproduce surround sound from stereo (two-speaker) headphones may use HRTFs. Some forms of HRTF-processing may have also been included in computer software to simulate surround sound playback from loudspeakers.
- Humans may estimate the location of a source by taking cues derived from one ear (monaural cues), and by comparing cues received at both ears (difference cues or binaural cues). Among the difference cues may be time differences of arrival and intensity differences.
- the monaural cues may come from the interaction between the sound source and the human anatomy, in which the original source sound may be modified before it enters the ear canal for processing by the auditory system. These modifications may encode the source location, and may be captured via an impulse response which may relate the source location and the ear location.
- This impulse response may be termed the head-related impulse response (HRIR): Convolution of an arbitrary source sound with the HRIR may convert the sound to that which would have been heard by the listener if it had been played at the source location, with the listener's ear at the receiver location. HRIRs may have been used to produce virtual surround sound.
- HRIR head-related impulse response
- the HRTF may be the Fourier transform of HRIR.
- the HRTF may also be referred to as the anatomical transfer function (ATF).
- HRTFs for left and right ear may describe the filtering of a sound source (x(t)) before it is perceived at the left and right ears as xL(t) and xR(t), respectively.
- the HRTF may also be described as the modifications to a sound from a direction in free air to the sound as it arrives at the eardrum. These modifications may include the shape of the listener's outer ear, the shape of the listener's head and body, the acoustical properties of the space in which the sound is played, and so on. All these properties may influence how (or whether) a listener may accurately tell what direction a sound is coming from.
- HRTFs may vary significantly from person to person. Perceptual distortions may occur when one listens to sounds spatialized with non-individualized HRTF. This focus on a "one size fits all" approach to HRTF may assume average physiology and morphology across all users.
- the size of the head and placement of the headphones or headset may be a critical determining factor in how the coefficients are created and applied to the filter.
- the shape of the ear and size of the ears may have a major impact on how sounds propagate from the drivers to the inner ear. As such, one size or type of filter does not fit all listeners, resulting i poor performance of the virtual surround sound system as it targets the average values found across all humans.
- a user may establish settings which may be unique to that user, and those settings may not persist across multiple systems and device chains, as the parameters established by the user during calibration remain locked into that particular device for which settings were configured.
- the user may calibrate the HRTF filter so that it works best for them, and then may save those settings to the client and mirroring them into the cloud for use on any client in the future.
- the HRTF calibration parameters used on one audio reproduction system such as a personal computer
- the user may configure or calibrate the HRTF algorithm on a singular device or system and have those systems persist across a multitude of devices and systems through software interface which authenticate them and transport their profile settings from the cloud into that system.
- a graphical user interface for the calibration of an HRTF and cloud services which synchronize a singular device or system with a multitude of devices and systems through the use of unique identifiers for the user and the device(s) and machine(s) used to calibrate the HRTF algorithm may be provided. This may be based on user input to determine offsets in the virtual speaker positions, selection of head size, ear size and ear shape, which determine the appropriate HRTF coefficients to apply within the audio filter. For example, HRTF coefficients may be stored in the cloud and downloaded to the client, or may be included in the installation and local-only. Once the user calibrates, and selections of HRTF coefficients are made, the user's chosen configuration may be stored in the cloud for use on any other PC client they log into.
- the first selection that a user may make is his or her head size, which may be a subjective selection given a set of options such as small, medium and large, corresponding to the circumference of the head as measured at the brow and around the largest area of the back of the head. This may be similar in approach to that of hat size measurement.
- the options may be relative to an average circumference of 58 cm, with a standard deviation of +/- 7 cm.
- the second selection that a user may make is his or her head shape, based on a set of provided options such as round, oval, and inverted egg.
- the third selection that a user may make is his or her ear size, which may be a subjective selection given a set of options such as small, medium and large, corresponding to the size of their outer ear (pinna, or auricle) relative to an average of 6 cm with a standard deviation of +/- 1.5 cm.
- the fourth selection that a user makes is his or her ear shape, which may be a subjective selection by the user of 8 common ear shape types.
- FIG. 5A is a diagram of an application program window 400 in accordance with one embodiment of the present invention.
- the application program window 400 includes a first sub-window 401, a second sub-window 402, a third sub-window 403 and a fourth sub- window 404.
- the first sub-window 401 provides means for the users to select their head size.
- the first sub-window 401 includes an input window for the user to input/type their head circumference size.
- the first sub-window 401 may contain a drop-down menu/list with preset head circumference sizes, e.g. from 51cm to 65cm, wherein the preset sizes are selectable by the user.
- the first sub-window 401 includes a plurality of images illustrating the different head sizes, e.g. a first image with a range of 51cm-55cm, a second image with a range of 56cm-60cm, and a third image with a range of 61cm-65cm.
- the users select the image with the closest range to their head size.
- the first sub- window may be a combination of the input window, the drop-down menu, or the plurality of images to allow flexibility in selecting the head size.
- the second sub-window 402 provides means for the users to select their head shape.
- the second sub-window 402 includes a drop-down menu/list with preset head shapes, e.g.
- the second sub-window 402 includes a plurality of images illustrating the different head shapes, and the users select the closest image to their head shape.
- the second sub-window 402 is a combination of the drop-down menu with preset head shapes and plurality of images with different head shapes.
- the third sub-window 403 provides means for the users to select their ear size.
- the third sub-window 403 includes an input window for the user to input/type their ear size.
- the third sub-window 403 may contain a drop-down menu/list with preset ear sizes, e.g. outer ear size of about 4.5cm to 7.5cm, wherein the preset sizes are selectable by the user.
- the third sub-window 403 includes a plurality of images illustrating the different ear sizes, e.g.
- the third sub-window 403 may be a combination of the input window, the drop-down menu, or the plurality of images.
- the fourth sub-window 404 provides means for the users to select their ear shape.
- the fourth sub-window 404 includes a plurality of images illustrating the different ear shapes to allow the users to select their ear shape that is closest to one of the images.
- FIG. 5B shows the fourth sub-window 404 with a plurality of images representing common ear shapes.
- further adjustments may be to the positioning of virtual surround sound speaker locations by the user, to personalize his or her listening experience.
- This method may enable the user to more fully realize the surround sound spatiality by making adjustments to the virtual speaker location in the graphical user interface, which may be translated into adjustments to the HRTF coefficients for each speaker positions.
- the results of such a graphical method of calibration may be immediately apparent to the user, as he or she may perceive the changes in virtual speaker positions during the calibration steps.
- the method may begin by instructing the user to place his or her preferred headphones onto his or her head, while an audio clip is playing which cycles through the default virtual speaker locations.
- the user may hear what these default positions sound like given his or her morphological parameters such as head size and ear shape, as well as the mechanical design and other characteristics of his or her preferred headphones.
- FIG. 6A shows a screen shot 601 of a graphical user interface for calibrating virtual speaker positions in accordance with an embodiment.
- the user may select one of the virtual speaker locations 611 to 617 and may be presented with a respective marker highlighting the position of the audio at the default angle (azimuth) relative to the head/user representation 670 at the center of the sound field.
- FIG. 6B shows a screen shot 602 of a graphical user interface in accordance with an embodiment, wherein a virtual speaker location marker 626 is shown when the virtual speaker location 616 is selected.
- the user may choose to adjust the position of this marker 626 that is associated only to virtual speaker location 616, which may result in adjustments made to the HRTF coefficients applied by the filter, in order to shift the perceived audio origination point (from the virtual speaker location 616) around the sound field. For example, by positioning of the virtual surround sound speaker locations by the user, the user may "freely" shift/move the 7 virtual speaker locations 61 1 to 617 within the geometric shape 650.
- the user may fully customize the sound field to his or her preference. According to various embodiments, this may enable the user to achieve a subjectively better virtual surround sound audio experience through real-time adjustments of the speaker positions with a synthesized multi-channel surround sound audio source playing through the HRTF filter being modified.
- the new HRTF coefficients may be saved for that particular user and optionally associated with his or her preferred headphones. Other HRTF calibrations may be performed for other headphones, enabling the user to customize and calibrate his or her HRTF filter library for a plurality of headphone or headset devices.
- the appropriate HRTF filter coefficients may be applied to a wide variety of applications, and may persist across multiple devices and systems used by the user. This may ensure the best possible virtual surround sound experience no matter which system the user is currently using.
- the current state of HRTF calibration may be limited to a standard set of predetermined filters, created based on objective morphological factors and may not present the user with affordances for calibrating their virtual surround sound experience using a graphical user interface to select their morphological parameters and control the positioning of virtual surround sound speaker positions.
- the state of the art may be advanced according to various embodiments by associating a unique profile created by the user for their device(s) and system(s) with their unique identification in the cloud service, enabling a consistent experience across multiple devices through connectivity of HRTF calibration software to the cloud service.
- FIG. 7 shows a screen shot 700 of a graphical user interface or application program window in accordance with an embodiment, wherein an audio output device may be set.
- the application program window 700 is generated by the application program operating on a computing device.
- the application program is connected to a remote server over a network (e.g. Internet cloud service), wherein the application program stores user profiles on the remote server or receives stored user profiles from the remote server.
- a network e.g. Internet cloud service
- the application program retrieves the stored user profiles associated with the user ID from the remote server and displays it in the on a profile sub-window 720 of the application program window 700.
- ID unique user identification
- the profile sub-window 720 displays a list of user profiles and also enables new profiles to be created, wherein the list of user profiles is retrievable from the remote server or from the local client (i.e. the computer device).
- a particular user profile e.g. "Profile” shown in FIG 7
- an audio device sub-window 730 displays a list of audio devices associated with the user profile. For example, all Razer analog headphones and headsets may be provided in the list of audio devices and one of them may be selected for calibration.
- the application program displays the name of the new headphone/headset in the audio device sub-window 730 if it is compatible with the application program.
- the application program window 700 comprises a top menu bar 710, including the functions "SETTINGS”, “AUDIO”, “EQ” and “CALIBRATION".
- FIG. 7 shows the application program window 700 when the "SETTINGS” 71 1 function is selected by the user.
- FIG. 8 shows a screen shot 800 of a graphical user interface or application program window in accordance with an embodiment, wherein general audio output parameters may be set.
- the application program window 800 displays an audio output sub-window 810 as well as the profile sub-window 720.
- the audio output sub-window 810 enables the user to adjust audio output parameters, such as but not limited to "BASS BOOST", “VOICE CLARITY”, “VOLUME NORMALIZATION” and "VOLUME LEVEL".
- the application program associates or stores the data of the desired audio output parameters to the selected profile in the profile sub-window 720.
- the profile and the associated audio output parameters can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
- FIG. 9 shows a screen shot 900 of a graphical user interface or application program window in accordance with an embodiment, wherein equalizer (EQ) parameters may be set.
- the application program window 900 displays a drop-down menu with preset EQ settings 910, wherein the drop-down menu may include common EQ settings such as but not limited to "Classical”, “Rock”, “Dance”, “Jazz”, etc.
- the application program window 900 includes a plurality of EQ frequency bars that enables user to configure the desired EQ settings.
- the application program window 900 includes the profile sub-window 720.
- the application program associates or stores the data of the desired EQ settings to the selected profile in the profile sub-window 720.
- the profile and the associated EQ settings can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
- FIG. 10 shows a screen shot 1000 of a graphical user interface or application program window in accordance with an embodiment, wherein the position of a virtual speaker may be adjusted.
- the application program window 1000 displays a representation of a plurality of speakers 1101-1107 arranged on a circular path 1002.
- a representation of a user 1001 is located at the central position of the circular path 1002.
- the default position of the speaker 1101 is about 0 degrees from the user 1001.
- speaker 1102 is positioned around 45 degrees, speaker 1 103 around 90 degrees, speaker 1104 around 135 degrees, speaker 1105 around 225 degrees, speaker 1 106 around 270 degrees and speaker 1 107 around 325 degrees relative to the user 1001 respectively.
- the 7 speakers 1101-1107 represent a 7.1 surround system but it can be appreciated that for other surround systems the number of speakers may vary, e.g. 5 speakers may be used to represent a 5.1 surround system.
- the circular path 1002 may take other forms/shapes, such as a square path of a rectangular path.
- a surround sound audio loop may be playing, such as a helicopter, which may move around all the virtual speaker positions.
- the user may click on a "Test All” button 1201 to replay this surround sound audio loop and listen to all the speaker positions with any changes made.
- a virtual speaker location for example speaker 1102
- the other speakers may fade away and the selected speaker may be highlighted, like shown in FIG. 1 1.
- An audio loop may play from the selected virtual speaker location.
- FIG. 1 1 shows a screen shot 1100 of a graphical user interface in accordance with an embodiment, wherein a marker 1 122 indicating an angle for the chosen virtual speaker 1 102 may be set.
- the user may click the calibration marker 1 122 (with the shape of a ball) and may drag it around the circular path 1002 to varying degrees based on the speaker 1 102 selected to adjust the position of the sound until it appears to originate from the virtual speaker location, or their desired location.
- the virtual speaker location 1 102 is not moveable.
- the calibration marker 1 122 may not end up directly on top of the speaker 1202 - it may simply be an offset of the sound to account for ear and head size, headphone type, for example.
- the user interface does not include the calibration marker 1 122 and the user may adjust the position of the sound from the speaker 1102 by clicking and dragging the speaker 1102 around the circular path 1002.
- the adjustment of the calibration marker 1 122 that is associated only to speaker 1 102 results in adjustments to the HRTF coefficient associated with speaker 1102.
- the user may fully customize the sound field to his or her preference.
- the new HRTF coefficients are stored with the desired profile and associated with his or her preferred headphones selected in the audio device sub-window 730 show in FIG 7.
- Other HRTF calibrations may be performed for other headphones, enabling the user to customize and calibrate his or her HRTF filter library for a plurality of headphone or headset devices.
- the calibration marker or speaker may be moved only to a certain degree, so as for example not to move a right speaker entirely to a left side or to move a front speaker to the rear. This helps users to maintain the audio fidelity while calibrating their headphone/headset.
- the azimuth for each speaker 1101-1 107 may be restricted such that the user may move a selected speaker about +/-15 to 20 degrees from its default/original position.
- the range of angles of azimuth of speaker 1 102 may be fixed to a corresponding zone/region 1 132 so that the speaker 1102 (or its associated marker) is restricted from being moved outside the zone 1132.
- the speaker 1102 (or its associated marker) is only slidable around the circular path 1002 within the zone 1 132, or moveable about +/- 20 degrees from its default angle of 45 degrees (i.e. from 25 degrees to 65 degrees).
- the application program determines a parameter for a HRTF based on the angle of the speaker 1 102 (or its associated calibration marker).
- the application program modifies an audio signal in accordance with the HRTF based on the determined parameter, resulting in the output of the modified audio signal from the speaker 1102 to the user.
- the application program may be configured to determine a parameter for HRTF based on the input indication of the angle using a lookup table storing a relation between angles and parameters for the HRTF.
- the application program associates or stores the parameter of the HRTF to the selected profile in the profile sub-window 720.
- the profile and the parameters of the HRTF can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
- the application program includes an interface for the user to select the size and shape of his/her head and ear (similar to FIG 5A as discussed above) prior to the calibration of the speakers 1101-1 107 shown in FIGS 10 and 1 1.
- the application program determines the parameter for the HRTF based on the received input indicating the size and shape of the head and ear.
- the application program modifies an audio signal in accordance with the HRTF based on the determined parameter, resulting in the output of the modified audio signal from the speakers 1 101-1 107 to the user.
- the application program may be configured to determine a parameter for HRTF based on the input indication of the head shape, head size, ear shape or ear size using a lookup table storing a relation between the head shape, head size, ear shape, ear size and parameters for the HRTF.
- the application program associates or stores the parameter of the HRTF to the selected profile in the profile sub-window 720.
- the profile and the parameters of the HRTF can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
- the application program window comprises a "Reset” button to allow the user to reset the speakers 1 101-1 107 to their default positions.
- a checkbox may be provided, and checking the checkbox may override calibration settings saved to the desired profile, e.g. "Profile” shown in the profile sub-window 720 in FIG. 12 and apply the calibration settings globally to all profiles associated to the unique user ID.
- Profile e.g. "Profile” shown in the profile sub-window 720 in FIG. 12
- FIG. 13 shows a screen shot 1300 of a graphical user interface or application program window showing alternative representation of the speakers.
- the speakers are represented by an "arrow-head” image instead of a loudspeaker image illustrated in FIG. 10.
- the "arrow-head” speakers 1311-1317 are moveable on a circular path 1301.
- the application program window includes another circular path 1302 concentric to circular path 1301.
- the circular path 1302 includes a plurality of calibration markers positioned adjacent to the speakers 1311-1317 in their default position.
- the speaker 1312 includes a corresponding calibration marker 1322 that is moveable within a zone/region 1332 of the circular path 1302.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Stereophonic System (AREA)
- Electrophonic Musical Instruments (AREA)
Abstract
Priority Applications (7)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| AU2012394979A AU2012394979B2 (en) | 2012-11-22 | 2012-11-22 | Method for outputting a modified audio signal and graphical user interfaces produced by an application program |
| CN201280078150.3A CN105027580B (zh) | 2012-11-22 | 2012-11-22 | 用于输出修改的音频信号的方法 |
| SG11201503926WA SG11201503926WA (en) | 2012-11-22 | 2012-11-22 | Method for outputting a modified audio signal and graphical user interfaces produced by an application program |
| EP12888662.9A EP2923500A4 (fr) | 2012-11-22 | 2012-11-22 | Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application |
| US14/646,677 US9569073B2 (en) | 2012-11-22 | 2012-11-22 | Method for outputting a modified audio signal and graphical user interfaces produced by an application program |
| PCT/SG2012/000439 WO2014081384A1 (fr) | 2012-11-22 | 2012-11-22 | Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application |
| TW102134986A TWI616810B (zh) | 2012-11-22 | 2013-09-27 | 輸出一調變音訊之方法及藉由一應用程式產生之圖形使用者介面 |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/SG2012/000439 WO2014081384A1 (fr) | 2012-11-22 | 2012-11-22 | Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2014081384A1 true WO2014081384A1 (fr) | 2014-05-30 |
Family
ID=50776418
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/SG2012/000439 Ceased WO2014081384A1 (fr) | 2012-11-22 | 2012-11-22 | Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application |
Country Status (7)
| Country | Link |
|---|---|
| US (1) | US9569073B2 (fr) |
| EP (1) | EP2923500A4 (fr) |
| CN (1) | CN105027580B (fr) |
| AU (1) | AU2012394979B2 (fr) |
| SG (1) | SG11201503926WA (fr) |
| TW (1) | TWI616810B (fr) |
| WO (1) | WO2014081384A1 (fr) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2017058097A1 (fr) * | 2015-09-28 | 2017-04-06 | Razer (Asia-Pacific) Pte. Ltd. | Ordinateurs, procédés de commande d'un ordinateur, et supports lisibles par ordinateur |
| CN107484097A (zh) * | 2017-10-10 | 2017-12-15 | 歌尔股份有限公司 | 耳机左右喇叭防呆检测机构 |
| WO2019166030A1 (fr) * | 2018-02-28 | 2019-09-06 | 成都星环科技有限公司 | Système d'adaptation de champ sonore intelligent, système d'étalonnage de champ sonore intelligent et système de traitement vocal intelligent |
| EP3554098A4 (fr) * | 2016-12-12 | 2019-12-18 | Sony Corporation | Procédé de mesure hrtf, dispositif de mesure hrtf et programme |
Families Citing this family (54)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US9219460B2 (en) | 2014-03-17 | 2015-12-22 | Sonos, Inc. | Audio settings based on environment |
| US9106192B2 (en) | 2012-06-28 | 2015-08-11 | Sonos, Inc. | System and method for device playback calibration |
| WO2014081384A1 (fr) * | 2012-11-22 | 2014-05-30 | Razer (Asia-Pacific) Pte. Ltd. | Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application |
| US20160147415A1 (en) * | 2013-08-01 | 2016-05-26 | Thales | Programming system for a situation analysis system on board a carrier comprising at least one onboard listening system |
| CN104197299A (zh) * | 2014-08-21 | 2014-12-10 | 浙江生辉照明有限公司 | 照明装置及基于该装置的语音播报系统及方法 |
| US9952825B2 (en) | 2014-09-09 | 2018-04-24 | Sonos, Inc. | Audio processing algorithms |
| US9544706B1 (en) * | 2015-03-23 | 2017-01-10 | Amazon Technologies, Inc. | Customized head-related transfer functions |
| US11113022B2 (en) * | 2015-05-12 | 2021-09-07 | D&M Holdings, Inc. | Method, system and interface for controlling a subwoofer in a networked audio system |
| US11209972B2 (en) | 2015-09-02 | 2021-12-28 | D&M Holdings, Inc. | Combined tablet screen drag-and-drop interface |
| US9609436B2 (en) * | 2015-05-22 | 2017-03-28 | Microsoft Technology Licensing, Llc | Systems and methods for audio creation and delivery |
| WO2017049169A1 (fr) | 2015-09-17 | 2017-03-23 | Sonos, Inc. | Faciliter l'étalonnage d'un dispositif de lecture audio |
| US9693165B2 (en) | 2015-09-17 | 2017-06-27 | Sonos, Inc. | Validation of audio calibration using multi-dimensional motion check |
| US10142755B2 (en) * | 2016-02-18 | 2018-11-27 | Google Llc | Signal processing methods and systems for rendering audio on virtual loudspeaker arrays |
| FI20165211A (fi) | 2016-03-15 | 2017-09-16 | Ownsurround Ltd | Järjestely HRTF-suodattimien valmistamiseksi |
| CN107204132A (zh) * | 2016-03-16 | 2017-09-26 | 中航华东光电(上海)有限公司 | 3d虚拟立体声空中预警系统 |
| US9860662B2 (en) | 2016-04-01 | 2018-01-02 | Sonos, Inc. | Updating playback device configuration information based on calibration data |
| US9763018B1 (en) | 2016-04-12 | 2017-09-12 | Sonos, Inc. | Calibration of audio playback devices |
| CN105933826A (zh) * | 2016-06-07 | 2016-09-07 | 惠州Tcl移动通信有限公司 | 一种自动设置声场的方法、系统及耳机 |
| US9794710B1 (en) | 2016-07-15 | 2017-10-17 | Sonos, Inc. | Spatial audio correction |
| US10372406B2 (en) | 2016-07-22 | 2019-08-06 | Sonos, Inc. | Calibration interface |
| US10459684B2 (en) | 2016-08-05 | 2019-10-29 | Sonos, Inc. | Calibration of a playback device based on an estimated frequency response |
| USD818488S1 (en) * | 2016-10-14 | 2018-05-22 | Keepitsafe, Inc. | Display screen or portion thereof with a graphical user interface |
| US10901681B1 (en) * | 2016-10-17 | 2021-01-26 | Cisco Technology, Inc. | Visual audio control |
| KR20190091445A (ko) * | 2016-10-19 | 2019-08-06 | 오더블 리얼리티 아이엔씨. | 오디오 이미지를 생성하는 시스템 및 방법 |
| CN107979807A (zh) * | 2016-10-25 | 2018-05-01 | 北京酷我科技有限公司 | 一种模拟环绕立体声的方法及系统 |
| US20180115854A1 (en) * | 2016-10-26 | 2018-04-26 | Htc Corporation | Sound-reproducing method and sound-reproducing system |
| US10375498B2 (en) | 2016-11-16 | 2019-08-06 | Dts, Inc. | Graphical user interface for calibrating a surround sound system |
| TWI630828B (zh) * | 2017-06-14 | 2018-07-21 | 趙平 | 供使用者定向交談之智慧型耳機裝置個人化系統及其使用方法 |
| US20180367935A1 (en) * | 2017-06-15 | 2018-12-20 | Htc Corporation | Audio signal processing method, audio positional system and non-transitory computer-readable medium |
| CN108196815B (zh) * | 2017-12-28 | 2020-01-03 | 维沃移动通信有限公司 | 一种通话声音的调节方法和移动终端 |
| US10412527B1 (en) * | 2018-01-09 | 2019-09-10 | Facebook Technologies, Llc | Head-related transfer function determination using base stations |
| US10225656B1 (en) * | 2018-01-17 | 2019-03-05 | Harman International Industries, Incorporated | Mobile speaker system for virtual reality environments |
| US10652686B2 (en) * | 2018-02-06 | 2020-05-12 | Sony Interactive Entertainment Inc. | Method of improving localization of surround sound |
| FI20185300A1 (fi) | 2018-03-29 | 2019-09-30 | Ownsurround Ltd | Järjestely päähän liittyvien siirtofunktiosuodattimien muodostamiseksi |
| JP7442494B2 (ja) | 2018-07-25 | 2024-03-04 | ドルビー ラボラトリーズ ライセンシング コーポレイション | 光学式捕捉によるパーソナライズされたhrtf |
| US10299061B1 (en) | 2018-08-28 | 2019-05-21 | Sonos, Inc. | Playback device calibration |
| WO2020044244A1 (fr) | 2018-08-29 | 2020-03-05 | Audible Reality Inc. | Système et procédé de commande d'un moteur audio tridimensionnel |
| US10856097B2 (en) | 2018-09-27 | 2020-12-01 | Sony Corporation | Generating personalized end user head-related transfer function (HRTV) using panoramic images of ear |
| US11113092B2 (en) | 2019-02-08 | 2021-09-07 | Sony Corporation | Global HRTF repository |
| US11451907B2 (en) | 2019-05-29 | 2022-09-20 | Sony Corporation | Techniques combining plural head-related transfer function (HRTF) spheres to place audio objects |
| US11347832B2 (en) | 2019-06-13 | 2022-05-31 | Sony Corporation | Head related transfer function (HRTF) as biometric authentication |
| US10734965B1 (en) | 2019-08-12 | 2020-08-04 | Sonos, Inc. | Audio calibration of a portable playback device |
| US11146908B2 (en) | 2019-10-24 | 2021-10-12 | Sony Corporation | Generating personalized end user head-related transfer function (HRTF) from generic HRTF |
| US11070930B2 (en) | 2019-11-12 | 2021-07-20 | Sony Corporation | Generating personalized end user room-related transfer function (RRTF) |
| USD958166S1 (en) * | 2019-11-19 | 2022-07-19 | Johnson Systems Inc. | Display screen with graphical user interface |
| USD926789S1 (en) * | 2019-11-19 | 2021-08-03 | Johnson Systems Inc. | Display screen with graphical user interface |
| CN115280799A (zh) * | 2019-12-31 | 2022-11-01 | 哈曼国际工业有限公司 | 利用不可见扬声器实现虚拟音效的系统和方法 |
| WO2022010453A1 (fr) * | 2020-07-06 | 2022-01-13 | Hewlett-Packard Development Company, L.P. | Annulation de traitement spatial dans des écouteurs |
| US11854555B2 (en) * | 2020-11-05 | 2023-12-26 | Sony Interactive Entertainment Inc. | Audio signal processing apparatus, method of controlling audio signal processing apparatus, and program |
| CN113365191A (zh) * | 2021-06-04 | 2021-09-07 | 广州酷狗计算机科技有限公司 | 音乐播放的方法、装置、设备及存储介质 |
| US20220407772A1 (en) * | 2021-06-16 | 2022-12-22 | Hewlett-Packard Development Company, L.P. | Configuration profiles |
| EP4409571B1 (fr) | 2021-09-30 | 2025-03-26 | Sonos Inc. | Gestion de conflit pour processus de détection de mot d'activation |
| CN116095595B (zh) * | 2022-08-19 | 2023-11-21 | 荣耀终端有限公司 | 音频处理方法和装置 |
| CN116744215B (zh) * | 2022-09-02 | 2024-04-19 | 荣耀终端有限公司 | 音频处理方法和装置 |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20060280323A1 (en) * | 1999-06-04 | 2006-12-14 | Neidich Michael I | Virtual Multichannel Speaker System |
| KR20080055622A (ko) * | 2006-12-16 | 2008-06-19 | 김종철 | 인간친화형 컴퓨터입출력 시스템 |
| US20120022842A1 (en) * | 2009-02-11 | 2012-01-26 | Arkamys | Test platform implemented by a method for positioning a sound object in a 3d sound environment |
| US20120093348A1 (en) * | 2010-10-14 | 2012-04-19 | National Semiconductor Corporation | Generation of 3D sound with adjustable source positioning |
| US20120201405A1 (en) * | 2007-02-02 | 2012-08-09 | Logitech Europe S.A. | Virtual surround for headphones and earbuds headphone externalization system |
Family Cites Families (31)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6046722A (en) * | 1991-12-05 | 2000-04-04 | International Business Machines Corporation | Method and system for enabling blind or visually impaired computer users to graphically select displayed elements |
| US5850455A (en) * | 1996-06-18 | 1998-12-15 | Extreme Audio Reality, Inc. | Discrete dynamic positioning of audio signals in a 360° environment |
| US6266571B1 (en) * | 1997-10-29 | 2001-07-24 | International Business Machines Corp. | Adaptively configuring an audio interface according to selected audio output device |
| US6067084A (en) * | 1997-10-29 | 2000-05-23 | International Business Machines Corporation | Configuring microphones in an audio interface |
| US6504553B2 (en) * | 1999-02-25 | 2003-01-07 | International Business Machines Corporation | Trouble shooting a wrong audio source |
| US6492999B1 (en) * | 1999-02-25 | 2002-12-10 | International Business Machines Corporation | Connecting and optimizing audio input devices |
| GB2374507B (en) * | 2001-01-29 | 2004-12-29 | Hewlett Packard Co | Audio user interface with audio cursor |
| US20030227476A1 (en) * | 2001-01-29 | 2003-12-11 | Lawrence Wilcock | Distinguishing real-world sounds from audio user interface sounds |
| GB2374772B (en) * | 2001-01-29 | 2004-12-29 | Hewlett Packard Co | Audio user interface |
| GB2376123B (en) * | 2001-01-29 | 2004-06-30 | Hewlett Packard Co | Facilitation of speech recognition in user interface |
| US7073125B1 (en) * | 2002-04-26 | 2006-07-04 | Microsoft Corporation | Integrated display of gauges |
| US20040264704A1 (en) | 2003-06-13 | 2004-12-30 | Camille Huin | Graphical user interface for determining speaker spatialization parameters |
| US7928311B2 (en) * | 2004-12-01 | 2011-04-19 | Creative Technology Ltd | System and method for forming and rendering 3D MIDI messages |
| FR2880755A1 (fr) | 2005-01-10 | 2006-07-14 | France Telecom | Procede et dispositif d'individualisation de hrtfs par modelisation |
| US7461345B2 (en) * | 2005-03-11 | 2008-12-02 | Adobe Systems Incorporated | System and method for displaying information using a compass |
| US20060277034A1 (en) | 2005-06-01 | 2006-12-07 | Ben Sferrazza | Method and system for processing HRTF data for 3-D sound positioning |
| US7549123B1 (en) * | 2005-06-15 | 2009-06-16 | Apple Inc. | Mixing input channel signals to generate output channel signals |
| WO2007048900A1 (fr) | 2005-10-27 | 2007-05-03 | France Telecom | Individualisation de hrtfs utilisant une modelisation par elements finis couplee a un modele correctif |
| US20080178088A1 (en) * | 2006-07-27 | 2008-07-24 | Personics Holdings Inc. | Method and device of customizing headphones |
| US7792674B2 (en) * | 2007-03-30 | 2010-09-07 | Smith Micro Software, Inc. | System and method for providing virtual spatial sound with an audio visual player |
| JP4355359B1 (ja) * | 2008-05-27 | 2009-10-28 | パナソニック株式会社 | マイクを外耳道開口部に設置する耳掛型補聴器 |
| TWI384889B (zh) * | 2008-12-26 | 2013-02-01 | Alpha Networks Inc | 聲音播放裝置參數調整方法 |
| US8923995B2 (en) * | 2009-12-22 | 2014-12-30 | Apple Inc. | Directional audio interface for portable media device |
| NZ587483A (en) * | 2010-08-20 | 2012-12-21 | Ind Res Ltd | Holophonic speaker system with filters that are pre-configured based on acoustic transfer functions |
| US9377941B2 (en) * | 2010-11-09 | 2016-06-28 | Sony Corporation | Audio speaker selection for optimization of sound origin |
| US20120124470A1 (en) * | 2010-11-17 | 2012-05-17 | The Johns Hopkins University | Audio display system |
| WO2012164346A1 (fr) * | 2011-05-27 | 2012-12-06 | Sony Ericsson Mobile Communications Ab | Sélection ou adaptation d'une fonction de transfert liée à la tête (hrtf) selon la taille de la tête |
| US8787584B2 (en) * | 2011-06-24 | 2014-07-22 | Sony Corporation | Audio metrics for head-related transfer function (HRTF) selection or adaptation |
| US9167368B2 (en) * | 2011-12-23 | 2015-10-20 | Blackberry Limited | Event notification on a mobile device using binaural sounds |
| EP2829050A1 (fr) * | 2012-03-23 | 2015-01-28 | Dolby Laboratories Licensing Corporation | Techniques pour mettre en valeur des interlocuteurs dans une scène de conférence bidimensionnelle (2d) ou tridimensionnelle (3d) |
| WO2014081384A1 (fr) * | 2012-11-22 | 2014-05-30 | Razer (Asia-Pacific) Pte. Ltd. | Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application |
-
2012
- 2012-11-22 WO PCT/SG2012/000439 patent/WO2014081384A1/fr not_active Ceased
- 2012-11-22 SG SG11201503926WA patent/SG11201503926WA/en unknown
- 2012-11-22 AU AU2012394979A patent/AU2012394979B2/en active Active
- 2012-11-22 EP EP12888662.9A patent/EP2923500A4/fr not_active Ceased
- 2012-11-22 CN CN201280078150.3A patent/CN105027580B/zh active Active
- 2012-11-22 US US14/646,677 patent/US9569073B2/en active Active
-
2013
- 2013-09-27 TW TW102134986A patent/TWI616810B/zh active
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20060280323A1 (en) * | 1999-06-04 | 2006-12-14 | Neidich Michael I | Virtual Multichannel Speaker System |
| KR20080055622A (ko) * | 2006-12-16 | 2008-06-19 | 김종철 | 인간친화형 컴퓨터입출력 시스템 |
| US20120201405A1 (en) * | 2007-02-02 | 2012-08-09 | Logitech Europe S.A. | Virtual surround for headphones and earbuds headphone externalization system |
| US20120022842A1 (en) * | 2009-02-11 | 2012-01-26 | Arkamys | Test platform implemented by a method for positioning a sound object in a 3d sound environment |
| US20120093348A1 (en) * | 2010-10-14 | 2012-04-19 | National Semiconductor Corporation | Generation of 3D sound with adjustable source positioning |
Non-Patent Citations (1)
| Title |
|---|
| See also references of EP2923500A4 * |
Cited By (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2017058097A1 (fr) * | 2015-09-28 | 2017-04-06 | Razer (Asia-Pacific) Pte. Ltd. | Ordinateurs, procédés de commande d'un ordinateur, et supports lisibles par ordinateur |
| CN108292142A (zh) * | 2015-09-28 | 2018-07-17 | 雷蛇(亚太)私人有限公司 | 计算机、控制计算机的方法及计算机可读介质 |
| US10356526B2 (en) | 2015-09-28 | 2019-07-16 | Razer (Asia-Pacific) Pte. Ltd. | Computers, methods for controlling a computer, and computer-readable media |
| CN108292142B (zh) * | 2015-09-28 | 2021-07-06 | 雷蛇(亚太)私人有限公司 | 计算机、控制计算机的方法及计算机可读介质 |
| EP3554098A4 (fr) * | 2016-12-12 | 2019-12-18 | Sony Corporation | Procédé de mesure hrtf, dispositif de mesure hrtf et programme |
| US11159906B2 (en) | 2016-12-12 | 2021-10-26 | Sony Corporation | HRTF measurement method, HRTF measurement device, and program |
| CN107484097A (zh) * | 2017-10-10 | 2017-12-15 | 歌尔股份有限公司 | 耳机左右喇叭防呆检测机构 |
| WO2019166030A1 (fr) * | 2018-02-28 | 2019-09-06 | 成都星环科技有限公司 | Système d'adaptation de champ sonore intelligent, système d'étalonnage de champ sonore intelligent et système de traitement vocal intelligent |
Also Published As
| Publication number | Publication date |
|---|---|
| SG11201503926WA (en) | 2015-06-29 |
| EP2923500A1 (fr) | 2015-09-30 |
| AU2012394979A1 (en) | 2015-06-04 |
| AU2012394979B2 (en) | 2016-07-14 |
| TWI616810B (zh) | 2018-03-01 |
| CN105027580B (zh) | 2017-05-17 |
| US20150293655A1 (en) | 2015-10-15 |
| US9569073B2 (en) | 2017-02-14 |
| TW201421354A (zh) | 2014-06-01 |
| CN105027580A (zh) | 2015-11-04 |
| EP2923500A4 (fr) | 2016-06-08 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| AU2012394979B2 (en) | Method for outputting a modified audio signal and graphical user interfaces produced by an application program | |
| EP3619921B1 (fr) | Processeur audio, système, procédé ainsi que programme d'ordinateur pour la reproduction audio | |
| JP6672425B2 (ja) | 反射音レンダリングのために上方発射ドライバをもつオーディオ・スピーカー | |
| AU2017279615B2 (en) | Method and device for rendering acoustic signal, and computer-readable recording medium | |
| CN111294724B (zh) | 多个音频流的空间重新定位 | |
| US10003905B1 (en) | Personalized end user head-related transfer function (HRTV) finite impulse response (FIR) filter | |
| CN104904235A (zh) | 用于使用向上发射驱动器的反射声音呈现的虚拟高度滤波器 | |
| US11102604B2 (en) | Apparatus, method, computer program or system for use in rendering audio | |
| CN115777203A (zh) | 信息处理装置、输出控制方法和程序 | |
| US12526597B2 (en) | Reproducing device, reproducing method, information processing device, information processing method, and program | |
| US20030099369A1 (en) | System for headphone-like rear channel speaker and the method of the same | |
| Cuevas-Rodríguez et al. | The 3D Tune-In Toolkit–3D audio spatialiser, hearing loss and hearing aid simulations | |
| CN108886652A (zh) | 考虑到个体要求的重现声音的方法 | |
| CN111756929A (zh) | 多屏终端音频播放方法、装置、终端设备以及存储介质 | |
| CN121002897A (zh) | 虚拟听觉显示滤波器与相关联系统、方法和非暂时性计算机可读介质 | |
| US7050596B2 (en) | System and headphone-like rear channel speaker and the method of the same | |
| EP4677864A1 (fr) | Systèmes et procédés pour audio spatial hybride | |
| US12348951B2 (en) | System and method for virtual sound effect with invisible loudspeaker(s) | |
| US6983054B2 (en) | Means for compensating rear sound effect | |
| EP4460986A1 (fr) | Appareil et procédé de mise en oeuvre d'un rendu d'objet audio polyvalent | |
| US20260025630A1 (en) | Methods, devices, and systems for reproducing spatial audio using binaural externalization processing extensions | |
| BE1028706B1 (nl) | Koptelefoon | |
| US20250016519A1 (en) | Audio device with head orientation-based filtering and related methods | |
| CN117395592A (zh) | 音频处理方法、系统及电子设备 | |
| JP3208565U (ja) | 音色が調節可能なイヤホン |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| WWE | Wipo information: entry into national phase |
Ref document number: 201280078150.3 Country of ref document: CN |
|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12888662 Country of ref document: EP Kind code of ref document: A1 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2012888662 Country of ref document: EP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 14646677 Country of ref document: US |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| ENP | Entry into the national phase |
Ref document number: 2012394979 Country of ref document: AU Date of ref document: 20121122 Kind code of ref document: A |