[go: up one dir, main page]

WO2014081384A1 - Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application - Google Patents

Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application Download PDF

Info

Publication number
WO2014081384A1
WO2014081384A1 PCT/SG2012/000439 SG2012000439W WO2014081384A1 WO 2014081384 A1 WO2014081384 A1 WO 2014081384A1 SG 2012000439 W SG2012000439 W SG 2012000439W WO 2014081384 A1 WO2014081384 A1 WO 2014081384A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
head
application program
input indicating
parameter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/SG2012/000439
Other languages
English (en)
Inventor
Min-Liang Tan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Razer Asia Pacific Pte Ltd
Original Assignee
Razer Asia Pacific Pte Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Razer Asia Pacific Pte Ltd filed Critical Razer Asia Pacific Pte Ltd
Priority to AU2012394979A priority Critical patent/AU2012394979B2/en
Priority to CN201280078150.3A priority patent/CN105027580B/zh
Priority to SG11201503926WA priority patent/SG11201503926WA/en
Priority to EP12888662.9A priority patent/EP2923500A4/fr
Priority to US14/646,677 priority patent/US9569073B2/en
Priority to PCT/SG2012/000439 priority patent/WO2014081384A1/fr
Priority to TW102134986A priority patent/TWI616810B/zh
Publication of WO2014081384A1 publication Critical patent/WO2014081384A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04847Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0486Drag-and-drop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S1/00Two-channel systems
    • H04S1/002Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
    • H04S1/005For headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/01Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]

Definitions

  • Various embodiments generally relate to methods for outputting a modified audio signal and to graphical user interfaces produced by an application program.
  • a head-reflectance transfer function (or head-related transfer function; HRTF) may be applied to an incoming analog stereo audio signal in order to create the illusion of a multi-channel audio system through typical stereo headphones.
  • This HRTF may have to be calibrated to a specific user.
  • a method for outputting a modified audio signal may be provided.
  • the method may include: receiving from a user an input indicating an angle; determining a parameter for a head-related transfer function based on the received input indicating the angle; modifying an audio signal in accordance with the head-related transfer function based on the determined parameter; and outputting the modified audio signal.
  • a graphical user interface produced by an application program may be provided.
  • the graphical user interface may include: an application program window generated by the application program, wherein the application program window may include: a visual representation of the user; a visual representation of a speaker on a geometric shape around the user; and an input for inputting an indication of an angle on the geometric shape with respect to the visual representation of the speaker.
  • FIG. 1 shows a flow diagram illustrating a method for outputting a modified audio signal in accordance with an embodiment
  • FIG. 2 shows an audio output device in accordance with an embodiment
  • FIG. 3 shows a graphical user interface in accordance with an embodiment
  • FIG. 4 shows a graphical user interface in accordance with an embodiment
  • FIG. 5 A shows a diagram of an application program window in accordance with one embodiment
  • FIG. 5B shows a plurality of ear shapes
  • FIG. 5 shows a screen shot of a graphical user interface for calibrating virtual speaker positions in accordance with an embodiment
  • FIG. 6A shows a screen shot of a graphical user interface for calibrating virtual speaker positions in accordance with an embodiment
  • FIG. 6B shows a screen shot of a graphical user interface in accordance with an embodiment, wherein a virtual speaker location marker is shown when the virtual speaker location 616 is selected;
  • FIG. 7 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein an audio output device may be set;
  • FIG. 8 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein general audio output parameters may be set;
  • FIG. 9 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein equalizer parameters may be set;
  • FIG. 10 shows a screen shot of a graphical user interface or application program window in accordance with an embodiment, wherein the position of a virtual speaker may be adjusted;
  • FIG. 11 shows a screen shot of a graphical user interface in accordance with an embodiment, wherein a marker indicating an angle for the chosen virtual speaker may be set;
  • FIG. 12 shows a screen shot of a graphical user interface in accordance with an embodiment, wherein a marker indicating an angle for the chosen virtual speaker may be set;
  • FIG. 13 shows a screen shot of a graphical user interface or application program window showing alternative representation of the speakers
  • FIG. 14 shows an application window according to an embodiment
  • the audio output device may include a memory which is for example used in the processing carried out by the audio output device.
  • a memory used in the embodiments may be a volatile memory, for example a DRAM (Dynamic Random Access Memory) or a non- volatile memory, for example a PROM (Programmable Read Only Memory), an EPROM (Erasable PROM), EEPROM (Electrically Erasable PROM), or a flash memory ⁇ e.g., a floating gate memory, a charge trapping memory, an MRAM (Magnetoresistive Random Access Memory) or a PCRAM (Phase Change Random Access Memory).
  • DRAM Dynamic Random Access Memory
  • PROM Programmable Read Only Memory
  • EPROM Erasable PROM
  • EEPROM Electrical Erasable PROM
  • flash memory e.g., a floating gate memory, a charge trapping memory, an MRAM (Magnetoresistive Random Access Memory) or a PCRAM (Phase Change Random Access Memory).
  • a “circuit” may be understood as any kind of a logic implementing entity, which may be special purpose circuitry or a processor executing software stored in a memory, firmware, or any combination thereof.
  • a “circuit” may be a hard-wired logic circuit or a programmable logic circuit such as a programmable processor, e.g. a microprocessor (e.g. a Complex Instruction Set Computer (CISC) processor or a Reduced Instruction Set Computer (RISC) processor).
  • a “circuit” may also be a processor executing software, e.g. any kind of computer program, e.g. a computer program using a virtual machine code such as e.g. Java.
  • circuit any other kind of implementation of the respective functions which will be described in more detail below may also be understood as a “circuit” in accordance with an alternative embodiment. It will be understood that what is described herein as circuits with different names (for example “circuit A” and “circuit B”) may also be provided in one physical circuit like described above.
  • a geometric shape may be or may include a conic section, for example a circle, an ellipse, a parabola or a hyperbola, or may include or may be a polygon, or may include or may be any other kind of geometric shape.
  • FIG. 1 shows a flow diagram 100 illustrating a method for outputting a modified audio signal in accordance with an embodiment.
  • an input indicating an angle may be received from a user.
  • a parameter (or a plurality of parameters) for a head-related transfer function may be determined based on the received input indicating the angle.
  • an audio signal may be modified in accordance with the head-related transfer function based on the determined parameter (or the plurality of determined parameters).
  • the modified audio signal may be output.
  • the input indicating the angle may be a graphical input indicating the angle by a point on a geometric shape.
  • the input indicating the angle may be a graphical input indicating the angle by a direction from a center of a geometric shape.
  • the input indicating the angle may be a real number indicating the angle.
  • the method may further include displaying a presently set angle.
  • the receiving the input indicating the angle from the user may include receiving an indication for increasing or decreasing the angle in response to the displaying.
  • the method may further include setting the angle based on the indication.
  • the method may further include receiving from the user an input indicating a head size of the user.
  • the parameter (or the plurality of parameters) for the head-related transfer function may be determined further based on the received input indicating the head size of the user.
  • the method may further include receiving from the user an input indicating a head shape of the user.
  • the parameter (or the plurality of parameters) for the head-related transfer function may be determined further based on the received input indicating the head shape of the user.
  • the method may further include receiving from the user an input indicating an ear size of the user.
  • the parameter (or the plurality of parameters) for the head- related transfer function may be determined further based on the received input indicating the ear size of the user.
  • the method may further include receiving from the user an input indicating an ear shape of the user.
  • the parameter (or the plurality of parameters) for the head-related transfer function may be determined further based on the received input indicating the ear shape of the user.
  • the receiving and the determining may be performed for a plurality of virtual speaker positions.
  • the method may further include sending the determined parameter (or plurality of determined parameters) to a se er in a cloud.
  • the method may further include: receiving a parameter (or a plurality of parameters) for the head-related transfer function from a server in a cloud; modifying the audio signal in accordance with the head-related transfer function based on the received parameter (or the plurality of received parameters); and outputting the modified audio signal.
  • the parameter (or the plurality of parameters) for the head-related transfer function may be determined based on the received input indicating the angle using a lookup table storing a relation between angles and parameters (or the plurality of parameters).
  • FIG. 2 shows an audio output device 200 in accordance with an embodiment.
  • the audio output device 200 may include an input circuit 202 configured to receive from a user an input indicating an angle.
  • the audio output device 200 may further include a determination circuit 204 configured to determine a parameter (or a plurality of parameters) for a head-related transfer function based on the received input indicating the angle.
  • the audio output device 200 may further include a modification circuit 206 , configured to modify an audio signal in accordance with the head-related transfer function based on the determined parameter (or the plurality of determined parameters).
  • the audio output device 200 may further include an output circuit 208 configured to output the modified audio signal.
  • the input circuit 202, the determination circuit 204, the modification circuit 206, and the output circuit 208 may be connected via a connection 210 (or a plurality of separate connections), for example an electrical or optical connection, for example any kind of cable or bus.
  • the input circuit 202 may be configured to receive a graphical input indicating the angle by a point on a geometric shape.
  • the input circuit 202 may be configured to receive a graphical input indicating the angle by a direction from a center of a geometric shape.
  • the input circuit 202 may be configured to receive a real number indicating the angle.
  • the audio output device 200 may further include a display circuit (not shown) configured to display a presently set angle.
  • the input circuit 202 may be configured to receive from the user an indication for increasing or decreasing the angle in response to the displayed presently set angle.
  • the audio output device 200 may further include a setting circuit (not shown) configured to set the angle based on the indication.
  • the input circuit 202 may further be configured to receive from the user an input indicating a head size of the user.
  • the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the head size of the user.
  • the input circuit 202 may further be configured to receive from the user an input indicating a head shape of the user.
  • the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the head shape of the user.
  • the input circuit 202 may further be configured to receive from the user an input indicating an ear size of the user.
  • the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the ear size of the user.
  • the input circuit 202 may further be configured to receive from the user an input indicating an ear shape of the user.
  • the determination circuit 204 may further be configured to determine the parameter (or the plurality of parameters) for the head- related transfer function based on the received input indicating the ear shape of the user.
  • the input circuit 202 and the determination circuit 204 may be configured to perform the receiving and the determining for a plurality of virtual speaker positions.
  • the audio output device 200 may further include a sending circuit (not shown) configured to send the determined parameter (or the plurality of determined parameters) to a server in a cloud.
  • a sending circuit (not shown) configured to send the determined parameter (or the plurality of determined parameters) to a server in a cloud.
  • the audio output device 200 may further include a receiving circuit (not shown) configured to receive a parameter (or a plurality of parameters) for the head-related transfer function from a server in a cloud.
  • the modification circuit 206 may be configured to modify the audio signal in accordance with the head-related transfer function based on the received parameter (or the plurality of received parameters).
  • the output circuit 208 may be configured to output the modified audio signal.
  • the determination circuit 204 may be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function based on the received input indicating the angle using a lookup table storing a relation between angles and parameters (or the plurality of parameters).
  • FIG. 3 shows a graphical user interface 300 in accordance with an embodiment.
  • the graphical user interface 300 may for example be displayed on a computer screen 302.
  • the graphical user interface 300 may include an application program window 304 generated by the application program.
  • the application program window 304 may include a visual representation 306 of the user (here shown as a geometric shape).
  • the application program window 304 may further include a visual representation 308 of a speaker (here shown as a geometric shape) on a geometric shape 310 around the user.
  • the geometric shape 310 in the graphical user interface may be displayed as any kind of geometrical form, for example an ellipse, wherein the user 306 is located at the center of the geometric shape, for example the ellipse 310.
  • the speaker 308 is movable along the geometric shape 310, wherein the positioning of the speaker 308 translates to inputting an indication of an angle on the geometric shape 310 with respect to the user 306.
  • the application program window 304 may further include an input 312 for inputting an indication of an angle on the geometric shape with respect to the visual representation of the speaker (here shown as a geometric shape that may be moved on the geometric shape 310), wherein the input 312 is associated to the speaker 308. The moving of the speaker 308 adjusts the angle.
  • FIG. 4 shows an alternative implementation 400.
  • a movable input 312 may be used to adjust the angle.
  • the visual representation 308 of the speaker may include an image of a speaker or other images of output speaker device.
  • the visual representation 308 of the speaker may include a ball.
  • the input 3 12 may include a marker configured to be moved on the geometric shape 3 10.
  • the input 312 may include the visual representation of the speaker configured to be moved on the geometric shape.
  • the visual representation 308 of the speaker or the input 312 may include a needle of a compass configured to be moved around the geometric shape 310 with respect to the user 306.
  • the graphical user interface 300 may be configured to send the input indication of the angle to the application program.
  • the application program may be configured to: determine a parameter (or a plurality of parameters) for a head-related transfer function based on the received input indicating the angle, modify an audio signal in accordance with the head-related transfer function based on the determined parameter (or the plurality of determined parameters); and output the modified audio signal.
  • the application program window 304 may further be configured to receive from the user an input indicating a head size of the user.
  • the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the head size of the user.
  • the application program window 304 may further be configured to receive from the user an input indicating a head shape of the user.
  • the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the head shape of the user.
  • the application program window 304 may further be configured to receive from the user an input indicating an ear size of the user.
  • the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the ear size of the user.
  • the application program window 304 may further be configured to receive from the user an input indicating an ear shape of the user.
  • the application program may further be configured to determine the parameter (or the plurality of parameters) for the head-related transfer function further based on the received input indicating the ear shape of the user.
  • the application program window 304 may include visual representations of a plurality of speakers and the input 312 may be for inputting an angle for each of the visual representation of each speaker.
  • the application program window 304 may further include a sender input for receiving an input for instructing the application program to send a parameter (or a plurality of parameters) determined based on the input angle to a server in a cloud.
  • the application program window 304 may further include a receiver for receiving a parameter (or a plurality of parameters) for a head-related transfer function from a server in a cloud.
  • the application program may be configured to modify the audio signal in accordance with the head-related transfer function based on the received parameter (or the plurality of received parameters) and to output the modified audio signal.
  • the application program may be configured to determine a parameter (or a plurality of parameters) for a head-related transfer function based on the input indication of the angle using a lookup table storing a relation between angles and parameters (or the plurality of parameters) for the head- related transfer function.
  • the seven speakers may represent a 7.1 sound system.
  • Typical sound systems may be 5.1 (e.g. cinema theatre).
  • the UI may be provided to (i) allow user to calibrate the sound setting of a 7.1 audio headset and/ or (ii) perform virtualization so that a 2.1 or 5.1 headset sounds like a 7.1 system to the user.
  • a method of HRTF calibration may be provided.
  • a head-reflectance transfer function (or head-related transfer function; HRTF) may be applied to an incoming analog stereo audio signal in order to create the illusion of a multi-channel audio system through typical stereo headphones.
  • HRTF head-related transfer function
  • a method of calibrating an HRTF system using a graphical user interface to position virtual speaker positions may be provided, in a way which is easy to understand and manipulate by a novice user with no prior experience in tuning an HRTF.
  • further the association of the HRTF calibration parameters determined by the user with a unique cloud identifier for that user, storing these settings for use across any device connecting to the cloud service may be provided. Cloud identification may enable not only the storage of an HRTF calibration profile for a particular user, but also the machine and devices in the audio reproduction environment such as the digital-to-analog converter (DAC), headphone amplifier, and the headphones or headset used to reproduce sound.
  • DAC digital-to-analog converter
  • a head-related transfer function may be a response that may describe how an ear receives a sound from a point in space; a pair of HRTFs for two ears may be used to synthesize a binaural sound that seems to come from a particular point in space. It may be a transfer function, describing how a sound from a specific point will arrive at the ear (generally at the outer end of the auditory canal). Some consumer home entertainment products designed to reproduce surround sound from stereo (two-speaker) headphones may use HRTFs. Some forms of HRTF-processing may have also been included in computer software to simulate surround sound playback from loudspeakers.
  • Humans may estimate the location of a source by taking cues derived from one ear (monaural cues), and by comparing cues received at both ears (difference cues or binaural cues). Among the difference cues may be time differences of arrival and intensity differences.
  • the monaural cues may come from the interaction between the sound source and the human anatomy, in which the original source sound may be modified before it enters the ear canal for processing by the auditory system. These modifications may encode the source location, and may be captured via an impulse response which may relate the source location and the ear location.
  • This impulse response may be termed the head-related impulse response (HRIR): Convolution of an arbitrary source sound with the HRIR may convert the sound to that which would have been heard by the listener if it had been played at the source location, with the listener's ear at the receiver location. HRIRs may have been used to produce virtual surround sound.
  • HRIR head-related impulse response
  • the HRTF may be the Fourier transform of HRIR.
  • the HRTF may also be referred to as the anatomical transfer function (ATF).
  • HRTFs for left and right ear may describe the filtering of a sound source (x(t)) before it is perceived at the left and right ears as xL(t) and xR(t), respectively.
  • the HRTF may also be described as the modifications to a sound from a direction in free air to the sound as it arrives at the eardrum. These modifications may include the shape of the listener's outer ear, the shape of the listener's head and body, the acoustical properties of the space in which the sound is played, and so on. All these properties may influence how (or whether) a listener may accurately tell what direction a sound is coming from.
  • HRTFs may vary significantly from person to person. Perceptual distortions may occur when one listens to sounds spatialized with non-individualized HRTF. This focus on a "one size fits all" approach to HRTF may assume average physiology and morphology across all users.
  • the size of the head and placement of the headphones or headset may be a critical determining factor in how the coefficients are created and applied to the filter.
  • the shape of the ear and size of the ears may have a major impact on how sounds propagate from the drivers to the inner ear. As such, one size or type of filter does not fit all listeners, resulting i poor performance of the virtual surround sound system as it targets the average values found across all humans.
  • a user may establish settings which may be unique to that user, and those settings may not persist across multiple systems and device chains, as the parameters established by the user during calibration remain locked into that particular device for which settings were configured.
  • the user may calibrate the HRTF filter so that it works best for them, and then may save those settings to the client and mirroring them into the cloud for use on any client in the future.
  • the HRTF calibration parameters used on one audio reproduction system such as a personal computer
  • the user may configure or calibrate the HRTF algorithm on a singular device or system and have those systems persist across a multitude of devices and systems through software interface which authenticate them and transport their profile settings from the cloud into that system.
  • a graphical user interface for the calibration of an HRTF and cloud services which synchronize a singular device or system with a multitude of devices and systems through the use of unique identifiers for the user and the device(s) and machine(s) used to calibrate the HRTF algorithm may be provided. This may be based on user input to determine offsets in the virtual speaker positions, selection of head size, ear size and ear shape, which determine the appropriate HRTF coefficients to apply within the audio filter. For example, HRTF coefficients may be stored in the cloud and downloaded to the client, or may be included in the installation and local-only. Once the user calibrates, and selections of HRTF coefficients are made, the user's chosen configuration may be stored in the cloud for use on any other PC client they log into.
  • the first selection that a user may make is his or her head size, which may be a subjective selection given a set of options such as small, medium and large, corresponding to the circumference of the head as measured at the brow and around the largest area of the back of the head. This may be similar in approach to that of hat size measurement.
  • the options may be relative to an average circumference of 58 cm, with a standard deviation of +/- 7 cm.
  • the second selection that a user may make is his or her head shape, based on a set of provided options such as round, oval, and inverted egg.
  • the third selection that a user may make is his or her ear size, which may be a subjective selection given a set of options such as small, medium and large, corresponding to the size of their outer ear (pinna, or auricle) relative to an average of 6 cm with a standard deviation of +/- 1.5 cm.
  • the fourth selection that a user makes is his or her ear shape, which may be a subjective selection by the user of 8 common ear shape types.
  • FIG. 5A is a diagram of an application program window 400 in accordance with one embodiment of the present invention.
  • the application program window 400 includes a first sub-window 401, a second sub-window 402, a third sub-window 403 and a fourth sub- window 404.
  • the first sub-window 401 provides means for the users to select their head size.
  • the first sub-window 401 includes an input window for the user to input/type their head circumference size.
  • the first sub-window 401 may contain a drop-down menu/list with preset head circumference sizes, e.g. from 51cm to 65cm, wherein the preset sizes are selectable by the user.
  • the first sub-window 401 includes a plurality of images illustrating the different head sizes, e.g. a first image with a range of 51cm-55cm, a second image with a range of 56cm-60cm, and a third image with a range of 61cm-65cm.
  • the users select the image with the closest range to their head size.
  • the first sub- window may be a combination of the input window, the drop-down menu, or the plurality of images to allow flexibility in selecting the head size.
  • the second sub-window 402 provides means for the users to select their head shape.
  • the second sub-window 402 includes a drop-down menu/list with preset head shapes, e.g.
  • the second sub-window 402 includes a plurality of images illustrating the different head shapes, and the users select the closest image to their head shape.
  • the second sub-window 402 is a combination of the drop-down menu with preset head shapes and plurality of images with different head shapes.
  • the third sub-window 403 provides means for the users to select their ear size.
  • the third sub-window 403 includes an input window for the user to input/type their ear size.
  • the third sub-window 403 may contain a drop-down menu/list with preset ear sizes, e.g. outer ear size of about 4.5cm to 7.5cm, wherein the preset sizes are selectable by the user.
  • the third sub-window 403 includes a plurality of images illustrating the different ear sizes, e.g.
  • the third sub-window 403 may be a combination of the input window, the drop-down menu, or the plurality of images.
  • the fourth sub-window 404 provides means for the users to select their ear shape.
  • the fourth sub-window 404 includes a plurality of images illustrating the different ear shapes to allow the users to select their ear shape that is closest to one of the images.
  • FIG. 5B shows the fourth sub-window 404 with a plurality of images representing common ear shapes.
  • further adjustments may be to the positioning of virtual surround sound speaker locations by the user, to personalize his or her listening experience.
  • This method may enable the user to more fully realize the surround sound spatiality by making adjustments to the virtual speaker location in the graphical user interface, which may be translated into adjustments to the HRTF coefficients for each speaker positions.
  • the results of such a graphical method of calibration may be immediately apparent to the user, as he or she may perceive the changes in virtual speaker positions during the calibration steps.
  • the method may begin by instructing the user to place his or her preferred headphones onto his or her head, while an audio clip is playing which cycles through the default virtual speaker locations.
  • the user may hear what these default positions sound like given his or her morphological parameters such as head size and ear shape, as well as the mechanical design and other characteristics of his or her preferred headphones.
  • FIG. 6A shows a screen shot 601 of a graphical user interface for calibrating virtual speaker positions in accordance with an embodiment.
  • the user may select one of the virtual speaker locations 611 to 617 and may be presented with a respective marker highlighting the position of the audio at the default angle (azimuth) relative to the head/user representation 670 at the center of the sound field.
  • FIG. 6B shows a screen shot 602 of a graphical user interface in accordance with an embodiment, wherein a virtual speaker location marker 626 is shown when the virtual speaker location 616 is selected.
  • the user may choose to adjust the position of this marker 626 that is associated only to virtual speaker location 616, which may result in adjustments made to the HRTF coefficients applied by the filter, in order to shift the perceived audio origination point (from the virtual speaker location 616) around the sound field. For example, by positioning of the virtual surround sound speaker locations by the user, the user may "freely" shift/move the 7 virtual speaker locations 61 1 to 617 within the geometric shape 650.
  • the user may fully customize the sound field to his or her preference. According to various embodiments, this may enable the user to achieve a subjectively better virtual surround sound audio experience through real-time adjustments of the speaker positions with a synthesized multi-channel surround sound audio source playing through the HRTF filter being modified.
  • the new HRTF coefficients may be saved for that particular user and optionally associated with his or her preferred headphones. Other HRTF calibrations may be performed for other headphones, enabling the user to customize and calibrate his or her HRTF filter library for a plurality of headphone or headset devices.
  • the appropriate HRTF filter coefficients may be applied to a wide variety of applications, and may persist across multiple devices and systems used by the user. This may ensure the best possible virtual surround sound experience no matter which system the user is currently using.
  • the current state of HRTF calibration may be limited to a standard set of predetermined filters, created based on objective morphological factors and may not present the user with affordances for calibrating their virtual surround sound experience using a graphical user interface to select their morphological parameters and control the positioning of virtual surround sound speaker positions.
  • the state of the art may be advanced according to various embodiments by associating a unique profile created by the user for their device(s) and system(s) with their unique identification in the cloud service, enabling a consistent experience across multiple devices through connectivity of HRTF calibration software to the cloud service.
  • FIG. 7 shows a screen shot 700 of a graphical user interface or application program window in accordance with an embodiment, wherein an audio output device may be set.
  • the application program window 700 is generated by the application program operating on a computing device.
  • the application program is connected to a remote server over a network (e.g. Internet cloud service), wherein the application program stores user profiles on the remote server or receives stored user profiles from the remote server.
  • a network e.g. Internet cloud service
  • the application program retrieves the stored user profiles associated with the user ID from the remote server and displays it in the on a profile sub-window 720 of the application program window 700.
  • ID unique user identification
  • the profile sub-window 720 displays a list of user profiles and also enables new profiles to be created, wherein the list of user profiles is retrievable from the remote server or from the local client (i.e. the computer device).
  • a particular user profile e.g. "Profile” shown in FIG 7
  • an audio device sub-window 730 displays a list of audio devices associated with the user profile. For example, all Razer analog headphones and headsets may be provided in the list of audio devices and one of them may be selected for calibration.
  • the application program displays the name of the new headphone/headset in the audio device sub-window 730 if it is compatible with the application program.
  • the application program window 700 comprises a top menu bar 710, including the functions "SETTINGS”, “AUDIO”, “EQ” and “CALIBRATION".
  • FIG. 7 shows the application program window 700 when the "SETTINGS” 71 1 function is selected by the user.
  • FIG. 8 shows a screen shot 800 of a graphical user interface or application program window in accordance with an embodiment, wherein general audio output parameters may be set.
  • the application program window 800 displays an audio output sub-window 810 as well as the profile sub-window 720.
  • the audio output sub-window 810 enables the user to adjust audio output parameters, such as but not limited to "BASS BOOST", “VOICE CLARITY”, “VOLUME NORMALIZATION” and "VOLUME LEVEL".
  • the application program associates or stores the data of the desired audio output parameters to the selected profile in the profile sub-window 720.
  • the profile and the associated audio output parameters can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
  • FIG. 9 shows a screen shot 900 of a graphical user interface or application program window in accordance with an embodiment, wherein equalizer (EQ) parameters may be set.
  • the application program window 900 displays a drop-down menu with preset EQ settings 910, wherein the drop-down menu may include common EQ settings such as but not limited to "Classical”, “Rock”, “Dance”, “Jazz”, etc.
  • the application program window 900 includes a plurality of EQ frequency bars that enables user to configure the desired EQ settings.
  • the application program window 900 includes the profile sub-window 720.
  • the application program associates or stores the data of the desired EQ settings to the selected profile in the profile sub-window 720.
  • the profile and the associated EQ settings can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
  • FIG. 10 shows a screen shot 1000 of a graphical user interface or application program window in accordance with an embodiment, wherein the position of a virtual speaker may be adjusted.
  • the application program window 1000 displays a representation of a plurality of speakers 1101-1107 arranged on a circular path 1002.
  • a representation of a user 1001 is located at the central position of the circular path 1002.
  • the default position of the speaker 1101 is about 0 degrees from the user 1001.
  • speaker 1102 is positioned around 45 degrees, speaker 1 103 around 90 degrees, speaker 1104 around 135 degrees, speaker 1105 around 225 degrees, speaker 1 106 around 270 degrees and speaker 1 107 around 325 degrees relative to the user 1001 respectively.
  • the 7 speakers 1101-1107 represent a 7.1 surround system but it can be appreciated that for other surround systems the number of speakers may vary, e.g. 5 speakers may be used to represent a 5.1 surround system.
  • the circular path 1002 may take other forms/shapes, such as a square path of a rectangular path.
  • a surround sound audio loop may be playing, such as a helicopter, which may move around all the virtual speaker positions.
  • the user may click on a "Test All” button 1201 to replay this surround sound audio loop and listen to all the speaker positions with any changes made.
  • a virtual speaker location for example speaker 1102
  • the other speakers may fade away and the selected speaker may be highlighted, like shown in FIG. 1 1.
  • An audio loop may play from the selected virtual speaker location.
  • FIG. 1 1 shows a screen shot 1100 of a graphical user interface in accordance with an embodiment, wherein a marker 1 122 indicating an angle for the chosen virtual speaker 1 102 may be set.
  • the user may click the calibration marker 1 122 (with the shape of a ball) and may drag it around the circular path 1002 to varying degrees based on the speaker 1 102 selected to adjust the position of the sound until it appears to originate from the virtual speaker location, or their desired location.
  • the virtual speaker location 1 102 is not moveable.
  • the calibration marker 1 122 may not end up directly on top of the speaker 1202 - it may simply be an offset of the sound to account for ear and head size, headphone type, for example.
  • the user interface does not include the calibration marker 1 122 and the user may adjust the position of the sound from the speaker 1102 by clicking and dragging the speaker 1102 around the circular path 1002.
  • the adjustment of the calibration marker 1 122 that is associated only to speaker 1 102 results in adjustments to the HRTF coefficient associated with speaker 1102.
  • the user may fully customize the sound field to his or her preference.
  • the new HRTF coefficients are stored with the desired profile and associated with his or her preferred headphones selected in the audio device sub-window 730 show in FIG 7.
  • Other HRTF calibrations may be performed for other headphones, enabling the user to customize and calibrate his or her HRTF filter library for a plurality of headphone or headset devices.
  • the calibration marker or speaker may be moved only to a certain degree, so as for example not to move a right speaker entirely to a left side or to move a front speaker to the rear. This helps users to maintain the audio fidelity while calibrating their headphone/headset.
  • the azimuth for each speaker 1101-1 107 may be restricted such that the user may move a selected speaker about +/-15 to 20 degrees from its default/original position.
  • the range of angles of azimuth of speaker 1 102 may be fixed to a corresponding zone/region 1 132 so that the speaker 1102 (or its associated marker) is restricted from being moved outside the zone 1132.
  • the speaker 1102 (or its associated marker) is only slidable around the circular path 1002 within the zone 1 132, or moveable about +/- 20 degrees from its default angle of 45 degrees (i.e. from 25 degrees to 65 degrees).
  • the application program determines a parameter for a HRTF based on the angle of the speaker 1 102 (or its associated calibration marker).
  • the application program modifies an audio signal in accordance with the HRTF based on the determined parameter, resulting in the output of the modified audio signal from the speaker 1102 to the user.
  • the application program may be configured to determine a parameter for HRTF based on the input indication of the angle using a lookup table storing a relation between angles and parameters for the HRTF.
  • the application program associates or stores the parameter of the HRTF to the selected profile in the profile sub-window 720.
  • the profile and the parameters of the HRTF can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
  • the application program includes an interface for the user to select the size and shape of his/her head and ear (similar to FIG 5A as discussed above) prior to the calibration of the speakers 1101-1 107 shown in FIGS 10 and 1 1.
  • the application program determines the parameter for the HRTF based on the received input indicating the size and shape of the head and ear.
  • the application program modifies an audio signal in accordance with the HRTF based on the determined parameter, resulting in the output of the modified audio signal from the speakers 1 101-1 107 to the user.
  • the application program may be configured to determine a parameter for HRTF based on the input indication of the head shape, head size, ear shape or ear size using a lookup table storing a relation between the head shape, head size, ear shape, ear size and parameters for the HRTF.
  • the application program associates or stores the parameter of the HRTF to the selected profile in the profile sub-window 720.
  • the profile and the parameters of the HRTF can then be stored on the remote server over the network so that it is retrievable by the user when he /she subsequently logins to the application program using the same user ID and password.
  • the application program window comprises a "Reset” button to allow the user to reset the speakers 1 101-1 107 to their default positions.
  • a checkbox may be provided, and checking the checkbox may override calibration settings saved to the desired profile, e.g. "Profile” shown in the profile sub-window 720 in FIG. 12 and apply the calibration settings globally to all profiles associated to the unique user ID.
  • Profile e.g. "Profile” shown in the profile sub-window 720 in FIG. 12
  • FIG. 13 shows a screen shot 1300 of a graphical user interface or application program window showing alternative representation of the speakers.
  • the speakers are represented by an "arrow-head” image instead of a loudspeaker image illustrated in FIG. 10.
  • the "arrow-head” speakers 1311-1317 are moveable on a circular path 1301.
  • the application program window includes another circular path 1302 concentric to circular path 1301.
  • the circular path 1302 includes a plurality of calibration markers positioned adjacent to the speakers 1311-1317 in their default position.
  • the speaker 1312 includes a corresponding calibration marker 1322 that is moveable within a zone/region 1332 of the circular path 1302.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)
  • Electrophonic Musical Instruments (AREA)

Abstract

L'invention concerne, dans divers modes de réalisation, un procédé d'émission d'un signal audio modifié. Le procédé peut comprendre les étapes consistant à: recevoir de la part d'un utilisateur une entrée indiquant un angle; déterminer un paramètre pour une fonction de transfert liée à la tête en se basant sur l'entrée reçue indiquant l'angle; modifier un signal audio d'après la fonction de transfert liée à la tête en se basant sur le paramètre déterminé; et émettre le signal audio modifié.
PCT/SG2012/000439 2012-11-22 2012-11-22 Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application Ceased WO2014081384A1 (fr)

Priority Applications (7)

Application Number Priority Date Filing Date Title
AU2012394979A AU2012394979B2 (en) 2012-11-22 2012-11-22 Method for outputting a modified audio signal and graphical user interfaces produced by an application program
CN201280078150.3A CN105027580B (zh) 2012-11-22 2012-11-22 用于输出修改的音频信号的方法
SG11201503926WA SG11201503926WA (en) 2012-11-22 2012-11-22 Method for outputting a modified audio signal and graphical user interfaces produced by an application program
EP12888662.9A EP2923500A4 (fr) 2012-11-22 2012-11-22 Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application
US14/646,677 US9569073B2 (en) 2012-11-22 2012-11-22 Method for outputting a modified audio signal and graphical user interfaces produced by an application program
PCT/SG2012/000439 WO2014081384A1 (fr) 2012-11-22 2012-11-22 Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application
TW102134986A TWI616810B (zh) 2012-11-22 2013-09-27 輸出一調變音訊之方法及藉由一應用程式產生之圖形使用者介面

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/SG2012/000439 WO2014081384A1 (fr) 2012-11-22 2012-11-22 Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application

Publications (1)

Publication Number Publication Date
WO2014081384A1 true WO2014081384A1 (fr) 2014-05-30

Family

ID=50776418

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/SG2012/000439 Ceased WO2014081384A1 (fr) 2012-11-22 2012-11-22 Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application

Country Status (7)

Country Link
US (1) US9569073B2 (fr)
EP (1) EP2923500A4 (fr)
CN (1) CN105027580B (fr)
AU (1) AU2012394979B2 (fr)
SG (1) SG11201503926WA (fr)
TW (1) TWI616810B (fr)
WO (1) WO2014081384A1 (fr)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017058097A1 (fr) * 2015-09-28 2017-04-06 Razer (Asia-Pacific) Pte. Ltd. Ordinateurs, procédés de commande d'un ordinateur, et supports lisibles par ordinateur
CN107484097A (zh) * 2017-10-10 2017-12-15 歌尔股份有限公司 耳机左右喇叭防呆检测机构
WO2019166030A1 (fr) * 2018-02-28 2019-09-06 成都星环科技有限公司 Système d'adaptation de champ sonore intelligent, système d'étalonnage de champ sonore intelligent et système de traitement vocal intelligent
EP3554098A4 (fr) * 2016-12-12 2019-12-18 Sony Corporation Procédé de mesure hrtf, dispositif de mesure hrtf et programme

Families Citing this family (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9219460B2 (en) 2014-03-17 2015-12-22 Sonos, Inc. Audio settings based on environment
US9106192B2 (en) 2012-06-28 2015-08-11 Sonos, Inc. System and method for device playback calibration
WO2014081384A1 (fr) * 2012-11-22 2014-05-30 Razer (Asia-Pacific) Pte. Ltd. Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application
US20160147415A1 (en) * 2013-08-01 2016-05-26 Thales Programming system for a situation analysis system on board a carrier comprising at least one onboard listening system
CN104197299A (zh) * 2014-08-21 2014-12-10 浙江生辉照明有限公司 照明装置及基于该装置的语音播报系统及方法
US9952825B2 (en) 2014-09-09 2018-04-24 Sonos, Inc. Audio processing algorithms
US9544706B1 (en) * 2015-03-23 2017-01-10 Amazon Technologies, Inc. Customized head-related transfer functions
US11113022B2 (en) * 2015-05-12 2021-09-07 D&M Holdings, Inc. Method, system and interface for controlling a subwoofer in a networked audio system
US11209972B2 (en) 2015-09-02 2021-12-28 D&M Holdings, Inc. Combined tablet screen drag-and-drop interface
US9609436B2 (en) * 2015-05-22 2017-03-28 Microsoft Technology Licensing, Llc Systems and methods for audio creation and delivery
WO2017049169A1 (fr) 2015-09-17 2017-03-23 Sonos, Inc. Faciliter l'étalonnage d'un dispositif de lecture audio
US9693165B2 (en) 2015-09-17 2017-06-27 Sonos, Inc. Validation of audio calibration using multi-dimensional motion check
US10142755B2 (en) * 2016-02-18 2018-11-27 Google Llc Signal processing methods and systems for rendering audio on virtual loudspeaker arrays
FI20165211A (fi) 2016-03-15 2017-09-16 Ownsurround Ltd Järjestely HRTF-suodattimien valmistamiseksi
CN107204132A (zh) * 2016-03-16 2017-09-26 中航华东光电(上海)有限公司 3d虚拟立体声空中预警系统
US9860662B2 (en) 2016-04-01 2018-01-02 Sonos, Inc. Updating playback device configuration information based on calibration data
US9763018B1 (en) 2016-04-12 2017-09-12 Sonos, Inc. Calibration of audio playback devices
CN105933826A (zh) * 2016-06-07 2016-09-07 惠州Tcl移动通信有限公司 一种自动设置声场的方法、系统及耳机
US9794710B1 (en) 2016-07-15 2017-10-17 Sonos, Inc. Spatial audio correction
US10372406B2 (en) 2016-07-22 2019-08-06 Sonos, Inc. Calibration interface
US10459684B2 (en) 2016-08-05 2019-10-29 Sonos, Inc. Calibration of a playback device based on an estimated frequency response
USD818488S1 (en) * 2016-10-14 2018-05-22 Keepitsafe, Inc. Display screen or portion thereof with a graphical user interface
US10901681B1 (en) * 2016-10-17 2021-01-26 Cisco Technology, Inc. Visual audio control
KR20190091445A (ko) * 2016-10-19 2019-08-06 오더블 리얼리티 아이엔씨. 오디오 이미지를 생성하는 시스템 및 방법
CN107979807A (zh) * 2016-10-25 2018-05-01 北京酷我科技有限公司 一种模拟环绕立体声的方法及系统
US20180115854A1 (en) * 2016-10-26 2018-04-26 Htc Corporation Sound-reproducing method and sound-reproducing system
US10375498B2 (en) 2016-11-16 2019-08-06 Dts, Inc. Graphical user interface for calibrating a surround sound system
TWI630828B (zh) * 2017-06-14 2018-07-21 趙平 供使用者定向交談之智慧型耳機裝置個人化系統及其使用方法
US20180367935A1 (en) * 2017-06-15 2018-12-20 Htc Corporation Audio signal processing method, audio positional system and non-transitory computer-readable medium
CN108196815B (zh) * 2017-12-28 2020-01-03 维沃移动通信有限公司 一种通话声音的调节方法和移动终端
US10412527B1 (en) * 2018-01-09 2019-09-10 Facebook Technologies, Llc Head-related transfer function determination using base stations
US10225656B1 (en) * 2018-01-17 2019-03-05 Harman International Industries, Incorporated Mobile speaker system for virtual reality environments
US10652686B2 (en) * 2018-02-06 2020-05-12 Sony Interactive Entertainment Inc. Method of improving localization of surround sound
FI20185300A1 (fi) 2018-03-29 2019-09-30 Ownsurround Ltd Järjestely päähän liittyvien siirtofunktiosuodattimien muodostamiseksi
JP7442494B2 (ja) 2018-07-25 2024-03-04 ドルビー ラボラトリーズ ライセンシング コーポレイション 光学式捕捉によるパーソナライズされたhrtf
US10299061B1 (en) 2018-08-28 2019-05-21 Sonos, Inc. Playback device calibration
WO2020044244A1 (fr) 2018-08-29 2020-03-05 Audible Reality Inc. Système et procédé de commande d'un moteur audio tridimensionnel
US10856097B2 (en) 2018-09-27 2020-12-01 Sony Corporation Generating personalized end user head-related transfer function (HRTV) using panoramic images of ear
US11113092B2 (en) 2019-02-08 2021-09-07 Sony Corporation Global HRTF repository
US11451907B2 (en) 2019-05-29 2022-09-20 Sony Corporation Techniques combining plural head-related transfer function (HRTF) spheres to place audio objects
US11347832B2 (en) 2019-06-13 2022-05-31 Sony Corporation Head related transfer function (HRTF) as biometric authentication
US10734965B1 (en) 2019-08-12 2020-08-04 Sonos, Inc. Audio calibration of a portable playback device
US11146908B2 (en) 2019-10-24 2021-10-12 Sony Corporation Generating personalized end user head-related transfer function (HRTF) from generic HRTF
US11070930B2 (en) 2019-11-12 2021-07-20 Sony Corporation Generating personalized end user room-related transfer function (RRTF)
USD958166S1 (en) * 2019-11-19 2022-07-19 Johnson Systems Inc. Display screen with graphical user interface
USD926789S1 (en) * 2019-11-19 2021-08-03 Johnson Systems Inc. Display screen with graphical user interface
CN115280799A (zh) * 2019-12-31 2022-11-01 哈曼国际工业有限公司 利用不可见扬声器实现虚拟音效的系统和方法
WO2022010453A1 (fr) * 2020-07-06 2022-01-13 Hewlett-Packard Development Company, L.P. Annulation de traitement spatial dans des écouteurs
US11854555B2 (en) * 2020-11-05 2023-12-26 Sony Interactive Entertainment Inc. Audio signal processing apparatus, method of controlling audio signal processing apparatus, and program
CN113365191A (zh) * 2021-06-04 2021-09-07 广州酷狗计算机科技有限公司 音乐播放的方法、装置、设备及存储介质
US20220407772A1 (en) * 2021-06-16 2022-12-22 Hewlett-Packard Development Company, L.P. Configuration profiles
EP4409571B1 (fr) 2021-09-30 2025-03-26 Sonos Inc. Gestion de conflit pour processus de détection de mot d'activation
CN116095595B (zh) * 2022-08-19 2023-11-21 荣耀终端有限公司 音频处理方法和装置
CN116744215B (zh) * 2022-09-02 2024-04-19 荣耀终端有限公司 音频处理方法和装置

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060280323A1 (en) * 1999-06-04 2006-12-14 Neidich Michael I Virtual Multichannel Speaker System
KR20080055622A (ko) * 2006-12-16 2008-06-19 김종철 인간친화형 컴퓨터입출력 시스템
US20120022842A1 (en) * 2009-02-11 2012-01-26 Arkamys Test platform implemented by a method for positioning a sound object in a 3d sound environment
US20120093348A1 (en) * 2010-10-14 2012-04-19 National Semiconductor Corporation Generation of 3D sound with adjustable source positioning
US20120201405A1 (en) * 2007-02-02 2012-08-09 Logitech Europe S.A. Virtual surround for headphones and earbuds headphone externalization system

Family Cites Families (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6046722A (en) * 1991-12-05 2000-04-04 International Business Machines Corporation Method and system for enabling blind or visually impaired computer users to graphically select displayed elements
US5850455A (en) * 1996-06-18 1998-12-15 Extreme Audio Reality, Inc. Discrete dynamic positioning of audio signals in a 360° environment
US6266571B1 (en) * 1997-10-29 2001-07-24 International Business Machines Corp. Adaptively configuring an audio interface according to selected audio output device
US6067084A (en) * 1997-10-29 2000-05-23 International Business Machines Corporation Configuring microphones in an audio interface
US6504553B2 (en) * 1999-02-25 2003-01-07 International Business Machines Corporation Trouble shooting a wrong audio source
US6492999B1 (en) * 1999-02-25 2002-12-10 International Business Machines Corporation Connecting and optimizing audio input devices
GB2374507B (en) * 2001-01-29 2004-12-29 Hewlett Packard Co Audio user interface with audio cursor
US20030227476A1 (en) * 2001-01-29 2003-12-11 Lawrence Wilcock Distinguishing real-world sounds from audio user interface sounds
GB2374772B (en) * 2001-01-29 2004-12-29 Hewlett Packard Co Audio user interface
GB2376123B (en) * 2001-01-29 2004-06-30 Hewlett Packard Co Facilitation of speech recognition in user interface
US7073125B1 (en) * 2002-04-26 2006-07-04 Microsoft Corporation Integrated display of gauges
US20040264704A1 (en) 2003-06-13 2004-12-30 Camille Huin Graphical user interface for determining speaker spatialization parameters
US7928311B2 (en) * 2004-12-01 2011-04-19 Creative Technology Ltd System and method for forming and rendering 3D MIDI messages
FR2880755A1 (fr) 2005-01-10 2006-07-14 France Telecom Procede et dispositif d'individualisation de hrtfs par modelisation
US7461345B2 (en) * 2005-03-11 2008-12-02 Adobe Systems Incorporated System and method for displaying information using a compass
US20060277034A1 (en) 2005-06-01 2006-12-07 Ben Sferrazza Method and system for processing HRTF data for 3-D sound positioning
US7549123B1 (en) * 2005-06-15 2009-06-16 Apple Inc. Mixing input channel signals to generate output channel signals
WO2007048900A1 (fr) 2005-10-27 2007-05-03 France Telecom Individualisation de hrtfs utilisant une modelisation par elements finis couplee a un modele correctif
US20080178088A1 (en) * 2006-07-27 2008-07-24 Personics Holdings Inc. Method and device of customizing headphones
US7792674B2 (en) * 2007-03-30 2010-09-07 Smith Micro Software, Inc. System and method for providing virtual spatial sound with an audio visual player
JP4355359B1 (ja) * 2008-05-27 2009-10-28 パナソニック株式会社 マイクを外耳道開口部に設置する耳掛型補聴器
TWI384889B (zh) * 2008-12-26 2013-02-01 Alpha Networks Inc 聲音播放裝置參數調整方法
US8923995B2 (en) * 2009-12-22 2014-12-30 Apple Inc. Directional audio interface for portable media device
NZ587483A (en) * 2010-08-20 2012-12-21 Ind Res Ltd Holophonic speaker system with filters that are pre-configured based on acoustic transfer functions
US9377941B2 (en) * 2010-11-09 2016-06-28 Sony Corporation Audio speaker selection for optimization of sound origin
US20120124470A1 (en) * 2010-11-17 2012-05-17 The Johns Hopkins University Audio display system
WO2012164346A1 (fr) * 2011-05-27 2012-12-06 Sony Ericsson Mobile Communications Ab Sélection ou adaptation d'une fonction de transfert liée à la tête (hrtf) selon la taille de la tête
US8787584B2 (en) * 2011-06-24 2014-07-22 Sony Corporation Audio metrics for head-related transfer function (HRTF) selection or adaptation
US9167368B2 (en) * 2011-12-23 2015-10-20 Blackberry Limited Event notification on a mobile device using binaural sounds
EP2829050A1 (fr) * 2012-03-23 2015-01-28 Dolby Laboratories Licensing Corporation Techniques pour mettre en valeur des interlocuteurs dans une scène de conférence bidimensionnelle (2d) ou tridimensionnelle (3d)
WO2014081384A1 (fr) * 2012-11-22 2014-05-30 Razer (Asia-Pacific) Pte. Ltd. Procédé d'émission d'un signal audio modifié et interface graphique d'utilisateur produite par un programme d'application

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060280323A1 (en) * 1999-06-04 2006-12-14 Neidich Michael I Virtual Multichannel Speaker System
KR20080055622A (ko) * 2006-12-16 2008-06-19 김종철 인간친화형 컴퓨터입출력 시스템
US20120201405A1 (en) * 2007-02-02 2012-08-09 Logitech Europe S.A. Virtual surround for headphones and earbuds headphone externalization system
US20120022842A1 (en) * 2009-02-11 2012-01-26 Arkamys Test platform implemented by a method for positioning a sound object in a 3d sound environment
US20120093348A1 (en) * 2010-10-14 2012-04-19 National Semiconductor Corporation Generation of 3D sound with adjustable source positioning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2923500A4 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017058097A1 (fr) * 2015-09-28 2017-04-06 Razer (Asia-Pacific) Pte. Ltd. Ordinateurs, procédés de commande d'un ordinateur, et supports lisibles par ordinateur
CN108292142A (zh) * 2015-09-28 2018-07-17 雷蛇(亚太)私人有限公司 计算机、控制计算机的方法及计算机可读介质
US10356526B2 (en) 2015-09-28 2019-07-16 Razer (Asia-Pacific) Pte. Ltd. Computers, methods for controlling a computer, and computer-readable media
CN108292142B (zh) * 2015-09-28 2021-07-06 雷蛇(亚太)私人有限公司 计算机、控制计算机的方法及计算机可读介质
EP3554098A4 (fr) * 2016-12-12 2019-12-18 Sony Corporation Procédé de mesure hrtf, dispositif de mesure hrtf et programme
US11159906B2 (en) 2016-12-12 2021-10-26 Sony Corporation HRTF measurement method, HRTF measurement device, and program
CN107484097A (zh) * 2017-10-10 2017-12-15 歌尔股份有限公司 耳机左右喇叭防呆检测机构
WO2019166030A1 (fr) * 2018-02-28 2019-09-06 成都星环科技有限公司 Système d'adaptation de champ sonore intelligent, système d'étalonnage de champ sonore intelligent et système de traitement vocal intelligent

Also Published As

Publication number Publication date
SG11201503926WA (en) 2015-06-29
EP2923500A1 (fr) 2015-09-30
AU2012394979A1 (en) 2015-06-04
AU2012394979B2 (en) 2016-07-14
TWI616810B (zh) 2018-03-01
CN105027580B (zh) 2017-05-17
US20150293655A1 (en) 2015-10-15
US9569073B2 (en) 2017-02-14
TW201421354A (zh) 2014-06-01
CN105027580A (zh) 2015-11-04
EP2923500A4 (fr) 2016-06-08

Similar Documents

Publication Publication Date Title
AU2012394979B2 (en) Method for outputting a modified audio signal and graphical user interfaces produced by an application program
EP3619921B1 (fr) Processeur audio, système, procédé ainsi que programme d'ordinateur pour la reproduction audio
JP6672425B2 (ja) 反射音レンダリングのために上方発射ドライバをもつオーディオ・スピーカー
AU2017279615B2 (en) Method and device for rendering acoustic signal, and computer-readable recording medium
CN111294724B (zh) 多个音频流的空间重新定位
US10003905B1 (en) Personalized end user head-related transfer function (HRTV) finite impulse response (FIR) filter
CN104904235A (zh) 用于使用向上发射驱动器的反射声音呈现的虚拟高度滤波器
US11102604B2 (en) Apparatus, method, computer program or system for use in rendering audio
CN115777203A (zh) 信息处理装置、输出控制方法和程序
US12526597B2 (en) Reproducing device, reproducing method, information processing device, information processing method, and program
US20030099369A1 (en) System for headphone-like rear channel speaker and the method of the same
Cuevas-Rodríguez et al. The 3D Tune-In Toolkit–3D audio spatialiser, hearing loss and hearing aid simulations
CN108886652A (zh) 考虑到个体要求的重现声音的方法
CN111756929A (zh) 多屏终端音频播放方法、装置、终端设备以及存储介质
CN121002897A (zh) 虚拟听觉显示滤波器与相关联系统、方法和非暂时性计算机可读介质
US7050596B2 (en) System and headphone-like rear channel speaker and the method of the same
EP4677864A1 (fr) Systèmes et procédés pour audio spatial hybride
US12348951B2 (en) System and method for virtual sound effect with invisible loudspeaker(s)
US6983054B2 (en) Means for compensating rear sound effect
EP4460986A1 (fr) Appareil et procédé de mise en oeuvre d'un rendu d'objet audio polyvalent
US20260025630A1 (en) Methods, devices, and systems for reproducing spatial audio using binaural externalization processing extensions
BE1028706B1 (nl) Koptelefoon
US20250016519A1 (en) Audio device with head orientation-based filtering and related methods
CN117395592A (zh) 音频处理方法、系统及电子设备
JP3208565U (ja) 音色が調節可能なイヤホン

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201280078150.3

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12888662

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2012888662

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 14646677

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2012394979

Country of ref document: AU

Date of ref document: 20121122

Kind code of ref document: A