[go: up one dir, main page]

WO2020062922A1 - Procédé de traitement d'effet sonore et produit associé - Google Patents

Procédé de traitement d'effet sonore et produit associé Download PDF

Info

Publication number
WO2020062922A1
WO2020062922A1 PCT/CN2019/090380 CN2019090380W WO2020062922A1 WO 2020062922 A1 WO2020062922 A1 WO 2020062922A1 CN 2019090380 W CN2019090380 W CN 2019090380W WO 2020062922 A1 WO2020062922 A1 WO 2020062922A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
determining
sound
mono data
dimensional coordinates
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2019/090380
Other languages
English (en)
Chinese (zh)
Inventor
严锋贵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Publication of WO2020062922A1 publication Critical patent/WO2020062922A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2400/00Details of stereophonic systems covered by H04S but not provided for in its groups
    • H04S2400/01Multi-channel, i.e. more than two input channels, sound reproduction with two speakers wherein the multi-channel information is substantially preserved

Definitions

  • the present application relates to the technical field of audio playback, and in particular, to a sound effect processing method and related products.
  • the sound volume is often based on the volume set by the user, and the sounding body uses a constant power corresponding to the volume set by the user to play the audio, so that the played sound meets the user's loudness requirements.
  • a playback method is too simple in form, which often brings sensory fatigue to users.
  • an embodiment of the present application provides a sound effect processing method, including:
  • an embodiment of the present application provides a sound effect processing device, including:
  • a determining unit configured to determine the three-dimensional coordinates of each of the plurality of sound sources corresponding to the electronic device and the mono data generated by each sound source to obtain a plurality of first three-dimensional coordinates and a plurality of mono data; The second three-dimensional coordinates of the target object corresponding to the electronic device;
  • a synthesizing unit is configured to synthesize the plurality of mono data according to the plurality of first three-dimensional coordinates and the second three-dimensional coordinate to obtain target two-channel data.
  • an embodiment of the present application provides an electronic device including a processor, a memory, a communication interface, and one or more programs, wherein the one or more programs are stored in the memory and configured to be processed by the foregoing. And the program includes instructions for some or all of the steps as described in the first aspect.
  • an embodiment of the present application provides a computer-readable storage medium, wherein the computer-readable storage medium stores a computer program, and the computer program causes a computer to execute the program as described in the first aspect of the embodiment of the application. Describe some or all of the steps.
  • an embodiment of the present application provides a computer program product, wherein the computer program product includes a non-transitory computer-readable storage medium storing a computer program, and the computer program is operable to cause a computer to execute Some or all of the steps described in the first aspect of the embodiment of the application.
  • the computer program product can be a software installation package.
  • FIG. 1A is a schematic structural diagram of an electronic device according to an embodiment of the present application.
  • FIG. 1B is a schematic diagram of a scene of a coordinate axis of an electronic device according to an embodiment of the present application
  • FIG. 2A is a schematic flowchart of a sound effect processing method according to an embodiment of the present application.
  • FIG. 2B is a schematic diagram of a multi-channel dual-channel data scenario according to an embodiment of the present application.
  • FIG. 3 is a schematic structural diagram of a sound effect processing device according to an embodiment of the present application.
  • FIG. 4 is a schematic structural diagram of another electronic device according to an embodiment of the present application.
  • the electronic devices involved in the embodiments of the present application may include various handheld devices (such as smart phones), vehicle-mounted devices, virtual reality (VR) / augmented reality (AR) devices with wireless communication functions, and may Wearable devices, computing devices or other processing devices connected to wireless modems, and various forms of user equipment (UE), mobile stations (MS), terminal devices, R & D / test platforms, Server and so on.
  • UE user equipment
  • MS mobile stations
  • terminal devices R & D / test platforms, Server and so on.
  • R & D / test platforms Server and so on.
  • FIG. 1A is a schematic structural diagram of an electronic device according to an embodiment of the present application.
  • the electronic device includes a control circuit and an input-output circuit, and the input-output circuit is connected to the control circuit.
  • the control circuit may include a storage and processing circuit.
  • the storage circuit in the storage and processing circuit may be a memory, such as a hard disk drive memory, a non-volatile memory (such as a flash memory or other electronic programmable read-only memory used to form a solid-state drive, etc.), a volatile memory (such as a static Or dynamic random access memory, etc.), this embodiment is not limited.
  • the processing circuit in the storage and processing circuit can be used to control the operation of the electronic device.
  • the processing circuit can be implemented based on one or more microprocessors, microcontrollers, digital signal processors, baseband processors, power management units, audio codec chips, application specific integrated circuits, display driver integrated circuits, and the like.
  • the storage and processing circuit can be used to run software in an electronic device, for example, playing an incoming call alert ringing application, playing a short message alert ringing application, playing an alarm alert ringing application, playing a media file application, an Internet protocol voice ( Voice over Internet Protocol (VOIP) telephone calling applications, operating system functions, etc.
  • These software can be used to perform some control operations, such as playing the call alert ring, playing the short message alert ring, playing the alarm alert ring, playing media files, making voice phone calls, and other functions in electronic devices.
  • the examples are not limited.
  • the input-output circuit can be used to enable the electronic device to implement data input and output, that is, to allow the electronic device to receive data from an external device and to allow the electronic device to output data from the electronic device to the external device.
  • the input-output circuit may further include a sensor.
  • the sensor may include an ambient light sensor, an infrared proximity sensor based on light and capacitance, an ultrasonic sensor, and a touch sensor (for example, a light touch sensor and / or a capacitive touch sensor, where the touch sensor may be part of a touch display screen, or Can be used independently as a touch sensor structure), acceleration sensor, gravity sensor, and other sensors.
  • the input-output circuit may further include an audio component, and the audio component may be used to provide audio input and output functions for the electronic device.
  • the audio component may also include a tone generator and other components for generating and detecting sound.
  • the senor further includes a three-axis acceleration sensor for measuring a posture and an inclination angle of the electronic device.
  • a three-axis acceleration sensor for measuring a posture and an inclination angle of the electronic device.
  • it can also be used as a motion offset compensation calculation when the global positioning system (GPS) signal is not good, which can fully and accurately reflect the motion properties of the object.
  • GPS global positioning system
  • FIG. 1B is a schematic diagram of a scene where a three-dimensional acceleration sensor determines a coordinate axis of an electronic device.
  • the x-axis, y-axis, and z-axis are relative to the body of the electronic device.
  • the y-axis body is upward
  • the x-axis body is right
  • the z-axis is perpendicular to the front of the fuselage, and the center
  • the horizontal component, vertical component, and vertical component are generally a unit of gravity (the size is 1g (m * m / s), the direction is perpendicular to the ground downward), and the projection on each axis. That is, the horizontal component is the corresponding value on the x-axis, the vertical component is the corresponding value on the y-axis, and the vertical component is the corresponding value on the z-axis.
  • the x-axis defaults to 0, the y-axis defaults to 0, and the z-axis defaults to 9.81; place the electronic device on the desktop with the z-axis at -9.81; tilt the electronic device to the left, x The axis is positive; tilt the electronic device to the right and the x-axis is negative; tilt the electronic device upwards and the y-axis is negative; tilt the electronic device downwards and the y-axis is positive; set the z-axis to less than -3 In this case, the touch screen of the electronic device faces downward.
  • the input-output circuit may also include one or more display screens.
  • the display screen may include one or a combination of a liquid crystal display, an organic light emitting diode display, an electronic ink display, a plasma display, and a display using other display technologies.
  • the display screen may include a touch sensor array (ie, the display screen may be a touch display screen).
  • the touch sensor can be a capacitive touch sensor formed by a transparent array of touch sensor electrodes (such as indium tin oxide (ITO) electrodes), or it can be a touch sensor formed using other touch technologies, such as sonic touch, pressure-sensitive touch, resistance Touch, optical touch, etc. are not limited in the embodiments of the present application.
  • the input-output circuit may further include a communication circuit for providing an electronic device with a capability to communicate with an external device.
  • the communication circuit may include analog and digital input-output interface circuits, and wireless communication circuits based on radio frequency signals and / or optical signals.
  • the wireless communication circuit in the communication circuit may include a radio frequency transceiver circuit, a power amplifier circuit, a low noise amplifier, a switch, a filter, and an antenna.
  • the wireless communication circuit in the communication circuit may include a circuit for supporting near field communication (NFC) by transmitting and receiving a near field coupled electromagnetic signal.
  • the communication circuit may include a near field communication antenna and a near field communication transceiver.
  • the communication circuit may also include a cellular phone transceiver and antenna, a wireless local area network transceiver circuit and antenna, and the like.
  • the input-output circuit may further include an input-output unit.
  • the input-output unit may include a button, a joystick, a click wheel, a scroll wheel, a touch pad, a keypad, a keyboard, a camera, a light emitting diode, and other status indicators.
  • the electronic device may further include a battery (not shown), and the battery is used to provide power to the electronic device.
  • an embodiment of the present application provides a schematic flowchart of a sound effect processing method, which is applied to an electronic device. Specifically, as shown in FIG. 2A, the method includes:
  • S201 Determine the three-dimensional coordinates of each sound source in the multiple sound sources corresponding to the electronic device and the mono data generated by each sound source to obtain multiple first three-dimensional coordinates and multiple mono data.
  • the embodiments of the present application can be applied to a virtual reality / augmented reality scene, or a three-dimensional (3D) recording scene.
  • the sound source may be a sounding body in a virtual scene, for example, an airplane in a game scene, and the sound source may be a fixed sound source or a mobile sound source.
  • the sound source corresponding to the electronic device can use the above-mentioned coordinate axis as a reference to determine the first three-dimensional coordinate corresponding to the sound source, and when the sound source emits sound, it can obtain the source Of mono data.
  • the electronic device may include multiple sound sources.
  • the sound source of the game scene includes airplanes, guns, rivers, etc.
  • the corresponding mono data is the gliding sound of the aircraft, the loading of guns, the sound of the fire, the sound of the water flowing in the river;
  • the sound source of the game scene can also include game players,
  • the corresponding mono data is footstep sounds, voice sounds, etc. of the game player, which is not limited here.
  • step S201 may include: determining multiple reference objects corresponding to the electronic device; determining behavior information of each reference object in the multiple reference objects to obtain multiple behavior information; and according to the The plurality of behavior information determines a plurality of sound sources corresponding to the electronic device in the plurality of reference objects; determining a coordinate position corresponding to each sound source in the plurality of sound sources to obtain the plurality of first three-dimensional coordinates; The behavior information corresponding to each of the plurality of sound sources determines the mono data of the sound source to obtain the plurality of mono data.
  • the reference object may be an object presented on a display page of the electronic device, for example, a house on the display page, a car in which a game player rides, or a firearm held.
  • the reference object may also be an object that is not presented on the display page, such as a nearby game player, a gun, a vehicle, or the like.
  • the behavior information is dynamic information of the reference object. It can be understood that different reference objects correspond to different types of behaviors. For example, guns can emit the sound of loading and firing guns, but not the sound of water flowing or talking. And each reference corresponds to a different sound when the behavior information is different. For example, the house will not emit sound under normal circumstances, but will emit a sound of bombardment when it is bombarded; the starting sound and driving sound will only occur when the car is moving. Sound; gamers produce voices when sending voices, footsteps when walking, and more. Therefore, in the possible examples described above, first determine multiple reference objects corresponding to the electronic device, and then obtain behavior information of each reference object, and determine whether the reference object is a sound source according to each behavior information, thereby improving the accuracy of determining the sound source. . Then, the coordinate position of each sound source is further determined to obtain a plurality of first three-dimensional coordinates, and then the mono data of the sound source is determined according to the behavior information corresponding to each sound source, which improves the accuracy of determining the mono data.
  • This application does not limit how to determine the coordinate position corresponding to the sound source. For example, if it corresponds to a game scene and the game scene corresponds to a three-dimensional map, the coordinate position corresponding to different sound sources can be determined according to the map, that is, for characters Determining the first three-dimensional coordinates at a specific location can improve the accuracy of determining the first three-dimensional coordinates, facilitate the improvement of the 3D sound effect of the target two-channel data, and allow the user to be immersed in the game and feel the game world more realistic.
  • the Determining the mono data of the sound source according to the behavior information corresponding to each of the multiple sound sources to obtain the multiple mono data includes: determining a sound type and a playback parameter corresponding to the target behavior information; The sound type and the playback parameter generate mono data of the target sound source.
  • the sound type is the sound type of the target behavior information corresponding to the sound.
  • the firearm includes the sound type of loading, firing, and hitting, and the playback parameters are loudness, frequency, and tone.
  • step S201 takes the target sound source as an example, determines the sound type and playback parameters of the target sound source according to the target behavior information corresponding to the target sound source, and then generates the target sound source according to the sound type and playback parameters.
  • Mono data can further improve the accuracy of determining mono data, improve the fit of mono data to application scenarios, and improve user experience.
  • S202 Determine a second three-dimensional coordinate of a target object corresponding to the electronic device.
  • the target object may be a game player corresponding to an electronic device in a game, a virtual reality or an augmented reality scene, or a target user corresponding to an electronic device in a 3D recording scene.
  • the target object may also correspond to a three-dimensional position, that is, a second three-dimensional position.
  • the first three-dimensional position is different from the second three-dimensional position.
  • reference may be made to the method of determining the first three-dimensional coordinate, that is, the second three-dimensional coordinate of the target object is determined using the coordinate axis shown in FIG. 1B as a reference, and details are not described herein again.
  • S203 Combining the plurality of mono data according to the plurality of first three-dimensional coordinates and the second three-dimensional coordinate to obtain target two-channel data.
  • the mono data corresponding to each sound source can be synthesized to obtain the target two-channel data.
  • a plurality of first three-dimensional coordinates, a second three-dimensional coordinate, and a plurality of mono data can be input into a Head Related Transformation Function (HRTF) model to obtain target binaural data.
  • HRTF Head Related Transformation Function
  • the electronic device can filter the audio data (sound from the sound source) using HRTF filters to obtain virtual surround sound, also called surround sound or panoramic sound, to achieve a three-dimensional stereo sound effect.
  • HRTF Head-related Impulse Response
  • BRIR Binaural Room Impulse Response
  • the electronic device may be based on the spatial three-dimensional coordinate position (x, y, z) of the sound source, and the position may be any coordinate.
  • the left and right channels are generated based on the mono data generated by the sound source.
  • the principle of the left and right channels is based on the time difference between the sound source and the listener (X, Y, Z). And the phase pressure difference generates a two-channel sound.
  • step S203 may include: determining the left ear three-dimensional coordinates and the right ear three-dimensional coordinates corresponding to the second three-dimensional coordinates; and according to the plurality of first three-dimensional coordinates, the left ear three-dimensional coordinates Determining the transmission paths between the plurality of sound sources and the target object by the coordinates and the three-dimensional coordinates of the right ear to obtain a plurality of transmission paths; determining each of the plurality of mono data according to the plurality of transmission paths The time and phase pressure of the mono data transmitted to the three-dimensional coordinates of the left ear and the three-dimensional coordinates of the right ear are obtained in multiple times and multiple phase pressures; and the multiple mono data are determined according to the multiple times.
  • a plurality of time differences are obtained for the time difference corresponding to each mono data, and a phase pressure difference corresponding to each mono data in the plurality of mono data is determined according to the plurality of phase pressures to obtain a plurality of phase pressure differences; Determining the delay parameters corresponding to each of the plurality of mono data by the plurality of time differences and the plurality of phase pressure differences to obtain a plurality of delay parameters; and according to the plurality of delays Parameter, processing each mono data in the plurality of mono data to obtain corresponding left channel parameters and right channel parameters; corresponding to each mono data in the plurality of mono data And synthesizing the plurality of mono data to obtain the target two-channel data.
  • the target object corresponds to a left ear three-dimensional coordinate and a right ear three-dimensional coordinate.
  • This application does not limit how to determine the three-dimensional coordinates of the left ear and the three-dimensional coordinates of the right ear, and may be determined according to the 3D character model of the target object, that is, according to the second three-dimensional coordinates of the target object and the right ear that are preset in the 3D character model.
  • the correlation between the three-dimensional coordinates, the second three-dimensional coordinates and the left ear three-dimensional coordinates is determined.
  • the time difference and the phase pressure difference are respectively the time difference and the phase pressure difference transmitted to the left ear three-dimensional coordinates and the right ear three-dimensional coordinates. That is, the time difference and phase pressure difference between the mono data corresponding to the sound source and the left and right ears of the target object are transmitted.
  • step S203 the left channel parameters and the right channel parameters of the mono data can be determined according to the delay parameters, and then synthesized to obtain the target two-channel data, which improves the playback effect of the audio data. Create immersive sensations to improve user experience.
  • This application does not limit how to determine the time and phase pressure.
  • This application takes the time and phase pressure corresponding to the three-dimensional coordinates of the target sound source to the left ear as an example.
  • the time and phase pressure corresponding to the three-dimensional coordinates of the target sound source to the left ear are transmitted. Refer to this method for the method of determining the phase pressure and the monophonic data corresponding to other sound sources other than the target sound source among multiple sound sources to the left ear three-dimensional coordinates and the right ear three-dimensional coordinates.
  • the plurality of sound sources includes a target sound source
  • the plurality of first three-dimensional coordinates includes a target first three-dimensional coordinate corresponding to the target sound source
  • the plurality of mono data includes the target sound Target mono data corresponding to the source
  • Obtaining multiple time and multiple phase pressures with time and phase pressure includes: obtaining a cross-section using the target first three-dimensional coordinate and the left ear three-dimensional coordinate as an axis; determining the target first three-dimensional coordinate and the left ear three-dimensional Blocking objects between coordinates; determining a plurality of reference transmission paths for transmitting the target mono data to the three-dimensional coordinates of the left ear according to the cross section and the occluding objects; determining the reference transmission paths according to the plurality of reference transmission paths
  • the propagation of the target mono data can include multiple reference transmissions. path.
  • the first three-dimensional coordinates of the target and the three-dimensional coordinates of the left ear are used as the cross-sections. Since the sound propagation direction is fixed, the propagation trajectory will also have a certain symmetry along a certain symmetry axis, and multiple transmissions can be obtained path. The propagation of sound will be dispersed and transmitted when it encounters occluded objects, so that multiple corresponding reference transmission paths are determined according to the cross-section and the occluded objects in the application scene.
  • the cross-section of a sound source data transmission is determined by using the first three-dimensional coordinates of the target and the three-dimensional coordinates of the left ear as axes, and then multiple reference transmission paths corresponding to the target mono data are determined according to the cross-section and the occluded object.
  • the multiple reference transmission paths determine the phase pressure transmitted by the target sound source to the left ear of the target object. That is, the possible multiple reference transmission paths are determined according to the cross-sections corresponding to the first three-dimensional coordinates of the target and the three-dimensional coordinates of the left ear, and then the time and phase pressure are determined according to the multiple reference transmission paths, which improves the accuracy of determining the time and phase pressure. Sex.
  • the determining the time and phase pressure of transmitting the target mono data to the three-dimensional coordinates of the left ear according to the multiple reference transmission paths includes: determining each of the multiple reference transmission paths The sound intensity and sound pressure corresponding to a transmission path are used to obtain a plurality of sound intensity and a plurality of sound pressures; and the target mono data is determined to be transmitted to the left ear three-dimensionally according to the sound intensity and the sound pressure. Coordinate time and phase pressure.
  • the sound intensity refers to the sound energy of a unit area per unit time, which is perpendicular to the direction of propagation, and the unit is W / m2.
  • Sound pressure is the increase in pressure due to the presence of sound waves, and the unit is Pa.
  • This application does not limit how to determine the phase pressure based on multiple sound pressures.
  • the preset weights can be determined according to the relative distance corresponding to the corresponding reference transmission path, and then multiple sound pressures and corresponding preset weights are weighted. Get the phase pressure.
  • each sound source data has a corresponding sound pressure. Therefore, first determining the sound pressure corresponding to each reference transmission path to obtain multiple sound pressures, and then determining the phase pressure of the target mono data transmitted to the three-dimensional coordinates of the left ear based on the multiple sound pressures can improve the accuracy of determining the phase pressure.
  • the three-dimensional coordinates of each sound source in a plurality of sound sources corresponding to the electronic device and the mono data generated by each sound source are determined to obtain a plurality of first three-dimensional coordinates and a plurality of sound sources.
  • the channel data and the second three-dimensional coordinates of the target object corresponding to the electronic device are determined.
  • the plurality of mono data is synthesized to obtain the target two-channel data.
  • the target two-channel data corresponding to the multiple sound sources is generated, thereby improving the playback effect of the audio data and generating an immersive feeling.
  • the method further includes: determining a target reverberation parameter corresponding to the target object; and processing the target two-channel data according to the target reverberation parameter to obtain reverberant two-channel data.
  • the target reverb parameters include input volume, low frequency cut, high frequency cut, early reflection time, spatial breadth, diffusion degree, low mixing ratio, crossover point, reverberation time, high frequency attenuation point, dry sound adjustment, and reverberation.
  • the volume, the amount of early reflected sound, the width of the sound field, the output sound field, and the tail sound are not limited here.
  • the target reverberation parameter corresponding to the target object is determined, and then the target two-channel data is processed according to the target reverberation parameter to obtain the reverberant two-channel data.
  • processing the target two-channel data according to the target object can further improve the playback effect of the audio data and improve the user experience.
  • determining the target reverberation parameter corresponding to the target object includes: obtaining multiple historical reverberations corresponding to the target object that are stored in advance. Playing records; acquiring listening parameters corresponding to each historical reverb playing record in the plurality of historical reverb playing records to obtain a plurality of listening parameters; and determining a target reverb parameter corresponding to the target object according to the plurality of listening parameters.
  • the listening parameters include audio type, playback duration, playback adjustment times, user mood parameters, and so on. It can be understood that obtaining multiple historical reverberation playback records corresponding to the target object in advance, obtaining listening parameters corresponding to each historical reverberation playback record, and then determining the target reverberation parameters based on the multiple listening parameters improves the determination of the target reverberation. The accuracy of the parameters is convenient for improving the playback effect.
  • determining the target reverberation parameter corresponding to the target object according to the multiple listening parameters includes: according to the Multiple listening parameters determine the evaluation value corresponding to each historical reverb play record in the multiple historical reverb play records to obtain multiple evaluation values; and use the historical reverb record corresponding to the maximum value among the multiple evaluation values as A target historical reverberation record; and a reverberation parameter corresponding to the target historical reverberation record as a target reverberation parameter corresponding to the target object.
  • the plurality of historical reverberation playback records includes a target historical reverberation playback record
  • taking the target historical reverberation playback record as an example, first determine an application scenario of the electronic device, and determine that the application scenario corresponds to Preset mood parameters. Then, an evaluation value corresponding to the target historical reverberation record is determined according to a difference between the preset mood parameter and the user mood parameter.
  • the preset mood parameters corresponding to the target object are different.
  • the preset mood parameters of the target object are determined according to the application scenario corresponding to the electronic device.
  • the evaluation value corresponding to the target historical reverberation record is determined according to the difference between the preset mood parameter and the user mood parameter.
  • the evaluation value corresponding to the historical reverberation record is determined according to the listening parameters of each historical reverberation record, and then the historical reverberation record corresponding to the largest evaluation value is selected as the target historical reverberation record, and then according to the target historical reverberation record.
  • the recorded listening parameters determine the target reverberation parameters, which improves the accuracy of determining the target reverberation parameters.
  • FIG. 3 is a schematic structural diagram of a sound effect processing device according to an embodiment of the present application.
  • the sound effect processing device 300 includes a determining unit 301 and a synthesizing unit 302, where:
  • the determining unit 301 is configured to determine the three-dimensional coordinates of each sound source in the multiple sound sources corresponding to the electronic device and the mono data generated by each sound source to obtain multiple first three-dimensional coordinates and multiple mono data; The second three-dimensional coordinates of the target object corresponding to the electronic device;
  • the synthesizing unit 302 is configured to synthesize the plurality of mono data according to the plurality of first three-dimensional coordinates and the second three-dimensional coordinate to obtain target binaural data.
  • the determining unit 301 is further configured to determine the left ear three-dimensional coordinates and the right ear three-dimensional coordinates corresponding to the second three-dimensional coordinates; and determine the multiple according to the plurality of first three-dimensional coordinates, the left ear three-dimensional coordinates, and the right ear three-dimensional coordinates.
  • a plurality of transmission paths are obtained from transmission paths between the sound sources and the target object; and each of the plurality of mono data is determined to be transmitted to the left ear three-dimensional coordinates according to the plurality of transmission paths.
  • time and phase pressure of the three-dimensional coordinate of the right ear to obtain multiple times and multiple phase pressures; determining a time difference corresponding to each of the plurality of mono data according to the plurality of times to obtain multiple Time difference, determining a phase pressure difference corresponding to each of the plurality of mono data according to the plurality of phase pressures to obtain a plurality of phase pressure differences; and according to the plurality of time differences and the plurality of phases,
  • the bit pressure difference determines a delay parameter corresponding to each of the plurality of mono data to obtain a plurality of delay parameters; according to the plurality of delay parameters, for each of the plurality of mono data,
  • the channel data is processed to obtain corresponding left channel parameters and right channel parameters; the synthesizing unit 302 is specifically configured to use the left channel parameters and the right channels corresponding to each of the plurality of mono data Channel parameters, synthesizing the plurality of mono data to obtain target two-channel data.
  • the plurality of sound sources includes a target sound source
  • the plurality of first three-dimensional coordinates includes a target first three-dimensional coordinate corresponding to the target sound source
  • the plurality of mono data includes all The target mono data corresponding to the target sound source
  • the determining unit 301 is specifically configured to obtain a cross-section using the target first three-dimensional coordinate and the left ear three-dimensional coordinate as an axis; determining the target first three-dimensional coordinate and the left ear three-dimensional Blocking objects between coordinates; determining a plurality of reference transmission paths for transmitting the target mono data to the three-dimensional coordinates of the left ear according to the cross section and the occluding objects; determining the reference transmission paths according to the plurality of reference transmission paths The time and phase pressure of the target mono data transmitted to the three-dimensional coordinates of the left ear.
  • the determining unit 301 is specifically configured to determine The sound intensity and sound pressure corresponding to each transmission path in the plurality of reference transmission paths obtain a plurality of sound intensity and a plurality of sound pressures; and determine the target mono data to be transmitted to the left according to the plurality of sound intensity.
  • the phase pressure at which the target mono data is transmitted to the three-dimensional coordinate of the left ear is determined according to the plurality of sound pressures.
  • the three-dimensional coordinates of each sound source in the plurality of sound sources corresponding to the electronic device and the mono data generated by each sound source are used to obtain multiple first three-dimensional coordinates and multiple mono channels.
  • the determining unit 301 is specifically configured to determine multiple reference objects corresponding to the electronic device; determine behavior information of each reference object in the multiple reference objects to obtain multiple behavior information; and determine based on the multiple behavior information
  • a plurality of sound sources corresponding to the electronic device in the plurality of reference objects determining a coordinate position corresponding to each sound source in the plurality of sound sources to obtain a plurality of first three-dimensional coordinates; and according to the plurality of sound sources,
  • the behavior information corresponding to each sound source determines the mono data of the sound source to obtain a plurality of mono data.
  • the multiple behavior information includes target behavior information corresponding to a target sound source; and determining the mono data of the sound source according to the behavior information corresponding to each sound source in the multiple sound sources
  • the determining unit 301 is specifically configured to determine a sound type and a playback parameter corresponding to the target behavior information; and generate a target sound source according to the sound type and the playback parameter.
  • the determination unit 301 is further configured to determine a target reverberation parameter corresponding to the target object; and the synthesis unit 302 is further configured to perform the target two-channel data on the target reverberation parameter according to the target reverberation parameter. Process it to get reverberant two-channel data.
  • the determining unit 301 is specifically configured to obtain multiple historical reverb play records corresponding to the target object stored in advance; obtain each historical reverb play in the multiple historical reverb play records Recording corresponding listening parameters to obtain multiple listening parameters; and determining a target reverberation parameter corresponding to the target object according to the multiple listening parameters.
  • the determining unit 301 is specifically configured to determine an evaluation value corresponding to each historical reverb play record in the multiple historical reverb play records according to the multiple listening parameters to obtain multiple evaluation values;
  • the historical reverberation record corresponding to the maximum value among the multiple evaluation values is used as the target historical reverberation record, and the reverberation parameter corresponding to the target historical reverberation record is used as the target reverberation parameter corresponding to the target object.
  • FIG. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
  • the electronic device 400 includes a processor 410, a memory 420, a communication interface 430, and one or more programs 440.
  • the one or more programs 440 are stored in the memory 420, and are configured by
  • the processor 410 executes, and the program 440 includes instructions for performing the following steps:
  • the program 440 in terms of synthesizing the plurality of mono data according to the plurality of first three-dimensional coordinates and the second three-dimensional coordinate to obtain target two-channel data, the program 440
  • the instructions in are specifically used to perform the following operations:
  • phase pressure difference corresponding to the channel data is used to obtain multiple phase pressure differences
  • the plurality of sound sources includes a target sound source
  • the plurality of first three-dimensional coordinates includes a target first three-dimensional coordinate corresponding to the target sound source
  • the plurality of mono data includes all The target mono data corresponding to the target sound source; and determining the time and phase at which each mono data of the plurality of mono data is transmitted to the left ear three-dimensional coordinate according to the plurality of transmission paths
  • the pressure obtains multiple time and multiple phase pressure aspects, and the instructions in the program 440 are specifically used to perform the following operations:
  • Time and phase pressures for transmitting the target mono data to the three-dimensional coordinates of the left ear according to the multiple reference transmission paths are provided.
  • the instructions in the program 440 are specifically used to do the following:
  • the three-dimensional coordinates of each sound source and the mono data generated by each sound source in the plurality of sound sources corresponding to the electronic device 400 are obtained to obtain a plurality of first three-dimensional coordinates and a plurality of mono sounds.
  • the instructions in the program 440 are specifically used to perform the following operations:
  • the mono data of the sound source is determined according to the behavior information corresponding to each of the multiple sound sources to obtain multiple mono data.
  • the multiple behavior information includes target behavior information corresponding to a target sound source; and determining the mono data of the sound source according to the behavior information corresponding to each sound source in the multiple sound sources
  • the instructions in the program 440 are specifically used to perform the following operations:
  • Mono data of the target sound source is generated according to the sound type and the playback parameter.
  • the instructions in the program 440 are further used to perform the following operations:
  • the instructions in the program 440 are specifically used to perform the following operations:
  • a target reverberation parameter corresponding to the target object is determined according to the plurality of listening parameters.
  • the instructions in the program 440 are specifically used to perform the following operations:
  • the historical reverberation record corresponding to the maximum value among the multiple evaluation values is used as the target historical reverberation record, and the reverberation parameter corresponding to the target historical reverberation record is used as the target reverberation parameter corresponding to the target object.
  • the sound processing method and the electronic device provided in the embodiments of the present application are similar to the method embodiments of the present application. Therefore, the implementation of the sound processing method and the electronic device can refer to the method implementation, the sound processing method and the electronic For the beneficial effects of the device, refer to the beneficial effects of the method. For brevity description, they are not repeated here.
  • An embodiment of the present application further provides a computer storage medium, wherein the computer storage medium stores a computer program for causing a computer to execute a part or all of the steps of any method as described in the method embodiment, and the computer includes an electronic device. device.
  • An embodiment of the present application further provides a computer program product.
  • the computer program product includes a non-transitory computer-readable storage medium storing the computer program, and the computer program is operable to cause a computer to execute a part of any method as described in the method embodiments. Or all steps.
  • the computer program product may be a software installation package, and the computer includes an electronic device.
  • the disclosed device may be implemented in other ways.
  • the device embodiments described above are only schematic, such as the division of units, which is only a logical function division. In actual implementation, there may be another division manner. For example, multiple units or components may be combined or integrated into Another system, or some features, can be ignored or not implemented.
  • the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be electrical or other forms.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, which may be located in one place, or may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objective of the solution of this embodiment.
  • each functional unit in each embodiment of the present application may be integrated into one processing unit, or each of the units may exist separately physically, or two or more units may be integrated into one unit.
  • the above integrated unit may be implemented in the form of hardware or in the form of a software program mode.
  • the integrated unit When the integrated unit is implemented in the form of a software program and sold or used as an independent product, it can be stored in a computer-readable memory.
  • the technical solution of the present application essentially or part that contributes to the existing technology or all or part of the technical solution can be embodied in the form of a software product, which is stored in a memory.
  • a computer device which may be a personal computer, a server, or a network device, etc.
  • the foregoing memory includes: a U disk, a read-only memory (ROM), a random access memory (RAM), a mobile hard disk, a magnetic disk, or an optical disk, and other media that can store program codes.
  • the program may be stored in a computer-readable memory, and the memory may include a flash disk. , ROM, RAM, disk or disc, etc.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Stereophonic System (AREA)

Abstract

L'invention concerne un procédé de traitement d'effet sonore et un produit associé. Le procédé consiste à : déterminer une coordonnée tridimensionnelle de chaque source sonore parmi une pluralité de sources sonores correspondant à un dispositif électronique, et des données à piste unique générées par chaque source sonore, de façon à obtenir une pluralité de premières coordonnées tridimensionnelles et une pluralité d'éléments de données à piste unique; déterminer une seconde coordonnée tridimensionnelle d'un objet cible correspondant au dispositif électronique; et, en fonction de la pluralité de premières coordonnées tridimensionnelles et de la seconde coordonnée tridimensionnelle, synthétiser la pluralité d'éléments de données à piste unique pour obtenir des données cibles à double piste. Au moyen de la présente invention, un effet de lecture de données audio peut être amélioré.
PCT/CN2019/090380 2018-09-25 2019-06-06 Procédé de traitement d'effet sonore et produit associé Ceased WO2020062922A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811118269.4 2018-09-25
CN201811118269.4A CN109246580B (zh) 2018-09-25 2018-09-25 3d音效处理方法及相关产品

Publications (1)

Publication Number Publication Date
WO2020062922A1 true WO2020062922A1 (fr) 2020-04-02

Family

ID=65056811

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/090380 Ceased WO2020062922A1 (fr) 2018-09-25 2019-06-06 Procédé de traitement d'effet sonore et produit associé

Country Status (2)

Country Link
CN (1) CN109246580B (fr)
WO (1) WO2020062922A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114404973A (zh) * 2021-12-28 2022-04-29 网易(杭州)网络有限公司 播放音频的方法、装置及电子设备
CN114594480A (zh) * 2022-03-11 2022-06-07 北京女娲补天科技信息技术有限公司 一种基于声波定位的投掷项目测试方法及装置

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108924705B (zh) * 2018-09-25 2021-07-02 Oppo广东移动通信有限公司 3d音效处理方法及相关产品
CN109246580B (zh) * 2018-09-25 2022-02-11 Oppo广东移动通信有限公司 3d音效处理方法及相关产品
CN113467603B (zh) * 2020-03-31 2024-03-08 抖音视界有限公司 音频处理方法、装置、可读介质及电子设备
CN116389982B (zh) * 2023-05-19 2024-10-11 零束科技有限公司 音频处理方法、装置、电子设备和存储介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060133628A1 (en) * 2004-12-01 2006-06-22 Creative Technology Ltd. System and method for forming and rendering 3D MIDI messages
CN103218198A (zh) * 2011-08-12 2013-07-24 索尼电脑娱乐公司 运动用户的声音定位
CN104869524A (zh) * 2014-02-26 2015-08-26 腾讯科技(深圳)有限公司 三维虚拟场景中的声音处理方法及装置
CN109246580A (zh) * 2018-09-25 2019-01-18 Oppo广东移动通信有限公司 3d音效处理方法及相关产品

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105159655B (zh) * 2014-05-28 2020-04-24 腾讯科技(深圳)有限公司 行为事件的播放方法和装置
CN106448687B (zh) * 2016-09-19 2019-10-18 中科超影(北京)传媒科技有限公司 音频制作及解码的方法和装置
CN107360494A (zh) * 2017-08-03 2017-11-17 北京微视酷科技有限责任公司 一种3d音效处理方法、装置、系统及音响系统
CN107608519A (zh) * 2017-09-26 2018-01-19 深圳传音通讯有限公司 一种声音调整方法及虚拟现实设备

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060133628A1 (en) * 2004-12-01 2006-06-22 Creative Technology Ltd. System and method for forming and rendering 3D MIDI messages
CN103218198A (zh) * 2011-08-12 2013-07-24 索尼电脑娱乐公司 运动用户的声音定位
CN104869524A (zh) * 2014-02-26 2015-08-26 腾讯科技(深圳)有限公司 三维虚拟场景中的声音处理方法及装置
CN109246580A (zh) * 2018-09-25 2019-01-18 Oppo广东移动通信有限公司 3d音效处理方法及相关产品

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114404973A (zh) * 2021-12-28 2022-04-29 网易(杭州)网络有限公司 播放音频的方法、装置及电子设备
CN114594480A (zh) * 2022-03-11 2022-06-07 北京女娲补天科技信息技术有限公司 一种基于声波定位的投掷项目测试方法及装置

Also Published As

Publication number Publication date
CN109246580A (zh) 2019-01-18
CN109246580B (zh) 2022-02-11

Similar Documents

Publication Publication Date Title
US10911882B2 (en) Methods and systems for generating spatialized audio
WO2020062922A1 (fr) Procédé de traitement d'effet sonore et produit associé
EP3039677B1 (fr) Système et procédé d'apprentissage virtuel multidimensionnel
US11109177B2 (en) Methods and systems for simulating acoustics of an extended reality world
EP3629145B1 (fr) Procédé de traitement d'effet audio 3d et produits associés
CN115250412B (zh) 音频处理方法、装置、无线耳机及计算机可读介质
CN109104687B (zh) 音效处理方法及相关产品
CN109254752B (zh) 3d音效处理方法及相关产品
CN109121069B (zh) 3d音效处理方法及相关产品
CN108924705B (zh) 3d音效处理方法及相关产品
CN113115175B (zh) 3d音效处理方法及相关产品
CN116421971A (zh) 空间音频信号的生成方法及装置、存储介质、电子设备
US20250324188A1 (en) Audio mixed reality and associated systems, methods, devices, and non-transitory computer-readable media
CN109327794B (zh) 3d音效处理方法及相关产品
CN109243413B (zh) 3d音效处理方法及相关产品
CN117476015A (zh) 音频处理方法、装置、设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19866526

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19866526

Country of ref document: EP

Kind code of ref document: A1