Czyzewski, 2003 - Google Patents
Automatic identification of sound source position employing neural networks and rough setsCzyzewski, 2003
View PDF- Document ID
- 18161334089115630365
- Author
- Czyzewski A
- Publication year
- Publication venue
- Pattern Recognition Letters
External Links
Snippet
Methods for the identification of direction of the incoming acoustical signal in the presence of noise and reverberation are investigated. Since the problem is a non-deterministic one, thus applications of two learning algorithms, namely neural networks and rough sets are …
- 230000001537 neural 0 title abstract description 31
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L21/0216—Noise filtering characterised by the method used for estimating noise
- G10L2021/02161—Number of inputs available containing the signal or the noise to be suppressed
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signal, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding, i.e. using interchannel correlation to reduce redundancies, e.g. joint-stereo, intensity-coding, matrixing
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/005—Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets providing an auditory perception; Electric tinnitus maskers providing an auditory perception
- H04R25/40—Arrangements for obtaining a desired directivity characteristic
- H04R25/407—Circuits for combining signals of a plurality of transducers
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Nguyen et al. | Robust source counting and DOA estimation using spatial pseudo-spectrum and convolutional neural network | |
| Asaei et al. | Structured sparsity models for reverberant speech separation | |
| Czyzewski | Automatic identification of sound source position employing neural networks and rough sets | |
| Smaragdis | Information theoretic approaches to source separation | |
| Neo et al. | Signal compaction using polynomial EVD for spherical array processing with applications | |
| Dadvar et al. | Robust binaural speech separation in adverse conditions based on deep neural network with modified spatial features and training target | |
| Song et al. | An integrated multi-channel approach for joint noise reduction and dereverberation | |
| Wang et al. | Dasformer: Deep alternating spectrogram transformer for multi/single-channel speech separation | |
| Scheibler et al. | Separake: Source separation with a little help from echoes | |
| Choi et al. | Multichannel signal separation for cocktail party speech recognition: A dynamic recurrent network | |
| Gul et al. | Integration of deep learning with expectation maximization for spatial cue-based speech separation in reverberant conditions | |
| Herzog et al. | Direction preserving wiener matrix filtering for ambisonic input-output systems | |
| Wang et al. | Multi-speaker speech separation under reverberation conditions using Conv-Tasnet | |
| Anemüller et al. | Adaptive separation of acoustic sources for anechoic conditions: A constrained frequency domain approach | |
| Aarabi et al. | Robust sound localization using conditional time–frequency histograms | |
| Luo et al. | Implicit filter-and-sum network for multi-channel speech separation | |
| Gul et al. | Preserving the beamforming effect for spatial cue-based pseudo-binaural dereverberation of a single source | |
| Dehghan Firoozabadi et al. | A novel nested circular microphone array and subband processing-based system for counting and DOA estimation of multiple simultaneous speakers | |
| Tran et al. | Automatic adaptive speech separation using beamformer-output-ratio for voice activity classification | |
| Asano et al. | Evaluation and real-time implementation of blind source separation system using time-delayed decorrelation | |
| Jing et al. | End-to-end doa-guided speech extraction in noisy multi-talker scenarios | |
| Corey et al. | Nonstationary source separation for underdetermined speech mixtures | |
| de Cheveigné | The cancellation principle in acoustic scene analysis | |
| Hammer et al. | FCN approach for dynamically locating multiple speakers | |
| Kealey et al. | Unsupervised improved mvdr beamforming for sound enhancement |