[go: up one dir, main page]

Czyzewski, 2003 - Google Patents

Automatic identification of sound source position employing neural networks and rough sets

Czyzewski, 2003

View PDF
Document ID
18161334089115630365
Author
Czyzewski A
Publication year
Publication venue
Pattern Recognition Letters

External Links

Snippet

Methods for the identification of direction of the incoming acoustical signal in the presence of noise and reverberation are investigated. Since the problem is a non-deterministic one, thus applications of two learning algorithms, namely neural networks and rough sets are …
Continue reading at multimed.org (PDF) (other versions)

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • G10L2021/02161Number of inputs available containing the signal or the noise to be suppressed
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signal, using source filter models or psychoacoustic analysis
    • G10L19/008Multichannel audio signal coding or decoding, i.e. using interchannel correlation to reduce redundancies, e.g. joint-stereo, intensity-coding, matrixing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0272Voice signal separating
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/005Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets providing an auditory perception; Electric tinnitus maskers providing an auditory perception
    • H04R25/40Arrangements for obtaining a desired directivity characteristic
    • H04R25/407Circuits for combining signals of a plurality of transducers

Similar Documents

Publication Publication Date Title
Nguyen et al. Robust source counting and DOA estimation using spatial pseudo-spectrum and convolutional neural network
Asaei et al. Structured sparsity models for reverberant speech separation
Czyzewski Automatic identification of sound source position employing neural networks and rough sets
Smaragdis Information theoretic approaches to source separation
Neo et al. Signal compaction using polynomial EVD for spherical array processing with applications
Dadvar et al. Robust binaural speech separation in adverse conditions based on deep neural network with modified spatial features and training target
Song et al. An integrated multi-channel approach for joint noise reduction and dereverberation
Wang et al. Dasformer: Deep alternating spectrogram transformer for multi/single-channel speech separation
Scheibler et al. Separake: Source separation with a little help from echoes
Choi et al. Multichannel signal separation for cocktail party speech recognition: A dynamic recurrent network
Gul et al. Integration of deep learning with expectation maximization for spatial cue-based speech separation in reverberant conditions
Herzog et al. Direction preserving wiener matrix filtering for ambisonic input-output systems
Wang et al. Multi-speaker speech separation under reverberation conditions using Conv-Tasnet
Anemüller et al. Adaptive separation of acoustic sources for anechoic conditions: A constrained frequency domain approach
Aarabi et al. Robust sound localization using conditional time–frequency histograms
Luo et al. Implicit filter-and-sum network for multi-channel speech separation
Gul et al. Preserving the beamforming effect for spatial cue-based pseudo-binaural dereverberation of a single source
Dehghan Firoozabadi et al. A novel nested circular microphone array and subband processing-based system for counting and DOA estimation of multiple simultaneous speakers
Tran et al. Automatic adaptive speech separation using beamformer-output-ratio for voice activity classification
Asano et al. Evaluation and real-time implementation of blind source separation system using time-delayed decorrelation
Jing et al. End-to-end doa-guided speech extraction in noisy multi-talker scenarios
Corey et al. Nonstationary source separation for underdetermined speech mixtures
de Cheveigné The cancellation principle in acoustic scene analysis
Hammer et al. FCN approach for dynamically locating multiple speakers
Kealey et al. Unsupervised improved mvdr beamforming for sound enhancement