• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • No language data
  • Tagged with
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • 1
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
1

Channel sparsity aware polynomial expansion filters for nonlinear acoustic echo cancellation

Vinith Vijayarajan (5930993) 16 January 2019 (has links)
<div> <div> <div> <p>Speech quality is a demand in voice commanded systems and in telephony. The voice communication system in real time often suffers from audible echoes. In order to cancel echoes, an acoustic echo cancellation system is designed and applied to increase speech quality both subjectively and objectively. </p> <p>In this research we develop various nonlinear adaptive filters wielding the new channel sparsity-aware recursive least squares (RLS) algorithms using a sequential update. The developed nonlinear adaptive filters using the sparse sequential RLS (S-SEQ-RLS) algorithm apply a discard function to disregard the coefficients which are not significant or close to zero in the weight vector for each channel in order to reduce the computational load and improve the algorithm convergence rate. The channel sparsity-aware algorithm is first derived for nonlinear system modeling or system identification, and then modified for application of echo cancellation. Simulation results demonstrate that by selecting a proper threshold value in the discard function, the proposed nonlinear adaptive filters using the RLS (S-SEQ-RLS) algorithm can achieve the similar performance as the nonlinear filters using the sequential RLS (SEQ-RLS) algorithm in which the channel weight vectors are sequentially updated. Furthermore, the proposed channel sparsity-aware RLS algorithms require a lower computational load in comparison with the non-sequential and non-sparsity algorithms. The computational load for the sparse algorithms can further be reduced by using data-selective strategies. </p> </div> </div> </div>

Page generated in 0.1606 seconds