Linear expansions for frequency selective channels in OFDM
MetadataShow full item record
Modeling the frequency selective fading channels as random processes we employ a linear expansion based on the Karhumen-Loeve (KL) series representation involving a complete set of orthogonal deterministic vectors with a corresponding uncorrelated random coefficients. Focusing on OFDM transmissions through frequency selective fading this paper pursues a computationally efficient pilot-aided linear minimum mean square error (MMSE) uncorrelated KL series expansion coefficients estimation algorithm. Based on such an expansion no matrix inversion is required in the proposed MMSE estimator. Moreover truncation in the linear expansion of channel is achieved by exploiting the optimal truncation property of the KL expansion resulting in a smaller computational load on the estimation algorithm. The performance of the proposed approach is studied through analytical and experimental results. We first exploit the performance of the MMSE channel estimator based on the evaluation of minimum Bayesian MSE. We also provide performance analysis results studying the influence of the effect of SNR and correlation mismatch on the estimator performance. Simulation results confirm our theoretical results and illustrate that the proposed algorithm is capable of tracking fast fading and improving performance. (c) 2005 Elsevier GmbH. All rights reserved.