Please use this identifier to cite or link to this item:
|dc.description.abstract||Recent enhancement of Long Term Evolution / Advanced (LTE/LTE-A) is aimed to increase the throughput of the Orthogonal Frequency Division Multiplexing (OFDM) system. This is quite obvious where both frequency and time division are employed in order to augment the system capacity. In LTE/LTE-A heterogeneous Networks (Het Nets), femto cells (He Node Bs) are the optimal choice to extend the coverage within indoor environment. Nonetheless, the deployment of such base stations, somewhat to certain extent, does create cross-tier and co-tier interferences with the macro cell users (UEmacro-eNB) that ostensibly can become a critical challenge. It is implied that the unutilized / idle spectrums reallocation might be one of the solutions to mitigate such challenge. However, the errorless spectrum sensing does pose another issue. This paper outlines the investigation findings where spectrum sensing techniques is deployed to maximize the spectrum detection capacity with minimum error in OFDM based Het Nets. In addition, in the attempt to increase the efficiency of spectrum resources, this paper proposes a sensing technique which is imposed over advanced energy detection technique to detect the idle spectrums. The result of the proposed scheme is evaluated using Monte Carlo simulation. © 2014 IEEE.|
|dc.title||Performance analysis of spectrum sensing methods: A numerical approach||en_US|
|Appears in Collections:||CSIT Scholarly Publication|
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.