One-bit massive MIMO precoding using unsupervised deep learning

buir.contributor.authorKazemi, Mohammad
buir.contributor.orcidKazemi, Mohammad|0000-0001-5177-1874
dc.citation.epage34680
dc.citation.spage34668
dc.citation.volumeNumber12
dc.contributor.authorHosseinzadeh, Mohsen
dc.contributor.authorAghaeinia, Hassan
dc.contributor.authorKazemi, Mohammad
dc.date.accessioned2025-02-19T13:41:26Z
dc.date.available2025-02-19T13:41:26Z
dc.date.issued2024-02-01
dc.departmentDepartment of Electrical and Electronics Engineering
dc.description.abstractThe recently emerged symbol-level precoding (SLP) technique is a promising solution in multi-user wireless communication systems due to its ability to transform harmful multi-user interference (MUI) into useful signals, thereby improving system performance. Conventional symbol-level precoding designs have a significant computational complexity that makes their practical implementation difficult and imposes excessive computational complexity on the system. To deal with this problem, we suggest a new deep learning (DL) based approach that utilizes low-complexity designs of symbol-level precoding. This paper focuses on DL-based one-bit precoding approaches for downlink massive multiple-input multiple-output (MIMO) systems, where one-bit digital-to-analog converters (DACs) are used to reduce cost and power. Unlike previous works, the optimized one-bit precoder for multiuser massive MIMO system (HDL-O1PmMIMO) for a wide range of signal-to-noise-ratio (SNR) has a low computational complexity, making it suitable for real precoding scenarios. In this paper, we first design an unsupervised DL-based precoder (UDL-O1PmMIMO) to address the low SNR scenarios, using which we then design a hybrid DL-based precoder (HDL-O1PmMIMO) to address both low and high SNR scenarios. The method suggested in this article utilizes a novel residual DL network structure, which helps overcome the problem of training very deep networks. Additionally, a novel customized cost function, specifically for one-bit precoding in massive MIMO systems, is introduced to optimize the performance of the system in handling interference. The results of an experiment conducted on a general test set using Python and MATLAB show that the proposed approach outperforms existing methods in three aspects: it has a lower bit error rate, it takes less time to generate the precoded vector, and it is more resistant to imperfect channel estimation.
dc.identifier.doi10.1109/ACCESS.2024.3360862
dc.identifier.eissn2169-3536
dc.identifier.urihttps://hdl.handle.net/11693/116451
dc.language.isoEnglish
dc.publisherInstitute of Electrical and Electronics Engineers
dc.relation.isversionofhttps://dx.doi.org/10.1109/ACCESS.2024.3360862
dc.rightsCC BY-NC-ND 4.0 DEED (Attribution-NonCommercial-NoDerivatives 4.0
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/
dc.source.titleIEEE Access
dc.subjectMassive MIMO
dc.subjectOne-bit DAC
dc.subjectPrecoding
dc.subjectUnsupervised deep learning
dc.titleOne-bit massive MIMO precoding using unsupervised deep learning
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
One-bit_massive_MIMO_precoding_using_unsupervised_deep_learning.pdf
Size:
1.85 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: