CSI COMPRESSION AND FEEDBACK METHOD BASED ON DEEP LEARNING
A channel state information CSI compression and feedback solution based on deep learning. In a large-scale multiple-input-multiple-output MIMO system, CSI is very important for the base station to achieve high performance gain. In a frequency division duplex FDD mode, user equipment UE needs to estimate CSI and then feed it back to the base station. Effective CSI compression will greatly reduce the cost of feedback. The present disclosure uses a two-dimensional sequence-to-sequence 2D Seq2Seq structure neural network to realize effective CSI compression and reconstruction. The base station uses CSI data to train the parameters of the neural network, which, after training, are deployed at the client and the base station for effective CSI compression and reconstruction. The present disclosure can realize high-precision compression and reconstruction of CSI, and provides an efficient and stable channel state information feedback method for a MIMO system.
The present application is a continuation of International Application No. PCT/CN2023/093244, filed on May 10, 2023, which claims priority to Chinese Application No. 202211065211.4, filed on Sep. 1, 2022, the contents of both of which are incorporated herein by reference in their entireties.
TECHNICAL FIELDThe present disclosure relates to the field of wireless communication, and in particular, to a CSI compression and feedback method based on deep learning.
BACKGROUNDMIMO (Multiple Input Multiple Output) antenna array and OFDM (Orthogonal Frequency Division Multiplexing) are two key technologies in 5G (Fifth Generation) and B5G (Beyond 5G) communication systems. In a MIMO-OFDM system, in order to bring sufficient performance gain, a base station needs accurate downlink CSI (Channel State Information) for beamforming, subcarrier allocation and power control. In a FDD (Frequency Division Duplex) mode, the downlink CSI needs to be estimated at a user equipment side, and fed back to the base station by the user equipment side. However, in the MIMO-OFDM system, the numbers of antennas and subcarriers are usually high, so that the data dimension of CSI is very high, which leads to high CSI feedback cost.
MIMO-OFDM CSI is usually expressed as one matrix in mathematics, in which row and column indexes correspond to antenna and subcarrier index information, respectively. Due to the similarity of electromagnetic wave transmission between different antennas and subcarriers, there is implicit correlation between elements in a CSI matrix, which makes the whole CSI matrix compressible. Therefore, how to effectively compress CSI to reduce the feedback cost has become an important topic in the field of wireless communication.
In recent years, deep learning technology has been widely used in the field of wireless communication, showing great application potential. However, most of the current deep learning methods are data-driven, which usually cannot fully tap the potential of deep learning technology in the current task.
At present, most CSI compression and feedback methods based on deep learning use a convolutional neural network CNN to compress and reconstruct CSI. However, due to the translation invariance and smoothness of convolutional neural networks, CSI compression reconstruction methods based on CNN usually encounter significant performance bottlenecks. Based on the physical characteristics of MIMO-OFDM CSI, how to design a neural network structure that is more suitable for the unique characteristics of MIMO-OFDM CSI to give full play to the potential of deep learning in CSI compression and reconstruction has become an important issue in the compression and feedback task of MIMO-OFDM CSI.
SUMMARYThe present disclosure aims to provide a more efficient CSI compression and reconstruction method based on the shortcomings of the prior art, so as to better complete the CSI feedback task.
The specific technical solution adopted by the present disclosure is as follows: a CSI compression and feedback method based on deep learning includes the following steps:
-
- S1, establishing and training a two-dimensional sequence-to-sequence structure neural network;
- S1.1, collecting, by a base station, CSI data from historical communication data as training data of the two-dimensional sequence-to-sequence structure neural network;
- S1.2, building, by the base station, the two-dimensional sequence-to-sequence structure neural network, and randomly initializing parameters of the two-dimensional sequence-to-sequence structure neural network, wherein the two-dimensional sequence-to-sequence structure neural network includes an encoder and a decoder, the encoder includes first cells and first sub-blocks, the decoder includes second cells and second sub-blocks, the encoder includes one two-dimensional recurrent neural network and one fully connected network in a cascade connection, the decoder includes one fully connected network and one two-dimensional recurrent neural network in a cascade connection; inputting, by the encoder, the first sub-blocks divided from CSI into the first cells, concatenating outputs of the first cells into a first matrix according to an order of the first cells, and compressing the first matrix into a first vector; and reshaping, by the decoder, the compressed first vector into a second matrix, dividing the second matrix into the second sub-blocks, input the second sub-blocks into the second cells, and concatenating outputs of the second cells according to an order of the second cells to obtain reconstructed CSI; and
- S1.3, training, by the base station, the two-dimensional sequence-to-sequence structure neural network in the sub-step S1.2 by using the collected CSI data, and continuously training the parameters of the two-dimensional sequence-to-sequence structure neural network through a gradient descent method until convergence;
- S2, distributing, by the base station, the encoder of the two-dimensional sequence-to-sequence structure neural network trained in the sub-step S1.3 to a client;
- S3, compressing, by the client, downlink CSI by using the encoder distributed in the step S2, and feeding the compressed CSI back to the base station; and
- S4, reconstructing, by the base station, the compressed CSI fed back by the client in the step S3 by using the decoder of the two-dimensional sequence-to-sequence structure neural network trained in the sub-step S1.3 to obtain complete downlink CSI.
Further, the two-dimensional recurrent neural network uses a two-dimensional long short-term memory network.
Further, a CSI division process in S1.2 is as follows: dividing the CSI into two-dimensional sequences including several sub-blocks for Nt×Ne-dimensional multi input MIMO-OFDM CSI, where Nt represents a number of base station antennas and Ne represents a number of subcarriers.
Further, a working process of the encoder is as follows: inputting each sub-block obtained after dividing the MIMO-OFDM CSI into a first cell corresponding to the two-dimensional long short-term memory network of the encoder, wherein an output size of the first cell in the two-dimensional long short-term memory network is same as a dimension of the sub-blocks; concatenating the outputs of all the first cells into a first matrix with a dimension size of Nt×Ne according to the order of the first cells, and stretching the first matrix into a first vector with a length of Nt×Ne; and inputting the first vector with the length of Nt×Ne into a fully connected layer to obtain a compressed vector hcom.
Further, a working process of the decoder is as follows: inputting the compressed vector hcom compressed by the encoder into a fully connected layer in the decoder to obtain a second vector with a length of Nt×Ne, reconstructing the second vector with the length of Nt×Ne into a second matrix with a dimension size of Nt×Ne, dividing the second matrix into the second sub-blocks in a same way as the CSI division process, sequentially inputting the second sub-blocks into the second cells corresponding to the two-dimensional long short-term memory network of the decoder, and concatenating the outputs of all the second cells into the second matrix with the dimension size of Nt×Ne according to the order of the second cells to obtain reconstructed MIMO-OFDM CSI.
Further, in a training method for the neural network in S1.3, a loss function Loss (+) of the neural network is defined as follows:
where Θ represents a parameter set of the two-dimensional sequence-to-sequence structure neural network, N represents a number of the training data, Hn represents a nth original training data, and Hn represents an output obtained by inputting Hn into the two-dimensional sequence to sequence structure neural network.
Further, the two-dimensional sequence-to-sequence structure neural network in the sub-step S1.3 is trained based on a optimizer based on gradient descent, including: performing, by the optimizer, gradient descent by deriving the loss function and backpropagating the gradient to generate a gradient corresponding to the parameters of the two-dimensional sequence-to-sequence structure neural network, to continuously optimize the parameters of the two-dimensional sequence-to-sequence structure neural network until the convergence.
The present disclosure has the beneficial effects that the CSI compression and feedback method based on 2D Seq2Seq can realize high-precision CSI feedback under the condition of using feedback vectors with fewer dimensions, and solve the performance bottleneck problem of other current CSI feedback methods based on deep learning. Moreover, the relationship between the computational complexity of this solution has a linear correlation with the number of antennas Nt and the number of subcarriers Ne, and the computational complexity is low, which is suitable for the actual deployment requirements.
In this embodiment, the base station of the large-scale access system is equipped with 32 antennas, and each user terminal is equipped with one antenna. The base station uses an OFDM mode with 256 subcarriers to serve the users.
Based on the base station, this embodiment provides a CSI feedback solution based on deep learning, which includes the following steps:
1) A base station collects 40,000 CSI data from historical communication data as training data of the neural network.
2) The base station builds a 2D Seq2Seq (two-dimensional sequence-to-sequence) structure neural network according to the structure shown in
2. a) The structure of 2dseq 2seq model is as follows: the 2D Seq2Seq model consists of an Encoder and a Decoder. The encoder consists of a two-layer 2D LSTM (long short-term memory) network and a fully connected network in cascade connection; the decoder consists of a fully connected network and a two-layer 2D LSTM network in cascade connection.; the two-layer 2D LSTM refers to two cascaded 2D LSTM networks, in which the hidden state information dimensions of the two cascaded 2D LSTM networks are equal; the overall and internal structure and calculation schematic diagram of the 2D LSTM network are shown in
where ∘ represents a Hadamard product, and W, U, V, b represent parameters of the neural network.
W represents a weight matrix of the input part, U represents a weight matrix of the first dimension state information, V represents a weight matrix of the second dimension state information, b represents bias; the subscripts i, f, o, λ, {tilde over (c)} represent the gates or memories to which the parameters W, U, V, b belong.
2.b) For the Nt×Ne-dimensional MIMO-OFDM CSI (Nt is the number of base station antennas and N is the number of subcarriers), it is divided into two-dimensional sequences consisting of several sub-blocks, as shown in
2.c) In the Encoder, firstly, the MIMO-OFDM CSI is divided into several sub-blocks as in step 2.b). Then each sub-block is input into the corresponding cell in the two-layer 2D LSTM of the Encoder, where the output size of the cell in the two-layer 2D LSTM is the same as the dimension of the sub-block. After two-layer of 2D LSTM, the outputs of all cells are spliced into a matrix with a size of Nt×Ne according to the order of corresponding cells, and the matrix is stretched into a vector with a length of Nt×Ne. Then, the vector with the length of Nt×Ne is input into the fully connected layer to obtain a compressed vector hcom. In this way, the Encoder completes the task of compressing the original MIMO-OFDM CSI into a vector with a lower dimension.
2.d) In the Decoder, firstly, the compressed vector hcom compressed by the Encoder is input into the fully connected layer in the Decoder to obtain a vector with a length of Nt×Ne, and then the vector with the length of Nt×Ne is reconstructed into a matrix with a dimension size of Nt×Ne. Then, the matrix is divided into several sub-blocks according to the same division method as the sub-blocks in step 2.c), and input them into the corresponding cells in the two-layer 2D LSTM of Decoder in turn, and the outputs of all cells are spliced into a matrix with a size of Nt×Ne according to the order of the corresponding cells, so as to obtain the reconstructed MIMO-OFDM CSI. In this way, the Decoder completes the task of reconstructing MIMO-OFDM CSI from compressed vector.
3) The base station uses the collected training data to train the neural network in step 2), and continuously trains the parameters of the 2D Seq2Seq neural network through a back propagation method until convergence;
3.a) The loss function (Loss) of the neural network is defined as follows:
where Θ represents s the parameter of the neural network, N represents the number of training data, Hn represents the nth original training data, and Hn represents an output obtained after inputting Hn into the 2D Seq2Seq network.
3.b) The training of the neural network is based on an Adaptive Moment Estimation (Adam) optimizer. The learning rate is set to 0.001, 800 cycles of training are carried out, and after the 500th training cycle, the learning rate is reduced to 1/5 of the previous one.
4) The base station sends the Encoder of the 2D Seq2Seq neural network trained in step 3) to the client.
5) The user compresses the downlink CSI using the Encoder issued in step 4), and feeds back the compressed information to the base station.
6) The base station reconstructs the compressed information fed back by the user in step 5) by using the Decoder of the 2D Seq2Seq neural network trained in step 3) to obtain a complete downlink CSI.
It can be seen from the computer simulation that as shown in
The above is only the preferred embodiment of the present disclosure, but the protection scope of the present disclosure is not limited thereto. Any change or substitution that can be easily conceived by a person skilled in the art within the technical scope disclosed in the present disclosure shall be included in the protection scope of the present disclosure. Therefore, the scope of protection of the present disclosure shall be defined based on the scope of protection of the claims.
Claims
1. A CSI compression and feedback method based on deep learning, comprising:
- step S1, establishing and training a two-dimensional sequence-to-sequence structure neural network;
- sub-step S1.1, collecting, by a base station, CSI data from historical communication data as training data of the two-dimensional sequence-to-sequence structure neural network;
- sub-step S1.2, building, by the base station, the two-dimensional sequence-to-sequence structure neural network, and randomly initializing parameters of the two-dimensional sequence-to-sequence structure neural network, wherein the two-dimensional sequence-to-sequence structure neural network comprises an encoder and a decoder, the encoder comprises first cells and first sub-blocks, the decoder comprises second cells and second sub-blocks, the encoder comprises one two-dimensional recurrent neural network and one fully connected network in a cascade connection, and the decoder comprises one fully connected network and one two-dimensional recurrent neural network in a cascade connection; inputting, by the encoder, the first sub-blocks divided from CSI into the first cells, concatenating outputs of the first cells into a first matrix according to an order of the first cells, and compressing the first matrix into a first vector; and reshaping, by the decoder, the compressed first vector into a second matrix, dividing the second matrix into the second sub-blocks, input the second sub-blocks into the second cells, and concatenating outputs of the second cells according to an order of the second cells to obtain reconstructed CSI; and
- sub-step S1.3, training, by the base station, the two-dimensional sequence-to-sequence structure neural network in the sub-step S1.2 by using the collected CSI data, and continuously training the parameters of the two-dimensional sequence-to-sequence structure neural network by a gradient descent method until convergence;
- step S2, distributing, by the base station, the encoder of the two-dimensional sequence-to-sequence structure neural network trained in the sub-step S1.3 to a client;
- step S3, compressing, by the client, downlink CSI by using the encoder distributed in the step S2, and feeding the compressed downlink CSI back to the base station; and
- step S4, reconstructing, by the base station, the compressed downlink CSI fed back by the client in the step S3 by using the decoder of the two-dimensional sequence-to-sequence structure neural network trained in the sub-step S1.3 to obtain complete downlink CSI.
2. The CSI compression and feedback method based on deep learning according to claim 1, wherein the two-dimensional recurrent neural network uses a two-dimensional long short-term memory network.
3. The CSI compression and feedback method based on deep learning according to claim 2, wherein a CSI division process in the sub-step S1.2 comprises:
- dividing the CSI into two-dimensional sequences comprising several sub-blocks for Nt×Ne-dimensional multi input MIMO-OFDM CSI, where Nt represents a number of base station antennas and Ne represents a number of subcarriers.
4. The CSI compression and feedback method based on deep learning according to claim 3, wherein a working process of the encoder comprises:
- inputting each sub-block obtained after dividing the MIMO-OFDM CSI into a first cell corresponding to the two-dimensional long short-term memory network of the encoder, wherein an output size of the first cell in the two-dimensional long short-term memory network is same as a dimension of the first sub-blocks; concatenating the outputs of all the first cells into a first matrix with a dimension size of Nt×Ne according to the order of the first cells, and stretching the first matrix into a first vector with a length of Nt×Ne and inputting the first vector with the length of Nt×Ne into a fully connected layer to obtain a compressed vector hcom.
5. The CSI compression and feedback method based on deep learning according to claim 3, wherein a working process of the decoder comprises:
- inputting the compressed vector hcom compressed by the encoder into a fully connected layer in the decoder to obtain a second vector with a length of Nt×Ne, reconstructing the second vector with the length of Nt×Ne into a second matrix with a dimension size of Nt×Ne, dividing the second matrix into the second sub-blocks in a same way as the CSI division process, sequentially inputting the second sub-blocks into the second cells corresponding to the two-dimensional long short-term memory network of the decoder, and concatenating the outputs of all the second cells into the second matrix with the dimension size of Nt×Ne according to the order of the second cells to obtain reconstructed MIMO-OFDM CSI.
6. The CSI compression and feedback method based on deep learning according to claim 1, wherein in a training method for the two-dimensional sequence-to-sequence structure neural network in the sub-step S1.3, a loss function Loss (Θ) of the two-dimensional sequence-to-sequence structure neural network is defined as follows: Loss ( Θ ) = 1 N ∑ n = 1 N H n - H n 2 2,
- where Θ represents a parameter set of the two-dimensional sequence-to-sequence structure neural network, N represents a number of the training data, Hn represents a nth original training data, and Hn represents an output obtained by inputting Hn into the two-dimensional sequence-to-sequence structure neural network.
7. The CSI compression and feedback method based on deep learning according to claim 1, wherein the two-dimensional sequence-to-sequence structure neural network in the sub-step S1.3 is trained based on a optimizer based on gradient descent, comprising: performing, by the optimizer, gradient descent by deriving the loss function and back propagating the gradient to generate a gradient corresponding to the parameters of the two-dimensional sequence-to-sequence structure neural network, to continuously optimize the parameters of the two-dimensional sequence-to-sequence structure neural network until the convergence.
Type: Application
Filed: Dec 5, 2024
Publication Date: Mar 20, 2025
Inventors: Zhaoyang ZHANG (Hangzhou), Zirui CHEN (Hangzhou)
Application Number: 18/969,330