Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
40 views8 pages

Deep Learning Based Channel Estimation Algorithm Over Time Selective Fading Channels

Uploaded by

aimad baigouar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
40 views8 pages

Deep Learning Based Channel Estimation Algorithm Over Time Selective Fading Channels

Uploaded by

aimad baigouar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

1

Deep Learning based Channel Estimation Algorithm


over Time Selective Fading Channels
Qinbo Bai, Jintao Wang, Senior Member, IEEE, Yue Zhang, Senior Member, IEEE, and Jian Song, Fellow, IEEE

Abstract—The research about deep learning application for possible [3]. In Orthogonal Frequency Division Multiplexing
physical layer has been received much attention in recent years. [4] system, the deep learning algorithm for joint channel
In this paper, we propose a Deep Learning(DL) based channel estimation and signal detection has been researched in [5].
arXiv:1908.11013v1 [eess.SP] 29 Aug 2019

estimator under time varying Rayleigh fading channel. We


build up, train and test the channel estimator using Neural To overcome the back-propagation problem in NN transmitter
Network(NN). The proposed DL-based estimator can dynamically when the channel is unknown, different methods are proposed.
track the channel status without any prior knowledge about Policy gradient algorithm in the reinforcement learning is
the channel model and statistic characteristics. The simulation used in [6]. A new deep learning technology, Conditional
results show the proposed NN estimator has better Mean Square Generative Adversarial Nets [7], is introduced in [8] to emulate
Error(MSE) performance compared with the traditional algo-
rithms and some other DL-based architectures. Furthermore, the unknown channel. Simultaneous Perturbation Stochastic
the proposed DL-based estimator also shows its robustness with Approximation [9] algorithm is utilized in [10] to give a direct
the different pilot densities. estimation of the channel gradient.
Index Terms—Deep learning, time varying channel, channel However, in order to make DL-based communication system
estimation, sliding structure, Neural Network meaningful in the practical system, complex channels need to
be considered. One kind of complex channel, which is difficult
to handle with a traditional algorithm, is the time selective
I. I NTRODUCTION
channel. Due to the movement of receiver, the channel status
S the machine learning technology and the perfor-
A mance of hardware develop rapidly in recent years,
Deep Learning(DL) has been successfully applied to many
will change in time domain. Research on such channel using
deep learning is somehow only a little. Sliding Bidirectional
Recurrent Neural Network(SBRNN) has been put forward in
fields, especially in Computer Version and Nature Language [11] and works as a detector to learn rapid varying optical
Processing(NLP). Such technology has been applied to the and molecular channel. A simple application of neural network
physical layer processing of communication systems in [1]. to Rayleigh fading channel is given in [12]. Multiple Layers
Since then, more research has been focusing on applying Perceptron(MLP) is used to undertake channel estimation for
learning algorithms to different communication user scenarios. the time selective channel [13] and doubly selective channel
In traditional communication system, it always consists of [14], respectively.
different modules such as source coding, channel coding, However, MLP is a memoryless structure. Thus, it can’t
modulation, demodulation, estimation, equalization, etc. And learn the relation of data in time domain well. Besides, linear
an end-to-end communication system under AWGN channel layers in MLP will result in the increasing of neurons size
is designed in [1]. Using fully connected NN, whose be- as input length increases. Despite that data can be divided
havior is similar to an autoencoder, it achieves the similar into blocks to avoid this problem, divided data may lead
performance to the tradition system with (7,4) Hamming code to the discontinuity of the channel estimation. Considering
and BPSK modulation. Such autoencoder learns how to get the similarity of this problem in NLP field, it is better to
an expression in a low dimension and the way to restore it. use Recurrent Neural Network(RNN) to get the estimation
And Convolution Neural Network(CNN) based model [2] has of channel. In this article, time varying Rayleigh fading
been developed to solve the dimensional explosion problem in channel is explored using the deep learning technology and
autoencoder and achieves better performance than traditional our contributions are summarized as below.
methods(64QAM+MMSE) under both AWGN and static fad- • Based on deep learning algorithm, the SBGRU channel
ing channel. Besides, a communication system with Software estimator is proposed to learn time varying Rayleigh
Defined Radio(SDR) only including NN are used to prove fading channel. Using RNN structure and sliding idea,
that transmission over the air with deep learning technology is SBGRU can handle the transmitted symbol with arbitrary
This work was supported in part by the National Key R&D Program of length and immediately provide the result as soon as the
China under Grant 2017YFE0112300 and Beijing National Research Center symbol arrived.
for Information Science and Technology under Grant BNR2019RC01014 and • Substantial simulations are provided in this paper to
BNR2019TD01001.(Corresponding author: Jintao Wang.)
Qinbo Bai, Jintao Wang, and Jian Song are with the Electronic Engineering analyze and explain NN estimator. The simulation result
Department, Tsinghua University, and Beijing National Research Center for shows the ability of SBGRU to track channel dynamically
Information Science and Technology (BNRist), Beijing 100084, China (e- and achieve better performance compared with traditional
mail:[email protected]).
Yue Zhang is with the Department of Engineering, University of Leicester, algorithms and other NNs. Besides, the SBGRU also has
Leicester, LE1 7RH, United Kingdom (e-mail: [email protected]). demonstrated the robustness with various pilot densities.
2

The rest parts of this article is arranged as follow: Section


II describes the basic channel model, data structure and the
signal flow model. Section III gives the deep learning based
algorithm in details for NN channel estimator. Section IV uses
quantities of simulation results to demonstrate the performance
of NN estimator. Finally, Section V concludes the paper and
gives some orientations for future work.
Notation: Bold lower-case letters and upper-case letters
denote vectors and matrices, respectively. The subscript on
Fig. 1. Data structure of transmitted signal in time domain
a lower-case letter xi represent i th element of vector x. E(·)
refers to the expectation. (·)T and (·) H refer to the transpose
and Hermite transpose of the vector. | · | represents for the
absolute value or amplitude for real number and complex
number, respectively. For two vectors or matrices a and
b,[a, b] is the matrix combing a and b. For two real numbers
a ≤ b, [a, b] is the set for all real numbers in range from a to
b. real(·) and image(·) are the functions giving the real and
imaginary part of complex vector for each element. Fig. 2. Signal flow model

II. S YSTEM M ODEL where φ d = rfds is the maximum Doppler frequency normalized
In this section, signal architecture and time varying Rayleigh by sampling rate. Besides, It is generally asked that the channel
fading channel model are firstly presented. Then, a signal flow has normalized gain E(|h[n]| 2 ) = 1 in order to simplify
model will be introduced. Denote the transmitted signal and following analysis.
received signal as x,y, respectively. Denote the Rayleigh time
varying channel as h. Considering a Linear Time Variant(LTV) B. Signal Architecture
model, the relation between input and output of channel is: Considering standard signal architecture, transmitted signals
y = h· x+ω (1) are generated as shown in Fig. 1. One single data frame
consists of K blocks. Due to multi-path not considering in
where ω is i,i,d Additive White Gaussian Noise(AWGN) this article, protection interval isn’t necessary. Each block has
vector, and ωi ∼ CN (0, σn2 ) Ns information symbols and Np pilot symbols. Thus, each
block has Ns + Np = N symbols and the whole frame has total
A. Time Varying Rayleigh Fading Channel Model L = NK symbols. Pilots are equally interval inserted in each
Np
block, and define Ns +N as the pilot density. Besides, pilots
Typically, wireless communication environment is generally p
in each block are the same, which results in repetition in time
modeled as Rayleigh fading channel. Multi-path will cause domain.
frequency selective fading and Doppler shifting will result in
time selective fading. However, in this paper, only time selec-
tive fading is considered in order to give the first exploration C. Signal Flow Model
of rapidly varying channel. The influence of multi-path will The signal flow model is shown in Fig. 2. At the transmitter
be researched in the future work. side, no deep learning technology is introduced. Information
Clarke’s model [15] is used in this paper to describe bits and pilot bits are combined to generated original signal.
time varying channel. In order to describe the time varying After modulating, transmitted signal x is sent to the channel
characteristic, Jakes Doppler Spectrum [16] is adopted here: and modulated pilots p are sent to NN estimator. At the
receiver side, NN channel estimator uses p and channel
1
S( f ) = q , | f | < fd (2) distorted signal plus the noise y to give the estimation of
π fd 1 − ( ffd )2 channel h.
Two things need to be notified. Firstly, due to no NN
where fd is the maximum Doppler shift. Given a speed v(m/s) introduced at transmitter, it is easy to add any traditional
and carrier frequency fc (Hz), fd = vcfc (c ≈ 3.0 ∗ 108 is the channel coding such as Low Density Parity Check(LDPC)
speed of light in free space). The autocorrelation of Jakes [17], to improve the performance against noise. Secondly,
Doppler Spectrum is: NN channel estimator doesn’t need any information about the
∫ fd channel. It means that the communication system is model
R(τ) = S( f ) exp( j2π f τ)df = J0 (2π fd τ) (3) free.
− fd

where J0 (·) is the first kind of Bessel function of 0 order and D. Traditional Algorithms For Channel Estimation
the discrete form of autocorrelation is:
In channel estimation, the most common estimators are
R[d] = J0 (2πφ d |d|) (4) Least Square(LS) [18] estimator and Minimal Mean Square
3

Error(MMSE) [19] estimator. According to (1), LS estimator The basic RNN cell will give the computation result as the
under the time varying channel is: following function.
y ht = T anh(Wih xt + bih + Whh ht−1 + bhh )
ĥ LS = (5) (9)
x
where T anh is hyperbolic tangent function and ht , ht−1 are
For those positions where pilots are inserted, above equation
the hidden states at time t and t − 1, respectively. xt is the
can be directly used to get the estimation. For other positions,
input at time t. Wih, Whh and bih, bhh are weights and biases,
linear interpolation is necessary. Denote j,k( j < k) to be po-
which need to be learned.
sitions of pilot nearest to the position i. Thus, the interpolated
However, the time varying channel h(t) has relation with
channel is:
both past and future channel states. Basic RNN cell is fed
k −i i−j forward only. Thus, bidirectional structure, as shown in Fig.
ĥi, LS = ĥ j, LS + ĥk, LS (6)
k−j k−j 3b, would have better performance. Blue blocks are forward
Due to the existence of noise, omitting the influence of cells and red blocks are backward cells. The data will not
interpolation, the expected Mean Square Error(MSE) of LS only be fed in forward direction, but fed backward again. The
0
estimator is: hidden states ht and ht are combined together to become the
input of a linear layer to give final results.
ω2 1 Another problem is that Basic RNN cell with (9) can’t
E(| ĥ LS − h| 2 ) = E( 2
)= (7)
x SN R capture long time information. To solve this problem, Long
Another traditional estimator would be MMSE estimator: Short Time Memory(LSTM) [20] cell has been put forward.
In this paper, Gated Recurrent Unit(GRU) [21] is used, one
σn2 −1
ĥ M M SE = Rhy R−1
yy y = R hh (R hh + I) ĥ LS (8) variation of LSTM, to replace basic RNN cell. The GRU will
σs2 give the result as the following function( [21],(5),(6),(7),(8))
where I represents unit matrix and Rhh = E(hh H ) represents z t = σ(Wz · [ht−1, xt ]) (10a)
correlation matrix:
rt = σ(Wr · [ht−1, xt ]) (10b)
 R[0] R[1] R[2] ··· R[L − 1] 
h t = T anh(W · [rt ∗ ht−1, xt ])
 
 R[1]
 R[0] R[1] ··· R[L − 2] 
 (10c)
Rhh = 
 R[2] R[1] R[0] ··· R[L − 3] 
.. .. .. .. ..
 ht = (1 − z t ) ∗ ht−1 + z t ∗ h t (10d)
. . . . .
 
 
where σ(·) refers to Sigmoid function fs (x) = 1+e1−x ,
 
 R[L − 1] R[L − 2] R[L − 3] ··· R[0] 
Wz , Wr , W are weights and ht , ht−1, xt have the same meaning
 
where R[·] can be calculated according to (4) as (9). Compared with basic RNN cell, GRU introduces 2
It should be noticed that the form of autocorrelation function gates, update gate z t and reset gate rt , to control the informa-
of channel and Doppler speed need to be given in advance tion flow. GRU has been proved to have similar performance
in order to undertake the MMSE estimation. However, real to LSTM on many tasks [22] and have higher speed due to
channel model and accurate statistic characteristic(Doppler less gate number.
speed here) are hard to know under practical application. Thus, Based on above discussion, BGRU cell will be used in NN
two methods for MMSE estimation are used in simulation. channel estimator. However, the result of simple BGRU is
Firstly, assuming above information already known, not good enough. The idea of Sliding BRNN(SBRNN) [11]
ĥ M M SE can be directly calculated according to (3) and (8). is considered to improve the performance further, and the
Thus, we call this method "MMSE theory". Secondly, after compare between BGRU and SBGRU will be given in section
getting LS estimation, ĥ LS can be used to calculate auto- IV.C.
correlation R[d] = n=0
Í L−1
ĥ LS [n] ĥ LS [n − d] and then use (8).
We call this method "MMSE sim" because the computation is
B. SBGRU structure
completed by simulation results.
SBRNN is put forward in [11] to work as a detector under
optical and molecule channel. Here, this structure is used
III. DL- BASED NN CHANNEL ESTIMATOR in estimation task under the time varying Rayleigh fading
To track a time varying channel, it is necessary to give neu- channel. A simple example of the sliding structure is given
ral network the ability of studying the behavior of correlation in Fig. 4. Each BGRU block in the figure has a fixed window
in time domain. Thus, a good choice to handle sequence data length W L . It should be stated that the selection of window
is using RNN. length has relationship with channel character. Due to the any
two moments of channel h is correlated, it is reasonable that
the longer the window is, the better the performance will be.
A. RNN structure The simulation about window length will be given in section
A simple example of 1 layer RNN is given in Fig. 3a. In IV.D.
this structure, the output of last time becomes one part of input SBGRU will be given W L symbols to undertake once com-
of this time. By this way, RNN can capture past information. putation, and will slide 1 symbol after each computation. Due
4

(a) (b)
Fig. 3. The structure of RNN (a) The structure of forward only RNN (b) The structure of bidirectional RNN

Fig. 4. Sliding structure of BGRU

to the sliding operation, most symbols in the sequence will be


estimated for several times. We take the average of all estima-
tion to give final results. Denote ht = fBGRU (xt , ht−1, ht+10 )

as the function of operation defined in (10) for bidirectional


version in BGRU layer. Denote S = { j | j ∈ Z, max(0, t − W L +
1) ≤ j ≤ min(t, L−1)} as the set including all starting positions
of BGRU for symbol xt , and final output of SBGRU for xt
is:
1 Õ j 0j
ht = fBGRU (xt , ht−1, ht+1 ) (11) Fig. 5. The structure of SBGRU channel estimator
|S| j ∈S
j 0j
where ht−1 and ht+1 are the hidden states for BGRU starting Considering the balance between accuracy and training
from j th symbol in time t − 1 for forward and t + 1 for time, here 2 layers BGRU are adopted to construct SBGRU
backward. layer. Denote function fSBGRU as the operation in SBGRU
layer defined by (11) and function fLinear as the operation in
C. Train and test NN estimator Linear layer defined as:
A final implement of SBGRU neural network is given in
Fig. 5. The input data to SBGRU consists of channel distorted fLinear (x) = W x + b (12)
signal y and the original pilot information p = [ p 1, p 2, ..., p K ],
where W, b are weight and bias in linear layer, respectively.
including K same pilot blocks p i with length N, and
The final estimation of channel, denoted as ĥ, can be expressed
p i = [p1, 01∗Ns , p2, 01∗Ns , ..., p N p , 01∗Ns ] as:
It means that the pilot sequence will have the same symbols ĥ = fLinear ( fSBGRU (Xin, θ S ), θ L ) (13)
as x in pilot positions and have 0 symbols in information
where θ S are the parameters of SBGRU and θ L are the
positions. Due to current deep learning platform only receiving
parameters of Linear layer.
real numbers, real part and image part of complex signal need
Denote θ = {θ S, θ L } to make notation clearly. To train the
to be separated firstly. Thus, the input data of the SBGRU will
NN estimator, a loss function, which can represent the system
be given as:
performance, needs to be constructed. And parameters θ need
Xin = [real(yT ), real(PT ), imag(yT ), imag(PT )]T to be optimized in order to minimize the loss function. Due
5

TABLE I
C HANNEL AND DATA PARAMETERS

Carrier frequency 5.2GHz


Sampling rate 0.25MHz
Receiver speed 10m/s
Signal Length 160 symbols
Pilot density 50%
Signal Block Length 16 symbols
Signal Block Number 10 symbols

to MSE always regarded as criterion in estimation problem,


MSE loss function is adopted, which can be expressed as
L

Loss(θ) = | ĥn − hn | 2 (14)
L n=1
Fig. 6. Performance compare with LS and MMSE estimator
Minimizing loss function can be completed by updating θ
iteratively. The most classical algorithm is Stochastic Gradient
Descant(SGD). Adam [23] optimization algorithm, which has B. Performance Comparison with the traditional algorithm
better performance in multiple tasks, is adopted here.
Here the proposed NN channel estimator is compared with
Testing data has the same structure and statistic character-
traditional algorithm, LS estimator and MMSE estimator. The
istics with training data. Trained parameters θ are loaded to
performance comparison is shown in Fig. 6. It is obvious
finish the computation of testing data and get the estimated
that "MMSE theory" achieves the best performance within
channel.
the testing SNR range. And the LS estimation is the worst
due to not considering the influence of the noise. And the
IV. S IMULATION R ESULTS simulation result does match the expected performance stated
in (7). "MMSE sim" estimation, stated in section II.E, has
In this section, we demonstrate the performance of NN small performance improvement compared with LS estimator
channel estimator under the time varying Rayleigh fading and the improvement decreases when SNR reaches high value.
channel and provide the explanation to the performance im- SBGRU estimator reaches the similar performance to "MMSE
provement through the simulation results. And the simulation theory" estimator and doesn’t need any channel knowledge.
setting for the NN estimator is firstly described. Then, four Besides, SBGRU estimator also greatly outperforms both LS
group simulation results of NN estimator have been presented and "MMSE sim". Such results prove that SBGRU estimator
and analyzed. is a best solution under the time varying channel.
To visualize how the SBGRU estimator work, the perfor-
mance of the channel tracking of the SBGRU and traditional
A. Simulation Setting
estimator is given in Fig. 7. In order to make the channel
In the following simulations, i.i.d. bit sequences are ran- varying significant in time domain, channel length is extended
domly generated, and QPSK modulation is used to map bits to 4000 symbols and SNR is set to 20dB. It’s easy to find
to symbols. According to the channel model given in section that SBGRU estimator can track the channel very well in
II.A and channel parameters given in Table I, 1200 channels most linear parts and has slight oscillation in non-linear parts.
are generated, 800 for training, 200 for validation and 200 However, in Fig. 7b, where white line represents real channel,
for testing. The selection of channel parameters and pilot both LS estimator and "MMSE sim" estimator vibrate heavily.
density is the same as [13] in order to undertake comparison
simulation in Section IV.C. Also, based on the data structure
in Fig. 1 and data parameters in Table I, 120000 sequences are
generated, 100000 for training, 10000 for validation and 10000 C. Performance Comparison with different structures of NN
for testing. When calculating the channel distorted signal, each When deep learning algorithms are used to undertake the
symbol sequence randomly choose one channel to send. channel estimation, different structures of neural network will
The default data and NN parameters of estimator, detector achieve different performances. Firstly, The enhancement of
and system are shown in Table II. the sliding operation for SBGRU is demonstrated in Fig. 8.
The proposed DL-based algorithm is implemented on a All settings are the same except that BGRU computes block
computer with an Intel (R) Corel (TM) i7-6700K CPU @ by block. The performance of BGRU decreases rapidly as
4.0GHz CPU, a NVIDIA GeForce GTX 1080 GPU and SNR increases because the introduction of sliding operation
16GB memory. Pytorch 1.0.0 and python 3.6 are used for the can utilizes the average channel information within a certain
estimation. time window.
6

TABLE II
NN PARAMETERS FOR SIMULATION

Parameter Estimator
NN architecture SBGRU
Number of hidden layers 2
Hidden size 40*2(2 for Bi-direction)
Window Length 40 symbols
Activation function Tanh for hidden layers & Relu for hidden layers
Loss function MSE
Optimizer Adam
Learning rate 0.001
Batch Size 128
Train SNR 20dB
Test SNR 5,10,15,20,25dB
Train number 100000
Validation number 10000
Test number 10000

(a) (b)
Fig. 7. Simulation results for Channel Tracking. (a) Tracking performance of SBGRU estimator (b) Tracking performance of LS and MMSE estimator

Besides, the channel estimation problem under similar time block length increases to 32, the performance increase a
varying channel has been researched in [13] by using MLP bit. However, a estimation block length of 40 will result in
neural network. Its basic idea is to include not only the channel performance decreased. It is because MLP with estimation
distorted data and pilot data but the estimated channel from last block length 40, which is not the integral multiple of original
block to get the better channel estimation performance. In its data block length 16, can’t fully explore the pilot information
simulations, it sets the estimation block length the same as the repeated in time domain. However, SBGRU estimator outper-
data structure. However, this estimation block length can be forms all above MLP estimator when SNR is above 5dB.
different. In order to compare the performance fairly, the NN Besides, thanks to the recurrent structure of RNN, previous
architecture in [13] is reconstructed, trained and tested using channel estimation doesn’t need to be inputed into neural
the same settings and simulation parameters as the SBGRU network. It can be captured by SBGRU automatically.
simulation. Besides, three different parameters 16, 32 and 40
are used to fully explore the influence of the estimation block D. Performance vs window length
length, .
Here the influence of sliding window length is explored.
The performance comparison between MLP and SBGRU is The performance among different window lengths is in Fig.
given in Fig. 9. MLP with estimation block length 16(same 10. The performance monotonically increases as the window
design as [13]) doesn’t work very well. It is possible that length getting longer. Except for window length of 16 symbols,
parameters in NN model is not enough so that the ability all 3 other window lengths have nearly the same performance.
to learn the nonlinear channel isn’t strong. When estimation It shows that the window length can’t be too short in or-
7

Fig. 8. Performance compare between Sliding BGRU and Non-sliding BGRU Fig. 10. The influence of sliding window length to SBGRU estimator

Fig. 9. Performance compare between SBGRU estimator and MLP estimator Fig. 11. The influence of pilot density to SBGRU estimator

der to have enough information to undertake the estimation. DL-based channel estimator can achieve better performance
However, the too long window length can’t bring much more than traditional algorithms and some NN estimators with dif-
improvement. Thus, selecting a suitable window length can ferent structures. Besides, the proposed NN channel estimator
achieve the balance between the accuracy and the speed of shows its ability to dynamically track the channel and its
training and testing. Overall, the setting of window length have robustness with pilot density.
the relation with channel characteristics. In the traditional communication, there are much more
complex traditional algorithms to complete channel estimation.
However, there are some unique advantages compared with the
E. Performance vs pilot density
traditional algorithms when deep learning algorithms are used.
Finally, the influence of pilot density is described to show
• Despite many estimation methods having been developed
the robustness of SBGRU estimator. The performance is
in traditional communication system, most of them al-
shown in Fig. 11. As the pilot density decreases, the MSE
ways assume the channel to be invariant in coherence
performance indeed decreases a little but not seriously. The
time. However, using deep learning algorithm, the prior
result is still much better than LS estimation and "MMSE sim"
knowledge about channel model and the channel invariant
estimation. Thus, SBGRU estimator shows the performance
in coherence time assumption aren’t needed during the
robustness with the different pilot densities.
training and testing, which shows the potential perfor-
mance of DL-based algorithm under the time varying
V. C ONCLUSION channel.
In this paper, a DL-based channel estimator is designed • The channel estimator designed in this paper can be easily
under the time varying Rayleigh fading channel. The proposed optimized by combining traditional algorithms. For exam-
8

ple, it’s convenient to insert the high performance channel [19] V. Charles Drastik, “Minimum mean squared error estimation,” Bulletin
coding before the modulation to protect the performance of The Australian Mathematical Society - BULL AUST MATH SOC,
vol. 30, 10 1984.
against Gaussian noise. Thus, the MSE performance can [20] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural
be further improved. computation, vol. 9, pp. 1735–80, 12 1997.
[21] K. Cho, B. van Merriënboer, C. Gulcehre, F. Bougares, H. Schwenk, and
In addition, there is still a lot work to do in applying Y. Bengio, “Learning phrase representations using rnn encoder-decoder
deep learning or machine learning technology to the physical for statistical machine translation,” 06 2014.
layer under time varying channel and here are some following [22] K. Greff, R. K. Srivastava, J. Koutník, B. R. Steunebrink, and J. Schmid-
huber, “Lstm: A search space odyssey,” IEEE Transactions on Neural
aspects. Networks and Learning Systems, vol. 28, no. 10, pp. 2222–2232, Oct
• Except for the channel estimation, it is also feasible to 2017.
[23] D. Kingma and J. Ba, “Adam: A method for stochastic optimization,”
construct a detector to undertake the equalization and International Conference on Learning Representations, 12 2014.
demodulation together using deep learning algorithm.
Thus, by connecting NN estimator and NN detector,
a wireless communication system can be constructed.
It is worth to explore whether such DL-based system
can achieve better bit error rate(BER) performance than
traditional system under the time varying channel and is
still robust with different pilot densities.

R EFERENCES
[1] T. O’Shea and J. Hoydis, “An introduction to deep learning for the
physical layer,” IEEE Transactions on Cognitive Communications and
Networking, vol. 3, no. 4, pp. 563–575, Dec 2017.
[2] B. Zhu, J. Wang, L. He, and J. Song, “Joint transceiver optimization for
wireless communication phy using neural network,” IEEE Journal on
Selected Areas in Communications, vol. 37, no. 6, pp. 1364–1373, June
2019.
[3] S. DÃűrner, S. Cammerer, J. Hoydis, and S. t. Brink, “Deep learning
based communication over the air,” IEEE Journal of Selected Topics in
Signal Processing, vol. 12, no. 1, pp. 132–143, Feb 2018.
[4] B. Le Floch, M. Alard, and C. Berrou, “Coded orthogonal frequency
division multiplex [tv broadcasting],” Proceedings of the IEEE, vol. 83,
no. 6, pp. 982–996, June 1995.
[5] H. Ye, G. Y. Li, and B. Juang, “Power of deep learning for channel
estimation and signal detection in ofdm systems,” IEEE Wireless Com-
munications Letters, vol. 7, no. 1, pp. 114–117, Feb 2018.
[6] B. Karanov, M. Chagnon, F. Thouin, T. A. Eriksson, H. BÃijlow,
D. Lavery, P. Bayvel, and L. Schmalen, “End-to-end deep learning
of optical fiber communications,” Journal of Lightwave Technology,
vol. 36, no. 20, pp. 4843–4855, Oct 2018.
[7] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” 11
2014.
[8] H. Ye, G. Y. Li, B. F. Juang, and K. Sivanesan, “Channel agnostic end-
to-end learning based communication systems with conditional gan,” in
2018 IEEE Globecom Workshops (GC Wkshps), Dec 2018, pp. 1–5.
[9] J. C. Spall, “An overview of the simultaneous perturbation method for
efficient optimization,” 02 2001.
[10] V. Raj and S. Kalyani, “Backpropagating through the air: Deep learn-
ing at physical layer without channel models,” IEEE Communications
Letters, vol. 22, no. 11, pp. 2278–2281, Nov 2018.
[11] N. Farsad and A. Goldsmith, “Neural network detection of data se-
quences in communication systems,” IEEE Transactions on Signal
Processing, vol. 66, no. 21, pp. 5663–5678, Nov 2018.
[12] S. Ganesh, V. Sayee Sunder, and A. Thakre, “Performance improvement
in rayleigh faded channel using deep learning,” in 2018 International
Conference on Advances in Computing, Communications and Informat-
ics (ICACCI), Sep. 2018, pp. 1307–1312.
[13] X. Ma, H. Ye, and Y. Li, “Learning assisted estimation for time-
varying channels,” in 2018 15th International Symposium on Wireless
Communication Systems (ISWCS), Aug 2018, pp. 1–5.
[14] Y. Yang, F. Gao, X. Ma, and S. Zhang, “Deep learning-based channel
estimation for doubly selective fading channels,” IEEE Access, vol. 7,
pp. 36 579–36 589, 2019.
[15] R. H. Clarke, “A statistical theory of mobile-radio reception,” The Bell
System Technical Journal, vol. 47, no. 6, pp. 957–1000, July 1968.
[16] M. J. Gans, “A power-spectral theory of propagation in the mobile-radio
environment,” IEEE Transactions on Vehicular Technology, vol. 21,
no. 1, pp. 27–38, Feb 1972.
[17] M. Livshitz, “Low density parity check (ldpc) code,” Patent, 11, 2013.
[18] S. A. van de Geer, Least Squares Estimation, 10 2005, vol. 2.

You might also like