A Novel Multilayer Neural Network Model For TOA-based Localization in Wireless Sensor Networks
A Novel Multilayer Neural Network Model For TOA-based Localization in Wireless Sensor Networks
net/publication/261351830
CITATIONS READS
16 145
2 authors:
All content following this page was uploaded by Reza Monir Vaghefi on 16 March 2015.
Abstract—A novel multilayer neural network model, called In SDP, the cost function is approximated and relaxed to
artificial synaptic network, was designed and implemented for a convex optimization problem and solved with efficient
single sensor localization with time-of-arrival (TOA) measure- algorithms that do not require any initialization. In MDS,
ments. In the TOA localization problem, the location of a source
sensor is estimated based on its distance from a number of the locations of the sources are estimated using data analysis
anchor sensors. The measured distance values are noisy and of the coordinates and distances. It is stated that the MDS
the estimator should be able to handle different amounts of approach is very sensitive to measurement noise and not
noise. Three neural network models: the proposed artificial applicable to low connectivity networks [7].
synaptic network, a multi-layer perceptron network, and a Linear estimators [8], [9], [10] are also investigated in a
generalized radial basis functions network were applied to
the TOA localization problem. The performance of the models number of studies. The model of the TOA localization, which
was compared with one another. The efficiency of the models is basically a non-linear problem, is linearized using some
was calculated based on the memory cost. The study result approximation. The least squares (LS) method is studied in
shows that the proposed artificial synaptic network has the [9]. The method is simple to implement but cannot handle
lowest RMS error and highest efficiency. The robustness of the measurement noise. The TOA measurement noise is
the artificial synaptic network was compared with that of
the least square (LS) method and the weighted least square usually modeled as a zero-mean Gaussian random variable
(WLS) method. The Cramer-Rao lower bound (CRLB) of TOA with a variance depending mostly on the distances (i.e., the
localization was used as a benchmark. The model’s robustness larger measured distances have higher noise than the shorter
in high noise is better than the WLS method and remarkably ones). To improve the performance of the LS method, the
close to the CRLB. weighted least squares (WLS) algorithm [11] is introduced
I. INTRODUCTION which can tolerate unequally sized noises. However, since
linearization of the TOA localization model is done under
W IRELESS sensor network (WSN) localization prob-
lem is one of the interesting subjects studied in recent
years. In this problem, the locations of anchor sensors are
the assumption that the measurement noise is sufficiently
small, the performance of LS and WLS declines considerably
as the measurement noise increases. Different extensions
known and the location of each source sensor is estimated
to the LS method such as the constrained weighted least
based on its distance from the anchor sensors. The approx-
squares (CWLS) method [8] and corrected least squares
imate distance of a source sensor from an anchor sensor
method [9] are also introduced. Although these algorithms
is obtained using different measurement methods such as
have better performance in high noise, their computation cost
time of arrival (TOA) [1], time difference of arrival [2],
and complexity is higher.
and received signal strength (RSS) [3]. Throughout this
In this project, the TOA-based sensor localization is tack-
work we assume that the distances are obtained using TOA
led using artificial neural network models. Shareef et al.
measurements, however, the problem can easily be extended
[12] have compared the performance of three types of neural
to the other methods.
networks namely RBF, MLP, and recurrent neural networks
The maximum likelihood (ML) estimator is the optimal
in sensor localization. Their result shows that the RBF
estimator when the number of data records is sufficiently
network performs better than the other networks but it has
large [1], [4]. However, the cost function of the ML estimator
higher memory and computation costs. On the other hand,
is non-linear and non-convex and finding the global minimum
the MLP network has the lowest memory and computation
requires convoluted computations. The solution of the ML
costs. Rahman et al. [13] have implemented a MLP network
estimator is computed using iterative optimization methods
for the WSN localization problem. The network has reached
[1]. Since the cost function has many saddle points and local
the RMSe of 0.54 meter for a 20𝑚 × 20𝑚 problem. The
minima, a good initialization guess is inevitable to make sure
localization problem in a UWB sensor network is tackled in
that the algorithm converges to the global minimum. To deal
[14]. A neural network model is introduced. The performance
with this behavior of the ML estimator, different methods
of the model has not been as good as the LS method. In
such as semidefinite programming (SDP) relaxation [5], [6]
[15], a neural network model is developed for identification
and multidimensional scaling (MDS) [7] are introduced.
of undetected direct paths (UDP) in TOA localization.
Sayed Yousef Monir Vaghefi is with the School of Computer Science The advantage of artificial neural network models is that
and Information Technology at Royal Melbourne Institute of Technology, they are adaptable to different conditions and situations.
Melbourne, Australia (email: [email protected]), Reza Monir
Vaghefi is with Department of Signals and Systems, Chalmers University Indeed, our proposed model is designed to tolerate a specific
of Technology, Gothenburg, Sweden (email: vaghefi@student.chalmers.se). condition or situation. For instance, if the sensor network is
II. NETWORKS
In this study, the TOA localization problem was considered
as a supervised learning problem, in which a neural network
is trained with a set of input-output data called the training
dataset, and tested with another set of input-output data called
the test dataset. Three neural network models: an artificial
synaptic network (ASN), a generalized radial basis func-
tions (GRBF) network, and a multilayer preceptron (MLP)
network were applied to the problem. The models were
implemented in C sharp . The performance and efficiency
of the models were studied. The best model was identified
and then tested for robustness. The designed ASN model is
comprised of a number of artificial synaptic networks - each
of them working on a data cluster. The architecture of each
network is presented in Fig. 1. The output of each network
is computed using:
𝑛
∑
𝑓 (𝑥) = 𝑤𝑖 ∗ 𝑥 𝑖 + 𝑤0 . (1)
𝑖=1
3080
20
3. For each center
Change the center to the centroid of the data points 15
assigned to that center
4. If not yet converged go to step 2 10
y coordinate [m]
b. Compute the networks output:
𝑛𝑒𝑡𝑤𝑜𝑟𝑘𝑜𝑢𝑡(𝑥) = 𝑐𝑜𝑚𝑝𝑢𝑡𝑒𝑜𝑢𝑡𝑝𝑢𝑡(𝑥) 0
−5
−10
−15
−20
−20 −15 −10 −5 0 5 10 15 20
x coordinate [m]
20
15
5
y coordinate [m]
−5
−10
−15
−20
−20 −15 −10 −5 0 5 10 15 20
x coordinate [m]
𝑛
∑
Fig. 3. GRBF Network Architecture
𝑛𝑒𝑡𝑤𝑜𝑟𝑘𝑜𝑢𝑡(𝑥) = 𝑤𝑖 ∗ 𝑥 𝑖 + 𝑤0 . (5)
𝑖=1
6. Compute the Root Mean Square Error:
√ where n is the number of inputs.
∑𝑛 2 The backpropagation(x:the data point) method:
𝑖=1 (𝑜𝑢𝑡𝑝𝑢𝑡𝑖 − 𝑛𝑒𝑡𝑤𝑜𝑟𝑘𝑜𝑢𝑡𝑖 )
𝑟𝑚𝑠𝑒 = (3) 1. Compute the output for the data point:
𝑛
𝑐𝑜𝑚𝑝𝑢𝑡𝑒𝑜𝑢𝑡𝑝𝑢𝑡(𝑥)
where n is the number of training data. 2. Backpropagate the error layer by layer
7. If (𝑟𝑚𝑠𝑒 > 𝑚𝑎𝑥𝑒𝑟𝑟𝑜𝑟) go to step 5 a. Output layer:
The computeoutput(x: the data point) method: 𝑑𝑒𝑙𝑡𝑎1 = 𝑜𝑢𝑡𝑝𝑢𝑡(𝑥) − 𝑛𝑒𝑡𝑤𝑜𝑟𝑘𝑜𝑢𝑡(𝑥)
1. Compute the Euclidean distance of x from all the centers b. Second layer:
2. Send x to the network with the closet center 𝑑𝑒𝑙𝑡𝑎2𝑖 = 𝑑𝑒𝑙𝑡𝑎1 ∗ 𝑥𝑖
3. Compute the output of the chosen network: c. Third layer:
𝑛 𝑑𝑒𝑙𝑡𝑎3𝑖𝑗 = 𝑑𝑒𝑙𝑡𝑎2𝑖 ∗ 𝑥𝑗
∑
𝑤𝑖 = 𝑤𝑖𝑗 ∗ 𝑥𝑗 + 𝑤𝑖0 . (4) 3. Update the weights of the deepest layer
𝑗=1 𝑙𝑎𝑦𝑒𝑟3𝑤𝑒𝑖𝑔ℎ𝑡𝑠 = 𝑙𝑎𝑦𝑒𝑟3𝑤𝑒𝑖𝑔ℎ𝑡𝑠 + 𝑒𝑡𝑎 ∗ 𝑑𝑒𝑙𝑡𝑎3 ∗ 𝑥
3081
TABLE I
The multilayer preceptron (MLP) network implemented T HE COMPARISON OF MODELS
for the problem has two hidden layers. The architecture of
RMSe Memory Cost Iterations
the network is presented in Fig. 2. The activation function of Generalized Radial Basis
the hidden neurons is a logistic function. The network was Functions (GRBF) Network 2.829 4*15+15=75 99917
Multilayer Preceptron (MLP)
trained using the backpropagation algorithm [16]. Network 3.204 4*7 + 7*7 +7 = 84 97186
The GRBF network [17] consists of a number of radial Artificial Synaptic Network
(ASN) 0.2999 3*4*5 +3*4= 72 91697
basis function neurons, each of them working on a center.
In the training, the fixed centers of the network are specified
using the k-means clustering algorithm. The network is then
trained using the gradient descent algorithm. Fig. 3 shows of different number of hidden neurons. The number of
the architecture of the network. hidden neurons in the first and second hidden layer is equal.
The networks were trained for 100 thousand iterations. By
III. DATA
increasing the number of hidden neurons, the RMSe first
The performance of a neural network model depends decreases. But then it increases because networks with more
on the density of the training data and complexity of the hidden neurons require more iterations to converge. Lower
problem. If the training data is not dense, the network does RMSe can be achieved, if the learning rate goes up. After
not have enough information to build the model. Considering changing the number of neurons to 12, the learning rate was
the modeling process as a hyperplane reconstruction process, increased. As a result the RMSe went down. To conclude,
the more complex the hyperplane is the more data is required the MLP network could not achieve better performance than
for reconstruction. the ASN model, even with more hidden neurons.
The datasets were generated in MATLAB. The training
dataset was a set of 144 data points evenly distributed over 5
the input space. Fig. 4 presents the training dataset. The 4.5
red points, diamonds, are the source sensors and the blue
4
points, squares, are the anchors. The test dataset was 300
data points randomly distributed over the input space. The 3.5
3082
4.5
each anchor was computed. A zero-mean Gaussian noise LS
WLS
was added to the distances. A set of 10 different four- 4
ASN−3
distance inputs were computed for each location. A total of CRLB
3.5
3000 inputs were sent to the ASN model and the RMSe
was computed. The experiment was repeated for different 3
RMSE [m]
2.5
20
2
15
1.5
10
1
5 0.5
y coordinate [m]
0
0 0 0.5 1 1.5 2 2.5
Standard deviation of measurement noise [m]
−5
Fig. 8. ASN model with 3 centers
−10
7
LS
−15 WLS
6 ASN−5
CRLB
−20
−20 −15 −10 −5 0 5 10 15 20 5
x coordinate [m] RMSE [m]
4
Fig. 7. Diamonds depict estimated locations and squares depict true
locations.
3
3083
7
LS Applied Mathematics, 2005, pp. 405–414. [Online]. Available:
WLS http://portal.acm.org/citation.cfm?id=1070432.1070488
6 ASN−5H [6] C. Meng, Z. Ding, and S. Dasgupta, “A semidefinite programming
CRLB
approach to source localization in wireless sensor networks,” Signal
5
Processing Letters, IEEE, vol. 15, pp. 253 –256, 2008.
[7] K. Cheung and H. So, “A multidimensional scaling framework for mo-
bile location using time-of-arrival measurements,” Signal Processing,
RMSE [m]
4 IEEE Transactions on, vol. 53, no. 2, pp. 460 – 470, feb 2005.
[8] K. W. Cheung, H. C. So, W.-K. Ma, and Y. T. Chan,
3
“A constrained least squares approach to mobile positioning:
algorithms and optimality,” EURASIP J. Appl. Signal Process.,
vol. 2006, pp. 150–150, January 2006. [Online]. Available:
2 http://dx.doi.org/10.1155/ASP/2006/20858
[9] Y. Chan and K. Ho, “A simple and efficient estimator for hyperbolic
1
location,” Signal Processing, IEEE Transactions on, vol. 42, no. 8, pp.
1905 –1915, aug 1994.
[10] R. Vaghefi, M. Gholami, and E. Strom, “Bearing-only target localiza-
0 tion with uncertainties in observer position,” in Personal, Indoor and
0 0.5 1 1.5 2 2.5 3 3.5 4
Standard deviation of measurement noise [m] Mobile Radio Communications Workshops (PIMRC Workshops), 2010
IEEE 21st International Symposium on, sept. 2010, pp. 238 –242.
[11] M. Spirito, “On the accuracy of cellular mobile station location
Fig. 10. Influence of lowering the expected training RMSe estimation,” Vehicular Technology, IEEE Transactions on, vol. 50,
no. 3, pp. 674 –685, may 2001.
7 [12] A. Shareef, Y. Zhu, and M. Musavi, “Localization
LS using neural networks in wireless sensor networks,” in
WLS
6 ASN−5T
Proceedings of the 1st international conference on MOBILe
CRLB Wireless MiddleWARE, Operating Systems, and Applications,
ser. MOBILWARE ’08. ICST, Brussels, Belgium, Belgium:
5 ICST (Institute for Computer Sciences, Social-Informatics and
Telecommunications Engineering), 2007, pp. 4:1–4:7. [Online].
Available: http://portal.acm.org/citation.cfm?id=1361492.1361497
RMSE [m]
4
[13] M. Rahman, Y. Park, and K.-D. Kim, “Localization of wireless
sensor network using artificial neural network,” in Communications
3 and Information Technology, 2009. ISCIT 2009. 9th International
Symposium on, sept. 2009, pp. 639 –642.
[14] S. Ergut, R. Rao, O. Dural, and Z. Sahinoglu, “Localization via tdoa
2
in a uwb sensor network using neural networks,” in Communications,
2008. ICC ’08. IEEE International Conference on, may 2008, pp. 2398
1 –2403.
[15] M. Heidari, N. Alsindi, and K. Pahlavan, “Udp identification and
error mitigation in toa-based indoor localization systems using neural
0
0 0.5 1 1.5 2 2.5 3 3.5 4 network architecture,” Wireless Communications, IEEE Transactions
Standard deviation of measurement noise [m] on, vol. 8, no. 7, pp. 3597 –3607, july 2009.
[16] S. Saarinen, R. Bramley, and G. Cybenko, “Neural networks, back-
Fig. 11. 5-center ASN model trained with noisy data propagation, and automatic differentiation,” in Automatic Differ-
entiation of Algorithms: Theory, Implementation, and Application,
A. Griewank and G. F. Corliss, Eds. Philadelphia, PA: SIAM, 1991,
pp. 31–42.
there is a limit in this improvement. If trained with noisy [17] S. Haykin, Neural Networks: A Comprehensive Foundation, 2nd ed.
Upper Saddle River, NJ, USA: Prentice Hall PTR, 1998.
data, the model can outperform the WLS method. In high
noise, the model performs better than the WLS method. On
the contrary, in low noise, the model fails to perform as well
as the WLS method.
R EFERENCES
[1] S. Kay, Fundamentals of statistical signal processing: estimation
theory. Prentice-Hall, 1993.
[2] K. Ho and Y. Chan, “Solution and performance analysis of geolocation
by tdoa,” Aerospace and Electronic Systems, IEEE Transactions on,
vol. 29, no. 4, pp. 1311 –1322, oct 1993.
[3] N. Patwari, I. Hero, A.O., M. Perkins, N. Correal, and R. O’Dea,
“Relative location estimation in wireless sensor networks,” Signal
Processing, IEEE Transactions on, vol. 51, no. 8, pp. 2137 – 2148,
aug. 2003.
[4] Y.-T. Chan, H. Yau Chin Hang, and P. chung Ching, “Exact and
approximate maximum likelihood localization algorithms,” Vehicular
Technology, IEEE Transactions on, vol. 55, no. 1, pp. 10 – 16, jan.
2006.
[5] A. M.-C. So and Y. Ye, “Theory of semidefinite programming for
sensor network localization,” in Proceedings of the sixteenth
annual ACM-SIAM symposium on Discrete algorithms, ser.
SODA ’05. Philadelphia, PA, USA: Society for Industrial and
3084
View publication stats