 Research
 Open Access
 Published:
Optimal deployment of largescale wireless sensor networks based on graph clustering and matrix factorization
EURASIP Journal on Advances in Signal Processing volume 2023, Article number: 33 (2023)
Abstract
It is undeniable that there are a large number of redundant nodes in a wireless sensor network. These redundant nodes cause a colossal waste of resources and seriously threaten the life of the sensor network. In this paper, we provide a sensor nodes optimization selection algorithm based on a graph for a largescale wireless sensor network. Firstly, we propose a representationclustering joint algorithm based on Graph Neural Network to partition the largescale graph into several subgraphs. Then, we use SingularValueQR Decomposition for the node selection of each subgraph and achieve the optimal deployment for a largescale wireless sensor network. We conduct the experiments on the CIMIS dataset. The results show that the mean square error between the reconstructed network and the original network is as low as 0.02433. Meanwhile, we also compare our algorithm with the classical optimization algorithm. The results imply that the mean square error of the proposed algorithm is lower and the distribution is more uniform. Further, we verify the scalability of the algorithm for the optimal deployment of the largescale wireless sensor network.
1 Introduction
A largescale wireless sensor network (LWSN) is a wireless network composed of a mass of sensors and is used to sense, collect, process, and transmit the information of the monitoring area [1]. An LWSN is characterized by a large number, wide distribution, dense deployment, frequent communication, and big data traffic [2, 3], which leads to huge communication loss among nodes, high energy consumption, bandwidth resource waste, and data redundancy [4]. It seriously affects the network life and costs expensive maintenance. Therefore, it is very necessary to optimize the deployment of nodes for LWSN.
In recent years, there are much research on the optimal deployment of the wireless sensor network(WSN) with an intelligence optimization algorithm, such as the bat algorithm [5], the improved whale algorithm [6], the virtual spring force algorithm [7] and the improved grey wolf optimization algorithm [8], et al.. For the optimal deployment of LWSN, Zhang et al. proposed an optimization strategy of LWSN based on matrix completion. Representative nodes were selected by defining the data information of nodes, and matrix completion was used to further reduce redundancy [9]. Jiang et al. proposed a mean filtering algorithm based on graphical node data to reduce the energy consumption and the data redundancy of LWSN [10]. For a network with a large number of device nodes, Bin Cao et al. divided the original highdimensional multiobjective optimization problem into several lowdimensional multiobjective optimizations and used a groupbased multiobjective evolutionary algorithm to close some nodes for saving resources [11].
As a common nonEuclidean data structure, the graph has the characteristics of diversity, irregularity, and large scale, which can efficiently and conveniently represent LWSN. In this paper, we proposed an optimal deployment algorithm in LWSN based on graph clustering and matrix decomposition to further optimize the sensor network randomly deployed in the monitoring area. Our contributions can be summarized as follows:
(1)We improve Deep Embedding Clustering with Graph Embedded Representation. It can learn the node’s embedded representation and the graph clustering jointly to enhance the learning of embedding and clustering at the same time.
(2)We adopt SingularValueQR Decomposition to select a subset of optimized nodes on each subnetwork, which achieves the optimal deployment and retains the crucial information of the LWSN.
The rest of this paper is organized as follows: Section 2 introduces the related work on the optimal deployment of sensor networks in recent years. Section 3 systematically expounds on our proposed theory and algorithm. Section 4 describes the experimental verification. Section 5 summarizes the work of this paper and suggests future research directions.
2 Related work
Some scholars use graphs as the medium and apply graph signal processing theory to the optimal deployment of sensor networks. Akie et al. proposed a deterministic graph signal sampling method, which used local operators implemented by Chebyshev polynomials to limit the coverage and select the optimal vertex subset based on the largest information [12]. On the basis of Akie’s research, Huan Xu considered the maximum information and residual energy of nodes and proposed a sampling set selection based on energyaware to achieve the balance of network optimization and energy constraints [13]. Paolo et al. introduced a probability sampling mechanism based on the adaptive least mean square algorithm so that each node of the graph was sampled with a given probability at each moment to obtain the optimal nodes [14]. Diego applied compressive sensing theory to graph signals and proposed an algorithm with local sampling and compressive sensing. This method calculated the signal coefficients random linear combination of signal coefficients between the node and its neighborhood and selected the optimal subset by comparing the coefficients [15].
These node optimization deployment methods are analyzed and processed on the entire graph. However, with the continuous increase in the number of sensor nodes, the execution efficiency of the algorithm also decreases sharply.
Some research is devoted to identifying crucial nodes in the graph based on the Graph Neural Network (GNN). Zhao G et al. proposed a method that transforms the crucial node identification in the graph into node classification, learns the representation of nodes through a graph convolution network, and then classifies influential nodes and nodes with less influence with a multilayer perception [16]. Ma M et al. proposed a crucial node identification model based on graph attention networks and reinforcement learning, where a graph attention network is used to obtain the embedded representation of each node, and combined with reinforcement learning, the node embedding is mapped to the corresponding node quality score. Then the ranking of crucial nodes is obtained [17].
The above methods focus on the supervised learning paradigm and use abundant groundtruth labels to train specific downstream tasks, which is hard to apply to realworld scenarios (realworld scenarios are usually unlabeled). More importantly, these methods are prone to fall into the situation of overfitting and weak robustness [18].
3 Theory and algorithm
The proposed algorithm is shown in Fig. 1. First, we introduced the NonNegative Kernel (NNK) [19] Regressive to construct the LWSN into a sparse graph topology driven by the measurement data. It fuses two classical graph construction methods, similaritybased and localityinducing [20], by the kernel method, where similaritybased methods rely on a kernel metric to measure the similarity among sensor nodes, and localityinducing methods depend on a regression objective to obtain the edge weights. And then we partition the largescale graph topology into several subgraphs through an endtoend representationclustering joint model based on GNN. Lastly, node selection is performed on each subgraph to achieve the optimal deployment of the LWSN based on singularvalueQR decomposition.
3.1 Subnetwork division
In the paper, we proposed the improved graph deep embedding clustering, an endtoend representationclustering joint algorithm, to partition the largescale graph into several subgraphs. In the improved graph deep embedding clustering, we propose the graph auto encoder coupled with Graph Attention Network(GAT) to learn the embedded representation of sensor nodes
where \(\varvec{X} = [{\varvec{x}}_{1}, {\varvec{x}}_{2}, \cdots , {\varvec{x}}_{i}, \cdots , {\varvec{x}}_{N}]^\text{T} \in \textbf{R}^{N \times d_\mathrm{{in}}}\) is the measurement matrix, where each row is the measurement values \({\varvec{x}} \in \textbf{R}^{d_\mathrm{{in}}}\) on a sensor node, and \(d_\mathrm{{in}}\) is the dimension of the measurement; \(\varvec{H} = [{\varvec{h}}_{1}, {\varvec{h}}_{2}, \cdots , {\varvec{h}}_{i}, \cdots , {\varvec{h}}_{N}]^\text{T} \in \textbf{R}^{N \times d_\mathrm{{out}}}\) is the representation of all sensor nodes, where each row \({\varvec{h}} \in \textbf{R}^{d_\mathrm{{out}}}\) is the embedded representation of a sensor node, and \(d_\mathrm{{out}}\) is the dimension of the embedded representation. Specifically, the embedded representation of the sensor node i at the lth layer is denoted as
where \(\sigma\) is an activation function; \(\mathcal {N}\) is the neighborhoods of sensor node i; \(W^{(l)}\) is the weight matrix; \(\alpha ^{(l)}_{ij}\) is the attention coefficient between sensor nodes i and j
where \({\varvec{a}}^{(l)}\) is a parameter vector; \(\Vert\) is the concatenation operation; \(\mathrm LReLU\) is Leaky Rectified Linear Unit, a nonsaturating activation function.
The decoder is used to reconstruct the adjacency matrix \(\hat{\varvec{A}}\)
The reconstruction loss function \(\mathcal {L}_{r}\) is defined as Binary Cross Entropy (BCE)
where \(A_{ij}\) is an element of the adjacency matrix; \(\hat{A}_{ij}\) is an element of the reconstruction adjacency matrix.
While obtaining a suitable node representation vector, we propose a Deep Embedding Clustering (DEC) algorithm for optimal sensor network division. First, we employ Kmeans on the initial node’s embedded representations \({\varvec{H}}^{(0)}\) to get R initial cluster centers \(\varvec{\mu } = [{\varvec{\mu }}_{1}, {\varvec{\mu }}_{2}, \cdots , {\varvec{\mu }}_{r}, \cdots , {\varvec{\mu }}_{R}]\). Then, the t distribution is used to measure the similarity between the node’s embedded representation vector \({\varvec{h}}_{i}\) and the subnetwork cluster center representation vector \({\varvec{\mu }}_{r}\). So the soft cluster assignment Q of sensor nodes is obtained
where \(q_{ir}\) can be seen as the probability of assigning node i to cluster r, if the probability is higher, the confidence of the assignment is higher. In other words, the sensor node is closer to the cluster center.
The auxiliary target distribution P is constructed based on Q distribution to improve cluster effect. First, we raise the assignment probability \(q_{ir}\) to the second power to emphasize the role of highconfidence nodes and reduce the influence of lowconfidence nodes. And then normalizing by frequency per cluster to prevent cluster degradation (all nodes belong to the same cluster)
where \(\sum \limits _{i} q_{ir}\) is soft cluster frequency and is the sum of the probabilities that the nodes belong to the rth cluster.
With the help of an auxiliary target distribution P, highconfidence assignments are used as soft labels to supervise the learning of Q to refine the clusters iteratively. To be specific, the training Q distribution is fitted to the P distribution by minimizing the Kullback–Leibler (KL) divergence between the P and Q
For a better subnetwork division, we joint the optimizations of representation learning and cluster assignment and define the total loss with reconstruction loss and clustering loss for optimization
where \(\gamma\) is the clustering coefficient for the balance of reconstruction loss and clustering loss.
In the following training, three parameters are needed to optimize and update: cluster centers \(\varvec{\mu }\), encoder’s weights W and auxiliary target distribution P.
First, fixing the auxiliary target distribution P, we use Stochastic Gradient Descent (SGD) to optimize the cluster centers and encoder weights jointly. The gradients of the clustering loss \(\mathcal {L}_{c}\) based on the embedded representation \({\varvec{h}}_{i}\) of sensor node i and cluster center representation \({\varvec{\mu }}_{r}\) can be computed as [21]:
Then the cluster center \({\varvec{\mu }}_{r}\) is updated by
where \(\lambda\) is the learning rate.
According to the chain rule, the gradient of the encoder weights can be computed as
Then the encoder weights W are updated by
While the auxiliary target distribution acts as a “groundtruth” label during training, it also depends on the current soft assignment distribution. In iterative training, the constant change of the target will hinder learning and convergence. Therefore, the auxiliary target distribution P will not be updated at every iteration but only updated at every T iteration by Eq.(7) to enhance the stability of the clustering process.
After the above learning and training, the cluster assignment probability vector tends to be onehot. Therefore, according to the last learned Q distribution, the cluster assignment label (i.e. the cluster assignment result of each sensor node i) of each sensor node can be obtained
We summarize the proposed improved graph deep embedding clustering in Algorithm 1.
3.2 Node optimization selection
For the subnetwork r, we construct the corresponding measurement matrix and adjacency matrix with \({\varvec{X}}_{r} ={\varvec{X}}\big _{{\varvec{y}} = r,:}\) and \({\varvec{A}}_{r} ={\varvec{A}}\big _{{\varvec{y}} = r,{\varvec{y}} = r}\), where \(r \in \varvec{y}\), \(\varvec{y}\) is the clustering results in Eq.(15), and perform Singular Value Decomposition(SVD) on \({\varvec{X}}_{r}\)
where \({\varvec{U}}_{r} \in \textbf{R}^{N_{r} \times N_{r}}\) is the left singular vector matrix of \({\varvec{X}}_{r}\), and \(N_{r}\) is the number of subnetwork nodes; \({\varvec{V}}_{r} \in \textbf{R}^{d_\text{in} \times d_\text{in}}\) is the right singular vector matrix of \({\varvec{X}}_{r}\); \({\varvec{\Sigma }}_{r}\) is the singular value diagonal matrix.
Then we partition \({\varvec{U}}_{r}\) as follows
where \({\varvec{U}}_{r1} \in \textbf{R}^{N_{r} \times M_{r}}\), \({\varvec{U}}_{r2} \in \textbf{R}^{N_{r} \times (N_{r}  M_{r})}\), \(M_{r}\) is the number of nodes after optimization selection.
Next, performing column pivoting QR decomposition on \({\varvec{U}}_{r1}\) and obtaining a permutation matrix \({\varvec{P}}_{r}\)
where \({\varvec{Q}}_{r}\) is a unitary matrix and \({\varvec{R}}_{r}\) is an upper triangular matrix.
The rows with the number 1 in the permutation matrix \({\varvec{P}}_{r}\) correspond to the \(M_{r}\) most important sensor nodes. According to the permutation matrix \({\varvec{P}}_{r}\), we can obtain the optimized measurement matrix \({\varvec{X}}_{r}^{'}\) and adjacency matrix \({\varvec{A}}_{r}^{'}\)
Obviously, the SVDQR method can select a set of independent nodes for each subnetwork to minimize the least square residual and achieve the optimal deployment of the largescale sensor network. Specifically, SVD can determine the principal components of the subnetwork by calculating the singular values of the measurement matrix. Based on the principal component, QR decomposition can distinguish the importance of each node, and choose to keep or discard it. Finally, the optimized measurement matrices and adjacency matrices of all subgraphs are concatenated together, and the optimization deployment of the LWSN is completed
where \(\big \Vert\) is the concatenation operation; R is the number of subnetwork.
According to the description of the above parts, the proposed optimal deployment algorithm for LWSN in the paper is shown in Algorithm 2.
3.3 Complexity analysis
Due to the large number of nodes in LWSN, many optimal deployment algorithms have high computational complexity and low execution efficiency. The algorithm proposed in this paper mainly focuses on two parts: improved graph deep embedding clustering and node optimization selection based on the SVDQR. The computational complexity of these two parts is significantly improved with the NNK graph. Therefore, this paper analyzes the computational complexity.
The graph auto encoder coupled with the GAT for coding learning. First, it needs to calculate the feature maps of all sensor nodes. The lth layer of the GAT needs to map the dimension of the representation vector \({\varvec{h}}_{i}^{(l1)}\) from \(d^{(l1)}\) to \(d^{(l)}\), and its computational complexity is \(O( Nd^{(l1)}d^{(l)})\). Therefore, the computational complexity of the feature mapping based on the GAT is \(O( Nd^{2}_\text{max})\), where N is the number of nodes, and \(d_\text{max} = \max \limits _{l} d^{(l)}\) is the maximum dimension of the hidden layers. Secondly, the attention coefficient needs to be calculated in the lth layer, and each attention coefficient needs to map a \(2d^{(l)}\) dimensional vector into a real number, and its computational complexity is \(O(d^{(l)})\). For the whole graph, the attention coefficient between each sensor node and its neighbor nodes needs to be calculated, the number of calculations is the same as the number of edges, so the computational complexity of the attention coefficient is \(O(Ed_\text{max})\), where E is the number of edges. Only weighted summation is involved in the next step of the GAT, and no complex multiplication operations are involved, so the computational complexity of the node representation part is \(O(Nd^{2}_\text{max} + Ed_\text{max})\).
In the deep embedding clustering step, we assign nodes to the clusters according to the probability distribution Q of the cluster centers by minimizing the KL divergence of the Q distribution and the P distribution. So its computational complexity is \(O(NRd_\text{out})\), where \(d_\text{out}\) is the embedded dimension of the representation vector, and R is the number of clusters. Therefore, the whole computational complexity in the sensor subnetwork division is \(O(Nd^{2}_\text{max} + Ed_\text{max} + NRd_\text{out})\).
For each subnetwork, the computational complexity of SVD is \(O(N_{r}^{3})\), and the computational complexity of QR decomposition is \(O(2N_{r}M_{r}^{2})\), where \(N_{r}\) is the number of nodes in the subnetwork, and \(M_{r}\) is the number of optimized nodes. Since \(M_{r} \ll N_{r}\), the computational complexity of optimizing selection for one subnetwork is \(O(N_{r}^{3})\), and the computational complexity of optimizing selection for R subnetworks is \(O(\displaystyle \sum \limits _{r=1}^{R}N_{r}^{3})\).
According to the previous analysis, the computational complexity of subnetwork division is \(O(Nd^{2}_\text{max} + Ed_\text{max} + NRd_\text{out})\). For a connected graph topology, the number of edges E is often much larger than the number of nodes N, so the edges number E has a significant influence on the computational complexity. And the edges number E in a sparse NNK graph does not exceed \(N\log N\), which effectively controls computational complexity. In addition, compared with directly performing matrix decomposition on the entire sensor network, dividing the sensor network into several subnetworks and separately performing matrix decomposition can reduce the computational complexity effectively. And according to the number of subnetworks, there are \(O(N) \leqslant O(\displaystyle \sum \limits _{r=1}^{R}N_{r}^{3}) \leqslant O(N^{3})\), where R is the subnetworks number.
4 Experiments
4.1 Datasets
We use the California Irrigation Management Information System(CIMIS)^{Footnote 1} dataset for the algorithm testing in the paper. The CIMIS composed of more than 140 active meteorological stations, and each station integrates seven sensor types of solar radiation, air temperature, soil temperature, relative humidity, wind speed, wind direction, and precipitation. The air temperature sensors were selected to build a WSN which contains 144 active nodes and the temperature values in one day are shown in Fig. 2.
4.2 Parameter setting
In the graph construction section, we set the maximum number of neighbors to 10 for each node and the weight zero threshold to \(3 \times 10^{3}\). In the clustering section, we divide the whole WSN into five subnetworks, and set the hidden dimension to 3, the embedded dimension to 5, and the learning rate to 0.0001, the weight decay to \(5 \times 10^{4}\). In order to improve the stability of clustering, the target distribution is updated every 5 iterations.
4.3 Evaluation metric
In order to compare the uniformity of the optimized deployment of the sensor network, we define the OptimizationOriginal Ratio(OOR) and distribution variance. OOR was the ratio of M, the number of optimally selected nodes, to N, the number of nodes in the original sensor network
The distribution variance was the OOR variance of each subnetwork, which was used to measure the uniformity of node distribution after optimization. Obviously, the smaller the distribution variance is, the more uniform of the optimally selected nodes are
where \(OOR_{r}\) is the OOR of subnetwork r, \(\overline{OOR}\) is the average OOR of each subnetwork; R is the number of subnetworks.
4.4 Experimental results and analysis
4.4.1 Basic experiment
We construct a sparse graph with 595 edges for the CIMIS dataset by the NNK algorithm (Fig. 3). The improved graph deep embedding clustering algorithm divides the sensor nodes into five clusters (Fig. 4).
For optimization selection, the optimal number of nodes starts from 5 and increases by 10. The relationship between the Mean Square Error (MSE) values of the WSN and the optimal number of nodes is shown in Fig. 5. To maximize the network life and reduce the maintenance cost, we need to reduce the scale of the WSN as much as possible within the acceptable range of data reconstruction precision (i.e. \(\text{MSE} < 0.1\)). Therefore, according to Fig. 5, we choose the optimal number as 15, and the MSE of the WSN is only 0.02433.
4.4.2 Comparative experiment
We have compared the proposed algorithm with five classic WSN optimization deployment algorithms, namely MinPinv [22], MaxVol [22], MinSpec [23], Entropy [24], and MI [25]. Figure 6 showed the Mean Square Error of the optimization algorithms. It can be seen that the proposed algorithm has a very obvious improvement in the optimal selection of sensor nodes compared with other algorithms.
When the optimal number is 15, we compared the distribution of optimal nodes and the distribution variance for different optimization algorithms, as shown in Figs. 7 and 8. The nodes selected by optimization are marked with color, and the remaining nodes are marked with gray.
It can be easily seen from Figs. 7 and 8 that the distribution of the optimal nodes with the proposed algorithm is more uniform and comprehensive than others, and its distribution variance is as low as \(2.68 \times 10^{4}\).
4.4.3 Scalability experiment
In order to illustrate the scalability of the proposed algorithm in this paper and its applicability to LWSN, we build several simulation datasets, which are as follows:
(1) Gaussian Simulation Dataset (GSD): In a unit area, sensor nodes are randomly deployed and the node signal values obey the N(0, 1) Gaussian distribution, where the network scales are 500, 1000, 2000, 5000 and, 10000 respectively.
(2) Uniform Simulation Dataset (USD): In a unit area, sensor nodes are randomly deployed and the signal values obey the uniform distribution of U(0, 1), where the network scales are 500, 1000, 2000, 5000 and, 10000 respectively.
The proposed algorithm is applied to the above simulation dataset, and the experimental results are shown in Fig. 9. The bar graph shows the relationship between the optimal number of nodes and the network size, and the line graph corresponds to the mean square error.
The results show that the proposed algorithm can significantly reduce the number of nodes required for LWSN without losing crucial information. It reflects that the proposed algorithm has good adaptability and scalability for LWSN.
5 Conclusion
In this paper, we innovatively built a largescale graph topology division algorithm based on deep embedding clustering and an optimal selection mechanism for sensor nodes based on SVDQR for the optimal deployment of LWSN. The experimental results verify the good applicability of the proposed algorithm. In conclusion, the proposed algorithm has the following advantages:

1.
The problem of node optimal deployment was investigated in subgraphs, which improved the algorithm scalability

2.
We applied GNN for graph representation and graph clustering in the same framework, which enhanced the adaptability of sensor node representation and clustering

3.
It made full use of the attribute information and structure information of the graph to carry out endtoend learning at the same time, which strengthened the learning quality of the GNN model

4.
Each step of the proposed algorithm could effectively improve the computational efficiency of subsequent steps, which greatly reduces the whole computational complexity. In the future, we will pay more attention to the optimal deployment of heterogeneous largescale sensor networks based on our proposed algorithm, and improve the execution efficiency of the proposed algorithm.
Availability of data and materials
The real dataset can be obtained from the link given in the footnote of the text, and simulation datasets can be generated as described in the paper.
Code availability
Not Applicable.
Abbreviations
 LWSN:

Largescale wireless sensor network
 WSN:

Wireless sensor network
 GNN:

Graph neural network
 NNK:

Nonnegative kernel
 GAT:

Graph attention network
 BCE:

Binary cross entropy
 DEC:

Deep embedding clustering
 KL:

Kullback–Leibler
 SGD::

Stochastic gradient descent
 SVD:

Singular value decomposition
 SVDQR:

SingularvalueQR decomposition
 CIMIS:

California irrigation management information system
 OOR:

Optimizationoriginal ratio
 MSE:

Mean square error
 GSD:

Gaussian simulation dataset
 USD:

Uniform simulation dataset
References
V. Patil, S. Deshpande, Design of fpga soft core based wsn node using customization paradigm. Wireless Pers. Commun. 122(1), 783–805 (2022). https://doi.org/10.1007/s1127702108925y
Y. Sangar, B. Krishnaswamy, WiChronos: energyefficient modulation for longrange, largescale wireless networks. Assoc. Comput. Machinery 10(11453372224), 3380898 (2020)
O.A. Khashan, R. Ahmad, N.M. Khafajah, An automated lightweight encryption scheme for secure and energyefficient communication in wireless sensor networks. Ad Hoc. Netw. 115, 102448 (2021). https://doi.org/10.1016/j.adhoc.2021.102448
G. Sahar, K.B.A. Bakar, F.T. Zuhra, S. Rahim, T. Bibi, S.H.H. Madni, Data redundancy reduction for energyefficiency in wireless sensor networks: a comprehensive review. IEEE Access 9, 157859–157888 (2021). https://doi.org/10.1109/ACCESS.2021.3128353
S.S. Mohar, S. Goyal, R. Kaur, Optimized sensor nodes deployment in wireless sensor network using bat algorithm. Wireless Pers. Commun. 116(4), 2835–2853 (2021). https://doi.org/10.1007/s1127702007823z
M. Toloueiashtian, M. Golsorkhtabaramiri, S.Y.B. Rad, An improved whale optimization algorithm solving the point coverage problem in wireless sensor networks. Telecommun. Syst. 79(3), 417–436 (2022). https://doi.org/10.2298/CSIS180103023W
X. Deng, Z. Yu, R. Tang, X. Qian, K. Yuan, S. Liu, An optimized node deployment solution based on a virtual spring force algorithm for wireless sensor network applications. Sensors 19(8), 1817 (2019). https://doi.org/10.3390/s19081817
K. Hegde, R. Dilli, Wireless sensor networks: network life time enhancement using an improved grey wolf optimization algorithm. Eng. Sci. (2022). https://doi.org/10.30919/es8d717
Z. Xiaohan, Y. Changchuan, W. Huarui, Energy optimization strategy for wireless sensor networks in largescale farmland habitat monitoring. Smart Agricul. 1(2), 55 (2019). https://doi.org/10.12133/j.smartag.2019.1.2.201812SA024
P. Jiang, Y. Li, F. Wu, S. Yu, H. Xu, Research on energy consumption and eliminate of large data redundancy method for large scale wireless sensor networks based on mean filter. Adv. Eng. Sci. 49(2), 145–151 (2017). https://doi.org/10.15961/j.jsuese.201601189
B. Cao, Q. Wei, Z. Lv, J. Zhao, A.K. Singh, Manyobjective deployment optimization of edge devices for 5G networks. IEEE Trans. Net. Sci. Eng. 7(4), 2117–2125 (2020). https://doi.org/10.1109/TNSE.2020.3008381
A. Sakiyama, Y. Tanaka, T. Tanaka, A. Ortega, Eigendecompositionfree sampling set selection for graph signals. IEEE Trans. Signal Process. 67(10), 2679–2692 (2019). https://doi.org/10.1109/TSP.2019.2908129
H. Xu, G. Li, G. Zhang (2019) Energyaware sampling set selection for signal reconstruction in internet of things. In: 2019 IEEE 5th international conference on computer and communications (ICCC), pp. 2146–2151. https://doi.org/10.1109/ICCC47050.2019.9064079. IEEE
P. Di Lorenzo, P. Banelli, E. Isufi, S. Barbarossa, G. Leus, Adaptive graph signal processing: Algorithms and optimal sampling strategies. IEEE Trans. Signal Process. 66(13), 3584–3598 (2018). https://doi.org/10.1109/TSP.2018.2835384
D. Valsesia, G. Fracastoro, E. Magli, Sampling of graph signals via randomized local aggregations. IEEE Trans. Signal Inform. Proc. Over Net. 5(2), 348–359 (2018). https://doi.org/10.1109/TSIPN.2018.2869354
G. Zhao, P. Jia, A. Zhou, B. Zhang, Infgcn: identifying influential nodes in complex networks with graph convolutional networks. Neurocomputing 414, 18–26 (2020). https://doi.org/10.1016/j.neucom.2020.07.028
M. Ma, X. Wang, Y. Li, Q. Zhang, C. Wang. Critical node identification based on attention flow networks. in: 2021 International conference on machine learning and intelligent systems engineering (MLISE), pp. 96–102 (2021). https://doi.org/10.1109/MLISE54096.2021.00025. IEEE
X. Liu, F. Zhang, Z. Hou, L. Mian, Z. Wang, J. Zhang, J. Tang, Selfsupervised learning: generative or contrastive. IEEE Trans. Knowl. Data Eng. 35(1), 857–876 (2021). https://doi.org/10.1109/TKDE.2021.3090866
S. Shekkizhar, A. Ortega. Graph construction from data by nonnegative kernel regression. In: ICASSP 20202020 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp. 3892–3896 (2020). https://doi.org/10.1109/ICASSP40776.2020.9054425. IEEE
L. Qiao, L. Zhang, S. Chen, D. Shen, Datadriven graph construction and graph learning: a review. Neurocomputing 312, 336–351 (2018). https://doi.org/10.1016/j.neucom.2018.05.084
J. Xie, R. Girshick, A. Farhadi. Unsupervised deep embedding for clustering analysis. In: international conference on machine learning, pp. 478–487 (2016). PMLR
M. Tsitsvero, S. Barbarossa, P. Di Lorenzo, Signals on graphs: uncertainty principle and sampling. IEEE Trans. Signal Process. 64(18), 4845–4860 (2016). https://doi.org/10.1109/TSP.2016.2573748
S. Chen, R. Varma, A. Sandryhaila, J. Kovačević, Discrete signal processing on graphs: sampling theory. IEEE Trans. Signal Process. 63(24), 6510–6523 (2015). https://doi.org/10.1109/TSP.2015.2469645
M.C. Shewry, H.P. Wynn, Maximum entropy sampling. J. Appl. Stat. 14(2), 165–170 (1987). https://doi.org/10.1080/02664768700000020
A. Krause, A. Singh, C. Guestrin. Nearoptimal sensor placements in gaussian processes: Theory, efficient algorithms and empirical studies. J. Machine Learn. Res. 9(2) (2008). https://doi.org/10.1145/1390681.1390689
Acknowledgements
This work is supported by National Natural Science Foundation of China (No.61731006, 61971310).
Funding
This work is supported by National Natural Science Foundation of China (No.61731006, 61971310).
Author information
Authors and Affiliations
Contributions
WW, as the corresponding author, provides research ideas, oversight, and leadership responsibility for the research activity planning and execution, including mentorship external to the core team. HG provides algorithm design and computer code implementation. QZ analyzes and synthesizes data. All authors read and approved by the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not Applicable.
Human and animal ethics
Not Applicable.
Consent for publication
There is the consent of all authors.
Conmpeting interests
The authors declare that they have no confict of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Gao, H., Zhu, Q. & Wang, W. Optimal deployment of largescale wireless sensor networks based on graph clustering and matrix factorization. EURASIP J. Adv. Signal Process. 2023, 33 (2023). https://doi.org/10.1186/s13634023009953
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13634023009953
Keywords
 Largescale wireless sensor network
 Deep embedding clustering
 Graph attention auto encoder
 SingularvalueQR decomposition