 Research
 Open access
 Published:
DQNbased resource allocation for NOMAMECaided multisource data stream
EURASIP Journal on Advances in Signal Processing volumeÂ 2023, ArticleÂ number:Â 44 (2023)
Abstract
This paper investigates a nonorthogonal multiple access (NOMA)aided mobile edge computing (MEC) network with multiple sources and one computing access point (CAP), in which NOMA technology is applied to transmit multisource data streams to CAP for computing. To measure the performance of the considered NOMAaided MEC network, we first design the system cost as a linear weighting function of energy consumption and delay under the NOMAaided MEC network. Moreover, we propose a deep Q network (DQN)based offloading strategy to minimize the system cost by jointly optimizing the offloading ratio and transmission power allocation. Finally, we design experiments to demonstrate the effectiveness of the proposed strategy. Specifically, the designed strategy can decrease the system cost by about 15% compared with local computing when the number of sources is 5.
1 Introduction
In the current society, with the advancement of wireless communication technology [1,2,3,4], the quantity of mobile device sources has skyrocketed, which results in the exponential growth of data to be handled [5,6,7]. However, the local computing capability of the device is often overwhelmed by the huge computing data stream, leading to the slow processing of data streams. To deal with this issue, the cloud server is used to assist devices in computing data stream [8,9,10], due to its advantages of much computing capacity compared to the local device. However, too many data streams are offloaded to the cloud, which may also bring a serious workload to the cloud server [11]. In addition, the wireless channel is vulnerable, which prolongs the communication delay and affects the systemâ€™s performance.
Based on the above local computing and cloud server problem, mobile edge computing (MEC) is designed to help to compute data stream [12,13,14]. In the MEC network, multisource data streams can be partially offloaded to the computing access point (CAP) to be computed [15]. Because the local device also has computing power, the local and cloud can perform data stream computing at the same time. Therefore, the offloading ratio of data streams becomes a key factor affecting the computing time. The authors in [16] presented an intelligent particle swarm optimization (PSO)based policy for MEC network unloading based on the cache mechanism, which employed the PSO algorithm to search for a suitable unloading ratio to achieve partial unloading. The PSO algorithm converged quickly and the algorithm was simple, but if the function had multiple local extrema, it was easily trapped in local extrema and cannot be got the optimal solution. The authors of [9, 17] studied a multiuser multiCAP MEC network with task offloading, where the network environment was timevarying, and the system cost was mainly determined by energy consumption and delay. Besides, a dynamic unloading policy based on DQN was devised. Users could dynamically adjust the unloading ratio to optimize the system cost to ensure the system performance of the MEC network.
Despite the above foundation, the MEC network with dynamical offloading still faces inherent limitations. Limited communication resources are difficult to support the orthogonal multiple access (OMA) of massive users. To deal with this issue, nonorthogonal multipleaccess (NOMA), which emerges as a new access technology, can help support the massive users. NOMA is a promising technology for reducing delay and energy consumption in MEC networks. The technology uses nonorthogonal transmission at the transmitter with the allocated source transmission power, introduces scrambling code information, and then removes the scrambling code information at the CAP through successive interference cancellation (SIC) to achieve correct demodulation [18]. Multiple sources share the same bandwidth to send the data stream simultaneously, and the CAP receives the transmitted information and then decodes it. This has a clear advantage in terms of increasing the transmission rate of the data stream [19, 20]. According to this principle, multiple sources can use the same bandwidth to offload data streams to the CAP simultaneously to decrease system energy consumption and delay.
So far, there has been a large number of investigations on the resource allocation of the NOMAMEC system. For example, the authors in [21] used reinforcement learning to optimize the computation and cache of the multiserver NOMAMEC system. The author of [22] studied a computing unloading system with the help of NOMA and dual connection (DC) and employed the deep learningbased intelligent unloading method to reduce the total system consumption. The authors in [23] designed a secure communication strategy for the NOMAassisted UAVMEC system for largescale access users. The author of [24] considered a NOMAMEC network, and jointly optimized the total system energy consumption through the convex theory and the iterative algorithm. Many current studies focus on optimizing the unloading ratio of the MEC network, but when NOMA technology is used to assist the unloading, the transmission power in the unloading stage also has a huge impact on the energy consumption and delay of the system.
On this basis, the author studied multisource MEC networks where CAPs are deployed at the edge. In the networks, computational data can be offloaded to adjacent CAPs through favorable data stream division and offloading to reach lowenergy consumption and low delay [5, 25, 26]. Meanwhile, an optimization strategy based on DQN is proposed for data stream offloading. A deep learningbased Qlearning algorithm integrates neural network techniques and value function approximation [27]. The neural network is trained by using target networks and empirical replay methods. The system cost is designed for a linear weighting function of delay and energy consumption. The offloading decision on MEC networks is modeled as the Markov decision process to employ reinforcement learning methods for resource allocation to improve network performance and reduce system cost [3, 7, 10]. Finally, the designed scheme is verified to be significantly superior through simulation experiments. The significant contributions of the paper are listed in the following:

We consider a NOMAaided MEC network with S sources and one CAP. Based on this, we propose a linear combination of energy consumption and delay in the system cost to measure the considered network performance. Meanwhile, the offload ratio and transmission power ratio are jointly optimized to lower the cost of the system.

We come up with a DQNbased data stream unloading optimization policy. In practical applications, the network environment is dynamic, which increases the optimization difficulty. Therefore, we use this strategy to dynamically obtain the allocation of the unloading ratio and transmission power ratio of the system.

We design experiments to compare different schemes, and the simulation experiment results indicate that the designed DQNbased strategy has a lower total system cost than other methods.
We have organized the rest of the paper as shown below. Following the introduction, we discuss the offloading model of the considered NOMAaided MEC network in Sect.Â and give the relevant calculation formula and model optimization formula of the system. After the discussion of the system model, the devised DQNbased method is shown in Sect.Â . SectionÂ presents the results of the simulation experiments. Finally, in Sect.Â , we conclude the whole work.
2 System model
As Fig.Â 1 shows, we explore the MEC network with S sources and one CAP, where the NOMA technology is applied to assist multisource data streams for transmission. Data streams are partially processed at local sources with limited computing capability and the other part of data streams are unloaded to CAP to be computed. Considering the performance of the MEC network, we use NOMA to help offload data streams to the CAP with sufficient computing capability through wireless links to accelerate computing. Concretely, the data stream sets of sources in the network are denoted by \(\left\{ D_s 1 \le s \le S \right\}\). Each source has different data streams \(D_s\) which has \(q_s\) number of bits and offloads a part of the data stream to the server to be computed. After calculating the offloaded data stream, the CAP returns the results to the source via dedicated feedback links. The following sections present the data stream offloading model, the local computing model, and the CAP computing model, respectively, specified in the following.
2.1 Data stream offloading model
In this part, we describe the data stream offloading model. When part of the data stream is offloaded to the CAP to be computed, multiple sources need to transmit the offloaded stream over the radio link using NOMA. The transmission rate of source \(D_s\) can be described as
where B is the bandwidth of wireless channel from \(D_s\) to CAP, \(P_s\) is the transmission power of source \(D_s\), and \(h_{s}^2\) is the channel gain of wireless channel from the source \(D_s\) to CAP. The symbol \(\sigma ^2\) stands for the noise of AWGN [28,29,30,31]. As previously stated, we assume \(P_{1}h_1^2 \le P_{2}h_2^2 \le \cdots \le P_{s}h_s^2\).
At each time slot, sources \(D_s\) have \(q_s\) bits of the data stream that need to be processed and offload a part of the data stream to CAP through the wireless link. The transmission delay of the offloaded data stream of sources \(D_s\) is [17, 32]
where \(\varvec{\beta _{s}} = [\beta _1,\beta _2,\ldots ,\beta _s]\) represents the percentage of source \(D_s\) to be unloaded to server which satisfies \(\beta _s \in [0,1]\). Since multisource data streams are unloaded in parallel, the total delay in the unloading phase is
In addition, the system energy consumption in the unloading phase can be obtained by
2.2 Local computing model
As mentioned above, some multisource data streams can be calculated locally. The local calculation delay of source \(D_s\) can express as
where \(f_s\) refers to local computing capability on source \(D_s\) and \(c_s\) represents the CPU cycle required for processing multisource data stream on source \(D_s\). The total local calculated time is
The total local calculated energy consumption is
2.3 CAP computing model
After part of the data stream on source \(D_s\) is successfully unloaded to the CAP through wireless links, the offloaded data stream will be computed at CAP. The computation delay at the CAP of source \(D_s\) is
where \(F_s\) denotes the computational power allocated to each source of the CAP. Different offloaded data streams are calculated in parallel at the CAP, so the total calculation delay at the CAP is
Meanwhile, the energy consumption produced at the CAP is
Since local operation and unloading can be performed at the same time, the total calculation delay of multisource data streams is
Moreover, the total energy consumption equation is
The total system energy consumption is the sum of energy consumption produced at each stage which is related to the delay at this stage. However, it is hard to measure the system behavior only by the total system energy consumption, because the total delay is the maximum of the local delay and the sum of delays produced at other stages. To comprehensively measure system performance, we devise the total system cost as a linear weighted function of total delay and total energy consumption [33], which can be described as
where \(\mu \in\) [0,1] is a weight factor. Notice that, when \(\mu\) = 0, the system cost consists directly of the system energy consumption, and we focus on the impact of system energy consumption on the considered system. When \(\mu\) = 1, the system cost consists only of the system delay, and we pay attention to the impact of delay on the considered MEC network. In the considered system, we can change the value of \(\mu\) to meet the requirements of different scenarios on energy consumption and time.
2.4 Problem formulation
As mentioned before, the system cost can be used to measure the systemâ€™s performance. To improve the system performance, the data stream needed to be processed with the minimum system cost. Therefore, we formulate this problem as minimizing the system cost by optimizing the unloading data stream and transmission power allocation, which can be expressed as
where \(C_1\) is a constraint on the unloading ratio, representing the limitation of the portion of the data stream unloaded to the CAP. Constraint \(C_2\) proposes the constraint of transmission power allocation, where \(P_{\max }^{s}\) represents the maximum transmission power. The \(\alpha _{s}\) is the transmission power allocation ratio between the sth source and the CAP. Due to the complexity of this problem, we employ DQN to solve this problem which is introduced in the next section. We summarize the notations mentioned in this section in TableÂ 1.
3 DQNbased offloading policy
The unloading policy determines the part of the data stream unloaded to the CAP and the transmission power allocation on the CAP, which significantly affects the performance of the system. In this section, we first formulate the optimization problem as a Markov decision process(MDP). Then we investigate a DQNbased optimization scheme to obtain the offload ratio and transmission power ratio to minimize the system cost.
In MDP, at time slot \(\tau\), the state of the environment is \(s_\tau\). The agents first obtain the action \(a_\tau\) from the policy \(\pi _\tau\) according to \(s_\tau\). Then the agents perform action \(a_\tau\) at the environment, resulting in the environment shift from \(s_\tau\) to \(s_{\tau +1}\) and obtaining the reward \(r_\tau\) from the environment. Concretely, the state space is
where \(\varvec{\beta }=\left\{ \beta _1,\beta _2,\beta _3,\ldots ,\beta _s \right\}\) is the unloading ratio of the multisource data stream, \(\varvec{\alpha }=\left\{ \alpha _1,\alpha _2,\alpha _3,\ldots ,\alpha _s \right\}\) is the transmission power allocation ratio at CAP. Besides, the action space is \({\varvec{A}}=\left\{ \delta _1,\delta _1^*,\delta _2,\delta _2^*,\delta _3,\delta _3^*,\ldots ,\delta _s,\delta _s^*, \varrho _1,\varrho _1^*,\varrho _2,\varrho _2^*,\varrho _3,\varrho _3^*,\ldots ,\varrho _s,\varrho _s^* \right\}\), where \(\delta _s=\theta\) and \(\delta _s^*=+\theta\) are the acts of adjusting the unloading ratio under the constraints \(C_1\), and \(\varrho _s=\theta\) and \(\varrho _s^*=+\theta\) are the actions to adapt the transmit power allocation ratio with the constraint \(C_2\). For minimizing the system cost of the considered NOMAaided MEC system, the reward is designed as
where \(\gamma _1> \gamma _2 >0\). Notice that, if the execution of \(a_\tau\) results in the reduction of system cost of the environment in time slot \(\tau +1\), the agents obtain a positive reward. On the contrary, the reward is negative. Moreover, the Q function which measures the performance of the action in the current environment is used in obtaining the best policy. It can be expressed as
As mentioned above, the optimization problem of the considered NOMAaided MEC network can be modeled as MDP. Therefore, we employ a deep learning method, DQN, to solve this problem.
As shown in Fig.Â 2, the DQN consists of two networks and one replay memory. The replay memory is used to store transition samples (\(s_\tau , a_\tau , r_\tau ,s_{\tau +1}\)). The evaluation network outputs the action \(a _ \tau \in {\varvec{A}}\) with the input state \(s _ \tau \in {\varvec{s}}\). When the replay memory has enough data, the evaluation network begins training and updates weights \(\vartheta\) every step. The target network, which helps train the evaluation network, is initialed as an evaluation network at the beginning and updated as the evaluation network at every certain step.
To avoid the network falling into local optimization, \(\varepsilon\)greedy strategy is used to help the agents to explore. It can be expressed as
where \(\vartheta\) is the weight of the evaluation network. We employ the temporal difference (TD) approach to support training DQN by defining the TDtarget obtained by the target network [33].
where \(\varphi\) is weight factor.
Moreover, we give the loss function [2, 34] based on TDtarget as
Based on the above, we summarize the DQNbased unloading policy in Algorithm 1
4 Results and discussion
This section demonstrates the advantage of the designed DQNbased unloading strategy in the considered NOMAaided MEC network through simulations. In the considered MEC network, with all channels experiencing Rayleigh flat fading [35,36,37]. We set the number of sources to 5 in our experiment. The size of multisource data streams is set to 14Â Mb, 3Â Mb, 16Â Mb, 8Â Mb, and 18Â Mb, and the computing power of each source is set to \(5 \times (10^7)\) cycle/s. Besides, the maximum transmission power and calculating power of each source are 1Â W and 1.5Â W. The computing power allocated to each source at the CAP is \(8 \times (10^7)\) cycle/s, and the calculating power is 1.5Â W. In addition, the total bandwidth is set to 6Â MHz. The detailed network parameter settings are shown in TableÂ 2.
FigureÂ 3 shows the reward of each episode during the training of the proposed DQNbased offloading policy, where \(\mu\) = 0.5 and the DQN has trained 300 episodes at all. From this figure, we can see that the reward grows rapidly in the previous 20 episodes. After about 50 episodes, the reward fluctuates slightly around 1950. It shows that our training converges, and helps to verify the efficacy of the designed DQNbased policy.
FigureÂ 4 plots system cost versus the number of training episodes under the three scenarios, where the number of sources is set to 5 and we train 150 episodes at all. For comparison, we also plot the cost of the other two scenarios. One is the local computation scenario, where each source data stream is computed locally; the other is the full offloading scenario, where data streams on each source are all offloaded to CAP for computing and the computing capability allocation on CAP is obtained by DQN. From this plot, we could notice that the system cost of the designed strategy declines sharply during the previous 100 episodes and converges at about 28.30 after 100 episodes. The system cost of the full offloading scenario decreases gradually during the previous 20 episodes and converges at about 31.6 after 20 episodes. On the contrary, the system cost of the local computation scenarios remains at 33.30 during the whole training. This result illustrates that the designed strategy has the best performance in reducing the system cost among the three proposed solutions. And the designed strategy can give a good unloading and resource allocation strategy for the considered NOMAaided MEC network.
FigureÂ 5 shows the relationship between the weighting factor \(\mu\) and the system cost \(\Theta\), where \(\mu\) varies from 0.1 to 0.9 and the number of sources is set to 5. We can see from Fig.Â 5 that the system cost obtained by the designed strategy is lower than that of the local computation and full offloading schemes with different values of \(\mu\). This indicates that the proposed scheme can significantly and efficiently improve the performance of the considered NOMAaided MEC network by reasonably allocating resources and unloading rate. Moreover, the total system cost of these three schemes decreases when \(\mu\) varies from 0.1 to 0.9. The main reason for the reduction of system cost is that the system cost is a linearly weighted function of time and energy consumption, where the impact of energy consumption is greater than the delay. The increase in \(\mu\) magnifies the impact of energy consumption, thus making the total system cost significantly lower.
FigureÂ 6 presents the influence of wireless bandwidth B on system cost, where \(S=5\) and the value of bandwidth varies from 5 to 9Â MHz. As shown in Fig.Â 6, the system cost of the designed strategy is lower than that of other schemes under the different values of bandwidth. This result indicates that our designed strategy outperforms other schemes under different communication environments. Moreover, the system cost of the designed scheme and the full offloading decrease, and the system cost of the local computation solutions stays at the same value as bandwidths increase. The reason is that the bigger bandwidth reduces the unloading cost of the designed policy and the full offloading scheme, while the data stream is calculated locally in the local computation scheme resulting in nonunloading cost.
FigureÂ 7 presents the impact on the system cost \(\Theta\) of the variation of the computing power allocated to each source of the CAP in three scenarios, where the number of sources is 5 and the computing power allocated to each source at the CAP changes from \(6 \times 10^7\) cycle/s to \(10 \times 10^7\) cycle/s. As shown in this figure, the system cost of the designed strategy and full offloading gradually decreases when the computing power allocated to each source at the CAP increases from \(6 \times 10^7\) cycle/s to \(10 \times 10^7\) cycle/s. The reason is that the CAP with more computational power can compute data streams faster, resulting in reduced system energy consumption and delay. Moreover, the system cost of the full offloading tends to decrease faster than that of the designed strategy. This is because the designed strategy is more robust in the various MEC environment and the full offloading solution is more impacted by the computational power at the CAP. This implies that the designed strategy obviously superior to the full offloading and local computation schemes.
FigureÂ 8 plots the system cost of the designed strategy versus the scale of the MEC network, where S changes from 2 to 6 and the local computing capability is set to \(5 \times 10^7\) cycle/s. From this plot, we can notice that the system cost of the designed strategy is always lower than that of local computing schemes and full offloading schemes as the value of S rises. It indicates that the designed strategy can improve the performance of considered NOMAaided MEC networks with different scales. Moreover, the system cost of the three schemes rises as the value of S varies from 2 to 6. This is due to the increasing number of sources leading to the computing data streams increasing for local and CAP. It leads to system delay grows.
5 Conclusion
This paper has investigated a NOMAaided MEC network with multisource and one CAP, in which multisource data streams were partially offloaded to CAP to accelerate computing. In the considered NOMAaided MEC network, we designed the system cost as the linear weighting function of energy consumption and delay produced in the unloading process. For reducing the cost of the considered network, we proposed the DQNbased offloading strategy for minimizing the system cost by optimizing the transmission power ratio and the offloading ratio during the unloading process. We compared the proposed DQNbased offloading strategy and other methods through experiments. The experimental results showed that the proposed method was more effective than other methods under different communication environments with various bandwidths. Moreover, in different NOMAaided MEC networks with different scales, different transmission capabilities, different bandwidth, or different computing capabilities, the proposed DQNbased offloading strategy is more robust than other methods with minimum system cost. Specifically, the designed strategy has been able to decrease the system cost by about \(15\%\) compared with local computing when the number of sources is 5.
Availability and data materials
The authors state the data available in this manuscript.
Abbreviations
 MEC:

Mobile edge computing
 NOMA:

Nonorthogonal multiple access
 DQN:

Deep Q network
 CAP:

Computing access point
 PSO:

Particle swarm optimization
 SIC:

Successive interference cancellation
 DC:

Dual connection
 UAV:

Unmanned aerial vehicle
 MDP:

Markov decision process
 TD:

Temporal difference
 AWGN:

Additive white Gaussian noise
References
W. Wu, F. Zhou, R.Q. Hu, B. Wang, Energyefficient resource allocation for secure nomaenabled mobile edge computing networks. IEEE Trans. Commun. 68(1), 493â€“505 (2020)
L. Chen, X. Lei, Relayassisted federated edge learning: performance analysis and system optimization. IEEE Trans. Commun. PP(99), 1â€“12 (2022)
R. Zhao, M. Tang, Profit maximization in cacheaided intelligent computing networks. Phys. Commun. PP(99), 1â€“10 (2022)
J. Ren, X. Lei, Z. Peng, X. Tang, O.A. Dobre, Risassisted cooperative NOMA with SWIPT. IEEE Wirel. Commun. Lett. (2023)
X. Liu, C. Sun, M. Zhou, C. Wu, B. Peng, P. Li, Reinforcement learningbased multislot doublethreshold spectrum sensing with Bayesian fusion for industrial big spectrum data. IEEE Trans. Ind. Inform. 17(5), 3391â€“3400 (2021)
Z. Na, B. Li, X. Liu, J. Wan, M. Zhang, Y. Liu, B. Mao, Uavbased widearea internet of things: An integrated deployment architecture. IEEE Netw. 35(5), 122â€“128 (2021)
W. Zhou, F. Zhou, Profit maximization for cacheenabled vehicular mobile edge computing networks. IEEE Trans. Veh. Technol. PP(99), 1â€“6 (2023)
W. Xu, Z. Yang, D.W.K. Ng, M. Levorato, Y.C. Eldar, M. Debbah, Edge learning for B5G networks with distributed signal processing: Semantic communication, edge computing, and wireless sensing. IEEE J. Sel. Top. Signal Process. arXiv:2206.00422 (2023)
X. Zheng, C. Gao, Intelligent computing for WPTMEC aided multisource data stream. to appear in EURASIP J. Adv. Signal Process. 2023(1) (2023)
S. Tang, L. Chen, Computational intelligence and deep learning for nextgeneration edgeenabled industrial IoT. IEEE Trans. Netw. Sci. Eng. 9(3), 105â€“117 (2022)
W. Wu, F. Zhou, B. Wang, Q. Wu, C. Dong, R.Q. Hu, Unmanned aerial vehicle swarmenabled edge computing: potentials, promising technologies, and challenges. IEEE Wirel. Commun. 29(4), 78â€“85 (2022)
W. Zhou, X. Lei, Priorityaware resource scheduling for uavmounted mobile edge computing networks. IEEE Trans. Veh. Technol. PP(99), 1â€“6 (2023)
L. Zhang, C. Gao, Deep reinforcement learning based IRSassisted mobile edge computing under physicallayer security. Phys. Commun. 55, 101896 (2022)
L. Chen, Physicallayer security on mobile edge computing for emerging cyber physical systems. Comput. Commun. 194(1), 180â€“188 (2022)
Y. Wu, C. Gao, Task offloading for vehicular edge computing with imperfect CSI: a deep reinforcement approach. Phys. Commun. 55, 101867 (2022)
W. Zhou, L. Chen, S. Tang, L. Lai, J. Xia, F. Zhou, L. Fan, Offloading strategy with PSO for mobile edge computing based on cache mechanism. Clust. Comput. 25(4), 2389â€“2401 (2022)
R. Zhao, C. Fan, J. Ou, D. Fan, J. Ou, M. Tang, Impact of direct links on intelligent reflect surfaceaided mec networks. Phys. Commun. 55, 101905 (2022)
Z. Ding, D.W.K. Ng, R. Schober, H.V. Poor, Delay minimization for NOMAMEC offloading. IEEE Signal Process. Lett. 25(12), 1875â€“1879 (2018)
X. Liu, Q. Sun, W. Lu, C. Wu, H. Ding, Bigdatabased intelligent spectrum sensing for heterogeneous spectrum communications in 5g. IEEE Wirel. Commun. 27(5), 67â€“73 (2020)
Z. Na, Y. Liu, J. Shi, C. Liu, Z. Gao, Uavsupported clustered NOMA for 6genabled internet of things: Trajectory planning and resource allocation. IEEE Internet Things J. 8(20), 15041â€“15048 (2021)
S. Li, B. Li, W. Zhao, Joint optimization of caching and computation in multiserver NOMAMEC system via reinforcement learning. IEEE Access 8, 112762â€“112771 (2020)
C. Li, H. Wang, R. Song, Intelligent offloading for nomaassisted MEC via dual connectivity. IEEE Internet Things J. 8(4), 2802â€“2813 (2021)
W. Lu, Y. Ding, Y. Gao, Y. Chen, N. Zhao, Z. Ding, A. Nallanathan, Secure nomabased UAVMEC network towards a flying eavesdropper. IEEE Trans. Commun. 70(5), 3364â€“3376 (2022)
L. Shi, Y. Ye, X. Chu, G. Lu, Computation energy efficiency maximization for a nomabased WPTMEC network. IEEE Internet Things J. 8(13), 10731â€“10744 (2021)
X. Liu, H. Ding, S. Hu, Uplink resource allocation for nomabased hybrid spectrum access in 6genabled cognitive internet of things. IEEE Internet Things J. 8(20), 15049â€“15058 (2021)
X. Liu, C. Sun, W. Yu, M. Zhou, Reinforcementlearningbased dynamic spectrum access for softwaredefined cognitive industrial internet of things. IEEE Trans. Ind. Inform. 18(6), 4244â€“4253 (2022)
B. Li, Z. Fei, J. Shen, X. Jiang, X. Zhong, Dynamic offloading for energy harvesting mobile edge computing: architecture, case studies, and future directions. IEEE Access 7, 79877â€“79886 (2019)
L. He, X. Tang, Learningbased MIMO detection with dynamic spatial modulation. IEEE Trans. Cogn. Commun. Netw PP(99), 1â€“12 (2023)
L. Zhang, S. Tang, Scoring Aided Federated Learning on Longtailed Data for Wireless IoMT based Healthcare System. IEEE J. Biomed. Health Inform. PP(99), 1â€“12 (2023)
J. Li, S. Dang, Y. Huang, Composite multiplemode orthogonal frequency division multiplexing with index modulation. IEEE Trans. Wirel. Commun. (2023)
S. Tang, X. Lei, Collaborative cacheaided relaying networks: performance evaluation and system optimization. IEEE J. Sel. Areas Commun. 41(3), 706â€“719 (2023)
J. Lu, M. Tang, Performance analysis for IRSassisted MEC networks with unit selection. Phys. Commun. 55, 101869 (2022)
C. Li, J. Xia, F. Liu, D. Li, L. Fan, G.K. Karagiannidis, A. Nallanathan, Dynamic offloading for multiuser muticap MEC networks: A deep reinforcement learning approach. IEEE Trans. Veh. Technol. 70(3), 2922â€“2927 (2021)
Y. Wu, C. Gao, Intelligent resource allocation scheme for cloudedgeend framework aided multisource data stream. EURASIP J. Adv. Signal Process. 2023(1) (2023, to appear)
W. Zhou, C. Li, M. Hua, Worstcase robust MIMO transmission based on subgradient projection. IEEE Commun. Lett. 25(1), 239â€“243 (2021)
J. Li, S. Dang, M. Wen, Index modulation multiple access for 6G communications: principles, applications, and challenges. IEEE Netw. (2023)
S. Tang, Dilated convolution based CSI feedback compression for massive MIMO systems. IEEE Trans. Veh. Technol. 71(5), 211â€“216 (2022)
Acknowledgements
None.
Funding
This work was supported by the KeyArea Research and Development Program of Guangdong Province, China (No. 2019B090904014), Science and Technology Projects in Guangzhou (No. 202102010412), and Yangcheng Scholars Research Project of Guangzhou (No. 202032832), and by Science and Technology Program of Guangzhou (No. 202201010047).
Author information
Authors and Affiliations
Contributions
JL designed the proposed framework and conducted the simulations, JX assisted in revising the manuscript for structure and grammar checking, VB helped to perfect the optimization method, F. Zhu assisted to optimize the design of the deep neural network, CG aided in conducting the simulations in this work, and SL helped to interpret the simulation results in this work. JX, FZ, and CG are the corresponding authors of this paper. All authors read and approved the final manuscript.
Corresponding authors
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing Interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ling, J., Xia, J., Zhu, F. et al. DQNbased resource allocation for NOMAMECaided multisource data stream. EURASIP J. Adv. Signal Process. 2023, 44 (2023). https://doi.org/10.1186/s13634023010052
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s13634023010052