Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
Multi-task learning provides plenty of room for performance improvement to single-task learning, when learned tasks are related and learned with mutual information. In this work, we analyze the efficiency of using a single-task reinforcement learning algorithm to mitigate jamming attacks with freque...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9528307/ |
id |
doaj-1b8440442b66493bb5e62fb7d3014052 |
---|---|
record_format |
Article |
spelling |
doaj-1b8440442b66493bb5e62fb7d30140522021-09-13T23:00:39ZengIEEEIEEE Access2169-35362021-01-01912319412320710.1109/ACCESS.2021.31098569528307Deep Multi-Task Conditional and Sequential Learning for Anti-JammingRobert Basomingera0https://orcid.org/0000-0003-4204-3554Young-June Choi1https://orcid.org/0000-0003-2240-0892Department of Computer Engineering, Ajou University, Suwon, South KoreaDepartment of Software and Computer Engineering, Ajou University, Suwon, South KoreaMulti-task learning provides plenty of room for performance improvement to single-task learning, when learned tasks are related and learned with mutual information. In this work, we analyze the efficiency of using a single-task reinforcement learning algorithm to mitigate jamming attacks with frequency hopping strategy. Our findings show that single-task learning implementations do not always guarantee optimal cumulative reward when some jammer’s parameters are unknown, notably the jamming time-slot length in this case. Therefore, to maximize packet transmission in the presence of a jammer whose parameters are unknown, we propose deep multi-task conditional and sequential learning (DMCSL), a multi-task learning algorithm that builds a transition policy to optimize conditional and sequential tasks. For the anti-jamming system, the proposed model learns two tasks: sensing time and transmission channel selection. DMCSL is a composite of the state-of-the-art reinforcement learning algorithms, multi-armed bandit and an extended deep-Q-network. To improve the chance of convergence and optimal cumulative reward of the algorithm, we also propose a continuous action-space update algorithm for sensing time action-space. The simulation results show that DMCSL guarantees better performance than single-task learning by relying on a logarithmically increased action-space sample. Against a random dynamic jamming time-slot, DMCSL achieves about three times better cumulative reward, and against a periodic dynamic jamming time-slot, it improves by 10% the cumulative reward.https://ieeexplore.ieee.org/document/9528307/Ad hoc networkdeep learningjamming attackmulti-armed banditspectrum sensing |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Robert Basomingera Young-June Choi |
spellingShingle |
Robert Basomingera Young-June Choi Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming IEEE Access Ad hoc network deep learning jamming attack multi-armed bandit spectrum sensing |
author_facet |
Robert Basomingera Young-June Choi |
author_sort |
Robert Basomingera |
title |
Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming |
title_short |
Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming |
title_full |
Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming |
title_fullStr |
Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming |
title_full_unstemmed |
Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming |
title_sort |
deep multi-task conditional and sequential learning for anti-jamming |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2021-01-01 |
description |
Multi-task learning provides plenty of room for performance improvement to single-task learning, when learned tasks are related and learned with mutual information. In this work, we analyze the efficiency of using a single-task reinforcement learning algorithm to mitigate jamming attacks with frequency hopping strategy. Our findings show that single-task learning implementations do not always guarantee optimal cumulative reward when some jammer’s parameters are unknown, notably the jamming time-slot length in this case. Therefore, to maximize packet transmission in the presence of a jammer whose parameters are unknown, we propose deep multi-task conditional and sequential learning (DMCSL), a multi-task learning algorithm that builds a transition policy to optimize conditional and sequential tasks. For the anti-jamming system, the proposed model learns two tasks: sensing time and transmission channel selection. DMCSL is a composite of the state-of-the-art reinforcement learning algorithms, multi-armed bandit and an extended deep-Q-network. To improve the chance of convergence and optimal cumulative reward of the algorithm, we also propose a continuous action-space update algorithm for sensing time action-space. The simulation results show that DMCSL guarantees better performance than single-task learning by relying on a logarithmically increased action-space sample. Against a random dynamic jamming time-slot, DMCSL achieves about three times better cumulative reward, and against a periodic dynamic jamming time-slot, it improves by 10% the cumulative reward. |
topic |
Ad hoc network deep learning jamming attack multi-armed bandit spectrum sensing |
url |
https://ieeexplore.ieee.org/document/9528307/ |
work_keys_str_mv |
AT robertbasomingera deepmultitaskconditionalandsequentiallearningforantijamming AT youngjunechoi deepmultitaskconditionalandsequentiallearningforantijamming |
_version_ |
1717380195209445376 |