Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming

Multi-task learning provides plenty of room for performance improvement to single-task learning, when learned tasks are related and learned with mutual information. In this work, we analyze the efficiency of using a single-task reinforcement learning algorithm to mitigate jamming attacks with freque...

Full description

Bibliographic Details
Main Authors: Robert Basomingera, Young-June Choi
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9528307/
id doaj-1b8440442b66493bb5e62fb7d3014052
record_format Article
spelling doaj-1b8440442b66493bb5e62fb7d30140522021-09-13T23:00:39ZengIEEEIEEE Access2169-35362021-01-01912319412320710.1109/ACCESS.2021.31098569528307Deep Multi-Task Conditional and Sequential Learning for Anti-JammingRobert Basomingera0https://orcid.org/0000-0003-4204-3554Young-June Choi1https://orcid.org/0000-0003-2240-0892Department of Computer Engineering, Ajou University, Suwon, South KoreaDepartment of Software and Computer Engineering, Ajou University, Suwon, South KoreaMulti-task learning provides plenty of room for performance improvement to single-task learning, when learned tasks are related and learned with mutual information. In this work, we analyze the efficiency of using a single-task reinforcement learning algorithm to mitigate jamming attacks with frequency hopping strategy. Our findings show that single-task learning implementations do not always guarantee optimal cumulative reward when some jammer’s parameters are unknown, notably the jamming time-slot length in this case. Therefore, to maximize packet transmission in the presence of a jammer whose parameters are unknown, we propose deep multi-task conditional and sequential learning (DMCSL), a multi-task learning algorithm that builds a transition policy to optimize conditional and sequential tasks. For the anti-jamming system, the proposed model learns two tasks: sensing time and transmission channel selection. DMCSL is a composite of the state-of-the-art reinforcement learning algorithms, multi-armed bandit and an extended deep-Q-network. To improve the chance of convergence and optimal cumulative reward of the algorithm, we also propose a continuous action-space update algorithm for sensing time action-space. The simulation results show that DMCSL guarantees better performance than single-task learning by relying on a logarithmically increased action-space sample. Against a random dynamic jamming time-slot, DMCSL achieves about three times better cumulative reward, and against a periodic dynamic jamming time-slot, it improves by 10% the cumulative reward.https://ieeexplore.ieee.org/document/9528307/Ad hoc networkdeep learningjamming attackmulti-armed banditspectrum sensing
collection DOAJ
language English
format Article
sources DOAJ
author Robert Basomingera
Young-June Choi
spellingShingle Robert Basomingera
Young-June Choi
Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
IEEE Access
Ad hoc network
deep learning
jamming attack
multi-armed bandit
spectrum sensing
author_facet Robert Basomingera
Young-June Choi
author_sort Robert Basomingera
title Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
title_short Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
title_full Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
title_fullStr Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
title_full_unstemmed Deep Multi-Task Conditional and Sequential Learning for Anti-Jamming
title_sort deep multi-task conditional and sequential learning for anti-jamming
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2021-01-01
description Multi-task learning provides plenty of room for performance improvement to single-task learning, when learned tasks are related and learned with mutual information. In this work, we analyze the efficiency of using a single-task reinforcement learning algorithm to mitigate jamming attacks with frequency hopping strategy. Our findings show that single-task learning implementations do not always guarantee optimal cumulative reward when some jammer’s parameters are unknown, notably the jamming time-slot length in this case. Therefore, to maximize packet transmission in the presence of a jammer whose parameters are unknown, we propose deep multi-task conditional and sequential learning (DMCSL), a multi-task learning algorithm that builds a transition policy to optimize conditional and sequential tasks. For the anti-jamming system, the proposed model learns two tasks: sensing time and transmission channel selection. DMCSL is a composite of the state-of-the-art reinforcement learning algorithms, multi-armed bandit and an extended deep-Q-network. To improve the chance of convergence and optimal cumulative reward of the algorithm, we also propose a continuous action-space update algorithm for sensing time action-space. The simulation results show that DMCSL guarantees better performance than single-task learning by relying on a logarithmically increased action-space sample. Against a random dynamic jamming time-slot, DMCSL achieves about three times better cumulative reward, and against a periodic dynamic jamming time-slot, it improves by 10% the cumulative reward.
topic Ad hoc network
deep learning
jamming attack
multi-armed bandit
spectrum sensing
url https://ieeexplore.ieee.org/document/9528307/
work_keys_str_mv AT robertbasomingera deepmultitaskconditionalandsequentiallearningforantijamming
AT youngjunechoi deepmultitaskconditionalandsequentiallearningforantijamming
_version_ 1717380195209445376