• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

Bi, J. (Bi, J..) | Xu, K. (Xu, K..) | Yuan, H. (Yuan, H..) | Zhou, M. (Zhou, M..)

Indexed by:

EI Scopus

Abstract:

Precise real-time prediction of the number of future network attacks cannot only prompt cloud infrastructures to fast respond to them and protect network security, but also prevents economic and business losses. In recent years, neural networks, e.g., Bi-direction Long and Short Term Memory (LSTM) and Temporal Convolutional Network (TCN), have been proven to be suitable for predicting time series data. Attention mechanisms are also widely used for the time series prediction. In this work, we propose a novel hybrid deep learning prediction method by combining the capabilities of a Savitzky-Golay (SG) filter, TCN, Multi-head self attention, and BiLSTM for the prediction of network attacks. This work first adopts a SG filter to eliminate noise in the raw data. It applies TCN to extract short-term features from the sequences. It then adopts multi-head self attention to capture intrinsic connections among features. Finally, this work adopts Bi-LSTM to extract bi-directional and long-term correlations in the sequences. Experimental results with a real-life dataset show that the proposed method outperforms several typical algorithms in terms of prediction accuracy.  © 2022 IEEE.

Keyword:

Savitzky-Golay filter Network attack prediction multi-head self attention long and short term memory and temporal convolutional network

Author Community:

  • [ 1 ] [Bi J.]Beijing University of Technology, Faculty of Information Technology, Beijing, 100124, China
  • [ 2 ] [Xu K.]Beijing University of Technology, Faculty of Information Technology, Beijing, 100124, China
  • [ 3 ] [Yuan H.]Beihang University, School of Automation Science and Electrical Engineering, Beijing, 100191, China
  • [ 4 ] [Zhou M.]New Jersey Institute of Technology, Department of Electrical and Computer Engineering, Newark, 07102, United States

Reprint Author's Address:

Email:

Show more details

Related Keywords:

Source :

ISSN: 1062-922X

Year: 2022

Volume: 2022-October

Page: 544-549

Language: English

Cited Count:

WoS CC Cited Count: 0

SCOPUS Cited Count: 3

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 2

Affiliated Colleges:

Online/Total:834/10617085
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.