Generating diverse and realistic malware variants is critical to improve the performance of deep learning-based Malware Detection Systems (MDSs) and fight against an increasing number of complex malware attacks. Traditionally, researchers construct the single generative model based on the whole dataset, which may suffer from mode collapse and scalability problem. In order to overcome these problems, we propose Multi-level Generative Pretrained Transformer (MLGPT) which organizes multiple GPTs in the tree. Each GPT in the tree can learn the unique pattern of malware language for one malware subfamily. Consequently, MLGPT has great potentials to produce more diverse and realistic malware variants than the single generative model. Experimental results show that performance improvement of MLGPT is statistically significant as compared to the single generative model while the construction time of MLGPT is comparable to the single generative model due to the parallel strategy.
Multi-Level Generative Pretrained Transformer for Improving Malware Detection Performance
Published 2024 in 2024 7th International Conference on Artificial Intelligence and Big Data (ICAIBD)
ABSTRACT
PUBLICATION RECORD
- Publication year
2024
- Venue
2024 7th International Conference on Artificial Intelligence and Big Data (ICAIBD)
- Publication date
2024-05-24
- Fields of study
Not labeled
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-25 of 25 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1