|
|
|
Sharoug Alzaidy and Hamad Binsalleeh
In the field of behavioral detection, deep learning has been extensively utilized. For example, deep learning models have been utilized to detect and classify malware. Deep learning, however, has vulnerabilities that can be exploited with crafted inputs,...
ver más
|
|
|
|
|
|
|
Hassan Khazane, Mohammed Ridouani, Fatima Salahdine and Naima Kaabouch
With the rapid advancements and notable achievements across various application domains, Machine Learning (ML) has become a vital element within the Internet of Things (IoT) ecosystem. Among these use cases is IoT security, where numerous systems are dep...
ver más
|
|
|
|
|
|
|
Yuwen Fu, E. Xia, Duan Huang and Yumei Jing
Machine learning has been applied in continuous-variable quantum key distribution (CVQKD) systems to address the growing threat of quantum hacking attacks. However, the use of machine learning algorithms for detecting these attacks has uncovered a vulner...
ver más
|
|
|
|
|
|
|
Afnan Alotaibi and Murad A. Rassam
Concerns about cybersecurity and attack methods have risen in the information age. Many techniques are used to detect or deter attacks, such as intrusion detection systems (IDSs), that help achieve security goals, such as detecting malicious attacks befo...
ver más
|
|
|
|
|
|
|
Lei Chen, Zhihao Wang, Ru Huo and Tao Huang
As an essential piece of infrastructure supporting cyberspace security technology verification, network weapons and equipment testing, attack defense confrontation drills, and network risk assessment, Cyber Range is exceptionally vulnerable to distribute...
ver más
|
|
|
|
|
|
|
James Msughter Adeke, Guangjie Liu, Junjie Zhao, Nannan Wu and Hafsat Muhammad Bashir
Machine learning (ML) models are essential to securing communication networks. However, these models are vulnerable to adversarial examples (AEs), in which malicious inputs are modified by adversaries to produce the desired output. Adversarial training i...
ver más
|
|
|
|
|
|
|
Dejian Guan, Wentao Zhao and Xiao Liu
Recent studies show that deep neural networks (DNNs)-based object recognition algorithms overly rely on object textures rather than global object shapes, and DNNs are also vulnerable to human-less perceptible adversarial perturbations. Based on these two...
ver más
|
|
|
|
|
|
|
Zesheng Chen, Li-Chi Chang, Chao Chen, Guoping Wang and Zhuming Bi
Speaker verification systems use human voices as an important biometric to identify legitimate users, thus adding a security layer to voice-controlled Internet-of-things smart homes against illegal access. Recent studies have demonstrated that speaker ve...
ver más
|
|
|
|
|
|
|
Weizhen Xu, Chenyi Zhang, Fangzhen Zhao and Liangda Fang
Adversarial attacks hamper the functionality and accuracy of deep neural networks (DNNs) by meddling with subtle perturbations to their inputs. In this work, we propose a new mask-based adversarial defense scheme (MAD) for DNNs to mitigate the negative e...
ver más
|
|
|
|
|
|
|
Fabio Carrara, Roberto Caldelli, Fabrizio Falchi and Giuseppe Amato
The adoption of deep learning-based solutions practically pervades all the diverse areas of our everyday life, showing improved performances with respect to other classical systems. Since many applications deal with sensible data and procedures, a strong...
ver más
|
|
|
|