Anomaly detection(AD) has been extensively studied and applied across various scenarios in recent years. However, gaps remain between the current performance and the desired recognition accuracy required for practical...
详细信息
Anomaly detection(AD) has been extensively studied and applied across various scenarios in recent years. However, gaps remain between the current performance and the desired recognition accuracy required for practical *** paper analyzes two fundamental failure cases in the baseline AD model and identifies key reasons that limit the recognition accuracy of existing approaches. Specifically, by Case-1, we found that the main reason detrimental to current AD methods is that the inputs to the recovery model contain a large number of detailed features to be recovered, which leads to the normal/abnormal area has not/has been recovered into its original state. By Case-2, we surprisingly found that the abnormal area that cannot be recognized in image-level representations can be easily recognized in the feature-level representation. Based on the above observations, we propose a novel recover-then-discriminate(ReDi) framework for *** takes a self-generated feature map(e.g., histogram of oriented gradients) and a selected prompted image as explicit input information to address the identified in Case-1. Additionally, a feature-level discriminative network is introduced to amplify abnormal differences between the recovered and input representations. Extensive experiments on two widely used yet challenging AD datasets demonstrate that ReDi achieves state-of-the-art recognition accuracy.
Matrix minimization techniques that employ the nuclear norm have gained recognition for their applicability in tasks like image inpainting, clustering, classification, and reconstruction. However, they come with inher...
详细信息
Matrix minimization techniques that employ the nuclear norm have gained recognition for their applicability in tasks like image inpainting, clustering, classification, and reconstruction. However, they come with inherent biases and computational burdens, especially when used to relax the rank function, making them less effective and efficient in real-world scenarios. To address these challenges, our research focuses on generalized nonconvex rank regularization problems in robust matrix completion, low-rank representation, and robust matrix regression. We introduce innovative approaches for effective and efficient low-rank matrix learning, grounded in generalized nonconvex rank relaxations inspired by various substitutes for the ?0-norm relaxed functions. These relaxations allow us to more accurately capture low-rank structures. Our optimization strategy employs a nonconvex and multi-variable alternating direction method of multipliers, backed by rigorous theoretical analysis for complexity and *** algorithm iteratively updates blocks of variables, ensuring efficient convergence. Additionally, we incorporate the randomized singular value decomposition technique and/or other acceleration strategies to enhance the computational efficiency of our approach, particularly for large-scale constrained minimization problems. In conclusion, our experimental results across a variety of image vision-related application tasks unequivocally demonstrate the superiority of our proposed methodologies in terms of both efficacy and efficiency when compared to most other related learning methods.
As a result of its aggressive nature and late identification at advanced stages, lung cancer is one of the leading causes of cancer-related deaths. Lung cancer early diagnosis is a serious and difficult challenge that...
详细信息
The increasing use of cloud-based image storage and retrieval systems has made ensuring security and efficiency crucial. The security enhancement of image retrieval and image archival in cloud computing has received c...
详细信息
The increasing use of cloud-based image storage and retrieval systems has made ensuring security and efficiency crucial. The security enhancement of image retrieval and image archival in cloud computing has received considerable attention in transmitting data and ensuring data confidentiality among cloud servers and users. Various traditional image retrieval techniques regarding security have developed in recent years but they do not apply to large-scale environments. This paper introduces a new approach called Triple network-based adaptive grey wolf (TN-AGW) to address these challenges. The TN-AGW framework combines the adaptability of the Grey Wolf Optimization (GWO) algorithm with the resilience of Triple Network (TN) to enhance image retrieval in cloud servers while maintaining robust security measures. By using adaptive mechanisms, TN-AGW dynamically adjusts its parameters to improve the efficiency of image retrieval processes, reducing latency and utilization of resources. However, the image retrieval process is efficiently performed by a triple network and the parameters employed in the network are optimized by Adaptive Grey Wolf (AGW) optimization. Imputation of missing values, Min–Max normalization, and Z-score standardization processes are used to preprocess the images. The image extraction process is undertaken by a modified convolutional neural network (MCNN) approach. Moreover, input images are taken from datasets such as the Landsat 8 dataset and the Moderate Resolution Imaging Spectroradiometer (MODIS) dataset is employed for image retrieval. Further, the performance such as accuracy, precision, recall, specificity, F1-score, and false alarm rate (FAR) is evaluated, the value of accuracy reaches 98.1%, the precision of 97.2%, recall of 96.1%, and specificity of 917.2% respectively. Also, the convergence speed is enhanced in this TN-AGW approach. Therefore, the proposed TN-AGW approach achieves greater efficiency in image retrieving than other existing
The use of management by objectives (MBOs) methodologies, particularly the objectives and key results (OKRs) framework, has gained widespread attention in recent years as a means of improving organizational performanc...
详细信息
The nonlinear filtering problem has enduringly been an active research topic in both academia and industry due to its ever-growing theoretical importance and practical *** main objective of nonlinear filtering is to i...
详细信息
The nonlinear filtering problem has enduringly been an active research topic in both academia and industry due to its ever-growing theoretical importance and practical *** main objective of nonlinear filtering is to infer the states of a nonlinear dynamical system of interest based on the available noisy measurements. In recent years, the advance of network communication technology has not only popularized the networked systems with apparent advantages in terms of installation,cost and maintenance, but also brought about a series of challenges to the design of nonlinear filtering algorithms, among which the communication constraint has been recognized as a dominating concern. In this context, a great number of investigations have been launched towards the networked nonlinear filtering problem with communication constraints, and many samplebased nonlinear filters have been developed to deal with the highly nonlinear and/or non-Gaussian scenarios. The aim of this paper is to provide a timely survey about the recent advances on the sample-based networked nonlinear filtering problem from the perspective of communication constraints. More specifically, we first review three important families of sample-based filtering methods known as the unscented Kalman filter, particle filter,and maximum correntropy filter. Then, the latest developments are surveyed with stress on the topics regarding incomplete/imperfect information, limited resources and cyber ***, several challenges and open problems are highlighted to shed some lights on the possible trends of future research in this realm.
We present a novel attention-based mechanism to learn enhanced point features for point cloud processing tasks, e.g., classification and segmentation. Unlike prior studies, which were trained to optimize the weights o...
详细信息
We present a novel attention-based mechanism to learn enhanced point features for point cloud processing tasks, e.g., classification and segmentation. Unlike prior studies, which were trained to optimize the weights of a pre-selected set of attention points, our approach learns to locate the best attention points to maximize the performance of a specific task, e.g., point cloud classification. Importantly, we advocate the use of single attention point to facilitate semantic understanding in point feature learning. Specifically,we formulate a new and simple convolution, which combines convolutional features from an input point and its corresponding learned attention point(LAP). Our attention mechanism can be easily incorporated into state-of-the-art point cloud classification and segmentation networks. Extensive experiments on common benchmarks, such as Model Net40, Shape Net Part, and S3DIS, all demonstrate that our LAP-enabled networks consistently outperform the respective original networks, as well as other competitive alternatives, which employ multiple attention points, either pre-selected or learned under our LAP framework.
Software security poses substantial risks to our society because software has become part of our life. Numerous techniques have been proposed to resolve or mitigate the impact of software security issues. Among them, ...
详细信息
Software security poses substantial risks to our society because software has become part of our life. Numerous techniques have been proposed to resolve or mitigate the impact of software security issues. Among them, software testing and analysis are two of the critical methods, which significantly benefit from the advancements in deep learning technologies. Due to the successful use of deep learning in software security, recently,researchers have explored the potential of using large language models(LLMs) in this area. In this paper, we systematically review the results focusing on LLMs in software security. We analyze the topics of fuzzing, unit test, program repair, bug reproduction, data-driven bug detection, and bug triage. We deconstruct these techniques into several stages and analyze how LLMs can be used in the stages. We also discuss the future directions of using LLMs in software security, including the future directions for the existing use of LLMs and extensions from conventional deep learning research.
Industrial cyber-physical systems closely integrate physical processes with cyberspace, enabling real-time exchange of various information about system dynamics, sensor outputs, and control decisions. The connection b...
详细信息
Industrial cyber-physical systems closely integrate physical processes with cyberspace, enabling real-time exchange of various information about system dynamics, sensor outputs, and control decisions. The connection between cyberspace and physical processes results in the exposure of industrial production information to unprecedented security risks. It is imperative to develop suitable strategies to ensure cyber security while meeting basic performance *** the perspective of control engineering, this review presents the most up-to-date results for privacy-preserving filtering,control, and optimization in industrial cyber-physical systems. Fashionable privacy-preserving strategies and mainstream evaluation metrics are first presented in a systematic manner for performance evaluation and engineering *** discussion discloses the impact of typical filtering algorithms on filtering performance, specifically for privacy-preserving Kalman filtering. Then, the latest development of industrial control is systematically investigated from consensus control of multi-agent systems, platoon control of autonomous vehicles as well as hierarchical control of power systems. The focus thereafter is on the latest privacy-preserving optimization algorithms in the framework of consensus and their applications in distributed economic dispatch issues and energy management of networked power systems. In the end, several topics for potential future research are highlighted.
The current large-scale Internet of Things(IoT)networks typically generate high-velocity network traffic *** use IoT devices to create botnets and launch attacks,such as DDoS,Spamming,Cryptocurrency mining,Phishing,**...
详细信息
The current large-scale Internet of Things(IoT)networks typically generate high-velocity network traffic *** use IoT devices to create botnets and launch attacks,such as DDoS,Spamming,Cryptocurrency mining,Phishing,*** service providers of large-scale IoT networks need to set up a data pipeline to collect the vast network traffic data from the IoT devices,store it,analyze it,and report the malicious IoT devices and types of ***,the attacks originating from IoT devices are dynamic,as attackers launch one kind of attack at one time and another kind of attack at another *** number of attacks and benign instances also vary from time to *** phenomenon of change in attack patterns is called concept ***,the attack detection system must learn continuously from the ever-changing real-time attack patterns in large-scale IoT network *** meet this requirement,in this work,we propose a data pipeline with Apache Kafka,Apache Spark structured streaming,and MongoDB that can adapt to the ever-changing attack patterns in real time and classify attacks in large-scale IoT *** concept drift is detected,the proposed system retrains the classifier with the instances that cause the drift and a representative subsample instances from the previous training of the *** proposed approach is evaluated with the latest dataset,IoT23,which consists of benign and several attack instances from various IoT *** classification accuracy is improved from 97.8%to 99.46%by the proposed *** training time of distributed random forest algorithm is also studied by varying the number of cores in Apache Spark environment.
暂无评论