The rise of deepfake technologies has posed significant challenges to privacy, security, and information integrity, particularly in audio and multimedia content. This paper introduces a Quantum-Trained Convolutional N...
详细信息
ISBN:
(数字)9798331519315
ISBN:
(纸本)9798331519322
The rise of deepfake technologies has posed significant challenges to privacy, security, and information integrity, particularly in audio and multimedia content. This paper introduces a Quantum-Trained Convolutional neural Network (QT-CNN) framework designed to enhance the detection of deepfake audio, leveraging the computational power of quantum machine learning (QML). The QT-CNN employs a hybrid quantum-classical approach, integrating Quantum neuralnetworks (QNNs) with classical neural architectures to optimize training efficiency while reducing the number of trainable parameters. Our method incorporates a novel quantum-to-classical parameter mapping that effectively utilizes quantum states to enhance the expressive power of the model, achieving up to 70% parameter reduction compared to classical models without compromising accuracy. Data pre-processing involved extracting essential audio features, label encoding, feature scaling, and constructing sequential datasets for robust model evaluation. Experimental results demonstrate that the QT-CNN achieves comparable performance to traditional CNNs, maintaining high accuracy during training and testing phases across varying configurations of QNN blocks. The QT framework’s ability to reduce computational overhead while maintaining performance underscores its potential for real-world applications in deepfake detection and other resource-constrained scenarios. This work highlights the practical benefits of integrating quantum computing into artificial intelligence, offering a scalable and efficient approach to advancing deepfake detection technologies.
As a fundamental task in the field of computer vision, super-resolution reconstruction has a wide range of applications on the industrial Internet. However, most of the existing super-resolution reconstruction methods...
详细信息
ISBN:
(数字)9781665460569
ISBN:
(纸本)9781665460569
As a fundamental task in the field of computer vision, super-resolution reconstruction has a wide range of applications on the industrial Internet. However, most of the existing super-resolution reconstruction methods are based on convolutional neuralnetworks for feature extraction, which have the problems of low model representation efficiency and insufficient explanation of the extraction process, which limit the performance of the model in industrial Internet scenarios. In recent years, although transformers have been proposed to solve the above problems well and achieve good results in image classification tasks, there is still room for improvement in the adaptability of super-resolution reconstruction tasks. Therefore, in this paper, an improved transformer-based image super-resolution reconstruction model is designed to effectively improve the performance of the image super-resolution reconstruction model. Experiments show that the adaptation of the Transformer mechanism on the super-resolution reconstruction task in this paper can practically improve the performance of the model on public datasets and industrial Internet scenario datasets.
In the Indian economy, Agriculture is considered one of the strongest pillars. Agriculture sector contributes significantly in our country, and it provides employability to many people live in a rural area. In our wor...
详细信息
ISBN:
(纸本)9780738146379
In the Indian economy, Agriculture is considered one of the strongest pillars. Agriculture sector contributes significantly in our country, and it provides employability to many people live in a rural area. In our work, various methods of imageprocessing are utilized for detecting healthy or unhealthy paddy leaves. This paper aims to carry out image analysis & classification methods to detect paddy leaf diseases and classification. We have built a system using imageprocessing for identifying spots in paddy images by applying segmentation techniques. This system mainly includes four components: pre-processing, segmentation to extract the infected region by adopting FCM and SLIC algorithm, Feature Extraction using statistical Gray-Level Co-Occurrence Matrix (GLCM) features, and color features are extracted using HSV planes. Finally, Classification is made by using an artificialneural network (ANN).
Monitoring water is a complex task due to its dynamic nature, added pollutants, and land build-up. The availability of high-resolution data by Sentinel-2 multispectral products makes implementing remote sensing applic...
详细信息
Deep learning (DL) algorithms are swiftly finding applications in computer vision and natural language processing. Nonetheless, they can also be employed for creating convincing deepfakes, which are challenging to dis...
详细信息
ISBN:
(数字)9798331509828
ISBN:
(纸本)9798331509835
Deep learning (DL) algorithms are swiftly finding applications in computer vision and natural language processing. Nonetheless, they can also be employed for creating convincing deepfakes, which are challenging to distinguish from reality. The advancements in image and video technology and tools, especially on social media platforms, potentially lead to misuse for malicious purposes like blackmail or defamation. To tackle this issue, several group of researchers tried upon spreading or creating awareness on real or fake data. The proposed approach involves combining Deepfake generation using GANs and Autoencoders with a Deepfake detection method. The aim of this initiative is exclusively to combat disinformation and online fraud for the welfare of the general population. Deepfakes, products of AI, have become increasingly realistic, rendering it nearly difficult to distinguish the content. Auto-encoders with sufficient time can achieve about 92 % accuracy. As the generator improves, the discriminator performance worsens as it struggles to differentiate real or fake data. A perfect generator results in 50% accuracy. With advancements in computational capacity and data availability, the proposed DDM (Deepfake Detection Model) has achieved greater accuracy rate of up to 92.3%.
Nowadays, computer vision is more and more widely applied in fitness and healthcare. However, little research has been done on yoga in this area, even though Yoga is a very popular sport. To fill this gap, this paper ...
详细信息
With the development of artificial intelligence (AI) technology, many applications are providing AI services. The key part of these AI services is the Deep neuralnetworks(DNNs) requiring a lot of computation. However...
详细信息
Although image super-resolution (SR) problem has ex-perienced unprecedented restoration accuracy with deep neuralnetworks, it has yet limited versatile applications due to the substantial computational costs. Since d...
详细信息
ISBN:
(数字)9798350353006
ISBN:
(纸本)9798350353013
Although image super-resolution (SR) problem has ex-perienced unprecedented restoration accuracy with deep neuralnetworks, it has yet limited versatile applications due to the substantial computational costs. Since differ-ent input images for SR face different restoration difficul-ties, adapting computational costs based on the input image, referred to as adaptive inference, has emerged as a promising solution to compress SR networks. Specifically, adapting the quantization bit-widths has successfully re-duced the inference and memory cost without sacrificing the accuracy. However, despite the benefits of the resul-tant adaptive network, existing works rely on time-intensive quantization-aware training with full access to the origi-nal training pairs to learn the appropriate bit allocation policies, which limits its ubiquitous usage. To this end, we introduce the first on-the-fly adaptive quantization frame-work that accelerates the processing time from hours to sec-onds. We formulate the bit allocation problem with only two bit mapping modules: one to map the input image to the image-wise bit adaptation factor and one to obtain the layer-wise adaptation factors. These bit mappings are cali-brated and fine-tuned using only a small number of calibration images. We achieve competitive performance with the previous adaptive quantization methods, while the processing time is accelerated by × 2000. Codes are available at https://***/Cheeun/AdaBM.
Ship recognition is a central component of the information perception for smart ships. Traditionally, the ship recognition mission is completed by remote sensing or synthetic aperture radar (SAR) or manual calibration...
Ship recognition is a central component of the information perception for smart ships. Traditionally, the ship recognition mission is completed by remote sensing or synthetic aperture radar (SAR) or manual calibration via maritime surveillance video. However, the difficulty arises in implementing the mission when it suffers the weather with a poor visibility. The recognition accuracy may deteriorate dramatically in such a hostile environment, especially in the heavy fog environment at sea when using conventional methods. In this case, we develop a novel framework for ship recognition based on a smart recognition system that relies on the dark channel defogging and the neuralnetworks. First, a dark-channel defogging algorithm is utilized to remove the adverse impact of poor visibility on recognition. After that, three different neuralnetworks are used to train the defogged images (Specifically: Faster-RCNN, SSD, YOLOv3). Finally, we present qualitative and quantitative analyses of three different neural network models. The results demonstrate the effectiveness of the proposed method in a hostile environment, especially in fog weather with poor visibility at sea.
In computer vision, neural network models typically require a large amount of manually annotated images or video data for training. To reduce annotation costs, self-supervised learning has gained significant attention...
详细信息
ISBN:
(数字)9798350374407
ISBN:
(纸本)9798350374414
In computer vision, neural network models typically require a large amount of manually annotated images or video data for training. To reduce annotation costs, self-supervised learning has gained significant attention. This paper proposes a self-supervised learning-based method by introducing an auxiliary task involving spatiotemporal context in videos—extracting video keyframes—to guide self-supervised learning. The neural network learns video stream features by completing this task, thereby understanding the high-level semantics of the video. Additionally, this auxiliary task achieves video content redundancy reduction and video compression through keyframe extraction. To quantitatively evaluate the effectiveness of this method, the neural network is transferred to the domain of action recognition and experiments are conducted on the UCF101 dataset, achieving an accuracy of 66.2% with fewer parameters and training data. The validation of the effectiveness of this method as an auxiliary task for video spatiotemporal features contributes to a better understanding of video semantic features by neuralnetworks and facilitates action recognition, thereby expanding the applications of self-supervised learning.
暂无评论