The proceedings contain 34 papers. The special focus in this conference is on image and Video Technology. The topics include: Spatial Variation Sequences for Remote Sensing applications with Small Sample Siz...
ISBN:
(纸本)9789819703753
The proceedings contain 34 papers. The special focus in this conference is on image and Video Technology. The topics include: Spatial Variation Sequences for Remote Sensing applications with Small Sample Sizes;exploring the Potential of High-Resolution Drone imagery for Improved 3D Human Avatar Reconstruction: A Comparative Study with Mobile images;point Cloud Novelty Detection Based on Latent Representations of a General Feature Extractor;Efficient 3Dconv Fusion of RGB and Optical Flow for Dynamic Hand Gesture Recognition and Localization;an Investigation of Video vision Transformers for Depression Severity Estimation from Facial Video Data;real-Time Automated Body Condition Scoring of Dairy Cows;Logo-SSL: Self-supervised Learning with Self-attention for Efficient Logo Detection;HAHANet: Towards Accurate image Classifiers with Less Parameters;evaluating Mammogram image Classification: Impact of Model Architectures, Pretraining, and Finetuning;melanoma Classification Using Deep Learning;3D Formation Control of Multiple Cooperating Autonomous Agents via Leader-Follower Strategy;LAPRNet: Lightweight Airborne Particle Removal Network for LiDAR Point Clouds;REAL-NET: A Monochromatic Depth Estimation Using REgional Attention and Local Feature Mapping;Spike-EFI: Spiking Neural Network for Event-Based Video Frame Interpolation;scrambleMix: A Privacy-Preserving imageprocessing for Edge-Cloud machine Learning;Comparison of Simplified SE-ResNet and SE-DenseNet for Micro-Expression Classification;facial Deepfake Detection Using Gaussian Processes;A Novel Steganography Scheme Using Logistic Map, BRISK Descriptor, and K-Means Clustering;a Holistic Approach to Elderly Safety: Sensor Fusion, Fall Detection, and Privacy-Preserving Techniques;cluster-Based Video Summarization with Temporal Context Awareness;On Deploying Mobile Deep Learning to Segment COVID-19 PCR Test Tube images;enhancing Safety During Surgical Procedures with Computer vision, Artificial Intelligence, and Natural
Synthetic media or "deepfakes" are making great advances in visual quality, diversity, and verisimilitude, empowered by large-scale publicly accessible datasets and rapid technical progress in deep generativ...
详细信息
Synthetic media or "deepfakes" are making great advances in visual quality, diversity, and verisimilitude, empowered by large-scale publicly accessible datasets and rapid technical progress in deep generative modeling. Heralding a paradigm shift in how online content is trusted, researchers in digital image forensics have responded with different proposals to reliably detect AI-generated images in the wild. However, binary classification of image authenticity is insufficient to regulate the ethical usage of deepfake technology as new applications are developed. This article provides an overview of the major innovations in synthetic forgery detection as of 2020, while highlighting the recent shift in research towards ways to attribute AI-generated images to their generative sources with evidence. We define the various categories of deepfakes in existence, the subtle processing traces and fingerprints that distinguish AI-generated images from reality and each other, and the different degrees of attribution possible with current understanding of generative algorithms. Additionally, we describe the limitations of synthetic image recognition methods in practice, the counter-forensic attacks devised to exploit these limitations, and directions for new research to assure the long-term relevance of deepfake forensics. Reliable, explainable, and generalizable attribution methods would hold malicious users accountable for AI-enabled disinformation, grant plausible deniability to appropriate users, and facilitate intellectual property protection of deepfake technology. This article is categorized under: Commercial, Legal, and Ethical Issues > Security and Privacy Algorithmic Development > Multimedia
There is a growing need for the development of computational methods and tools for automated, objective, and quantitative analysis of biomedical signal and image data to facilitate disease and treatment monitoring, ea...
There is a growing need for the development of computational methods and tools for automated, objective, and quantitative analysis of biomedical signal and image data to facilitate disease and treatment monitoring, early diagnosis, and scientific discovery. Recent advances in artificial intelligence and machine learning, particularly in deep learning, have revolutionized computer vision and image analysis for many application areas. While processing of non-biomedical signal, image, and video data using deep learning methods has been very successful, high-stakes biomedical applications present unique challenges such as different image modalities, limited training data, need for explainability and interpretability etc. that need to be addressed. In this dissertation, we developed novel, explainable, and attention-based deep learning frameworks for objective, automated, and quantitative analysis of biomedical signal, image, and video data. The proposed solutions involve multi-scale signal analysis for oraldiadochokinesis studies; ensemble of deep learning cascades using global soft attention mechanisms for segmentation of meningeal vascular networks in confocal microscopy; spatial attention and spatio-temporal data fusion for detection of rare and short-term video events in laryngeal endoscopy videos; and a novel discrete Fourier transform driven class activation map for explainable-AI and weakly-supervised object localization and segmentation for detailed vocal fold motion analysis using laryngeal endoscopy videos. Experiments conducted on the proposed methods showed robust and promising results towards automated, objective, and quantitative analysis of biomedical data, that is of great value for potential early diagnosis and effective disease progress or treatment monitoring.
Micro defects, such as casting pores in industrial products, have been detected by human visual inspection using X-ray CT images and imageprocessing tools. Although recent deep model-based methods achieve high anomal...
详细信息
Micro defects, such as casting pores in industrial products, have been detected by human visual inspection using X-ray CT images and imageprocessing tools. Although recent deep model-based methods achieve high anomaly detection performances, the detection of micro defects is challenging because metrics for anomaly detection are dominated by low-frequency information. To overcome the problem, we propose introducing frequency-dependent losses to capture reconstruction errors appearing around micro defects and frequency-dependent data augmentation to improve the sensitivity against the errors. We demonstrate the effectiveness of the proposed method through experiments with MVTec AD dataset especially on the detection of micro defects.
Traditional remote sensing imageprocessing is not able to provide timely information for near real-time applications due to the hysteresis of satellite-ground mutual communication and low processing efficiency. On-bo...
Traditional remote sensing imageprocessing is not able to provide timely information for near real-time applications due to the hysteresis of satellite-ground mutual communication and low processing efficiency. On-board intelligent processing is an important approach to improve the efficiency and intelligence of remote sensing satellites. This paper takes convolutional neural network (CNN) based on-board processing as the focus. Firstly, the basic workflow of CNN based on-board processing system is illustrated. Afterwards, the applications of lightweight CNN based on-board processing are thoroughly reviewed. The used CNN models are further analyzed to compare the advantages and disadvantages. Finally, current challenges are summarized and future works concerned with artificial intelligence are concluded.
Super-resolution (SR) is a fascinating frontier in medical ultrasound (US) imaging offering the possibility of studying biological activity at spatiotemporal scales beyond the classical diffraction limit [1]. The key ...
详细信息
ISBN:
(数字)9781665466578
ISBN:
(纸本)9781665466578
Super-resolution (SR) is a fascinating frontier in medical ultrasound (US) imaging offering the possibility of studying biological activity at spatiotemporal scales beyond the classical diffraction limit [1]. The key to SR is reliable detection and subsequent tracking of centroids of US contrast agents, over thousands of frames [1]. However, methods to overcome motion artefacts and background tissue speckle impose computational overhead [2];in addition to physical tradeoffs in data acquisition [1][3];thereby limiting biological applications to larger vessels with high blood flow rates [1]. The real-time or online nature of ultrasound imaging is sacrificed due to the offline nature of super-resolution processing methods [1]. In this work, we explore combinations of current machinevision algorithms, popular for similar object detection and tracking problems in optical imaging [4] - towards near real-time [5] super-resolution ultrasound imaging. We report encouraging results motivating further work towards improving state-of-the-art machinevision models designed for online, real-time, detection and tracking for ultrasound super-resolution.
In agricultural settings, handling of soft fruit is critical to ensuring quality and safety. This study introduces a novel opto-tactile sensing approach designed to enhance the handling and assessment of soft fruit, w...
详细信息
In agricultural settings, handling of soft fruit is critical to ensuring quality and safety. This study introduces a novel opto-tactile sensing approach designed to enhance the handling and assessment of soft fruit, with a case example of strawberries. Our approach utilises a Robotiq 2F-85 gripper equipped with the DIGIT vision-Based Tactile Sensor (VBTS) and attached to a Universal Robot UR10e. In contrast to force-based approaches, we introduce a novel purely image-based processing software pipeline for quantifying localised surface deformations in soft fruit. The system integrates fast and explainable imageprocessing techniques applying image differencing, denoising, K-means clustering for unsupervised classification, morphological operations, and connected components analysis (CCA) to quantify surface deformations accurately. A calibration of the imageprocessing pipeline using a rubber ball showed that the system effectively captured and analysed the rubber ball's surface deformations, benefiting from its uniform elasticity and predictable response to compression. As a soft fruit case example, the imageprocessing pipeline was subsequently applied to strawberries, blueberries, and raspberries, demonstrating that the calibration parameters derived from the rubber ball could effectively assess surface deformations in soft fruits. Despite the complex, nonlinear deformation characteristics inherent to strawberries, blueberries, and raspberries, the pipeline exhibited robust performance, capturing and quantifying subtle surface changes. These findings underscore the system's capacity for precise deformation analysis in delicate materials, offering major potential for further refinement and adaptation. This novel approach of proposing and testing an imageprocessing pipeline lays the groundwork for enhancing the handling and assessment of materials with intricate mechanical properties, paving the way for broader applications in sensitive agricultural and industrial
Camera traps have quickly transformed the way in which many ecologists study the distribution of wildlife species, their activity patterns and interactions among members of the same ecological community. Although they...
详细信息
Camera traps have quickly transformed the way in which many ecologists study the distribution of wildlife species, their activity patterns and interactions among members of the same ecological community. Although they provide a cost-effective method for monitoring multiple species over large spatial and temporal scales, the time required to process the data can limit the efficiency of camera-trap surveys. Thus, there has been considerable attention given to the use of artificial intelligence (AI), specifically deep learning, to help process camera-trap data. Using deep learning for these applications involves training algorithms, such as convolutional neural networks (CNNs), to use particular features in the camera-trap images to automatically detect objects (e.g. animals, humans, vehicles) and to classify species. To help overcome the technical challenges associated with training CNNs, several research communities have recently developed platforms that incorporate deep learning in easy-to-use interfaces. We review key characteristics of four AI platforms-Conservation AI, MegaDetector, MLWIC2: machine Learning for Wildlife image Classification and Wildlife Insights-and two auxiliary platforms-Camelot and Timelapse-that incorporate AI output for processing camera-trap data. We compare their software and programming requirements, AI features, data management tools and output format. We also provide R code and data from our own work to demonstrate how users can evaluate model performance. We found that species classifications from Conservation AI, MLWIC2 and Wildlife Insights generally had low to moderate recall. Yet, the precision for some species and higher taxonomic groups was high, and MegaDetector and MLWIC2 had high precision and recall when classifying images as either 'blank' or 'animal'. These results suggest that most users will need to review AI predictions, but that AI platforms can improve efficiency of camera-trap-data processing by allowing users to filt
Chest x-ray studies can be automatically detected and their locations located using artificial intelligence (AI) in healthcare. To detect the location of findings, additional annotation in the form of bounding boxes i...
详细信息
Architecture, classification, and major applications of Generative AI interfaces, specifically chatbots, are presented in this paper. Research paper details how the Generative AI interfaces work with various Generativ...
详细信息
暂无评论