In this research paper, an overview of computer methods for segmenting continuous-tone images into meaningful parts and characterizing these parts with 'features' is presented. image segmentation is an essenti...
详细信息
The rise of cultural tourism in India and massive digitization over the last decade has necessitated preserving Indian art forms. Recent advances in artificial intelligence (AI) have provided the tools and techniques ...
详细信息
The rapid expansion of autonomous driving technologies necessitates the development of robust systems for accurate road surface identification and classification to ensure safe and reliable driving. This review articl...
详细信息
The proceedings contain 34 papers. The special focus in this conference is on image and Video Technology. The topics include: Spatial Variation Sequences for Remote Sensing applications with Small Sample Siz...
ISBN:
(纸本)9789819703753
The proceedings contain 34 papers. The special focus in this conference is on image and Video Technology. The topics include: Spatial Variation Sequences for Remote Sensing applications with Small Sample Sizes;exploring the Potential of High-Resolution Drone imagery for Improved 3D Human Avatar Reconstruction: A Comparative Study with Mobile images;point Cloud Novelty Detection Based on Latent Representations of a General Feature Extractor;Efficient 3Dconv Fusion of RGB and Optical Flow for Dynamic Hand Gesture Recognition and Localization;an Investigation of Video vision Transformers for Depression Severity Estimation from Facial Video Data;real-Time Automated Body Condition Scoring of Dairy Cows;Logo-SSL: Self-supervised Learning with Self-attention for Efficient Logo Detection;HAHANet: Towards Accurate image Classifiers with Less Parameters;evaluating Mammogram image Classification: Impact of Model Architectures, Pretraining, and Finetuning;melanoma Classification Using Deep Learning;3D Formation Control of Multiple Cooperating Autonomous Agents via Leader-Follower Strategy;LAPRNet: Lightweight Airborne Particle Removal Network for LiDAR Point Clouds;REAL-NET: A Monochromatic Depth Estimation Using REgional Attention and Local Feature Mapping;Spike-EFI: Spiking Neural Network for Event-Based Video Frame Interpolation;scrambleMix: A Privacy-Preserving imageprocessing for Edge-Cloud machine Learning;Comparison of Simplified SE-ResNet and SE-DenseNet for Micro-Expression Classification;facial Deepfake Detection Using Gaussian Processes;A Novel Steganography Scheme Using Logistic Map, BRISK Descriptor, and K-Means Clustering;a Holistic Approach to Elderly Safety: Sensor Fusion, Fall Detection, and Privacy-Preserving Techniques;cluster-Based Video Summarization with Temporal Context Awareness;On Deploying Mobile Deep Learning to Segment COVID-19 PCR Test Tube images;enhancing Safety During Surgical Procedures with Computer vision, Artificial Intelligence, and Natural
Traditional remote sensing imageprocessing is not able to provide timely information for near real-time applications due to the hysteresis of satellite-ground mutual communication and low processing efficiency. On-bo...
Traditional remote sensing imageprocessing is not able to provide timely information for near real-time applications due to the hysteresis of satellite-ground mutual communication and low processing efficiency. On-board intelligent processing is an important approach to improve the efficiency and intelligence of remote sensing satellites. This paper takes convolutional neural network (CNN) based on-board processing as the focus. Firstly, the basic workflow of CNN based on-board processing system is illustrated. Afterwards, the applications of lightweight CNN based on-board processing are thoroughly reviewed. The used CNN models are further analyzed to compare the advantages and disadvantages. Finally, current challenges are summarized and future works concerned with artificial intelligence are concluded.
There is a growing need for the development of computational methods and tools for automated, objective, and quantitative analysis of biomedical signal and image data to facilitate disease and treatment monitoring, ea...
There is a growing need for the development of computational methods and tools for automated, objective, and quantitative analysis of biomedical signal and image data to facilitate disease and treatment monitoring, early diagnosis, and scientific discovery. Recent advances in artificial intelligence and machine learning, particularly in deep learning, have revolutionized computer vision and image analysis for many application areas. While processing of non-biomedical signal, image, and video data using deep learning methods has been very successful, high-stakes biomedical applications present unique challenges such as different image modalities, limited training data, need for explainability and interpretability etc. that need to be addressed. In this dissertation, we developed novel, explainable, and attention-based deep learning frameworks for objective, automated, and quantitative analysis of biomedical signal, image, and video data. The proposed solutions involve multi-scale signal analysis for oraldiadochokinesis studies; ensemble of deep learning cascades using global soft attention mechanisms for segmentation of meningeal vascular networks in confocal microscopy; spatial attention and spatio-temporal data fusion for detection of rare and short-term video events in laryngeal endoscopy videos; and a novel discrete Fourier transform driven class activation map for explainable-AI and weakly-supervised object localization and segmentation for detailed vocal fold motion analysis using laryngeal endoscopy videos. Experiments conducted on the proposed methods showed robust and promising results towards automated, objective, and quantitative analysis of biomedical data, that is of great value for potential early diagnosis and effective disease progress or treatment monitoring.
Super-resolution (SR) is a fascinating frontier in medical ultrasound (US) imaging offering the possibility of studying biological activity at spatiotemporal scales beyond the classical diffraction limit [1]. The key ...
详细信息
ISBN:
(数字)9781665466578
ISBN:
(纸本)9781665466578
Super-resolution (SR) is a fascinating frontier in medical ultrasound (US) imaging offering the possibility of studying biological activity at spatiotemporal scales beyond the classical diffraction limit [1]. The key to SR is reliable detection and subsequent tracking of centroids of US contrast agents, over thousands of frames [1]. However, methods to overcome motion artefacts and background tissue speckle impose computational overhead [2];in addition to physical tradeoffs in data acquisition [1][3];thereby limiting biological applications to larger vessels with high blood flow rates [1]. The real-time or online nature of ultrasound imaging is sacrificed due to the offline nature of super-resolution processing methods [1]. In this work, we explore combinations of current machinevision algorithms, popular for similar object detection and tracking problems in optical imaging [4] - towards near real-time [5] super-resolution ultrasound imaging. We report encouraging results motivating further work towards improving state-of-the-art machinevision models designed for online, real-time, detection and tracking for ultrasound super-resolution.
Detecting salient objects in an image has a vast number of applications across the web, and mostly it is done manually. Since the internet is growing faster than ever, it is not a feasible solution for high-scale dyna...
详细信息
Detecting salient objects in an image has a vast number of applications across the web, and mostly it is done manually. Since the internet is growing faster than ever, it is not a feasible solution for high-scale dynamic applications. Hence, imageprocessing automation with computer vision and machine learning has become a burning topic of research in the past few years. Saliency based image cropping is a task of identifying the notable segments in an image and be able to accommodate these parts in the crop ratio for any given specific viewport. It helps not only in image cropping but also with object recognition, visual tracking and visual restoration, etc. To address this, a novel approach for saliency detection is proposed based on heat map analysis and boundary prior. In this approach, the images go through heat map anal-ysis to identify salient objects based on machine learning model. After that, the images are segmented based on color and texture difference and then major contours are identified. The contours and the obtained saliency coordinates are accommodated inside the crop for each viewport requirement. The crop coordinates then go through two more processes, firstly a shift center process where the crop center is moved towards the important but lesser salient object, and then an inclusivity rule checks that the image is not cropped at coordinates without any salient objects. The simulation results reveal that the proposed algorithm attains better results than the cutting-edge algorithm of Twitter with similarity index of 89.89%.(c) 2022 Elsevier Ltd. All rights *** and peer-review under responsibility of the scientific committee of the 3rd International Con-ference on "Advancement in Nanoelectronics and Communication Technologies".
Camera traps have quickly transformed the way in which many ecologists study the distribution of wildlife species, their activity patterns and interactions among members of the same ecological community. Although they...
详细信息
Camera traps have quickly transformed the way in which many ecologists study the distribution of wildlife species, their activity patterns and interactions among members of the same ecological community. Although they provide a cost-effective method for monitoring multiple species over large spatial and temporal scales, the time required to process the data can limit the efficiency of camera-trap surveys. Thus, there has been considerable attention given to the use of artificial intelligence (AI), specifically deep learning, to help process camera-trap data. Using deep learning for these applications involves training algorithms, such as convolutional neural networks (CNNs), to use particular features in the camera-trap images to automatically detect objects (e.g. animals, humans, vehicles) and to classify species. To help overcome the technical challenges associated with training CNNs, several research communities have recently developed platforms that incorporate deep learning in easy-to-use interfaces. We review key characteristics of four AI platforms-Conservation AI, MegaDetector, MLWIC2: machine Learning for Wildlife image Classification and Wildlife Insights-and two auxiliary platforms-Camelot and Timelapse-that incorporate AI output for processing camera-trap data. We compare their software and programming requirements, AI features, data management tools and output format. We also provide R code and data from our own work to demonstrate how users can evaluate model performance. We found that species classifications from Conservation AI, MLWIC2 and Wildlife Insights generally had low to moderate recall. Yet, the precision for some species and higher taxonomic groups was high, and MegaDetector and MLWIC2 had high precision and recall when classifying images as either 'blank' or 'animal'. These results suggest that most users will need to review AI predictions, but that AI platforms can improve efficiency of camera-trap-data processing by allowing users to filt
Chest x-ray studies can be automatically detected and their locations located using artificial intelligence (AI) in healthcare. To detect the location of findings, additional annotation in the form of bounding boxes i...
详细信息
暂无评论