Cardiovascular diseases, which are currently the major causes of death globally, can be largely ameliorated through early detection and categorization. Electrocardiogram (ECG) tests have emerged as widely employed, lo...
详细信息
Cardiovascular diseases, which are currently the major causes of death globally, can be largely ameliorated through early detection and categorization. Electrocardiogram (ECG) tests have emerged as widely employed, low-cost and non-invasive procedures for evaluating electrical activities of the heart and diagnosing cardiovascular ailments. In this research, by using deep learning techniques to detect specific cardiac disorders like cardiac myocardial infarction(MI), arrhythmia, past history of myocardial infarction(PMI) and normal ECG patterns on a dataset containing patients with heart disease. We propose ECGConVT framework that combines Convolutional Neural Network (CNN) module for extracting local features, and vision Transformer (ViT) module for capturing global features. The final classification is achieved by combining the two using Multilayer Perceptron (MLP) module. The experimental results indicate promise of ECGConVT in ECG image classification where it outperforms other approaches showing an average accuracy of 98.5%, F1-score: 98.7%, Recall: 98.8% and Precision: 98.5%. In order to meet the practical needs of clinical applications, we implemented a lightweight post-processing step to reduce the size of the model.
imageprocessing with computer vision, particularly in the realm of projective geometry, offers remarkable potential for various applications. Through the lens of projective geometry, images can be transformed, augmen...
详细信息
IoT and edge devices dedicated to run machinevision algorithms are usually few years lagging currently available state-of-the-art technologies for hardware accelerators. This is mainly due to the non-negligible time ...
详细信息
IoT and edge devices dedicated to run machinevision algorithms are usually few years lagging currently available state-of-the-art technologies for hardware accelerators. This is mainly due to the non-negligible time delay required to implement and assess related algorithms. Among possible hardware platforms which are potentially being explored to handle real-time machinevision tasks, multi-core CPU and Graphical processing Unit (GPU) platforms remain the most widely used ones over Field Programmable Gate Array (FPGA) and Application Specific Integrated Circuit (ASIC)-based platforms. This is mainly due to the availability of powerful and user friendly software development tools, in addition to their lower cost, and obviously their high computation power with reasonable form factor and power consumption. Nevertheless, the trend now is towards a System-On-Chip (SOC) processors which combine ASIC/FPGA accelerators with GPU/multicore CPUs. This paper presents different state of the art IoT and edge machinevision technologies along with their performance and limitations. It can be a good reference for researchers involved in designing state of the art IoT embedded systems for machinevisionapplications.
Traditional remote sensing imageprocessing is not able to provide timely information for near real-time applications due to the hysteresis of satellite-ground mutual communication and low processing efficiency. On-bo...
详细信息
Particulate matter in the atmosphere obscures the visibility of the atmosphere, causing a condition known as haze. Other natural phenomena like mist, fog, and dust also obscure the vision;this is because of scattering...
详细信息
The use of AI technologies in remote sensing (RS) tasks has been the focus of many individuals in both the professional and academic domains. Having more accessible interfaces and tools that allow people of little or ...
详细信息
The use of AI technologies in remote sensing (RS) tasks has been the focus of many individuals in both the professional and academic domains. Having more accessible interfaces and tools that allow people of little or no experience to intuitively interact with RS data of multiple formats is a potential provided by this integration. However, the use of AI and AI agents to help automate RS-related tasks is still in its infancy stage, with some frameworks and interfaces built on top of well-known vision language models (VLM) such as GPT-4, segment anything model (SAM), and grounding DINO. These tools do promise and draw guidelines on the potentials and limitations of existing solutions concerning the use of said models. In this work, the state of the art AI foundation models (FM) are reviewed and used in a multi-modal manner to ingest RS imagery input and perform zero-shot object detection using natural language. The natural language input is then used to define the classes or labels the model should look for, then, both inputs are fed to the pipeline. The pipeline presented in this work makes up for the shortcomings of the general knowledge FMs by stacking pre-processing and post-processingapplications on top of the FMs;these applications include tiling to produce uniform patches of the original image for faster detection, outlier rejection of redundant bounding boxes using statistical and machine learning methods. The pipeline was tested with UAV, aerial and satellite images taken over multiple areas. The accuracy for the semantic segmentation showed improvement from the original 64% to approximately 80%-99% by utilizing the pipeline and techniques proposed in this work. GitHub Repository: MohanadDiab/LangRS.
Quick and reliable measurement of wood chip moisture content is an everlasting problem for numerous forest-reliant industries such as biofuel, pulp and paper, and bio-refineries. Moisture content is a critical attribu...
详细信息
Quick and reliable measurement of wood chip moisture content is an everlasting problem for numerous forest-reliant industries such as biofuel, pulp and paper, and bio-refineries. Moisture content is a critical attribute of wood chips due to its direct relationship with the final product quality. Conventional techniques for determining moisture content, such as oven-drying, possess some drawbacks in terms of their time-consuming nature, potential sample damage, and lack of real-time feasibility. Furthermore, alternative techniques, including NIR spectroscopy, electrical capacitance, X-rays, and microwaves, have demonstrated potential;nevertheless, they are still constrained by issues related to portability, precision, and the expense of the required equipment. Hence, there is a need for a moisture content determination method that is instant, portable, non-destructive, inexpensive, and precise. This study explores the use of deep learning and machinevision to predict moisture content classes from RGB images of wood chips. A large-scale image dataset comprising 1,600 RGB images of wood chips has been collected and annotated with ground truth labels, utilizing the results of the oven-drying technique. Two high-performing neural networks, MoistNetLite and MoistNetMax, have been developed leveraging Neural Architecture Search (NAS) and hyperparameter optimization. The developed models are evaluated and compared with state-of-the-art deep learning models. Results demonstrate that MoistNetLite achieves 87% accuracy with minimal computational overhead, while MoistNetMax exhibits exceptional precision with a 91% accuracy in wood chip moisture content class prediction. With improved accuracy (9.6% improvement in accuracy by MoistNetMax compared to the best baseline model ResNet152V2) and faster prediction speed (MoistNetLite being twice as fast as MobileNet), our proposed MoistNet models hold great promise for the wood chip processing industry to be efficiently deployed on p
When considering sparse motion capture marker data, one typically struggles to balance its overfitting via a high dimensional blendshape system versus underfitting caused by smoothness constraints. With the current tr...
详细信息
When considering sparse motion capture marker data, one typically struggles to balance its overfitting via a high dimensional blendshape system versus underfitting caused by smoothness constraints. With the current trend towards using more and more data, our aim is not to fit the motion capture markers with a parameterized (blendshape) model or to smoothly interpolate a surface through the marker positions, but rather to find an instance in the high resolution dataset that contains local geometry to fit each marker. Just as is true for typical machine learning applications, this approach benefits from a plethora of data, and thus we also consider augmenting the dataset via specially designed physical simulations that target the high resolution dataset such that the simulation output lies on the same so-called manifold as the data targeted.
In this study, we investigate the Deep image Prior (DIP) in enhancing image smoothing, a crucial component in numerous computer vision and graphics applications. Although deep learning has demonstrated remarkable achi...
详细信息
ISBN:
(纸本)9798350351439;9798350351422
In this study, we investigate the Deep image Prior (DIP) in enhancing image smoothing, a crucial component in numerous computer vision and graphics applications. Although deep learning has demonstrated remarkable achievements in these domains, it often falls short in flexibility and controllability, in contrast to traditional methods, which are more adaptable and typically exhibit subpar performance. Notably, some end-to-end deep learning models offer control over edge preservation, yet their performance remains marginally suboptimal. To address this shortcoming, we introduce an innovative network architecture that diverges from the traditional U-Net model, featuring a Laplacian pyramid as the encoder and a deep decoder as the decoding component, integrated with a bilateral filter loss to improve DIP. This design aids the network in rapidly assimilating essential low-frequency information. Our approach excels in retaining texture details, significantly improving image smoothing and related tasks beyond the capabilities of standard DIP methods. Moreover, our technique outperforms the leading unsupervised method, pyramid texture filtering, in texture filtering tasks and other applications.
image captured under poor-illumination conditions often display attributes of having poor contrasts, low brightness, a narrow gray range, colour distortions and considerable interference, which seriously affect the qu...
详细信息
image captured under poor-illumination conditions often display attributes of having poor contrasts, low brightness, a narrow gray range, colour distortions and considerable interference, which seriously affect the qualitative visual effects on human eyes and severely restrict the efficiency of several machinevision systems. In addition, underwater images often suffer from colour shift and contrast degradation because of an absorption and scattering of light while travelling in water. These unpleasant effects limits visibility, reduce contrast and even generate colour casts that limits the use of underwater images and videos in marine archaeology and biology. In medical imaging applications, medical images are important tools for detecting and diagnosing several medical conditions and ailments. However, the quality of medical images can often be degraded during image acquisition due to factors such as noise interference, artefacts, and poor illumination. This may lead to the misdiagnosis of medical conditions, which can further aggravate life threatening situations. image enhancement is one of the most important technologies in the field of imageprocessing, and its purpose is to improve the quality of images for specific applications. In general, the basic principle of image enhancement is to improve the quality and visual interpretability of an image so that it is more suitable for the specific applications and the observers. Over the last few decades, numerous image enhancement techniques have been proposed in the literature This study covers a systematic survey on existing state-of-the-art image enhancement techniques into broad classification of their algorithms. In addition, this paper summarises the datasets utilised in the literature for performing the experiments. Furthermore, an attention has been drawn towards several evaluation parameters for quantitative evaluation and compared different state-of-the-art algorithms for performance analysis on benchmark
暂无评论