In this study, an automatic monitoring system based on convolutional neural network (CNN) is proposed to address the automation and accuracy of remote sensing imageprocessing. With the maturity and wide application o...
详细信息
imageprocessing (IP) technology has emerged on the basis of AI, digital imaging technology, and multimedia technology, and people began to use computers to process images to improve image quality and improve human vi...
详细信息
This research paper introduces a Rule-Based Expert System designed for the automated editing of documents in PDF and PPT formats. The system employs a set of predefined rules, extracted from guideline documents using ...
详细信息
ISBN:
(数字)9783031585616
ISBN:
(纸本)9783031585609;9783031585616
This research paper introduces a Rule-Based Expert System designed for the automated editing of documents in PDF and PPT formats. The system employs a set of predefined rules, extracted from guideline documents using a Large Language Model (LLM), to execute tasks such as redaction of sensitive text/logo detection and annotation of text elements that deviate from prescribed font size guidelines. Following the detection and annotation process, the system further enhances documents by resizing the detected text elements based on the predefined rules. To achieve these editing tasks, the system integrates advanced imageprocessing techniques, leveraging fine-tuned Optical Character Recognition (OCR) for accurate text extraction from document images. Furthermore, Natural Language processing (NLP) algorithms are utilized to analyze and interpret textual content. The combination of imageprocessing, OCR, NLP, and rule extraction using LLM ensures a comprehensive approach to document editing, enhancing efficiency and accuracy. The proposed system addresses the need for automated and rule-driven document editing, contributing to advancements in information security and document standardization.
This work offers a thorough method for real-time dehazing of drone-captured images by different filtering techniques with post-processing improvements. Enhancing visibility and picture clarity in hazy situations is th...
详细信息
Here we consider the problem of denoising features associated to complex data, modeled as signals on a graph, via a smoothness prior. This is motivated in part by settings such as single-cell RNA where the data is ver...
详细信息
ISBN:
(纸本)9798350369298;9798350369304
Here we consider the problem of denoising features associated to complex data, modeled as signals on a graph, via a smoothness prior. This is motivated in part by settings such as single-cell RNA where the data is very high-dimensional, but its structure can be captured via an affinity graph. This allows us to utilize ideas from graph signal processing. In particular, we present algorithms for the cases where the signal is perturbed by Gaussian noise, dropout, and uniformly distributed noise. The signals are assumed to follow a prior distribution defined in the frequency domain which favors signals which are smooth across the edges of the graph. By pairing this prior distribution with our three models of noise generation, we propose Maximum A Posteriori (M.A.P.) estimates of the true signal in the presence of noisy data and provide algorithms for computing the M.A.P. Finally, we demonstrate the algorithms' ability to effectively restore signals from white noise on image data and from severe dropout in single-cell RNA sequence data.
Computer vision, driven by artificial intelligence, has become pervasive in diverse applications such as self-driving cars and law enforcement. However, the susceptibility of these systems to attacks has raised signif...
详细信息
作者:
Wanjari, KetanVerma, Prateek
Department of Computer Science and Engineering Faculty of Engineering and Technology Maharashtra Wardha442001 India
Department of Artificial Intelligence and Data Science Faculty of Engineering and Technology Maharashtra Wardha442001 India
Modern image recognition has experienced dramatic improvements because of Machine Learning and Deep Learning algorithms together. This study investigates CNNs and SVMs for recognition enhancement while reviewing image...
详细信息
—This paper is based on the application of lossless compression algorithms in image compression, aiming to solve problems such as insufficient storage space, low transmission efficiency, and heavy data processing bur...
详细信息
Medical imaging is crucial for heart diagnosis, but outdated algorithms and hardware result in delayed processing and low accuracy. Using NVIDIA Clara, a GPU-accelerated platform, the study proposes real-time cardiac ...
详细信息
The field of artificial intelligence (AI) holds a variety of algorithms designed with the goal of achieving high accuracy at low computational cost and latency. One popular algorithm is the vision transformer (ViT), w...
详细信息
ISBN:
(纸本)9798350383638;9798350383645
The field of artificial intelligence (AI) holds a variety of algorithms designed with the goal of achieving high accuracy at low computational cost and latency. One popular algorithm is the vision transformer (ViT), which excels at various computer vision tasks for its ability to capture long-range dependencies effectively. This paper analyzes a computing paradigm, namely, spatial transformer networks (STN), in terms of accuracy and hardware complexity for image classification tasks. The paper reveals that for 2D applications, such as image recognition and classification, STN is a great backbone for AI algorithms for its efficiency and fast inference time. This framework offers a promising solution for efficient and accurate AI for resource-constrained Internet of Things (IoT) and edge devices. The comparative analysis of STN implementations on the central processing unit (CPU), Raspberry Pi (RPi), and Resistive Random Access Memory (RRAM) architectures reveals nuanced performance variations, providing valuable insights into their respective computational efficiency and energy utilization.
暂无评论