We propose an end-To-end learned image data hiding framework that embeds and extracts secrets in the latent representations of a generic neural compressor. By leveraging a perceptual loss function in conjunction with ...
详细信息
The localization of image splicing involves identifying pixels in an image that have been spliced from other images, necessitating the discernment of splicing features. Despite significant advancements driven by the r...
详细信息
ISBN:
(纸本)9798350349405;9798350349399
The localization of image splicing involves identifying pixels in an image that have been spliced from other images, necessitating the discernment of splicing features. Despite significant advancements driven by the rise of social media and deep learning, existing methods exhibit limitations, often neglecting the integration of coarse and precise features and lacking the ability to understand objects. This leads to erroneous predictions in identifying spliced regions. This paper proposes Segment Anything Model with Integrated compression and Edge artifacts (SAM-ICE) for the localization of image splicing, addressing these limitations by fusing forged edge features and compression artifact features. Leveraging SAM's object understanding ability, our method identifies spliced regions using the fused features as guidance. Specifically, we employ Edge Artifact Extractor (EAE) to extract fine high-frequency edge splicing features and compression Artifact Extractor (CAE) to extract coarse compression artifact features. By combining these features, our method utilizes coarse-fine features to accurately pinpoint the spliced portions of the image. Experimental results demonstrate the superior accuracy, robustness, and generalizability of our method compared to the state-of-the-arts.
The intra block copy (IBC) mode introduced can significantly improve the coding efficiency for screen content. However, with the mixed content, the coding performance improvement of IBC is limited and while also intro...
详细信息
ISBN:
(纸本)9798331529543;9798331529550
The intra block copy (IBC) mode introduced can significantly improve the coding efficiency for screen content. However, with the mixed content, the coding performance improvement of IBC is limited and while also introducing the increases of the coding complexity. This paper proposes an early determination scheme for IBC prediction to reduce the encoding complexity over mixed content sequences containing captured content regions. Our method is based on a refined CTU-level screen content detection method, which can better distinguish screen content in mixed scenes. Experimental results show that the proposed method reduces encoding complexity with moderate coding performance loss on mixed content sequences.
Wavelet-based totally records compression is a form of information compression used to method far flung sensing and picture processing. This technique makes use of wavelets, that are mathematical features that divide ...
详细信息
How to maximize embedding capacity is one of the current challenges in the field of reversible data hiding. A reversible data hiding scheme is proposed based on the rearrangement and compression of prediction error bi...
详细信息
Lookup tables (LUTs) are commonly used to speed up image processing by handling complex mathematical functions like sine and exponential calculations. They are used in various applications such as camera image process...
详细信息
ISBN:
(纸本)9798331529543;9798331529550
Lookup tables (LUTs) are commonly used to speed up image processing by handling complex mathematical functions like sine and exponential calculations. They are used in various applications such as camera image processing, high-dynamic range imaging, and edge-preserving filtering. However, due to the increasing gap between computing and input/output performance, LUTs are becoming less effective. Even though specific circuits like SIMD can improve LUT efficiency, they still need to bridge the performance gap fully. The gap makes it difficult to choose between direct numerical and LUT calculations. For this problem, a register-LUTs method with the nearest neighbor was proposed;however, it is limited for functions with narrow-range values approaching zero. In this paper, we propose a method for using register LUTs to process images efficiently over a wide range of values. Our contributions include proposing register-LUT with linear interpolation for efficient computation, using a smaller data type for further efficiency, and suggesting an efficient data retrieving method.
Deep neural networks (DNN) are used to analyze images, videos, signals and texts require a lot of memory and intensive computing power. For example, the very successful GPT4 model contains more than a few trillion par...
详细信息
ISBN:
(纸本)9798350374520;9798350374513
Deep neural networks (DNN) are used to analyze images, videos, signals and texts require a lot of memory and intensive computing power. For example, the very successful GPT4 model contains more than a few trillion parameters. Although such models are of great impact, but they have been used very little in real-world applications, including industrial Internet of Things, self-driving cars, algorithmic health monitoring for use in limited mobile or edge devices. The requirement to run large models on resource-constrained peripherals has led to significant research interest in compressing DNN models. Signal processing researchers have traditionally advocated data (image/video/audio) compression, and by the way, many of these techniques are used for DNN compression. For example, source coding is a basic technique that has been widely used to compress various DNN models. In this paper, we present our views on the use of signal processing methods for DNN model compression.
In this paper, we propose a source coding scheme that represents data from unknown distributions through frequency and support information. Existing encoding schemes often compress data by sacrificing computational ef...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
In this paper, we propose a source coding scheme that represents data from unknown distributions through frequency and support information. Existing encoding schemes often compress data by sacrificing computational efficiency or by assuming the data follows a known distribution. We take advantage of the structure that arises within the spatial representation and utilize it to encode run-lengths within this representation using Golomb coding. Through theoretical analysis, we show that our scheme yields an overall bit rate that nears entropy without a computationally complex encoding algorithm and verify these results through numerical experiments.
This research study introduces a compressive method for securing Hindi text data, utilizing a combination of Base64 encoding and Fernet cryptography followed by ZIP compression. Initially, the Hindi text undergoes tra...
详细信息
ISBN:
(纸本)9798350391558;9798350379990
This research study introduces a compressive method for securing Hindi text data, utilizing a combination of Base64 encoding and Fernet cryptography followed by ZIP compression. Initially, the Hindi text undergoes transformation into plain text format through Base64 encoding, ensuring compatibility with subsequent encryption processes. The encoded text is then subjected to encryption using Fernet cryptography, a highly secure symmetric encryption technique known for its robust protection and data integrity capabilities. Subsequently, the encrypted text is compressed using ZIP compression, aimed at enhancing storage efficiency and facilitating faster transmission speeds. The resultant cipher data, incorporating both encryption and compression, offers a secure and optimized solution for storing and transmitting sensitive Hindi text data. This method effectively safeguards confidentiality, integrity, and efficiency in managing digital communications, addressing the critical need for secure handling of textual information in today's data-driven environments.
In this research, we conducted a preprocessing step on pure-tone audiometry image data to determine the presence or absence of hearing loss in patients, and designed machine learning models for hearing loss classifica...
详细信息
ISBN:
(纸本)9798350367164;9798350367157
In this research, we conducted a preprocessing step on pure-tone audiometry image data to determine the presence or absence of hearing loss in patients, and designed machine learning models for hearing loss classification using the preprocessed data. The dataset utilized consisted of a total of 18,530 data (9,265 normal, 9,265 hearing loss), and detailed parameter configuration was performed using randomly extracted training and validation data. This research involved converting image data into a CSV file format during the preprocessing stage, and the preprocessed data was then utilized to design and construct the Logistic Regression and Decision Tree Classifier machine learning models for patient classification. These models achieved an accuracy of 85.61% and 85.25%, respectively, in automatically determining the presence or absence of hearing loss.
暂无评论