Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (173)

Search Parameters:
Keywords = JPEG compression

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 3407 KB  
Article
HT-NRC: A High-Throughput and Noise-Resilient Lossless Image Compression Architecture for Deep-Space CMOS Cameras
by Haoyu Wu, Yonglin Bai and Jiarui Gao
Appl. Sci. 2026, 16(6), 2873; https://doi.org/10.3390/app16062873 - 17 Mar 2026
Abstract
Lossless image compression is pivotal for deep-space exploration. Considering the requirements of deep-space exploration for a high compression ratio and real-time processing, traditional image compression algorithms have garnered significant attention. However, existing algorithms struggle with real-time processing speed and compression degradation in high-noise [...] Read more.
Lossless image compression is pivotal for deep-space exploration. Considering the requirements of deep-space exploration for a high compression ratio and real-time processing, traditional image compression algorithms have garnered significant attention. However, existing algorithms struggle with real-time processing speed and compression degradation in high-noise regions, failing to meet the throughput demands of next-generation sensors. To address these challenges, this paper proposes a high-throughput and noise-resilient lossless image compression architecture, named HT-NRC, for deep-space CMOS cameras. First, to overcome the throughput bottleneck, we introduce a parallel processing method, which is built on index-based dispatch and Reorder mechanism. This is achieved by dynamically distributing pixel streams into parallel cores and utilizing a Reorder Buffer for sequence restoration. Second, to mitigate low compression efficiency in noisy backgrounds, we present a Heterogeneous Dual-Path Coding scheme. This system adaptively separates structural information for predictive coding and stochastic noise for raw packing with Bit-Plane Slicing (BPS) strategy. The proposed architecture was implemented on a Xilinx Virtex-7 FPGA (Xilinx, Inc., San Jose, CA, USA). Operating at 100 MHz, the system achieves a processing throughput of 414.7 Mpixel/s and a high average compression ratio under deep-space image datasets, while consuming an estimated total on-chip power of only 2.1 W. Experimental results show that our proposed method substantially outperforms existing baseline methods. Specifically, compared to the optimized serial JPEG-LS implementation processing one pixel per clock cycle, our parallel architecture achieves an approximately 314.7% increase in processing throughput. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

22 pages, 66701 KB  
Article
AVIF as an Alternative to JPEG and GPU Texture Compression Schemes for Texture Storage in 3D Computer Graphics
by Maria Grazia Corino, Tiziano Leidi and Achille Peternier
Appl. Sci. 2026, 16(5), 2541; https://doi.org/10.3390/app16052541 - 6 Mar 2026
Viewed by 260
Abstract
This article explores the potential of the emerging image compression standard AV1 Image File Format (AVIF) as a format for storing 2D texture data in 3D computer graphics, aiming to assess its suitability for graphics applications. It presents a comparative performance evaluation, focusing [...] Read more.
This article explores the potential of the emerging image compression standard AV1 Image File Format (AVIF) as a format for storing 2D texture data in 3D computer graphics, aiming to assess its suitability for graphics applications. It presents a comparative performance evaluation, focusing on image quality, compression efficiency, and processing times, by comparing AVIF with the traditional format JPEG and the texture compression schemes BPTC and S3TC. To conduct the evaluation, a selected set of test images is compressed into the specified formats, loaded as textures, and assessed in a mockup 3D application to evaluate their visual performance in a realistic rendering context. The results show that AVIF delivers better fidelity to the original image compared to JPEG, BPTC, and S3TC, while also yielding a smaller file size. It outperforms JPEG by 9.2 dB in visual quality and by 174.4% in compression ratio, on average. However, this comes at the cost of longer processing times, with AVIF taking 126 times longer than JPEG and 185 times longer than S3TC to encode an image. AVIF also showed a 536% increase in decoding time compared to JPEG. BPTC produced high-fidelity images, second only to AVIF, but it required longer encoding times, depending on the quality settings. However, unlike AVIF, it offers GPU optimization benefits. Full article
(This article belongs to the Special Issue Advances in Computer Graphics and 3D Technologies)
Show Figures

Figure 1

14 pages, 7061 KB  
Article
Robust Image Steganography in Online Social Networks via Neural Style Transfer
by Peng Luo, Jia Liu, Qian Dang and Dejun Mu
Mathematics 2026, 14(4), 629; https://doi.org/10.3390/math14040629 - 11 Feb 2026
Viewed by 371
Abstract
Existing style-transfer steganography schemes suffer from three critical limitations: insufficient robustness against online social network (OSN) processing pipelines, susceptibility to steganalytic detection, and degraded visual quality. To address these challenges holistically, we propose StegTransfer—a unified framework that integrates: (1) forward non-differentiable distortion simulation, [...] Read more.
Existing style-transfer steganography schemes suffer from three critical limitations: insufficient robustness against online social network (OSN) processing pipelines, susceptibility to steganalytic detection, and degraded visual quality. To address these challenges holistically, we propose StegTransfer—a unified framework that integrates: (1) forward non-differentiable distortion simulation, which emulates realistic OSN operations to enhance robustness; (2) adversarially hardened embedding through joint training with steganalyzers to improve security; and (3) payload-preserving style enhancement that optimizes visual aesthetics without sacrificing embedding capacity. Experimental evaluations demonstrate that StegTransfer achieves superior performance in visual fidelity (NIMA score: 6.32), robustness (PSNR up to 30.2 dB under JPEG compression), and security (detection rates as low as 15.5% and 62.3% under StegExpose and SiaStegNet, respectively. Full article
(This article belongs to the Special Issue Information Security and Image Processing)
Show Figures

Figure 1

17 pages, 3074 KB  
Article
Dual-Modal Vision–Sonar Object Detection for Underwater Robots Based on Deep Learning
by Xiaoming Wang, Zhenyu Wang and Dexue Bi
J. Mar. Sci. Eng. 2026, 14(4), 338; https://doi.org/10.3390/jmse14040338 - 10 Feb 2026
Viewed by 394
Abstract
Applying state-of-the-art RGB object detectors (e.g., YOLOv8) to underwater scenes often yields unstable performance due to scattering, absorption, illumination deficiency, and bandwidth-limited transmission that severely corrupt image contrast and details. Forward-looking sonar (FLS) remains informative in turbid or low-visibility water, yet its low [...] Read more.
Applying state-of-the-art RGB object detectors (e.g., YOLOv8) to underwater scenes often yields unstable performance due to scattering, absorption, illumination deficiency, and bandwidth-limited transmission that severely corrupt image contrast and details. Forward-looking sonar (FLS) remains informative in turbid or low-visibility water, yet its low resolution and weak semantics make conventional fusion architectures costly and difficult to deploy on resource-constrained robots. This paper proposes a paired-sample-free RGB–FLS joint training paradigm based on parameter sharing, where RGB and FLS images from different datasets are jointly used during training without any frame-level pairing or architectural modification. The resulting model preserves the original detector parameter scale and inference cost, and requires only RGB input at test time. Experiments on the SeaClear and Marine Debris FLS datasets under six representative underwater degradation factors (contrast loss, blur, resolution reduction, color cast, and JPEG compression) show consistent robustness gains over RGB-only training. In particular, under severe low-contrast corruption, the proposed training strategy improves mAP50 by more than 14 percentage points compared with the RGB-only baseline. These results indicate that sonar-domain supervision functions as an auxiliary structural constraint during optimization, rather than a conventional multi-source data enlargement. By forcing a shared-parameter detector to fit a texture-poor, geometry-dominant sonar domain, the learned representation is biased away from color/texture shortcuts and becomes more stable under adverse underwater degradations, without increasing deployment complexity. Full article
(This article belongs to the Special Issue Advances in Marine Autonomous Vehicles)
Show Figures

Figure 1

22 pages, 4477 KB  
Article
Robust Detection and Localization of Image Copy-Move Forgery Using Multi-Feature Fusion
by Kaiqi Lu and Qiuyu Zhang
J. Imaging 2026, 12(2), 75; https://doi.org/10.3390/jimaging12020075 - 10 Feb 2026
Viewed by 340
Abstract
Copy-move forgery detection (CMFD) is a crucial image forensics analysis technique. The rapid development of deep learning algorithms has led to impressive advancements in CMFD. However, existing models suffer from two key limitations: Their feature fusion modules insufficiently exploit the complementary nature of [...] Read more.
Copy-move forgery detection (CMFD) is a crucial image forensics analysis technique. The rapid development of deep learning algorithms has led to impressive advancements in CMFD. However, existing models suffer from two key limitations: Their feature fusion modules insufficiently exploit the complementary nature of features from the RGB domain and noise domain, resulting in suboptimal feature representations. During decoding, they simply classify pixels as authentic or forged, without aggregating cross-layer information or integrating local and global attention mechanisms, leading to unsatisfactory detection precision. To overcome these limitations, a robust detection and localization approach to image copy-move forgery using multi-feature fusion is proposed. Firstly, a Multi-Feature Fusion Network (MFFNet) was designed. Within its feature fusion module, features from both the RGB domain and noise domain were fused to enable mutual complementarity between distinct characteristics, yielding richer feature information. Then, a Lightweight Multi-layer Perceptron Decoder (LMPD) was developed for image reconstruction and forgery localization map generation. Finally, by aggregating information from different layers and combining local and global attention mechanisms, more accurate prediction masks were obtained. The experimental results demonstrate that the proposed MFFNet model exhibits enhanced robustness and superior detection and localization performance compared to existing methods when faced with JPEG compression, noise addition, and resizing operations. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

23 pages, 69888 KB  
Article
Patched-Based Swin Transformer Hyperprior for Learned Image Compression
by Sibusiso B. Buthelezi and Jules R. Tapamo
J. Imaging 2026, 12(1), 12; https://doi.org/10.3390/jimaging12010012 - 26 Dec 2025
Viewed by 496
Abstract
We present a hybrid end-to-end learned image compression framework that combines a CNN-based variational autoencoder (VAE) with an efficient hierarchical Swin Transformer to address the limitations of existing entropy models in capturing global dependencies under computational constraints. Traditional VAE-based codecs typically rely on [...] Read more.
We present a hybrid end-to-end learned image compression framework that combines a CNN-based variational autoencoder (VAE) with an efficient hierarchical Swin Transformer to address the limitations of existing entropy models in capturing global dependencies under computational constraints. Traditional VAE-based codecs typically rely on CNN-based priors with localized receptive fields, which are insufficient for modelling the complex, high-dimensional dependencies of the latent space, thereby limiting compression efficiency. While fully global transformer-based models can capture long-range dependencies, their high computational complexity makes them impractical for high-resolution image compression. To overcome this trade-off, our approach couples a CNN-based VAE with a patch-based hierarchical Swin Transformer hyperprior that employs shifted window self-attention to effectively model both local and global contextual information while maintaining computational efficiency. The proposed framework tightly integrates this expressive entropy model with an end-to-end differentiable quantization module, enabling joint optimization of the complete rate-distortion objective. By learning a more accurate probability distribution of the latent representation, the model achieves improved bitrate estimation and a more compact latent representation, resulting in enhanced compression performance. We validate our approach on the widely used Kodak, JPEG AI, and CLIC datasets, demonstrating that the proposed hybrid architecture achieves superior rate-distortion performance, delivering higher visual quality at lower bitrates compared to methods relying on simpler CNN-based entropy priors. This work demonstrates the effectiveness of integrating efficient transformer architectures into learned image compression and highlights their potential for advancing entropy modelling beyond conventional CNN-based designs. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

26 pages, 16853 KB  
Article
Semi-Fragile Watermarking Scheme for High-Resolution Color Images: Tamper Identification, Ownership Authentication, and Self-Recovery
by Manuel Cedillo-Hernandez, Antonio Cedillo-Hernandez, Francisco Javier Garcia-Ugalde and Juan Carlos Sanchez-Garcia
Algorithms 2026, 19(1), 28; https://doi.org/10.3390/a19010028 - 26 Dec 2025
Viewed by 570
Abstract
The advancements in communication and information technologies have substantially enabled the extensive distribution and modification of high-resolution color images. Although this accessibility provides many advantages, it also presents risks related to security. Specifically, when image modification is conducted with malicious intent, exceeding typical [...] Read more.
The advancements in communication and information technologies have substantially enabled the extensive distribution and modification of high-resolution color images. Although this accessibility provides many advantages, it also presents risks related to security. Specifically, when image modification is conducted with malicious intent, exceeding typical artistic or enhancement objectives, it can cause significant moral or economic harm to the image owner. To address this security requirement, this study presents an innovative semi-fragile watermarking algorithm designed specifically for high-resolution color images. The proposed method utilizes Discrete Cosine Transform domain watermarking implemented via Quantization Index Modulation with Dither Modulation. It incorporates several elements, such as convolutional encoding, a denoising convolutional neural network, and a very deep super-resolution neural network. This comprehensive strategy aims to provide ownership verification using a logo watermark, in conjunction with tamper detection and content self-recovery mechanisms. The self-recovery criterion is determined using a thumbnail image, created by downscaling to standard definition and applying JPEG2000 lossy compression. The resultant multifunctional design enhances the overall security of the information. Experimental validation confirms the enhanced imperceptibility, robustness, and capacity of the proposed method. Its efficacy was additionally corroborated through comparative analyses using contemporary state-of-the-art algorithms. Full article
Show Figures

Figure 1

23 pages, 10651 KB  
Article
Noise-Aware Hybrid Compression of Deep Models with Zero-Shot Denoising and Failure Prediction
by Lizhe Zhang, Quan Zhou, Ruihua Liu, Lang Huyan, Juanni Liu and Yi Zhang
Appl. Sci. 2025, 15(24), 12882; https://doi.org/10.3390/app152412882 - 5 Dec 2025
Viewed by 625
Abstract
Deep learning-based image compression achieves remarkable average rate-distortion performance but is prone to failure on noisy, high-frequency, or high-entropy inputs. This work systematically investigates these failure cases and proposes a noise-aware hybrid compression framework to address them. A High-Frequency Vulnerability Index (HFVI) is [...] Read more.
Deep learning-based image compression achieves remarkable average rate-distortion performance but is prone to failure on noisy, high-frequency, or high-entropy inputs. This work systematically investigates these failure cases and proposes a noise-aware hybrid compression framework to address them. A High-Frequency Vulnerability Index (HFVI) is proposed, integrating frequency energy, encoder Jacobian sensitivity, and texture entropy into a unified measure of degradation susceptibility. Guided by HFVI, the system incorporates a selective zero-shot denoising module (P2PA) and a lightweight hybrid codec selector that determines, for each image, whether P2PA is necessary and selecting the more reliable codec (a learning-based model or JPEG2000) accordingly, without retraining any compression backbones. Experiments span a 200,000-image cross-domain benchmark incorporating general datasets, synthetic noise (eight levels), and real-noise datasets demonstrate that the proposed pipeline improves PSNR by up to 1.28 dB, raises SSIM by 0.02, reduces LPIPS by roughly 0.05, and decreases the failure-case rate by 6.7% over the best baseline (Joint-IC). Additional intensity-profile and cross-validation analyses further validate the robustness and deployment readiness of the method, showing that the hybrid selector provides a practical path toward reliable, noise-adaptive deep image compression. Full article
Show Figures

Figure 1

17 pages, 3136 KB  
Article
A Robust Image Watermarking Scheme via Two-Stage Training and Differentiable JPEG Compression
by Hu Deng, Feng Chen, Pei Gan, Rongtao Liao and Xuehu Yan
Electronics 2025, 14(22), 4510; https://doi.org/10.3390/electronics14224510 - 18 Nov 2025
Cited by 1 | Viewed by 928
Abstract
Digital image watermarking is a vital tool for copyright protection and content authentication. However, most existing methods perform well only under single noise types, while real-world applications often involve composite noises with multiple distortions, leading to poor robustness. To address this issue, we [...] Read more.
Digital image watermarking is a vital tool for copyright protection and content authentication. However, most existing methods perform well only under single noise types, while real-world applications often involve composite noises with multiple distortions, leading to poor robustness. To address this issue, we propose a robust image watermarking scheme. To improve performance under combined noise conditions, a two-stage training strategy is introduced: in the first stage, noise intensity increases gradually to stabilize training; in the second stage, mixed strong noises are applied to enhance generalization against complex attacks. Specifically, a strength-balanced watermark optimization algorithm is employed during the testing stage to improve visual quality while maintaining strong robustness. Furthermore, to improve robustness against JPEG compression, we adopt a differentiable fine-grained JPEG module that accurately simulates real compression and enables gradient backpropagation during training. Experimental results demonstrate the superiority of the proposed method under various single and combined distortions. Under noise-free conditions, it achieves 0% bit error rate and 53.55 dB PSNR. Under composite distortions, our scheme maintains a low average BER of 2.40% and a PSNR of 42.70 dB. Full article
(This article belongs to the Topic Artificial Intelligence Models, Tools and Applications)
Show Figures

Figure 1

14 pages, 2365 KB  
Article
Seam Carving Forgery Detection Through Multi-Perspective Explainable AI
by Miguel José das Neves, Felipe Rodrigues Perche Mahlow, Renato Dias de Souza, Paulo Roberto G. Hernandes, José Remo Ferreira Brega and Kelton Augusto Pontara da Costa
J. Imaging 2025, 11(11), 416; https://doi.org/10.3390/jimaging11110416 - 18 Nov 2025
Viewed by 705
Abstract
This paper addresses the critical challenge of detecting content-aware image manipulations, specifically focusing on seam carving forgery. While deep learning models, particularly Convolutional Neural Networks (CNNs), have shown promise in this area, their black-box nature limits their trustworthiness in high-stakes domains like digital [...] Read more.
This paper addresses the critical challenge of detecting content-aware image manipulations, specifically focusing on seam carving forgery. While deep learning models, particularly Convolutional Neural Networks (CNNs), have shown promise in this area, their black-box nature limits their trustworthiness in high-stakes domains like digital forensics. To address this gap, we propose and validate a framework for interpretable forgery detection, termed E-XAI (Ensemble Explainable AI). Conceptually inspired by Ensemble Learning, our framework’s novelty lies not in combining predictive models, but in integrating a multi-perspective ensemble of explainability techniques. Specifically, we combine SHAP for fine-grained, pixel-level feature attribution with Grad-CAM for region-level localization to create a more robust and holistic interpretation of a single, custom-trained CNN’s decisions. Our approach is validated on a purpose-built, balanced, binary-class dataset of 10,300 images. The results demonstrate high classification performance on an unseen test set, with a 95% accuracy and a 99% precision for the forged class. Furthermore, we analyze the model’s robustness against JPEG compression, a common real-world perturbation. More importantly, the application of the E-XAI framework reveals how the model identifies subtle forgery artifacts, providing transparent, visual evidence for its decisions. This work contributes a robust end-to-end pipeline for interpretable image forgery detection, enhancing the trust and reliability of AI systems in information security. Full article
Show Figures

Figure 1

28 pages, 16265 KB  
Article
ADPGAN: Anti-Compression Attention-Based Diffusion Pattern Steganography Model Using GAN
by Zhen-Qiang Chen, Yu-Hang Huang, Xin-Yuan Chen and Sio-Long Lo
Electronics 2025, 14(22), 4426; https://doi.org/10.3390/electronics14224426 - 13 Nov 2025
Viewed by 768
Abstract
Image steganography is often employed in information security and confidential communications, yet it typically faces challenges of imperceptibility and robustness during transmission. Meanwhile, insufficient attention has been paid to preserving the quality of the secret image after JPEG compression at the receiver, which [...] Read more.
Image steganography is often employed in information security and confidential communications, yet it typically faces challenges of imperceptibility and robustness during transmission. Meanwhile, insufficient attention has been paid to preserving the quality of the secret image after JPEG compression at the receiver, which limits the effectiveness of steganography. In this study, we propose an anti-compression attention-based diffusion pattern steganography model using GAN (ADPGAN). ADPGAN leverages dense connectivity to fuse shallow and deep image features with secret data, achieving high robustness against JPEG compression. Meanwhile, an enhanced attention module and a discriminator are employed to minimize image distortion caused by data embedding, thereby significantly improving the imperceptibility of the host image. Based on ADPGAN, we propose a novel JPEG-compression-resistant image framework that improves the quality of the recovered image by ensuring that the degradation of the reconstructed image primarily stems from sampling rather than JPEG compression. Unlike direct embedding of full-size secret images, we downsample the secret image into a secret data stream and embed it into the cover image via ADPGAN, demonstrating high distortion resistance and high-fidelity recovery of the secret image. Ablation studies validate the effectiveness of ADPGAN, achieving a 0-bit error rate (BER) under JPEG compression at a quality factor of 20, yielding an average Peak Signal-to-Noise Ratio (PSNR) of 39.70 dB for the recovered images. Full article
(This article belongs to the Section Electronic Multimedia)
Show Figures

Figure 1

35 pages, 7763 KB  
Article
Cryptosystem for JPEG Images with Encryption Before and After Lossy Compression
by Manuel Alejandro Cardona-López, Juan Carlos Chimal-Eguía, Víctor Manuel Silva-García and Rolando Flores-Carapia
Mathematics 2025, 13(21), 3482; https://doi.org/10.3390/math13213482 - 31 Oct 2025
Cited by 2 | Viewed by 730
Abstract
JPEG images are widely used in multimedia transmission, such as on social media platforms, owing to their efficiency for reducing storage and transmission requirements. However, because such images may contain sensitive information, encryption is essential to ensure data privacy. Traditional image encryption schemes [...] Read more.
JPEG images are widely used in multimedia transmission, such as on social media platforms, owing to their efficiency for reducing storage and transmission requirements. However, because such images may contain sensitive information, encryption is essential to ensure data privacy. Traditional image encryption schemes face challenges when applied to JPEG images, as maintaining compatibility with the JPEG structure and managing the effects of lossy compression can distort encrypted data. Existing JPEG-compatible encryption methods, such as Encryption-then-Compression (EtC) and Compression-then-Encryption (CtE), typically employ a single encryption stage, either before or after compression, and often involve trade-offs between security, storage efficiency, and visual quality. In this work, an Encryption–Compression–Encryption algorithm is presented that preserves full JPEG compatibility while combining the advantages of both EtC and CtE schemes. In the proposed method, pixel-block encryption is first applied prior to JPEG compression, followed by selective coefficient encryption after compression, in which the quantized DC coefficient differences are permuted. Experimental results indicate that the second encryption stage enhances the entropy achieved in the first stage, with both stages complementing each other in terms of resistance to attacks. The addition of this second layer does not significantly impact storage efficiency or the visual quality of the decompressed image; however, it introduces a moderate increase in computational time due to the two-stage encryption process. Full article
(This article belongs to the Special Issue Applied Cryptography and Information Security with Application)
Show Figures

Figure 1

20 pages, 934 KB  
Article
Non-Uniform Entropy-Constrained L Quantization for Sparse and Irregular Sources
by Alin-Adrian Alecu, Mohammad Ali Tahouri, Adrian Munteanu and Bujor Păvăloiu
Entropy 2025, 27(11), 1126; https://doi.org/10.3390/e27111126 - 31 Oct 2025
Viewed by 724
Abstract
Near-lossless coding schemes traditionally rely on uniform quantization to control the maximum absolute error (L norm) of residual signals, often assuming a parametric model for the source distribution. This paper introduces a novel design framework for non-uniform, entropy-aware L-oriented [...] Read more.
Near-lossless coding schemes traditionally rely on uniform quantization to control the maximum absolute error (L norm) of residual signals, often assuming a parametric model for the source distribution. This paper introduces a novel design framework for non-uniform, entropy-aware L-oriented scalar quantizers that leverages a tight and differentiable approximation of the L distortion metric and does not require any parametric density function formulations. The framework is evaluated on both synthetic parametric sources and real-world medical depth map video datasets. For smoothly decaying distributions, such as the continuous Laplacian or discrete two-sided geometric distributions, the proposed method naturally converges to near-uniform quantizers, consistent with theoretical expectations. In contrast, for sparse or irregular sources, the algorithm produces highly non-uniform bin allocations that adapt to the local distribution structure and improve rate-distortion efficiency. When embedded in a residual-based near-lossless compression scheme, the resulting codec consistently outperforms versions equipped with uniform or piecewise-uniform quantizers, as well as state-of-the-art near-lossless schemes such as JPEG-LS and CALIC. Full article
(This article belongs to the Special Issue Information Theory and Data Compression)
Show Figures

Figure 1

32 pages, 3306 KB  
Article
AMSEANet: An Edge-Guided Adaptive Multi-Scale Network for Image Splicing Detection and Localization
by Yuankun Yang, Yueshun He, Xiaohui Ma, Wei Lv, Jie Chen and Hongling Wang
Sensors 2025, 25(20), 6494; https://doi.org/10.3390/s25206494 - 21 Oct 2025
Viewed by 1040
Abstract
In image splicing tamper detection, forgery operations simultaneously introduce macroscopic semantic inconsistencies and microscopic tampering artifacts. Conventional methods often treat semantic understanding and low-level artifact perception as separate tasks, which impedes their effective synergy. Meanwhile, frequency-domain information, a crucial clue for identifying traces [...] Read more.
In image splicing tamper detection, forgery operations simultaneously introduce macroscopic semantic inconsistencies and microscopic tampering artifacts. Conventional methods often treat semantic understanding and low-level artifact perception as separate tasks, which impedes their effective synergy. Meanwhile, frequency-domain information, a crucial clue for identifying traces of tampering, is frequently overlooked. However, a simplistic fusion of frequency-domain and spatial features can lead to feature conflicts and information redundancy. To resolve these challenges, this paper proposes an Adaptive Multi-Scale Edge-Aware Network (AMSEANet). This network employs a synergistic enhancement cascade architecture, recasting semantic understanding and artifact perception as a single, frequency-aware process guided by deep semantics. It leverages data-driven adaptive filters to precisely isolate and focus on edge artifacts that signify tampering. Concurrently, the dense fusion and enhancement of cross-scale features effectively preserve minute tampering clues and boundary details. Extensive experiments demonstrate that our proposed method achieves superior performance on several public datasets and exhibits excellent robustness against common attacks, such as noise and JPEG compression. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

13 pages, 317 KB  
Article
Enhancing JPEG XL’s Weighted Average Predictor: Genetic Algorithm Optimization of Expanded Sub-Predictor Ensemble
by Xavier Hill Roy and Mahmoud R. El-Sakka
Electronics 2025, 14(20), 4116; https://doi.org/10.3390/electronics14204116 - 21 Oct 2025
Viewed by 716
Abstract
Lossless image compression relies heavily on prediction algorithms to reduce spatial redundancy before entropy coding. The JPEG XL standard employs a weighted average predictor that combines four sub-predictors with adaptive weighting; however, it uses fixed initial scaling factors regardless of the image content. [...] Read more.
Lossless image compression relies heavily on prediction algorithms to reduce spatial redundancy before entropy coding. The JPEG XL standard employs a weighted average predictor that combines four sub-predictors with adaptive weighting; however, it uses fixed initial scaling factors regardless of the image content. This study introduces WOP8 (weighted optimization predictor for 8 sub-predictors), which extends the predictor diversity and optimizes initial weights using a genetic algorithm. Four additional predictors were incorporated—adaptive MED (JPEG-LS), enhanced adaptive median, Paeth (PNG), and GAP-based (CALIC)—forming an eight-predictor ensemble. A genetic algorithm with a population of 30 and 24 generations optimized the weight configurations by minimizing the compressed file size of the training data. Experiments were conducted on the Kodak and Tecnick datasets to evaluate performance and generalizability. The Kodak color dataset showed notable gains: with the weighted average predictor in isolation, WOP8 achieved a 0.24 BPP reduction (2.7% improvement) at high effort levels. Under standard JPEG XL operation mode, improvements were minor but consistent. These results confirm the value of targeted predictor optimization and demonstrate that genetic algorithms can effectively discover dataset-specific weighting patterns, offering a foundation for future component-level enhancements in JPEG XL. Full article
(This article belongs to the Special Issue Artificial Intelligence in Graphics and Images)
Show Figures

Figure 1

Back to TopTop