Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (43)

Search Parameters:
Keywords = state codebook

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 2669 KB  
Article
Extensible Heterogeneous Collaborative Perception in Autonomous Vehicles with Codebook Compression
by Babak Ebrahimi Soorchaei, Arash Raftari and Yaser Pourmohammadi Fallah
Robotics 2025, 14(12), 186; https://doi.org/10.3390/robotics14120186 - 10 Dec 2025
Viewed by 99
Abstract
Collaborative perception can mitigate occlusion and range limitations in autonomous driving, but deployment remains constrained by strict bandwidth budgets and heterogeneous agent stacks. We propose a communication-efficient and backbone-agnostic framework in which each agent’s encoder is treated as a black box, and a [...] Read more.
Collaborative perception can mitigate occlusion and range limitations in autonomous driving, but deployment remains constrained by strict bandwidth budgets and heterogeneous agent stacks. We propose a communication-efficient and backbone-agnostic framework in which each agent’s encoder is treated as a black box, and a lightweight interpreter maps its intermediate features into a canonical space. To reduce transmission cost, we integrate codebook-based compression that sends only compact discrete indices, while a prompt-guided decoder reconstructs semantically aligned features on the ego vehicle for downstream fusion. Training follows a two-phase strategy: Phase 1 jointly optimizes interpreters, prompts, and fusion components for a fixed set of agents; Phase 2 enables plug-and-play onboarding of new agents by tuning only their specific prompts. Experiments on OPV2V and OPV2VH+ show that our method consistently outperformed early-, intermediate-, and late-fusion baselines under equal or lower communication budgets. With a codebook of size 128, the proposed pipeline preserved over 95% of the uncompressed detection accuracy while reducing communication cost by more than two orders of magnitude. The model also maintained strong performance under bandwidth throttling, missing-agent scenarios, and heterogeneous sensor combinations. Compared to recent state-of-the-art methods such as PolyInter, MPDA, and PnPDA, our framework achieved higher AP while using significantly smaller message sizes. Overall, the combination of prompt-guided decoding and discrete Codebook compression provides a scalable, bandwidth-aware, and heterogeneity-resilient foundation for next-generation collaborative perception in connected autonomous vehicles. Full article
Show Figures

Figure 1

22 pages, 4609 KB  
Article
Statistical CSI-Based Beamspace Transmission for Massive MIMO LEO Satellite Communications
by Qian Dong, Yafei Wang, Nan Hu, Yiming Zhu, Wenjin Wang and Li Chai
Entropy 2025, 27(12), 1214; https://doi.org/10.3390/e27121214 - 28 Nov 2025
Viewed by 308
Abstract
In multibeam low-Earth-orbit (LEO) satellite systems, precoding has emerged as a key technology for mitigating co-channel interference (CCI) and for improving spectral efficiency (SE). However, its practical implementation is challenged by the difficulty of acquiring reliable instantaneous channel state information (iCSI) and by [...] Read more.
In multibeam low-Earth-orbit (LEO) satellite systems, precoding has emerged as a key technology for mitigating co-channel interference (CCI) and for improving spectral efficiency (SE). However, its practical implementation is challenged by the difficulty of acquiring reliable instantaneous channel state information (iCSI) and by the high computational complexity induced by large-scale antenna arrays, making it incompatible with fixed codebook-based beamforming schemes commonly adopted in operational systems. In this analysis, we propose a beamspace transmission framework leveraging statistical CSI (sCSI) and achieves reduced computational complexity compared with antenna-domain precoding designs. Specifically, we first propose a low-complexity beam selection algorithm that selects a small subset of beams for each user terminal (UT) from a fixed beamforming codebook, using only the UTs’ two-dimensional (2D) angular information. To suppress CCI among beams, we then derive a beamspace weighted minimum mean square error (WMMSE) precoding scheme based on the equivalent beamspace channel matrix. The derivation employs an sCSI-based WMMSE (sWMMSE) formulation derived from an upper bound approximation of the ergodic sum rate, which provides a tighter estimate than the expected mean square error (MSE)-based lower bound approximation. Simulation results demonstrate that the proposed sCSI-based beamspace transmission scheme achieves a favorable trade-off between performance and computational complexity. Full article
(This article belongs to the Topic Advances in Sixth Generation and Beyond (6G&B))
Show Figures

Figure 1

9 pages, 216 KB  
Article
Discussion of the Fetus in Fetal Cardiology Consultations: A Qualitative Study
by Samantha Syme, Kelsey Schweiberger, Judy C. Chang, Ann Kavanaugh-McHugh, Abdesalam Soudi, Justin T. Clapp, Nadine A. Kasparian, Robert M. Arnold and Kelly W. Harris
Children 2025, 12(9), 1211; https://doi.org/10.3390/children12091211 - 10 Sep 2025
Viewed by 1037
Abstract
Background: While prenatal diagnosis of congenital heart disease is increasingly common, and communication is essential to minimizing familial stress, little is known about how the fetus is discussed in this setting. This study observed how clinicians and families refer to the fetus during [...] Read more.
Background: While prenatal diagnosis of congenital heart disease is increasingly common, and communication is essential to minimizing familial stress, little is known about how the fetus is discussed in this setting. This study observed how clinicians and families refer to the fetus during initial fetal cardiology consultations. Methods: Initial fetal cardiology consultations from one institution were recorded and transcribed verbatim. A codebook was developed and used to code the transcripts. Codes included any reference to the fetus and any attribution of agency or mental states to the fetus. Results: Nineteen consultations performed by five clinicians from one academic institution were included. Clinicians and families most frequently referred to the fetus using personal terminology (e.g., third-person pronouns, a given name, or “son” or “daughter”). Impersonal terminology (e.g., “baby”) was used less frequently, followed by medical terminology (e.g., “fetus”), which was only used in two consultations. In about half of the consultations, clinicians conferred agency or mental states on the fetus by attributing actions, emotions, or knowledge to the fetus. Conclusions: Fetal cardiology clinicians primarily use personal terminology when referring to the fetus during initial consultations. Familial preferences need to be evaluated to optimize communication and support. Full article
(This article belongs to the Section Pediatric Cardiology)
35 pages, 954 KB  
Article
Beyond Manual Media Coding: Evaluating Large Language Models and Agents for News Content Analysis
by Stavros Doropoulos, Elisavet Karapalidou, Polychronis Charitidis, Sophia Karakeva and Stavros Vologiannidis
Appl. Sci. 2025, 15(14), 8059; https://doi.org/10.3390/app15148059 - 20 Jul 2025
Cited by 3 | Viewed by 2126
Abstract
The vast volume of media content, combined with the costs of manual annotation, challenges scalable codebook analysis and risks reducing decision-making accuracy. This study evaluates the effectiveness of large language models (LLMs) and multi-agent teams in structured media content analysis based on codebook-driven [...] Read more.
The vast volume of media content, combined with the costs of manual annotation, challenges scalable codebook analysis and risks reducing decision-making accuracy. This study evaluates the effectiveness of large language models (LLMs) and multi-agent teams in structured media content analysis based on codebook-driven annotation. We construct a dataset of 200 news articles on U.S. tariff policies, manually annotated using a 26-question codebook encompassing 122 distinct codes, to establish a rigorous ground truth. Seven state-of-the-art LLMs, spanning low- to high-capacity tiers, are assessed under a unified zero-shot prompting framework incorporating role-based instructions and schema-constrained outputs. Experimental results show weighted global F1-scores between 0.636 and 0.822, with Claude-3-7-Sonnet achieving the highest direct-prompt performance. To examine the potential of agentic orchestration, we propose and develop a multi-agent system using Meta’s Llama 4 Maverick, incorporating expert role profiling, shared memory, and coordinated planning. This architecture improves the overall F1-score over the direct prompting baseline from 0.757 to 0.805 and demonstrates consistent gains across binary, categorical, and multi-label tasks, approaching commercial-level accuracy while maintaining a favorable cost–performance profile. These findings highlight the viability of LLMs, both in direct and agentic configurations, for automating structured content analysis. Full article
(This article belongs to the Special Issue Natural Language Processing in the Era of Artificial Intelligence)
Show Figures

Figure 1

32 pages, 1277 KB  
Article
Distributed Prediction-Enhanced Beamforming Using LR/SVR Fusion and MUSIC Refinement in 5G O-RAN Systems
by Mustafa Mayyahi, Jordi Mongay Batalla, Jerzy Żurek and Piotr Krawiec
Appl. Sci. 2025, 15(13), 7428; https://doi.org/10.3390/app15137428 - 2 Jul 2025
Viewed by 971
Abstract
Low-latency and robust beamforming are vital for sustaining signal quality and spectral efficiency in emerging high-mobility 5G and future 6G wireless networks. Conventional beam management approaches, which rely on periodic Channel State Information feedback and static codebooks, as outlined in 3GPP standards, are [...] Read more.
Low-latency and robust beamforming are vital for sustaining signal quality and spectral efficiency in emerging high-mobility 5G and future 6G wireless networks. Conventional beam management approaches, which rely on periodic Channel State Information feedback and static codebooks, as outlined in 3GPP standards, are insufficient in rapidly varying propagation environments. In this work, we propose a Dominance-Enforced Adaptive Clustered Sliding Window Regression (DE-ACSW-R) framework for predictive beamforming in O-RAN Split 7-2x architectures. DE-ACSW-R leverages a sliding window of recent angle of arrival (AoA) estimates, applying in-window change-point detection to segment user trajectories and performing both Linear Regression (LR) and curvature-adaptive Support Vector Regression (SVR) for short-term and non-linear prediction. A confidence-weighted fusion mechanism adaptively blends LR and SVR outputs, incorporating robust outlier detection and a dominance-enforced selection regime to address strong disagreements. The Open Radio Unit (O-RU) autonomously triggers localised MUSIC scans when prediction confidence degrades, minimising unnecessary full-spectrum searches and saving delay. Simulation results demonstrate that the proposed DE-ACSW-R approach significantly enhances AoA tracking accuracy, beamforming gain, and adaptability under realistic high-mobility conditions, surpassing conventional LR/SVR baselines. This AI-native modular pipeline aligns with O-RAN architectural principles, enabling scalable and real-time beam management for next-generation wireless deployments. Full article
Show Figures

Figure 1

17 pages, 688 KB  
Article
Task-Based Quantizer for CSI Feedback in Multi-User MISO VLC/RF Systems
by Fugui He, Congcong Wang, Yao Nie, Xianglin Fan, Chensitian Zhang and Yang Yang
Electronics 2025, 14(11), 2277; https://doi.org/10.3390/electronics14112277 - 3 Jun 2025
Viewed by 797
Abstract
The performance of multiple-input single-output (MISO) transmission is highly dependent on the accuracy of the channel state information (CSI) at the base station (BS), which necessitates precise CSI estimation and reliable feedback from the user equipment. However, the overhead of the CSI feedback [...] Read more.
The performance of multiple-input single-output (MISO) transmission is highly dependent on the accuracy of the channel state information (CSI) at the base station (BS), which necessitates precise CSI estimation and reliable feedback from the user equipment. However, the overhead of the CSI feedback occupies substantial uplink bandwidth resources. To alleviate the overhead, this paper proposes a novel task-based quantizer for uplink MISO visible light communication (VLC) systems. In particular, a hybrid radio frequency (RF)/VLC system is considered, where VLC links are mainly used for large-volume downlink transmissions and RF links are used for uplink CSI feedback. Since the RF bandwidth resources are limited, the CSI is quantified to reduce the uplink resource requirements, which, however, inevitably causes CSI estimation errors at the BS. To guarantee the CSI estimation accuracy while minimizing the RF resource cost, a task-based quantization scheme for channel estimation (TQ-CE) is proposed. In the TQ-CE, both the quantized codebook and the post-processing matrix are optimized to minimize the mean square error (MSE) of the channel estimation. Taking the minimum MSE as the target task, the TQ-CE leverages vector quantization (VQ) to generate a codebook, which is designed to reduce the feedback overhead without compromising the precision of the channel estimation. Then, an optimal closed-form solution of the post-processing matrix is derived based on the minimum mean square error (MMSE) criterion. The simulation results demonstrate that the proposed TQ-CE achieved 0.25Mbit/s and 0.62Mbit/s higher data rates compared with the conventional scalar quantization-based channel estimation (SQ-CE) schemes and vector quantization-based channel estimation (VQ-CE) schemes, respectively. Moreover, in terms of the feedback overhead, compared with the 18-bit SQ-CE, the 4-bit TQ-CE achieved a 22.2% reduction in uplink bits. Full article
Show Figures

Figure 1

19 pages, 2588 KB  
Article
Multi-User MIMO Downlink Precoding with Dynamic User Selection for Limited Feedback
by Mikhail Bakulin, Taoufik Ben Rejeb, Vitaly Kreyndelin, Denis Pankratov and Aleksei Smirnov
Sensors 2025, 25(3), 866; https://doi.org/10.3390/s25030866 - 31 Jan 2025
Cited by 2 | Viewed by 1787
Abstract
In modern (5G) and future Multi-User (MU) wireless communication systems Beyond 5G (B5G) using Multiple-Input Multiple-Output (MIMO) technology, base stations with a large number of antennas communicate with many mobile stations. This technology is becoming especially relevant in modern multi-user wireless sensor networks [...] Read more.
In modern (5G) and future Multi-User (MU) wireless communication systems Beyond 5G (B5G) using Multiple-Input Multiple-Output (MIMO) technology, base stations with a large number of antennas communicate with many mobile stations. This technology is becoming especially relevant in modern multi-user wireless sensor networks in various application scenarios. The problem of organizing an MU mode on the downlink has arisen, which can be solved by precoding at the Base Station (BS) without using additional channel frequency–time resources. In order to utilize an efficient precoding algorithm at the base station, full Channel State Information (CSI) is needed for each mobile station. Transmitting this information for massive MIMO systems normally requires the allocation of high-speed channel resources for the feedback. With limited feedback, reduced information (partial CSI) is used, for example, the codeword from the codebook that is closest to the estimated channel vector (or matrix). Incomplete (or inaccurate) CSI causes interference from the signals, transmitted to neighboring mobile stations, that ultimately results in a decrease in the number of active users served. In this paper, we propose a new downlink precoding approach for MU-MIMO systems that also uses codebooks to reduce the information transmitted over a feedback channel. A key aspect of the proposed approach, in contrast to the existing ones, is the transmission of new, uncorrelated information in each cycle, which allows for accumulating CSI with higher accuracy without increasing the feedback overhead. The proposed approach is most effective in systems with dynamic user selection. In such systems, increasing the accuracy of CSI leads to an increase in the number of active users served, which after a few cycles, can reach a maximum value determined by the number of transmit antennas at the BS side. This approach appears to be promising for addressing the challenges associated with current and future massive MIMO systems, as evidenced by our statistical simulation results. Various methods for extracting and transmitting such uncorrelated information over a feedback channel are considered. In many known publications, the precoder, codebooks, CSI estimation methods and other aspects of CSI transmission over a feedback channel are separately optimized, but a comprehensive approach to jointly solving these problems has not yet been developed. In our paper, we propose to fill this gap by combining a new approach of precoding and CSI estimation with CSI accumulation and transmission over a feedback channel. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

20 pages, 7907 KB  
Article
Channel Code-Book (CCB): Semantic Image-Adaptive Transmission in Satellite–Ground Scenario
by Hui Cao, Shujun Han, Rui Meng, Xiaodong Xu and Ping Zhang
Sensors 2025, 25(1), 269; https://doi.org/10.3390/s25010269 - 6 Jan 2025
Viewed by 2252
Abstract
Satellite–ground communication is a critical component in the global communication system, significantly contributing to environmental monitoring, radio and television broadcasting, aerospace operations, and other domains. However, the technology encounters challenges in data transmission efficiency, due to the drastic alterations in the communication channel [...] Read more.
Satellite–ground communication is a critical component in the global communication system, significantly contributing to environmental monitoring, radio and television broadcasting, aerospace operations, and other domains. However, the technology encounters challenges in data transmission efficiency, due to the drastic alterations in the communication channel caused by the rapid movement of satellites. In comparison to traditional transmission methods, semantic communication (SemCom) technology enhances transmission efficiency by comprehending and leveraging the intrinsic meaning of information, making it ideal for image transmission in satellite communications. Nevertheless, current SemCom methods still struggle to adapt to varying channel conditions. To address this, we propose a SemCom transmission model based on a Channel Code-Book (CCB) for adaptive image transmission in diverse channel environments. Our model reconstructs and restores the original image by documenting fading and noise states under various channel conditions and dynamically adjusting the denoiser’s model parameters. Extensive experimental results demonstrate that our CCB model outperforms three representative baseline models, including Deep JSCC, ASCN, and WITT in various environments and task conditions, achieving an advantage of more than 10 dB under high signal-to-noise ratio conditions. Full article
Show Figures

Figure 1

40 pages, 4416 KB  
Review
A Review on Millimeter-Wave Hybrid Beamforming for Wireless Intelligent Transport Systems
by Waleed Shahjehan, Rajkumar Singh Rathore, Syed Waqar Shah, Mohammad Aljaidi, Ali Safaa Sadiq and Omprakash Kaiwartya
Future Internet 2024, 16(9), 337; https://doi.org/10.3390/fi16090337 - 14 Sep 2024
Cited by 16 | Viewed by 8676
Abstract
As the world braces for an era of ubiquitous and seamless connectivity, hybrid beamforming stands out as a beacon guiding the evolutionary path of wireless communication technologies. Several hybrid beamforming technologies are explored for millimeter-wave multiple-input multi-output (MIMO) communication. The aim is to [...] Read more.
As the world braces for an era of ubiquitous and seamless connectivity, hybrid beamforming stands out as a beacon guiding the evolutionary path of wireless communication technologies. Several hybrid beamforming technologies are explored for millimeter-wave multiple-input multi-output (MIMO) communication. The aim is to provide a roadmap for hybrid beamforming that enhances wireless fidelity. In this systematic review, a detailed literature review of algorithms/techniques used in hybrid beamforming along with performance metrics, characteristics, limitations, as well as performance evaluations are provided to enable communication compatible with modern Wireless Intelligent Transport Systems (WITSs). Further, an in-depth analysis of the mmWave hybrid beamforming landscape is provided based on user, link, band, scattering, structure, duplex, carrier, network, applications, codebook, and reflecting intelligent surfaces to optimize system design and performance across diversified user scenarios. Furthermore, the current research trends for hybrid beamforming are provided to enable the development of advanced wireless communication systems with optimized performance and efficiency. Finally, challenges, solutions, and future research directions are provided so that this systematic review can serve as a touchstone for academics and industry professionals alike. The systematic review aims to equip researchers with a deep understanding of the current state of the art and thereby enable the development of next-generation communication in WITSs that are not only adept at coping with contemporary demands but are also future-proofed to assimilate upcoming trends and innovations. Full article
Show Figures

Figure 1

19 pages, 6601 KB  
Article
An Innovative Recompression Scheme for VQ Index Tables
by Yijie Lin, Jui-Chuan Liu, Ching-Chun Chang and Chin-Chen Chang
Future Internet 2024, 16(8), 297; https://doi.org/10.3390/fi16080297 - 19 Aug 2024
Cited by 2 | Viewed by 1078
Abstract
As we move into the digital era, the pace of technological advancement is accelerating rapidly. Network traffic often becomes congested during the transmission of large data volumes. To mitigate this, data compression plays a crucial role in minimizing transmitted data. Vector quantization (VQ) [...] Read more.
As we move into the digital era, the pace of technological advancement is accelerating rapidly. Network traffic often becomes congested during the transmission of large data volumes. To mitigate this, data compression plays a crucial role in minimizing transmitted data. Vector quantization (VQ) stands out as a potent compression technique where each image block is encoded independently as an index linked to a codebook, effectively reducing the bit rate. In this paper, we introduce a novel scheme for recompressing VQ indices, enabling lossless restoration of the original indices during decoding without compromising visual quality. Our method not only considers pixel correlations within each image block but also leverages correlations between neighboring blocks, further optimizing the bit rate. The experimental results demonstrated the superior performance of our approach over existing methods. Full article
Show Figures

Figure 1

21 pages, 2816 KB  
Article
Adaptive Hybrid Beamforming Codebook Design Using Multi-Agent Reinforcement Learning for Multiuser Multiple-Input–Multiple-Output Systems
by Manasjyoti Bhuyan, Kandarpa Kumar Sarma, Debashis Dev Misra, Koushik Guha and Jacopo Iannacci
Appl. Sci. 2024, 14(16), 7109; https://doi.org/10.3390/app14167109 - 13 Aug 2024
Cited by 1 | Viewed by 3375
Abstract
This paper presents a novel approach to designing beam codebooks for downlink multiuser hybrid multiple-input–multiple-output (MIMO) wireless communication systems, leveraging multi-agent reinforcement learning (MARL). The primary objective is to develop an environment-specific beam codebook composed of non-interfering beams, learned by cooperative agents within [...] Read more.
This paper presents a novel approach to designing beam codebooks for downlink multiuser hybrid multiple-input–multiple-output (MIMO) wireless communication systems, leveraging multi-agent reinforcement learning (MARL). The primary objective is to develop an environment-specific beam codebook composed of non-interfering beams, learned by cooperative agents within the MARL framework. Machine learning (ML)-based beam codebook design for downlink communications have been based on channel state information (CSI) feedback or only reference signal received power (RSRP), consisting of an offline training and user clustering phase. In massive MIMO, the full CSI feedback data is of large size and is resource-intensive to process, making it challenging to implement efficiently. RSRP alone for a stand-alone base station is not a good marker of the position of a receiver. Hence, in this work, uplink CSI estimated at the base station along with feedback of RSRP and binary acknowledgment of the accuracy of received data is utilized to design the beamforming codebook at the base station. Simulations using sub-array antenna and ray-tracing channel demonstrate the proposed system’s ability to learn topography-aware beam codebook for arbitrary beams serving multiple user groups simultaneously. The proposed method extends beyond mono-lobe and fixed beam architectures by dynamically adapting arbitrary shaped beams to avoid inter-beam interference, enhancing the overall system performance. This work leverages MARL’s potential in creating efficient beam codebooks for hybrid MIMO systems, paving the way for enhanced multiuser communication in future wireless networks. Full article
(This article belongs to the Special Issue New Challenges in MIMO Communication Systems)
Show Figures

Figure 1

31 pages, 28677 KB  
Article
Color Image Encryption Based on an Evolutionary Codebook and Chaotic Systems
by Yuan Cao and Yinglei Song
Entropy 2024, 26(7), 597; https://doi.org/10.3390/e26070597 - 12 Jul 2024
Cited by 2 | Viewed by 1586
Abstract
Encryption of images is an important method that can effectively improve the security and privacy of crucial image data. Existing methods generally encrypt an image with a combination of scrambling and encoding operations. Currently, many applications require highly secure results for image encryption. [...] Read more.
Encryption of images is an important method that can effectively improve the security and privacy of crucial image data. Existing methods generally encrypt an image with a combination of scrambling and encoding operations. Currently, many applications require highly secure results for image encryption. New methods that can achieve improved randomness for both the scrambling and encoding processes in encryption are thus needed to further enhance the security of a cipher image. This paper proposes a new method that can securely encrypt color images. As the first step of the proposed method, a complete bit-level operation is utilized to scramble the binary bits in a color image to a full extent. For the second step, the bits in the scrambled image are processed with a sweeping operation to improve the encryption security. In the final step of encryption, a codebook that varies with evolutionary operations based on several chaotic systems is utilized to encrypt the partially encrypted image obtained in the second step. Experimental results on benchmark color images suggest that this new approach can securely encrypt color images and generate cipher images that remain secure under different types of attacks. The proposed approach is compared with several other state-of-the-art encryption approaches and the results show that it can achieve improved encryption security for cipher images. Experimental results thus suggest that this new approach can possibly be utilized practically in applications where color images need to be encrypted for content protection. Full article
Show Figures

Figure 1

17 pages, 5104 KB  
Article
Hierarchical Vector-Quantized Variational Autoencoder and Vector Credibility Mechanism for High-Quality Image Inpainting
by Cheng Li, Dan Xu and Kuai Chen
Electronics 2024, 13(10), 1852; https://doi.org/10.3390/electronics13101852 - 9 May 2024
Cited by 2 | Viewed by 3699
Abstract
Image inpainting infers the missing areas of a corrupted image according to the information of the undamaged part. Many existing image inpainting methods can generate plausible inpainted results from damaged images with the fast-developed deep-learning technology. However, they still suffer from over-smoothed textures [...] Read more.
Image inpainting infers the missing areas of a corrupted image according to the information of the undamaged part. Many existing image inpainting methods can generate plausible inpainted results from damaged images with the fast-developed deep-learning technology. However, they still suffer from over-smoothed textures or textural distortion in the cases of complex textural details or large damaged areas. To restore textures at a fine-grained level, we propose an image inpainting method based on a hierarchical VQ-VAE with a vector credibility mechanism. It first trains the hierarchical VQ-VAE with ground truth images to update two codebooks and to obtain two corresponding vector collections containing information on ground truth images. The two vector collections are fed to a decoder to generate the corresponding high-fidelity outputs. An encoder then is trained with the corresponding damaged image. It generates vector collections approximating the ground truth by the help of the prior knowledge provided by the codebooks. After that, the two vector collections pass through the decoder from the hierarchical VQ-VAE to produce the inpainted results. In addition, we apply a vector credibility mechanism to promote vector collections from damaged images and approximate vector collections from ground truth images. To further improve the inpainting result, we apply a refinement network, which uses residual blocks with different dilation rates to acquire both global information and local textural details. Extensive experiments conducted on several datasets demonstrate that our method outperforms the state-of-the-art ones. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Image and Video Processing)
Show Figures

Figure 1

13 pages, 310 KB  
Article
Lossy Compression of Individual Sequences Revisited: Fundamental Limits of Finite-State Encoders
by Neri Merhav
Entropy 2024, 26(2), 116; https://doi.org/10.3390/e26020116 - 28 Jan 2024
Cited by 2 | Viewed by 1692
Abstract
We extend Ziv and Lempel’s model of finite-state encoders to the realm of lossy compression of individual sequences. In particular, the model of the encoder includes a finite-state reconstruction codebook followed by an information lossless finite-state encoder that compresses the reconstruction codeword with [...] Read more.
We extend Ziv and Lempel’s model of finite-state encoders to the realm of lossy compression of individual sequences. In particular, the model of the encoder includes a finite-state reconstruction codebook followed by an information lossless finite-state encoder that compresses the reconstruction codeword with no additional distortion. We first derive two different lower bounds to the compression ratio, which depend on the number of states of the lossless encoder. Both bounds are asymptotically achievable by conceptually simple coding schemes. We then show that when the number of states of the lossless encoder is large enough in terms of the reconstruction block length, the performance can be improved, sometimes significantly so. In particular, the improved performance is achievable using a random-coding ensemble that is universal, not only in terms of the source sequence but also in terms of the distortion measure. Full article
(This article belongs to the Collection Feature Papers in Information Theory)
Show Figures

Figure 1

20 pages, 6535 KB  
Article
An Innovative Information Hiding Scheme Based on Block-Wise Pixel Reordering
by Jui-Chuan Liu, Heng-Xiao Chi, Ching-Chun Chang and Chin-Chen Chang
Future Internet 2024, 16(1), 34; https://doi.org/10.3390/fi16010034 - 22 Jan 2024
Cited by 2 | Viewed by 2164
Abstract
Information has been uploaded and downloaded through the Internet, day in and day out, ever since we immersed ourselves in the Internet. Data security has become an area demanding high attention, and one of the most efficient techniques for protecting data is data [...] Read more.
Information has been uploaded and downloaded through the Internet, day in and day out, ever since we immersed ourselves in the Internet. Data security has become an area demanding high attention, and one of the most efficient techniques for protecting data is data hiding. In recent studies, it has been shown that the indices of a codebook can be reordered to hide secret bits. The hiding capacity of the codeword index reordering scheme increases when the size of the codebook increases. Since the codewords in the codebook are not modified, the visual performance of compressed images is retained. We propose a novel scheme making use of the fundamental principle of the codeword index reordering technique to hide secret data in encrypted images. By observing our experimental results, we can see that the obtained embedding capacity of 197,888 is larger than other state-of-the-art schemes. Secret data can be extracted when a receiver owns a data hiding key, and the image can be recovered when a receiver owns an encryption key. Full article
Show Figures

Figure 1

Back to TopTop