Next Article in Journal
Noise Analysis of a Passive Resonant Laser Gyroscope
Previous Article in Journal
A High-Efficiency Driver Circuit for a Gas-Sensor Microheater Based on a Switch-Mode DC-to-DC Converter
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HFR Projector Camera Based Visible Light Communication System for Real-Time Video Streaming

1
Department of System Cybernetics, Graduate School of Engineering, Hiroshima University, Hiroshima 739-0046, Japan
2
Digital Manufacturing Education Research Center, Hiroshima University, Hiroshima 739-0046, Japan
3
Graduate School of Advanced Science and Engineering, Hiroshima University, Hiroshima 739-0046, Japan
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5368; https://doi.org/10.3390/s20185368
Submission received: 28 August 2020 / Revised: 15 September 2020 / Accepted: 16 September 2020 / Published: 19 September 2020
(This article belongs to the Section Physical Sensors)

Abstract

:
This study develops a projector–camera-based visible light communication (VLC) system for real-time broadband video streaming, in which a high frame rate (HFR) projector can encode and project a color input video sequence into binary image patterns modulated at thousands of frames per second and an HFR vision system can capture and decode these binary patterns into the input color video sequence with real-time video processing. For maximum utilization of the high-throughput transmission ability of the HFR projector, we introduce a projector–camera VLC protocol, wherein a multi-level color video sequence is binary-modulated with a gray code for encoding and decoding instead of pure-code-based binary modulation. Gray code encoding is introduced to address the ambiguity with mismatched pixel alignments along the gradients between the projector and vision system. Our proposed VLC system consists of an HFR projector, which can project 590 × 1060 binary images at 1041 fps via HDMI streaming and a monochrome HFR camera system, which can capture and process 12-bit 512 × 512 images in real time at 3125 fps; it can simultaneously decode and reconstruct 24-bit RGB video sequences at 31 fps, including an error correction process. The effectiveness of the proposed VLC system was verified via several experiments by streaming offline and live video sequences.

1. Introduction

With the recent rapid advances in computer and image sensor technologies, many high frame rate (HFR) vision systems that can capture and process images simultaneously at thousands of frames per second have been developed [1,2,3,4,5]; many tracking algorithms, such as optical flow estimation [6,7], cam-shift tracking [8], and feature-point tracking [9], have been accelerated by the parallel implementation of these algorithms on field programmable gate array (FPGA) and graphics processing units. These HFR vision systems can have a large bandwidth that can simultaneously recognize high-speed phenomena, which are too fast to be recognized by the naked human eyes and standard video cameras operating at dozens of frames per second. Many vision-based dynamic sensing systems have been developed for the human-invisible dynamics of objects such as drone tracking [10,11], motion-blur-free video shooting [12,13,14,15], vibration analysis [16,17], and microscopic sensing [18,19,20,21]. In addition to these HFR vision systems that can capture the dynamic phenomena vibrating at hundreds or thousands of hertz, HFR projector systems based on the digital micro-mirror device (DMD) technology [22,23] can project binary image patterns at thousands of frames per second or more; several types of HFR projector–camera systems [24,25,26] have been reported in various applications, such as structured light based 3D sensing [27,28,29,30] and simultaneous projection mapping [31,32,33,34]. If a real-time HFR vision system could function as a communication receiver to perfectly capture and encode the HFR-blinking high-space-resolution image patterns with real-time video processing at thousands of frames per second, which are too fast for human eyes to see, and high-thorough visible light communication (VLC) could be realized for broadband video streaming by utilizing the high-resolution in mega-pixel order and high-frequency band in the kHz order in the HFR projection of an HFR projector.
In this study, we develop an HFR-projector–camera-based VLC system for broadband video streaming, in which a projector and vision system function as an image transmitter and image receiver operating at thousands of frames per second, respectively. We implemented video encoding and decoding processes based on a camera-projector-based VLC protocol for real-time video streaming so that the pixel-wise high-throughput light patterns projected from the HFR projector can be maximally utilized without losing any information when the spatial resolution of the HFR vision system is similar to that of the projector. The remainder of this paper is organized as follows. Section 2 summarizes the related works of VLC along with the problems in the conventional vision-based VLC system for broadband video streaming. Section 3 proposes the concept of the HFR-projector–camera-based VLC system. Section 4 and Section 5 provide an outline of the transmitter and receiver systems of the HFR-projector–camera-based VLC system, describing its VLC protocol that involves the video encoding and decoding processes based on gray code-based binary modulated light patterns which are robust to changing external illumination and unmatched pixel alignment between the vision system and projector, respectively. Section 6 describes the image quality parameter, and Section 7 presents the performance of real-time experiments conducted on several video sequences and live-camera video streaming, in which an HFR vision system can capture and process 512 × 512 images in real time at 3125 fps for 590 × 1060 binary images, which are modulated and projected from an HFR projector at 1041 fps. This system can simultaneously transmit and receive a 24-bit RGB 590 × 1060 video sequence at maximum 31 fps.

2. Related Works

VLC has emerged as an alternative technique to accommodate the exponentially increasing demands of radio frequency-based wireless communication [35,36,37,38]. The visible light corresponds to a band of frequencies ranging between 400 THz (750 nm) and 800 THz (375 nm) and is used as a source in VLC systems for transmitting encoded information using air as a transporting medium and decoded using an appropriate photoreceiver. The intensity of the light source is modulated according to the input data at a high rate which is faster than the persistence of human vision. However, a sensitive photodiode or an image sensor is used to detect the embedded information by decoding the on–off behavior of the light emitting diode (LED) [39,40,41,42,43,44]. In a VLC system, the image sensor has an advantage over the photodiode; it can separate information spatially and temporally from the light source, whereas the photodiode-based systems are highly sensitive to light and inexpensive, but require additional equipment for setting up a system. With the ability of image sensors to capture light, a new type of optical wireless communication is introduced known as camera communication, where image sensors are used for sensing the light intensity emitted from a light source [45,46]. Many potential applications of camera-based VLC systems include automotive systems [47,48,49], mobile phone-camera communications [50,51,52], indoor wireless communications [53,54,55], LED camera-based VLC [56,57], and image recognition and light signaling [58].
Image-sensors-based VLC systems have been developed to decode the information transmitted from different light sources, such as LEDs, display screens, and projectors. Various studies contributing to LED-camera based communication systems have been conducted, focusing on the rate of data transfer from the LED-to-camera and LED-based position location detection systems [59,60,61,62,63]. In addition, traffic signal LEDs are used for estimating the position of a vehicle using an in-vehicle camera and an LED-based VLC system [64,65,66,67]. The accuracy of LED-based systems depends on the number of LEDs used, focal length of the lens, pixel size, and frame rate of the camera receiver. To avoid the complexity of building an LED source circuit for transmission, a display screen or projector has been used as an alternative solution to increase the bit-rate transfer and overall speed of indoor-based VLC systems [68,69,70,71]. Display monitors and LCD panels modulate the screen intensities using different encoding techniques and are accordingly decoded using a camera at the receiver [72,73,74,75]. The data communication between a screen and a camera does not necessarily depend on the content of the screen, and it can be completely hidden from the user by integrating the information and content onto the screen, which further limits their application scenarios. The display screens and projectors with low frame rates make the communication systems slower and limited; this issue can be resolved by using HFR projectors that provide a high data transmission rate, in contrast to commercial projectors that support low frame rate projection and lack the controlling parameters.
The major drawback of conventional cameras and the cameras integrated in smartphones and tablets is that they all operate at a low frame rate due to which the communication bandwidth of VLC systems becomes low, which can be overcome by using an HFR camera. Therefore, we propose a VLC system with an HFR projector and HFR camera that can provide a higher communication bandwidth and better performance, as well as minimize the loss of information. This research mainly focuses on the spatio-temporal information which is similar to transmitting spatial information such as quick response codes (QR codes) and bar-codes; however, for transmitting the temporal information, LED-to-camera communication is considered, which decodes the data temporally. The processing of data in real time involves challenges, and, hence, additional information is embedded with transmitted video sequences for proper decoding at the receiver side. Thus, the spatio-temporal information is transmitted using an HFR projector which is then decoded spatially and temporally by an HFR camera that can be used in our system for transmitting real-time videos using VLC.

3. HFR Projector-Camera-Based VLC System

3.1. VLC System

This study introduces an HFR-projector–camera-based system for streaming videos in real time using VLC. There is some research based on projector–camera, but the projected content is perceptible to human eye, whereas the hidden encoded data is imperceptible. The drawback of these system is that they work at a low frame rate due to which the communication system becomes slower. The advantage of our system is that the data rate becomes higher by using the HFR projector and HFR camera. The system also explains the advantage of using gray-code coding over pure binary based coding and robustness to the ambient light. An overall block diagram of the proposed VLC system is shown in Figure 1, where an HFR projector can project the encoded stored color video sequences or universal serial bus (USB) camera videos into binary-modulated images that can be decoded using a monochrome HFR camera. The binary-modulated images are effectively decoded using additional information appended to each binary image as a header block that contains the current image information such as the frame number, starting of a new image, and channel bit plane information. In addition, the system can eliminate any ambiguities associated with mismatched pixel alignment along the gradient between the HFR projector and HFR camera using gray-code encoding instead of pure-binary-code-based image projection. At the receiver, the frame rate of the monochrome HFR camera is set to be thrice that of the HFR projector considering the Nyquist sampling rate so that the original projected image can be retrieved without any loss. The monochrome HFR camera captures the binary images to reconstruct the original image and background subtraction is performed for every captured binary image to make the system more robust against different textured backgrounds. In addition, the content of the cumulatively projected HFR binary images is imperceptible to human eyes, which results in secure data transmission.

3.2. System Configuration

An HFR projector is used as a transmitter to establish a high-speed VLC communication system with a high projection rate and control. The digital light processing (DLP) LightCrafter 4500 HFR projector uses a transmitter that provides a high projection rate of up to 4000 fps with bit plane projection control. DLP LightCrafter 4500 is a projection system with a two-dimensional array of electrically addressable and mechanically tiltable micro-mirrors to represent each pixel, known as digital mirror devices (DMDs), that are widely used in consumer electronics [76,77,78]. The DLP projector does not modulate the emitted wavelength of the projected light to reproduce the color intensity; instead, it reproduces by modulating the exposure time of the mirrors over a specific operating refresh time based on the projected frame bit planes. This projector supports 1-bit to 8-bit images with a resolution of 912 × 1140, and each pixel corresponds to a micro-mirror on the DMD. This feature helps with projecting data at the pixel level and transforms the image to be used for pixel-wise binary projection for the VLC system.
Dynamic changes related to HFR projection are imperceptible to human eyes and conventional cameras are unable to detect high-speed data or events. Therefore, to monitor high-speed phenomena continuously, we need HFR cameras to improve the shooting speed and performance. In this study, the proposed system consists of a monochrome HFR camera system that is an extension of Fastcam SA-X2 developed by Co. Photron and Hiroshima University; it provides a complementary metal oxide semiconductor (CMOS) sensor-based super high-speed vision platform that enables real-time image processing more than 10,000 fps against a megapixel image and global electronic shutter with excellent light sensitivity [4]. This camera is used as a receiver with an embedded external board that has an onboard FPGA for image processing; it produces output images with a resolution of 512 × 512 with a 12-bit dynamic range at 3125 fps in real time. This HFR camera system provides a high frame rate image capturing to meet the requirements of the proposed VLC system.

4. Transmitter Encoding System

The transmitter encoding system in the proposed VLC system has three stages: encoding image from pure-binary-code into gray-code, addition of header information, and binary image or bit-plane projection, as shown in Figure 2. The input RGB video is initially encoded frame-by-frame into gray-code from a pure-binary-code, and additional information, such as the frame number, along with other necessary information, is appended to the current image in the form of header information, which is then fed to the HFR projector, where it is deconstructed into binary images for projection.

4.1. Header Information

The communication link between the transmitter and receiver is established by appending additional information of the image in the form of blocks of pixels as header information to the transmitting image. In the header, four blocks of pixels represent information about the current image, as shown in Figure 3, where the first block S0, whose all pixel values are set to a maximum value of 255 for an 8-bit pixel, is used for determining the start of a new image and software-based synchronization. The next five blocks of pixels (that is, F4, F3, F2, F1, and F0) in the header represent a 5-bit frame number ranging from 0 to 31, which is assigned to each frame continuously. Thereafter, 2-bit channel information is added using the C1 and C0 blocks to represent the red, green, and blue channels of an image, whereas the last 3-bits (that is B2, B1, and B0) represent eight bit-planes of a single channel that will be used to determine any loss in the bit-plane in an RGB channel of the image. The last five blocks of pixels help in determining the sequence of binary images for reconstructing an image. Therefore, let I t ( x , y ) be the input image, which is combined with the header information of size I h ( w , y ) to form a combined image, I r g b ( m , n ) , before passing it to the HFR projector for binary image projection, as expressed in Equation (1):
I r g b ( m , n ) = I t ( x , y ) + I h ( w , y )

4.2. Projection Pattern

The spatio-temporal projection of binary images by an HFR projector is achieved by decomposing a given packed 24-bit RGB image into its equivalent twenty-four 1-bit binary images. The HFR projector supports 2 8 = 256 intensity levels for an 8-bit channel and the decomposition of a 24-bit RGB color image is demonstrated in Figure 4a, where I r g b ( m , n ) is a three-channel 24-bit color image that is split into three single-channel 8-bit images, I r ( m , n ) , I g ( m , n ) and I b ( m , n ) . The 8-bit single-channel image is converted into binary images by the HFR projector as eight 1-bit images, where B r t ( m , n ) , B g t ( m , n ) and B b t ( m , n ) represents the t t h 1-bit image of the red, green, and blue channels, respectively; “t” represents the bit-plane number ranging between 0 and 7 for an 8-bit image. The projection sequence of binary images is defined by the users in the HFR projector controlling software, and the projection of a new image is triggered by vertical synchronization (vsync) signals. The pattern sequence for binary image projection is shown in Figure 4b, where the total duration of exposure for all patterns should be less than or equal to the vsync duration. The HFR projector introduces a sequence of blank images when the duration of all projection patterns is not equal to vsync.

4.3. Gray-Code Encoding

The image reconstructed with pure-binary-code has ambiguities along the gradients due to mismatched pixel alignment between the HFR projector and HFR camera which is overcome by gray-code-based projection. The ambiguities observed in the images reconstructed using pure-binary-code includes ringing artifacts as shown in Figure 5, which are reduced by gray-code-based image projection. Let I t ( x , y ) be the input RGB color image having three 8-bit channels as red I r ( x , y ) , green I g ( x , y ) , and blue I b ( x , y ) channel as expressed in Equation (2):
I t ( x , y ) = I r ( x , y ) I g ( x , y ) I b ( x , y ) ,
The pixel-value P of an input image is represented by a sequence of binary values ( b n 1 , …, b 1 , b 0 ) based on Equation (3). In an 8-bit image, each pixel is represented as eight 1-bit binary images where the higher bit-planes contain more significant visual information and the lower bit-plane shows more details. Using Equation (4), the gray-code representation of a binary pixel value P, is ( g n 1 , …, g 1 , g 0 ), which is used to convert the pure-binary-code images of red I r ( x , y ) , green I g ( x , y ) , and blue I b ( x , y ) channels into gray-code as I g r a y r ( x , y ) , I g r a y g ( x , y ) , and I g r a y b ( x , y ) , respectively, which are combined to make one 24-bit gray-code color image, I g r a y t ( x , y ) , as shown in Equation (5). The gray-code image I g r a y t ( x , y ) is then combined with the header information I h ( w , y ) , to form I g r a y r g b ( m , n ) , as shown in Equation (6), for transmission through the HFR projector as binary images:
P = i = 0 n 1 b i 2 i = b 0 2 0 + b 1 2 1 + + b n 1 2 n 1 ,
g i = b i i = n 1 b i b i + 1 0 i n 2 ,
I g r a y r ( x , y ) I g r a y g ( x , y ) I g r a y b ( x , y ) = I g r a y t ( x , y ) ,
I h ( w , y ) + I g r a y t ( x , y ) = I g r a y r g b ( x , y ) ,

5. Receiver Decoding System

The receiver uses a monochrome HFR camera to decode the transmitted binary images of a 24-bit RGB image; its mechanism is shown in Figure 6. The transmitter and receiver systems are two separate systems, and they do not implement any hardware-based synchronization; therefore, software-based synchronization is used to synchronize them. After achieving synchronization, background subtraction is used to eliminate the ambient light on the projector screen in an indoor office room to extract the projected light intensity. The camera-projector alignment is corrected using camera calibration in post-processing to correct the orientation of the reconstructed image.

5.1. Software-Based Synchronization

The HFR projector and HFR camera need to be synchronized to decode and reconstruct image sequences by capturing the binary images without any loss in pixel information. The HFR projector and HFR camera operate on their respective internal system clocks and are not connected to any common hard-wired external trigger; therefore, software-based synchronization is achieved at the receiver using the HFR camera. Using the Nyquist sampling theorem, which states that a continuous-time signal can be sampled and perfectly reconstructed from its samples if the waveform is sampled over twice as fast as its highest frequency component, software-based synchronization is achieved by setting the frame rate of the HFR camera to three times that of the HFR projector. Figure 7 describes the software-based synchronization method where three images are captured for a projected binary image and a total of three cases are observed. In case-1, the HFR camera starts capturing images at the same moment as the HFR projector projects images; thus, we can observe the satisfactory brightness of the first two images. In case-2, the HFR camera starts capturing images with a delay. Consequently, we obtain good brightness in the first two images. However, in case-3, the HFR camera starts capturing images before the HFR projector starts projecting; therefore, we can observe the satisfactory brightness of the second and third images. Thus, we selected the second image to reconstruct the original image because it has significant brightness compared to the other two images that are produced during the transitional stage.

5.2. Background Subtraction

The ambient light effect on the projector screen is eliminated by the background subtraction method with thresholding. In this method, a reference image is subtracted from the input image where the reference image is estimated using the global thresholding method by projecting the maximum and minimum intensities through the HFR projector onto the screen. Let C i n ( u , v ) be the input image captured by the HFR camera, C t h r ( u , v ) be a reference or threshold image, and C b i n ( u , v ) be the binary image obtained after background subtraction. C b i n ( u , v ) is calculated using Equation (7), where L ( m , n ) is the pixel value at ( m , n ) of C i n ( u , v ) and t h r ( m , n ) is the threshold value at ( m , n ) of C t h r ( u , v ) :
C b i n ( m , n ) = 1 for L ( m , n ) t h r ( m , n ) 0 for L ( m , n ) < t h r ( m , n ) ,
The threshold value t h r ( m , n ) at ( m , n ) is calculated using Equation (8), where B ( m , n ) is the pixel value at ( m , n ) of C i n ( u , v ) , captured after projecting its maximum brightness, and D ( m , n ) is the pixel value at ( m , n ) of C i n ( u , v ) , captured after projecting a black image:
t h r ( m , n ) = B ( m , n ) 2 + D ( m , n ) ,
To evaluate the effectiveness of background subtraction, we used plain and patterned backgrounds as the projection screens. Initially, we projected the maximum and minimum brightness onto the projection screen to evaluate the background scene, which was subtracted from the input image. Figure 8a shows the input image used for projection, Figure 8b is the background used, Figure 8c shows the binarized image projected onto the background surface, Figure 8d illustrates the reconstructed image without background subtraction, and Figure 8e shows the reconstructed image with background subtraction. When using a plain white background, the global threshold value does not affect the entire reconstructed image because there is uniform reflectance of light throughout the surface. However, when a colored patterned background is used the threshold limit for each pixel varies owing to the reflectance of light, depending on the color it falls on. Therefore, we cannot use a global thresholding system. Figure 8d shows the reconstructed image, where a global thresholding technique is used; that is, a single threshold value is used for the entire image instead of a single pixel individually. The background subtraction method described above is used at the pixel level, where the threshold value is calculated for each pixel and, then, the image is reconstructed accordingly, as shown in Figure 8e.

5.3. Synthesizing 24-Bit RGB Image

The synthesizing or reconstruction of the original image is achieved by software-based synchronization of HFR projector–camera, background subtraction, and checking the header information. A threshold value T is required to extract data from the header information blocks which is constant and does not change dynamically as the threshold t h r ( m , n ) . The threshold value T determines the “0” and “1” bits of the header information and is calculated using Equation (9), where B m a x is the maximum brightness of a pixel in an image when projecting white light and D m i n defines the minimum brightness of a pixel in an image when projecting a black image:
T = B m a x + D m i n 2 ,
To explain the process of synthesizing a 24-bit RGB color image, consider a gray-level input image C i n ( u , v ) captured by the HFR camera and its corresponding binarized images of three channels as C b i n r ( t ) ( u , v ) , C b i n g ( t ) ( u , v ) and C b i n b ( t ) ( u , v ) image which is then combined to form a single 24-bit RGB color image C R G B ( u , v ) as shown in Equation (10), where t is the bitplane number of 8-bit channels. The C R G B ( u , v ) image is an encoded gray-code-based image that is further decoded to a pure-binary-code-based image by using Equation (11) at the pixel level to obtain the reconstructed RGB color image I R G B ( u , v ) :
C b i n r ( t ) ( u , v ) C b i n g ( t ) ( u , v ) C b i n b ( t ) ( u , v ) = C R G B ( u , v ) 0 t 7 ,
b i = g i i = n 1 b i + 1 g i 0 i n 2 ,

6. Image Quality in VLC

The image quality is a characteristic of an image that analyzes a set of measurable image quality attributes, such as image degradation and the amount of distortion or artifacts. Various physical properties, such as lens blur, display resolution, and refresh rate, affect the image quality, but are unlikely to change for a particular system. The perceived image quality in our system is compromised owing to the ambient light and the optics of the projector and camera systems. Image quality assessment is generally categorized into subjective and objective methods; for the proposed system, the objective method-based full reference metrics image quality assessment is used to evaluate the performance. For this image, registration is required between the reconstructed image and its reference image to evaluate the pixel-wise relationship between them. Therefore, image alignment or image registration is performed by warping the reconstructed images so that the features of the two images align perfectly. We used a plane projection surface for all experiments with different patterned backgrounds; only the geometric distortion was corrected and radiometric compensation was not considered. Quality measures, such as the peak-signal-to-noise ratio (PSNR), mean structural similarity index (MSSIM), and multi-scale structural similarity index (MS-SSIM) [79], were used to assess the image quality. PSNR was used to compare images with different dynamic ranges; it can be defined as the ratio of the maximum possible power of a signal and distortion. It has been expressed in Equation (12), where MSE is the mean-squared error and M A X I is the dynamic range of allowable pixel intensities:
P S N R = 10 · log 10 M A X I 2 M S E .
PSNR is easy to compute and has a good reduced reference model, but it does not match well with the human visual perceived quality. Here, the higher the PSNR value, the better the quality of the estimated image. Other methods based on the human visual system (HVS), such as SSIM and MS-SSIM, provide accurate results because they consider the human perception of image quality. The SSIM algorithm extracts the structural information from the field of view based on the HVS assumption. The pixels of the original image carry strong dependencies of the structure of a scene, which is independent of local luminance and contrast. Conversely, MSSIM is derived from SSIM by taking the mean of the SSIM index to evaluate the overall quality of the image:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 ) ,
M S S I M ( x , y ) = 1 M j = 1 M S S I M ( x j , y j ) ,
SSIM is a single-scale approach, but its performance depends mostly on appropriate viewing angles and the resolution of the display; it can be calculated using Equation (13), and Equation (14) represents the mean SSIM. This drawback of SSIM can be overcome by MS-SSIM, which is a novel synthesis-based approach to calibrate the parameters that weigh the relative importance between different scales; however, it is not very useful for badly blurred images. Equation (15) represents the MS-SSIM approach for image comparison at different scales. The measured error lies between 0 and 1, and the best quality value is 1. We used the PSNR and MS-SSIM methods to evaluate the image quality for our system:
M S S S I M ( x , y ) = [ l M ( x , y ) ] α M · j = 1 M [ c j ( x , y ) ] β j · [ s j ( x , y ) ] γ j ,
To evaluate the efficiency of the reconstructed images, a 5-bit frame number in the header information was used by assigning the frame number to each input frame, ranging from 1 to 32, thereby making a packet of 32 frames. These frame numbers were extracted at the receiver and checked for any loss within a packet of 32 images, which was calculated using Equation (16), where F r is the frame reconstruction efficiency and S r represents the successful frame reconstructed out of the total number of frames, F t , within one packet of 32 frames. Thus, using the image quality assessment method, we can define the quality of images reconstructed at the receiver. In addition, the frame reconstruction efficiency explains the number of frames being reconstructed at the receiver and those being lost due to the bandwidth of the system and luminescence of the HFR-projector:
F r [ % ] = S r F t × 100 ,

7. Experiments

The HFR-projector–camera system was set up in a controlled laboratory environment, and the corresponding experiments were conducted to evaluate the performance and image quality of the proposed VLC system. The projected video 590 × 1080 is a combination of 590 × 1060 gray-code images and 590 × 20 header information, projected in a bit plane sequence using the HFR projector. The bit plane sequence used for binary projection is shown in Figure 4b, where the green channel is projected first in a bit-plane sequence, followed by red and blue channel and the duration of exposure for each pattern is 960 μ s. Therefore, the total duration for all bit-plane images is 23,040 μ s, which is less than the v s y n c duration of the input video to avoid any frame loss. A 50-mm lens was mounted on the HFR camera, which was set to a maximum frame rate of 3125 fps. Therefore, the maximum frame rate of the HFR projector that can be used for projection for our system is 1041 fps, which is one-third of the HFR camera frame rate required for software-based synchronization. The experimental setup is shown in Figure 9a, where the distance between the HFR projector and screen is 950 mm and the projection display onto the screen is 448 mm × 415 mm. The distance between the HFR camera and screen is 1130 mm to ensure that the overall area of the projected video on the screen is captured by the camera. The experiments were performed on plain and patterned backgrounds, as shown in Figure 9b, for the proposed system for (a) a stored video sequence and (b) live video streaming from a USB camera. On the patterned background, the header information projected on a white background for the proper detection of header information. In addition, the indoor environment was illuminated with three different luminescence values (i.e., 0, 150, and 300 lux), using an external light source to evaluate the robustness of our system with respect to the ambient light.

7.1. Real-time Video Streaming—Stored Video Sequence

For a real-time video streaming experiment with a stored video sequence, we used the movie “BigBuckBunny” [80]. This experiment was performed to evaluate the performance and effectiveness of the binary and gray-code based encoding, and the background subtraction method. First, we estimated the background scene by projecting the maximum and minimum brightness for background subtraction. The pure-binary-code input images of 24-bit 1920 × 1080 RGB-color video were resized to 590 × 1060 which was then encoded to 590 × 1060 gray-code images, along with the addition of 590 × 20 header information, and projected using bit-plane or binary images at 1041 fps. The HFR camera captures 512 × 512 images and reconstructs the output image with a resolution of 510 × 459 by combining all bit-planes of a 24-bit RGB image sequentially. Figure 10 shows the comparison of the input image with binary-code-based projection and gray-code-based projection with background subtraction on a plain background. Figure 10a shows the full high definition input image 1920 × 1080 at 31 fps. Figure 10b,d depicts the reconstructed images 510 × 459 using pure-binary-code and gray-code respectively, without background subtraction. Figure 10c,e depicts the reconstructed images 510 × 459 using pure-binary-code and gray-code respectively, with background subtraction. Similarly, the experiments were performed for the patterned background, as shown in Figure 11. The images reconstructed with pure-binary-code exhibited artifacts due to the ambiguity of pixels with high spatial frequency; these artifacts were removed in the images reconstructed using gray-code-based transmission.
The image quality analysis and performance evaluation of the system measured under different on-screen luminescence of 0, 150, and 300 lux for three different input frame rates 11, 21, and 31 fps, for approximately a hundred consecutive frames are shown in Figure 12, Figure 13 and Figure 14.
Figure 12 and Figure 13 shows the image quality by measuring the PSNRs and MS-SSIMs, where gray-code-based video reconstruction with background subtraction has a better quality index in comparison to others with respect to different luminescence values. We observed that the image quality was reduced on the patterned background compared to the plain background; however, the luminescence was increased to 300 lux, the patterned background with a slightly darker shade showed a better reconstructed image quality than the image reconstructed with gray-code on a plain background. Figure 14a,b show that, on plain and patterned backgrounds, the reconstructed image without background subtraction slightly differs from the image reconstructed with background subtraction, and there is a marginal difference between the images reconstructed using pure-binary-code and gray-code. However, as we increase the transmission frame rate, the reconstruction frame rate also starts dropping owing to the limited transmission bandwidth and due to the mixing of the channels of two consecutive frames at HFR generated by the HFR projector. Thus, we are discarding the images reconstructed using different frame numbers for an RGB channel sequence. Figure 15 and Figure 16 show the images reconstructed at different luminescence values on plain and patterned backgrounds, respectively. It is evident that the background subtraction method is significantly effective even when the luminescence is increased.

7.2. Real-Time Video Streaming—USB Camera

The USB camera experiment was performed to verify the efficiency and performance of the real-time video streaming which can transmit the real-world information through the camera and verify its reconstruction at the receiver in real-time. In this experiment, the input video sequence was obtained from a USB camera (XIMEA, MQ003CG-CM), which is a 24-bit color camera, and its image resolution was set to 640 × 480 at 30 fps for transmission considering the conventional USB camera parameters. The experimental setup is shown in Figure 17. The experimental scene consists of a person throwing a football on the floor, and the HFR projector is set to 1041 fps with the same binary projection sequence as in Figure 4b, along with an HFR camera frame rate of 3125 fps. Figure 18 and Figure 19 show the comparison between pure-binary-code based and gray-code based reconstructed image sequences on the plain background, respectively. From Figure 18 and Figure 19, we can observe the reduction in artifacts when using gray-code-based encoding to reconstruct the image. Similarly, Figure 20 and Figure 21 show the comparison between pure-binary-code based and gray-code based reconstructed image sequences on the patterned background, respectively. The effectiveness of background subtraction method is evident from the reconstructed images in Figure 20 and Figure 21. Figure 22 shows the performance evaluation for the reconstructed USB camera video with three different input frame rates: 11, 21, and 31 fps for ambient luminescence of 0, 150, and 300 lux considering 100 consecutive images; their image qualities were measured using PSNR and MS-SSIM, as shown in Figure 23 and Figure 24 for the plain and patterned backgrounds, respectively. Figure 22 depicts that, with the increase in ambient light, a slight increase in frame loss can be observed at the receiver. Overall, the images reconstructed using gray-code with background subtraction have a higher image quality compared to other methods, and almost no frame loss is observed at 0 lux.

8. Conclusions

In this study, we developed a real-time video broadcasting system using VLC that can transmit saved and real-time USB camera videos through an HFR projector, operating at 1041 fps, and reconstruct the output color video using a monochrome HFR-camera at 3125 fps via software-based synchronization. In the proposed system, we evaluated the advantages of reconstructing the output images using gray-code over pure-binary-code based video transmission by removing the ambiguity occurring at gradients with pixels having a higher frequency component. Software-based synchronization is used to overcome the synchronization error between the HFR projector and HFR camera by considering the Nyquist sampling theorem. The use of thresholding-based background subtraction is efficient for eliminating the effect of ambient light and patterned background. Various experiments were conducted for real-time video broadcasting systems to evaluate the frame reconstruction at different fps and lux, wherein the frame loss was slightly increased with an increase in the frame rate and lux. However, the image quality of the reconstructed image was reduced as the luminescence of the ambient was increased, which was verified by comparing the image quality metrics, PSNR and MS-SSIM. The background subtraction method was found to be more effective for the patterned background than the plain background. Based on the experimental results, the system has limited bandwidth due to software-based synchronization, which can be increased in the future by perfectly synchronizing the HFR projector–camera system using an external trigger or visual feedback for the HFR camera.

Author Contributions

All authors contributed to the study design and manuscript preparation. I.I. contributed to the concept of HFR-vision-based visible light communication. S.R., K.S., and T.S. designed the high-speed camera-projector system for visible light communication. A.S. developed a visible-light communication algorithm for real-time video streaming, implemented it on a high-speed camera-projector system, and evaluated its performance for real-time video streaming. All authors have read and agreed to the published version of the manuscript.

Funding

The research has not been externally funded.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Watanabe, Y.; Komuro, T.; Ishikawa, M. 955-fps real-time shape measurement of a moving/deforming object using high-speed vision for numerous-point analysis. In Proceedings of the IEEE International Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007; pp. 3192–3197. [Google Scholar]
  2. Ishii, I.; Taniguchi, T.; Sukenobe, R.; Yamamoto, K. Development of high-speed and real-time vision platform, H3 vision. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), St. Louis, MO, USA, 10–15 October 2009; pp. 3671–3678. [Google Scholar]
  3. Ishii, I.; Tatebe, T.; Gu, Q.; Moriue, Y.; Takaki, T.; Tajima, K. 2000 fps real-time vision system with high-frame-rate video recording. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Anchorage, AK, USA, 3–7 May 2010; pp. 1536–1541. [Google Scholar]
  4. Sharma, A.; Shimasaki, K.; Gu, Q.; Chen, J.; Aoyama, T.; Takaki, T.; Ishii, I.; Tamura, K.; Tajima, K. Super high-speed vision platform that can process 1024 × 1024 images in real time at 12,500 fps. In Proceedings of the IEEE/SICE International Symposium on System Integration, Sapporo, Japan, 13–15 December 2016; pp. 544–549. [Google Scholar]
  5. Yamazaki, T.; Katayama, H.; Uehara, S.; Nose, A.; Kobayashi, M.; Shida, S.; Odahara, M.; Takamiya, K.; Hisamatsu, Y.; Matsumoto, S.; et al. A 1ms high-speed vision chip with 3D-stacked 140GOPS column-parallel PEs for spatio-temporal image processing. In Proceedings of the IEEE International Solid-State Circuits Conference (ISSCC), San Francisco, CA, USA, 5–9 February 2017; pp. 82–83. [Google Scholar]
  6. Ishii, I.; Taniguchi, T.; Yamamoto, K.; Takaki, T. High-frame-rate optical flow system. IEEE Trans. Circ. Sys. Video Tech. 2012, 22, 105–112. [Google Scholar] [CrossRef]
  7. Gu, Q.; Nakamura, N.; Aoyama, T.; Takaki, T.; Ishii, I. A full-pixel optical flow system using a GPU-based high-frame-rate vision. In Proceedings of the 2015 Conference on Advances In Robotics, Goa, India, 2–4 July 2015. Article 52. [Google Scholar]
  8. Ishii, I.; Tatebe, T.; Gu, Q.; Takaki, T. Color-histogram-based tracking at 2000 fps. J. Electron. Imaging 2012, 21, 1–14. [Google Scholar] [CrossRef]
  9. Gu, Q.; Raut, S.; Okumura, K.; Aoyama, T.; Takaki, T.; Ishii, I. Real-time image mosaicing system using a high-frame-rate video sequence. J. Robot. Mechatronics 2015, 27, 204–215. [Google Scholar] [CrossRef]
  10. Jiang, M.; Aoyama, T.; Takaki, T.; Ishii, I. Pixel-level and robust vibration source sensing in high-frame-rate video analysis. Sensors 2016, 16, 1842. [Google Scholar] [CrossRef] [PubMed]
  11. Jiang, M.; Gu, Q.; Aoyama, T.; Takaki, T.; Ishii, I. Real-time vibration source tracking using high-speed vision. IEEE Sens. J. 2017, 17, 1513–1527. [Google Scholar] [CrossRef]
  12. Ueno, T.; Gu, Q.; Aoyama, T.; Takaki, T.; Ishii, I.; Kawahara, T. Motion-blur-free microscopic video shooting based on frame-by-frame intermittent tracking. In Proceedings of the IEEE Conference on Automation Science and Engineering, Gothenburg, Sweden, 24–28 August 2015; pp. 837–842. [Google Scholar]
  13. Hayakawa, T.; Watanabe, T.; Ishikawa, M. Real-time high-speed motion blur compensation system based on back-and-forth motion control of galvanometer mirror. Opt. Express 2015, 23, 31648–31661. [Google Scholar] [CrossRef]
  14. Hayakawa, T.; Ishikawa, M. Development of motion-blur-compensated high-speed moving visual inspection vehicle for tunnels. Int. J. Civ. Struct. Eng. Res. 2016, 5, 151–155. [Google Scholar] [CrossRef]
  15. Inoue, M.; Gu, Q.; Jiang, M.; Takaki, T.; Ishii, I.; Tajima, K. Motion-blur-free high-speed video shooting using a resonant mirror. Sensors 2017, 17, 2483. [Google Scholar] [CrossRef] [Green Version]
  16. Yang, H.; Gu, Q.; Aoyama, T.; Takaki, T.; Ishii, I. Dynamics-based stereo visual inspection using multidimensional modal analysis. IEEE Sens. J. 2013, 13, 4831–4843. [Google Scholar] [CrossRef]
  17. Aoyama, T.; Li, L.; Jiang, M.; Inoue, K.; Takaki, T.; Ishii, I.; Yang, H.; Umemoto, C.; Matsuda, H.; Chikaraishi, M.; et al. Vibration sensing of a bridge model using a multithread active vision system. IEEE/ASME Trans. Mechatronics 2018, 23, 179–189. [Google Scholar] [CrossRef]
  18. Oku, H.; Ishii, I.; Ishikawa, M. Tracking a protozoon using high-speed visual feedback. In Proceedings of the IEEE Conference on Microtechnologies in Medicine and Biology, Lyon, France, 12–14 October 2000; pp. 156–159. [Google Scholar]
  19. Sakuma, S.; Kuroda, K.; Tsai, C.; Fukui, W.; Arai, F.; Kaneko, M. Red blood cell fatigue evaluation based on the close-encountering point between extensibility and recoverability. Lab Chip 2014, 14, 1135–1141. [Google Scholar] [CrossRef] [PubMed]
  20. Gu, Q.; Aoyama, T.; Takaki, T.; Ishii, I. Simultaneous vision-based shape and motion analysis of cells fast-flowing in a microchannel. IEEE Trans. Autom. Sci. Eng. 2015, 12, 204–215. [Google Scholar] [CrossRef]
  21. Gu, Q.; Kawahara, T.; Aoyama, T.; Takaki, T.; Ishii, I.; Takemoto, A.; Sakamoto, N. LOC-based high-throughput cell morphology analysis system. IEEE Trans. Autom. Sci. Eng. 2015, 12, 1346–1356. [Google Scholar] [CrossRef]
  22. Hornbeck, L.J. Digital light processing and MEMS: Timely convergence for a bright future. In Proceedings of the Plenary Session, SPIE Micromachining and Microfabrication’95, Austin, TX, USA, 24 October 1995. [Google Scholar]
  23. Younse, J.M. Projection display systems based on the Digital Micromirror Device (DMD). In Proceedings of the SPIE Conference on Microelectronic Structures and Microelectromechanical Devices for Optical Processing and Multimedia Applications, Austin, TX, USA, 24 October 1995; Volume 2641, pp. 64–75. [Google Scholar]
  24. Bimber, O.; Iwai, D.; Wetzstein, G.; Grundhöfer, A. The visual computing of projector–camera systems. In Proceedings of the SIGGRAPH ’08 ACM, Los Angeles, CA, USA, 11–15 August 2008. [Google Scholar]
  25. Takei, J.; Kagami, S.; Hashimoto, K. 3000-fps 3-D shape measurement using a high-speed camera-projector system. In Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Diego, CA, USA, 29 October–2 November 2007. [Google Scholar]
  26. Kagami, S. High-speed vision systems and projectors for real-time perception of the world. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, San Francisco, CA, USA, 13–18 June 2010; pp. 100–107. [Google Scholar]
  27. Gao, H.; Aoyama, T.; Takaki, T.; Ishii, I. A Self-Projected Light-Section Method for Fast Three-Dimensional Shape Inspection. Int. J. Optomechatronics 2012, 6, 289–303. [Google Scholar] [CrossRef] [Green Version]
  28. Liu, Y.; Gao, H.; Gu, Q.; Aoyama, T.; Takaki, T.; Ishii, I. High-frame-rate structured light 3-D vision for fast moving objects. J. Robot. Mechatronics 2014, 26, 311–320. [Google Scholar] [CrossRef]
  29. Li, B.; An, Y.; Cappelleri, D.; Xu, J.; Zhang, S. High-accuracy, high-speed 3D structured light imaging techniques and potential applications to intelligent robotics. Int. J. Intell. Robot. Appl. 2017, 1, 86–103. [Google Scholar] [CrossRef]
  30. Moreno, D.; Calakli, F.; Taubin, G. Unsynchronized structured light. ACM Trans. Graph. 2015, 34, 178. [Google Scholar] [CrossRef]
  31. Chen, J.; Yamamoto, T.; Aoyama, T.; Takaki, T.; Ishii, I. Simultaneous projection mapping using high-frame-rate depth vision. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014; pp. 4506–4511. [Google Scholar]
  32. Watanabe, Y.; Narita, G.; Tatsuno, S.; Yuasa, T.; Sumino, K.; Ishikawa, M. High-speed 8-bit image projector at 1000 fps with 3 ms delay. In Proceedings of the International Display Workshops (IDW2015), Shiga, Japan, 11 December 2015; pp. 1064–1065. [Google Scholar]
  33. Narita, G.; Watanabe, Y.; Ishikawa, M. Dynamic projection mapping onto deforming non-rigid surface using deformable dot cluster marker. IEEE Trans. Vis. Comput. Graph. 2017, 23, 1235–1248. [Google Scholar] [CrossRef]
  34. Fleischmann, O.; Koch, R. Fast projector–camera calibration for interactive projection mapping. In Proceedings of the 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 3798–3803. [Google Scholar]
  35. Cevik, T.; Yilmaz, S. An overview of visible light communication systems. IJCNC 2015, 7, 139–150. [Google Scholar] [CrossRef]
  36. Bhalerao, M.; Sonavane, S.; Kumar, V. A survey of wireless communication using visible light. Int. J. Adv. Eng. Technol. 2013, 5, 188–197. [Google Scholar]
  37. Jovicic, A.; Li, J.; Richardson, T. Visible light communication: Opportunities, challenges and the path to market. IEEE Commun. Mag. 2013, 51, 26–32. [Google Scholar] [CrossRef]
  38. Fath, T.; Haas, H. Performance comparison of mimo techniques for optical wireless communications in indoor environments. IEEE Trans. Commun. 2013, 6, 733–742. [Google Scholar] [CrossRef]
  39. Kumar, N.; Lourenco, N.R. Led-based visible light communication system: A brief survey and investigation. J. Eng. Appl. Sci. 2010, 5, 296–307. [Google Scholar] [CrossRef] [Green Version]
  40. Komine, T.; Nakagawa, M. Fundamental analysis for visible-light communication system using LED lights. IEEE Trans. Consum. Electron. 2004, 50, 100–107. [Google Scholar] [CrossRef]
  41. Bui, T.; Kiravittaya, S.; Sripimanwat, K.; Nguyen, N. A comprehensive lighting configuration for efficient indoor visible light communication networks. Int. J. Opt. 2016, 2016, 1–9. [Google Scholar] [CrossRef] [Green Version]
  42. Sindhubala, K.; Vijayalakshmi, B. Ecofriendly data transmission in visible light communication. In Proceedings of the Third International Conference on Computer, Communication, Control and Information Technology (C3IT), Hooghly, India, 7–8 February 2015; pp. 1–4. [Google Scholar]
  43. Zafar, F.; Karunatilaka, D.; Parthiban, R. Dimming schemes for visible light communication: The state of research. IEEE Wirel. Commun. 2015, 22, 29–35. [Google Scholar] [CrossRef]
  44. Rajagopal, S.; Roberts, R.D.; Lim, S.K. IEEE 802.15.7 visible light communication: Modulation schemes and dimming support. IEEE Commun. Mag. 2012, 50, 72–82. [Google Scholar] [CrossRef]
  45. Takai, I.; Ito, S.; Yasutomi, K.; Kagawa, K.; Andoh, M.; Kawahito, S. LED and CMOS image sensor based optical wireless communication system for automotive applications. IEEE Photonics J. 2013, 5, 6801418–6801418. [Google Scholar] [CrossRef]
  46. Takai, I.; Harada, T.; Andoh, M.; Yasutomi, K.; Kagawa, K.; Kawahito, S. Optical vehicle-to-vehicle communication system using LED transmitter and camera receiver. IEEE Photonics J. 2014, 6, 1–14. [Google Scholar] [CrossRef]
  47. Kasashima, T.; Yamazato, T.; Okada, H.; Fujii, T.; Yendo, T.; Arai, S. Interpixel interference cancellation method for road-to-vehicle visible light communication. In Proceedings of the IEEE 5th International Symposium on Wireless Vehicular Communications (WiVeC), Dresden, Germany, 2–3 June 2013; pp. 1–5. [Google Scholar]
  48. Chinthaka, H.; Premachandra, N.; Yendo, T.; Yamasato, T.; Fujii, T.; Tanimoto, M.; Kimura, Y. Detection of LED traffic light by image processing for visible light communication system. In Proceedings of the 2009 IEEE Intelligent Vehicles Symposium, Xi’an, China, 3–5 June 2009; pp. 179–184. [Google Scholar]
  49. Yamazato, T.; Takai, I.; Okada, H.; Fujii, T.; Yendo, T.; Arai, S.; Andoh, M.; Harada, T.; Yasutomi, K.; Kagawa, K.; et al. Image-sensor-based visible light communication for automotive applications. IEEE Commun. Mag. 2014, 52, 88–97. [Google Scholar] [CrossRef]
  50. Rajagopal, N.; Lazik, P.; Rowe, A. Visual light landmarks for mobile devices. In Proceedings of the 13th International Symposium on Information Processing in Sensor Networks, Berlin, Germany, 15–17 April 2014; pp. 249–260. [Google Scholar]
  51. Boubezari, R.; Le Minh, H.; Ghassemlooy, Z.; Bouridane, A.; Pham, A. Data detection for Smartphone visible light communications. In Proceedings of the 9th International Symposium on Communication Systems, Networks and Digital Signal Processing (CSNDSP), Manchester, UK, 23–25 July 2014; pp. 1034–1038. [Google Scholar]
  52. Corbellini, G.; Akşit, K.; Schmid, S.; Mangold, S.; Gross, T. Connecting networks of toys and smartphones with visible light communication. IEEE Commun. Mag 2014, 52, 72–78. [Google Scholar] [CrossRef]
  53. Wang, M.; Wu, J.; Yu, W.; Wang, H.; Li, J.; Shi, J.; Luo, C. Efficient coding modulation and seamless rate adaptation for visible light communications. IEEE Wirel. Commun. 2015, 22, 86–93. [Google Scholar] [CrossRef]
  54. Li, T.; An, C.; Tian, Z.; Campbell, A.T.; Zhou, X. Human sensing using visible light communication. In Proceedings of the MobiCom’15, Paris, France, 7–11 September 2015. [Google Scholar]
  55. Danakis, C.; Afgani, M.; Povey, G.; Underwood, I.; Haas, H. Using a CMOS camera sensor for visible light communication. In Proceedings of the IEEE GlobecomWorkshops (GC Wkshps), Anaheim, CA, USA, 3–7 December 2012; pp. 1244–1248. [Google Scholar]
  56. Wang, J.; Kang, Z.; Zou, N. Research on indoor visible light communication system employing white LED lightings. In Proceedings of the IET International Conference on Communication Technology and Application (ICCTA 2011), Beijing, China, 14–16 October 2011; pp. 934–937. [Google Scholar]
  57. Bui, T.C.; Kiravittaya, S. Demonstration of using camera communication based infrared LED for uplink in indoor visible light communication. In Proceedings of the IEEE Sixth International Conference on Communications and Electronics (ICCE), Ha Long, Vietnam, 27–29 July 2016; pp. 71–76. [Google Scholar]
  58. Chow, C.; Chen, C.; Chen, S. Enhancement of signal performance in LED visible light communications using mobile phone camera. IEEE Photonics J. 2015, 7, 1–7. [Google Scholar] [CrossRef]
  59. Xu, Y.; Zhao, J.; Shi, J.; Chi, N. Reversed three-dimensional visible light indoor positioning utilizing annular receivers with multi-photodiodes. Sensors 2016, 16, 1254. [Google Scholar] [CrossRef] [Green Version]
  60. Kuo, Y.; Pannuto, P.; Hsiao, K.; Dutta, P. Luxapose: Indoor positioning with mobile phones and visible light. In Proceedings of the 20th Annual International Conference on Mobile Computing and Networking, Maui, HI, USA, 7–11 September 2014; pp. 447–458. [Google Scholar]
  61. Jerome, K.; Tony, V.; Vinayak, R.; Dhanaraj, K.J. Indoor navigation using visible light communication. In Proceedings of the 2014 Texas Instruments India Educators’ Conference (TIIEC), Bangalore, India, 4–5 April 2014; pp. 46–52. [Google Scholar]
  62. Ganti, D.; Zhang, W.; Kavehrad, M. VLC-based indoor positioning system with tracking capability using Kalman and particle filters. In Proceedings of the 2014 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 10–13 January 2014; pp. 476–477. [Google Scholar]
  63. Do, T.; Yoo, M. An in-depth survey of visible light communication based positioning systems. Sensors 2016, 16, 678. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  64. Zhao, X.; Lin, J. Maximum likelihood estimation of vehicle position for outdoor image sensor-based visible light positioning system. Opt. Eng. 2016, 55, 1–8. [Google Scholar] [CrossRef] [Green Version]
  65. Do, T.; Yoo, M. Performance analysis of visible light communication using CMOS sensors. Sensors 2016, 16, 309. [Google Scholar] [CrossRef] [PubMed]
  66. Nguyen, T.; Hong, C.H.; Le, N.T.; Jang, Y.M. High-speed asynchronous optical camera communication using LED and rolling shutter camera. In Proceedings of the Seventh International Conference on Ubiquitous and Future Networks (ICUFN), Sapporo, Japan, 7–10 July 2015; pp. 214–219. [Google Scholar]
  67. Liu, Y.F.; Chen, H.; Liang, K.J.; Hsu, C.; Chow, C.; Yeh, C. Visible light communication using receivers of camera image sensor and solar Cell. IEEE Photonics J. 2016, 8, 1–7. [Google Scholar] [CrossRef]
  68. Hao, T.; Zhou, R.; Xing, G. Cobra: Color barcode streaming for smartphone systems. In Proceedings of the MobiSys 2012, Low Wood Bay, Lake District, UK, 25–29 June 2012; pp. 85–98. [Google Scholar]
  69. Hu, W.; Gu, H.; Pu, Q. Lightsync: Unsynchronized visual communication over screen-camera links. In Proceedings of the MobiCom 2013, Miami, FL, USA, 30 September–4 October 2013; pp. 15–26. [Google Scholar]
  70. Perli, S.D.; Ahmed, N.; Katabi, D. PixNet: LCD-Camera pairs as communication links. In Proceedings of the SIGCOMM ’10, New Delhi, India, 30 August–2 September 2010. [Google Scholar]
  71. Gao, Z.; Zhai, G.; Wu, X.; Min, X.; Zhi, C. DLP based anti-piracy display system. In Proceedings of the IEEE VCIP’14, Valletta, Malta, 7–10 December 2014. [Google Scholar]
  72. Dai, J.; Chung, R. Embedding imperceptible codes into video projection and applications in robotics. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura, Portugal, 7–12 October 2012; pp. 4399–4404. [Google Scholar]
  73. Zhang, B.; Ren, K.; Xing, G.; Fu, X.; Wang, C. SBVLC: Secure barcode-based visible light communication for smartphones. IEEE Trans. Mob. Comput. 2016, 15, 432–446. [Google Scholar] [CrossRef] [Green Version]
  74. Wang, A.; Li, Z.; Peng, C.; Shen, G.; Fang, G.; Zeng, B. InFrame++: Achieve Simultaneous Screen-Human Viewing and Hidden Screen-Camera Communication. In Proceedings of the 13th Annual International Conference on Mobile Systems, Applications, and Services (MobiSys ’15), New York, NY, USA, May 2015; pp. 181–195. [Google Scholar]
  75. Wang, A.; Peng, C.; Zhang, O.; Shen, G.; Zeng, B. InFrame: Multiflexing full-frame visible communication channel for humans and devices. In Proceedings of the HotNets-XIII Proceedings of the 13th ACM Workshop on Hot Topics in Networks, Los Angeles, CA, USA, 27–28 October 2014. [Google Scholar]
  76. Hornbeck, L.J. Digital light processing: A new MEMS-based display technology. In Proceedings of the Technical Digest of the IEEJ 14th Sensor Symposium, Kawasaki, Japan, 4–5 June 1996; pp. 297–304. [Google Scholar]
  77. Gove, R.J. DMD Display Systems: The Impact of an All-digital Display. Available online: https://www.semanticscholar.org/paper/DMD-Display-Systems-%3A-The-Impact-of-an-All-Digital-Gove/e5167d04802842fda09251429636d7300d340146 (accessed on 18 September 2020).
  78. Hornbeck, L.J. Digital light processing and MEMS: An overview. In Proceedings of the Digest IEEE/Leos 1996 Summer Topical Meeting. Advanced Applications of Lasers in Materials and Processing, Keystone, CO, USA, 5–9 August 1996; pp. 7–8. [Google Scholar]
  79. Wang, Z.; Bovik, A.C.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  80. Big Buck Bunny. Available online: http://www.bigbuckbunny.org (accessed on 18 September 2020).
Figure 1. (a) block diagram and (b) configuration of the proposed VLC system.
Figure 1. (a) block diagram and (b) configuration of the proposed VLC system.
Sensors 20 05368 g001
Figure 2. Transmitter.
Figure 2. Transmitter.
Sensors 20 05368 g002
Figure 3. Header information.
Figure 3. Header information.
Sensors 20 05368 g003
Figure 4. (a) decomposition of an RGB image into binary bit-plane images and (b) bit-plane projection pattern for a single RGB image.
Figure 4. (a) decomposition of an RGB image into binary bit-plane images and (b) bit-plane projection pattern for a single RGB image.
Sensors 20 05368 g004
Figure 5. (a) original image; (b) reconstructed image with pure-binary-code; (c) reconstructed image with gray-code.
Figure 5. (a) original image; (b) reconstructed image with pure-binary-code; (c) reconstructed image with gray-code.
Sensors 20 05368 g005
Figure 6. Receiver.
Figure 6. Receiver.
Sensors 20 05368 g006
Figure 7. Image selection for software-based synchronization.
Figure 7. Image selection for software-based synchronization.
Sensors 20 05368 g007
Figure 8. (a) original image; (b) background pattern; (c) projection on a background pattern; (d) reconstructed image without background subtraction; and (e) reconstructed image with background subtraction.
Figure 8. (a) original image; (b) background pattern; (c) projection on a background pattern; (d) reconstructed image without background subtraction; and (e) reconstructed image with background subtraction.
Sensors 20 05368 g008
Figure 9. (a) overview of the HFR projector–camera system; (b) plain and patterned background.
Figure 9. (a) overview of the HFR projector–camera system; (b) plain and patterned background.
Sensors 20 05368 g009
Figure 10. Reconstructed saved image sequence on plain background: (a) 1920 × 1080 input image; (b) 510 × 459 binary-code image without background subtraction; (c) 510 × 459 binary-code image with background subtraction; (d) 510 × 459 gray-code image without background subtraction; and (e) 510 × 459 gray-code image with background subtraction.
Figure 10. Reconstructed saved image sequence on plain background: (a) 1920 × 1080 input image; (b) 510 × 459 binary-code image without background subtraction; (c) 510 × 459 binary-code image with background subtraction; (d) 510 × 459 gray-code image without background subtraction; and (e) 510 × 459 gray-code image with background subtraction.
Sensors 20 05368 g010
Figure 11. Reconstructed saved image sequence on a patterned background: (a) 1920 × 1080 input image; (b) 510 × 459 binary-code image without background subtraction; (c) 510 × 459 binary-code image with background subtraction; (d) 510 × 459 gray-code image without background subtraction; and (e) 510 × 459 gray-code image with background subtraction.
Figure 11. Reconstructed saved image sequence on a patterned background: (a) 1920 × 1080 input image; (b) 510 × 459 binary-code image without background subtraction; (c) 510 × 459 binary-code image with background subtraction; (d) 510 × 459 gray-code image without background subtraction; and (e) 510 × 459 gray-code image with background subtraction.
Sensors 20 05368 g011
Figure 12. (a) PSNRs and (b) MS-SSIMs when a stored video sequence is streamed with pure-binary-code and gray-code images on a plain background.
Figure 12. (a) PSNRs and (b) MS-SSIMs when a stored video sequence is streamed with pure-binary-code and gray-code images on a plain background.
Sensors 20 05368 g012
Figure 13. (a) PSNRs and (b) MS-SSIMs when a stored video sequence is streamed with pure-binary-code and gray-code images on a patterned background.
Figure 13. (a) PSNRs and (b) MS-SSIMs when a stored video sequence is streamed with pure-binary-code and gray-code images on a patterned background.
Sensors 20 05368 g013
Figure 14. Frame reconstruction ratio when a stored movie is streaming on (a) plain background and (b) patterned background.
Figure 14. Frame reconstruction ratio when a stored movie is streaming on (a) plain background and (b) patterned background.
Sensors 20 05368 g014
Figure 15. Plain background: (a) experiment scene at different illuminance levels; (b) 1920 × 1080 input images; (c) 510 × 459 images reconstructed using pure-binary-code without background subtraction; (d) 510 × 459 reconstructed images with pure-binary-code with background subtraction; (e) 510 × 459 reconstructed images using gray-code without background subtraction; and (f) 510 × 459 reconstructed images using gray-code with background subtraction.
Figure 15. Plain background: (a) experiment scene at different illuminance levels; (b) 1920 × 1080 input images; (c) 510 × 459 images reconstructed using pure-binary-code without background subtraction; (d) 510 × 459 reconstructed images with pure-binary-code with background subtraction; (e) 510 × 459 reconstructed images using gray-code without background subtraction; and (f) 510 × 459 reconstructed images using gray-code with background subtraction.
Sensors 20 05368 g015
Figure 16. Pattern background: (a) experiment scene at different illuminance levels; (b) 1920 × 1080 input images; (c) 510 × 459 images reconstructed using pure-binary-code without background subtraction; (d) 510 × 459 reconstructed images with pure-binary-code with background subtraction; (e) 510 × 459 reconstructed images using gray-code without background subtraction; and (f) 510 × 459 reconstructed images using gray-code with background subtraction.
Figure 16. Pattern background: (a) experiment scene at different illuminance levels; (b) 1920 × 1080 input images; (c) 510 × 459 images reconstructed using pure-binary-code without background subtraction; (d) 510 × 459 reconstructed images with pure-binary-code with background subtraction; (e) 510 × 459 reconstructed images using gray-code without background subtraction; and (f) 510 × 459 reconstructed images using gray-code with background subtraction.
Sensors 20 05368 g016
Figure 17. Experiment setup for HFR-projector–camera system using a USB camera as input.
Figure 17. Experiment setup for HFR-projector–camera system using a USB camera as input.
Sensors 20 05368 g017
Figure 18. Reconstructed USB camera input image sequence on the plain background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Figure 18. Reconstructed USB camera input image sequence on the plain background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Sensors 20 05368 g018
Figure 19. Reconstructed USB camera input image sequence on the plain background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Figure 19. Reconstructed USB camera input image sequence on the plain background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Sensors 20 05368 g019
Figure 20. Reconstructed USB camera input image sequence on the pattern background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Figure 20. Reconstructed USB camera input image sequence on the pattern background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Sensors 20 05368 g020
Figure 21. Reconstructed USB camera input image sequence on the pattern background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Figure 21. Reconstructed USB camera input image sequence on the pattern background: (a) 640 × 480 input image; (b) 510 × 459 binary-code image without background subtraction; and (c) 510 × 459 binary-code image with background subtraction.
Sensors 20 05368 g021
Figure 22. Frame reconstruction ratio when USB camera video is streaming on (a) plain background and (b) patterned background.
Figure 22. Frame reconstruction ratio when USB camera video is streaming on (a) plain background and (b) patterned background.
Sensors 20 05368 g022
Figure 23. (a) PSNRs and (b) MS-SSIMs when USB camera video sequence is streamed with pure-binary-code and gray-code images on the plain background.
Figure 23. (a) PSNRs and (b) MS-SSIMs when USB camera video sequence is streamed with pure-binary-code and gray-code images on the plain background.
Sensors 20 05368 g023
Figure 24. (a) PSNRs and (b) MS-SSIMs when USB camera video sequence is streamed with pure-binary-code and gray-code images on the patterned background.
Figure 24. (a) PSNRs and (b) MS-SSIMs when USB camera video sequence is streamed with pure-binary-code and gray-code images on the patterned background.
Sensors 20 05368 g024

Share and Cite

MDPI and ACS Style

Sharma, A.; Raut, S.; Shimasaki, K.; Senoo, T.; Ishii, I. HFR Projector Camera Based Visible Light Communication System for Real-Time Video Streaming. Sensors 2020, 20, 5368. https://doi.org/10.3390/s20185368

AMA Style

Sharma A, Raut S, Shimasaki K, Senoo T, Ishii I. HFR Projector Camera Based Visible Light Communication System for Real-Time Video Streaming. Sensors. 2020; 20(18):5368. https://doi.org/10.3390/s20185368

Chicago/Turabian Style

Sharma, Atul, Sushil Raut, Kohei Shimasaki, Taku Senoo, and Idaku Ishii. 2020. "HFR Projector Camera Based Visible Light Communication System for Real-Time Video Streaming" Sensors 20, no. 18: 5368. https://doi.org/10.3390/s20185368

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop