Photometric Calibration and Image Stitching for a Large Field of View Multi-Camera System

A new compact large field of view (FOV) multi-camera system is introduced. The camera is based on seven tiny complementary metal-oxide-semiconductor sensor modules covering over 160° × 160° FOV. Although image stitching has been studied extensively, sensor and lens differences have not been considered in previous multi-camera devices. In this study, we have calibrated the photometric characteristics of the multi-camera device. Lenses were not mounted on the sensor in the process of radiometric response calibration to eliminate the influence of the focusing effect of uniform light from an integrating sphere. Linearity range of the radiometric response, non-linearity response characteristics, sensitivity, and dark current of the camera response function are presented. The R, G, and B channels have different responses for the same illuminance. Vignetting artifact patterns have been tested. The actual luminance of the object is retrieved by sensor calibration results, and is used to blend images to make panoramas reflect the objective luminance more objectively. This compensates for the limitation of stitching images that are more realistic only through the smoothing method. The dynamic range limitation of can be resolved by using multiple cameras that cover a large field of view instead of a single image sensor with a wide-angle lens. The dynamic range is expanded by 48-fold in this system. We can obtain seven images in one shot with this multi-camera system, at 13 frames per second.


Introduction
The omnidirectional camera is a type of ideal imaging device that has been increasingly used in recent years. The major advantage is its wide field of view (FOV), which allows it to capture a view of the entire scene. The synthesis of small patches into a large field of view has other advantages, such as reducing lens design difficulties. In addition, the dynamic range of a single sensor with a certain exposure setting has difficulties in meeting an extremely high dynamic range with wide FOV. The omnidirectional camera brings new prospects in various areas such as surveillance, tracking, visual navigation, localization and SLAM, structure from motion, photogrammetry, reconstruction of cultural heritage, and so on. A polycamera is a kind of non-central omnidirectional camera, comprising a cluster of conventional cameras pointing to different directions in a particular configuration. Recent attempts have been made to develop this type of camera. The PANOPTIC [1] is a hemispherical multiple camera system that uses 44 cameras covering 360˝ˆ180˝FOV. The AWARE [2] series camera consists of a multiscale optical layer, an optoelectronic readout layer, and an image presentation and interaction software layer. It uses hundreds of image sensors to address the conflict between FOV and resolution. The design methodology is described in detail in [3][4][5]. In [6], a cylindrical distributed polycamera was introduced based on FPGA hardware, using five digital cameras with wide-angle lens that provide a field of view with 360˝horizontal azimuth.
These cameras can generate realistic panoramas for distant scenes. A polycamera is a kind of non-central omnidirectional camera; parallax exists between sub-cameras, and CRFs of the sensors are different. Low-pass filtering and smoothing methods are used in image blending for realistic visual effects. The simplest way to create a final composite is to simply take an average value at each pixel, but this method usually does not work effectively because of exposure difference, misregistrations, and scene movement. A better weight averaging method is the weight decrease from pixels near the center of the image to pixels near the edge [7]. Gain compensation and the multiband blending method [8,9] can effectively resolve exposure differences, misregistrations, and other factors. However, these methods do not consider the actual luminance of an object, but only achieve vision requirements.
In this study, a calibration method of the photometric characteristic of cameras is introduced. Dark current, radiometric response, and chromatic response difference of sensors of the large FOV multi-camera are measured accurately. Vignetting (decrease in intensity toward image edges) artifact patterns have been tested using an integrating sphere uniform light source. The actual luminance of an object is retrieved through results of obtained sensor calibration, and simultaneously removal of saturated pixels and correcting the nonlinear response of the sensor. The corrected image data are used in blending of images to make panoramas more objectively reflect the objective luminance. The calibration result not only enhances panorama realism but also is an essential step in photometric 3D reconstruction, structure from motion and photogrammetry.
The rest of the paper is structured as follows: Section 2 introduces the hardware of the multi-camera system with large field of view. Section 3 describes the photometric modeling and calibration method. Section 4 presents our calibration and image stitching results. In Section 5, we present conclusions and ideas for future work.

Overview of the Large FOV Multi-Camera System
A state-of-the-art CMOS sensor is used to build our camera system. Field Programmable Gate Array (FPGA) is used to set the exposure time of the sensors, acquire RAW image data, and transfer data to PC. The computer host is used to collect image RAW data, demosaic the Bayer image, restore the image, and process the data further. The multi-camera system is compact and flexible.

Sensor and Lens
The sensor module utilized in our system is an OmniVision OV3640, which supports quadruple extended graphic array (QXGA) resolution, i.e., 2048ˆ1536 and any size that scales down from QXGA, such as 1080 p and VGA [10]. Charge-coupled device (CCD) and CMOS detectors each have strengths and weaknesses coming from their architecture or their fabrication process [11,12]. With consumer digital camera equipment extensively used and semiconductor technology advances, the image quality of the CMOS image sensors has been greatly improved since they first appeared, and it provides a digital interface rather than an analog one. The Serial Camera Control Bus (SCCB) is a universal control protocol for the OmniVision image sensor and the details are described in [13]. It can be used to communicate between FPGA and OV3640 sensors, controlling data formats, gains, offsets, exposure times, and so on. The sensor module has a 3.1 mm focal length and a f/2.8 f-number, and a relatively wide field of view of 65˝. Seven sensors are mounted on a 3D-printed structure. The optical axis angle between the central sensor and outer sensor is 40˝. In addition, the outer sensors are evenly distributed along a circle.

Established Multi-Camera System
Hardware platform is needed for sensor exposure setting and image data acquisition, comprising central FPGA [14], DDR3 SDRAM frame cache, and camera link interface to transmit data. The data acquisition function is implemented in the central FPGA and involves four sub-blocks, which are sensor controllers, data transmit multiplexer, DDR3 memory controller, and camera link interface.
The sensor controller module sets the exposure setting of each sensor, which are unique exposure times and other common parameters, namely, window, frame rate, and gain. The data cache includes two types, the line FIFO cache and the multiple frame DDR3 SDRAM cache. The image sensor outputs data line-by-line, and the line FIFO cache is used to match data rate between the image sensor and DDR3 memory controller, which are asynchronous modules. Exposure times are not equal; thus, shutters shoot at the same time but frames may not be output simultaneously. To obtain the fastest possible frame rate from the camera system, the slowest frame time or the longest exposure time is used as the base. Seven sensors shot at the same time, and the frame of the sensor with shorter exposure time is output earlier and then cached in DDR3 SDRAM; the frame of the sensor with longer exposure time is output later and cached in DDR3 SDRAM. Redundant faster frames are discarded during image frame with longer exposure time output. Multiple camera frames are cached in the same DDR3 SDRAM. The bandwidth of "full" camera link mode can reach 5.6 Gb/s. All data generated by the seven sensors are cached in DDR3 SDRAM, and then read out in accordance with the camera link specification. This multi-camera system can operate at 13 fps. The complete camera system, which includes seven sensors, a FPGA hardware platform, and camera link image data transfer interface, is shown in Figure 1. The sensor controller module sets the exposure setting of each sensor, which are unique exposure times and other common parameters, namely, window, frame rate, and gain. The data cache includes two types, the line FIFO cache and the multiple frame DDR3 SDRAM cache. The image sensor outputs data line-by-line, and the line FIFO cache is used to match data rate between the image sensor and DDR3 memory controller, which are asynchronous modules. Exposure times are not equal; thus, shutters shoot at the same time but frames may not be output simultaneously. To obtain the fastest possible frame rate from the camera system, the slowest frame time or the longest exposure time is used as the base. Seven sensors shot at the same time, and the frame of the sensor with shorter exposure time is output earlier and then cached in DDR3 SDRAM; the frame of the sensor with longer exposure time is output later and cached in DDR3 SDRAM. Redundant faster frames are discarded during image frame with longer exposure time output. Multiple camera frames are cached in the same DDR3 SDRAM. The bandwidth of "full" camera link mode can reach 5.6 Gb/s. All data generated by the seven sensors are cached in DDR3 SDRAM, and then read out in accordance with the camera link specification. This multi-camera system can operate at 13 fps. The complete camera system, which includes seven sensors, a FPGA hardware platform, and camera link image data transfer interface, is shown in Figure 1.

Photometric Modeling and Calibration Method
A digital camera includes three parts, namely, the optical lens, image sensor, and electronic system. Vignetting of the lens decreases intensity toward image edges. The photosensitive material and manufacturing technique of the sensor determine the area of light sensitivity as well as its sensitivity level. Dark current, fixed pattern noise and shot noise are also different due to manufacturing processes. The R, G, B channels of the color sensor have different response characteristics. Different sensors also exhibit similar characteristics. Therefore, each channel of each sensor must be calibrated separately and then normalized in a radiometric calibration process. Optoelectronic converted charge must be amplified, quantized, and read out. Amplifier noise, quantization noise, and noise caused by data transmission line and PCB result in differences in obtained and actual values. A complete calibration of the sensor requires corrections in the domains mentioned, and is applied to measurement equation. Linear and nonlinear response of total dynamic range is a critical factor in the system, and is solved in the calibration process.

Vignetting and Radiometric Response Modeling
As described, a digital number of image pixels is determined by the input light energy, optical lens, opto-electronics converted charge, and electronic system. Vignetting is the cosine-4th law of illuminance falloff [15]: Figure 1. The multi-camera system adapted to shoot images with a large field of view.

Photometric Modeling and Calibration Method
A digital camera includes three parts, namely, the optical lens, image sensor, and electronic system. Vignetting of the lens decreases intensity toward image edges. The photosensitive material and manufacturing technique of the sensor determine the area of light sensitivity as well as its sensitivity level. Dark current, fixed pattern noise and shot noise are also different due to manufacturing processes. The R, G, B channels of the color sensor have different response characteristics. Different sensors also exhibit similar characteristics. Therefore, each channel of each sensor must be calibrated separately and then normalized in a radiometric calibration process. Opto-electronic converted charge must be amplified, quantized, and read out. Amplifier noise, quantization noise, and noise caused by data transmission line and PCB result in differences in obtained and actual values. A complete calibration of the sensor requires corrections in the domains mentioned, and is applied to measurement equation. Linear and nonlinear response of total dynamic range is a critical factor in the system, and is solved in the calibration process.

Vignetting and Radiometric Response Modeling
As described, a digital number of image pixels is determined by the input light energy, optical lens, opto-electronics converted charge, and electronic system. Vignetting is the cosine-4th law of illuminance falloff [15]: where E 1 0 is illuminance of on-axis, E 1 is illuminance of off-axis, and ω 1 is off-axis angle. The off-axis illuminance attenuation caused by the cosine-4th effect of the imaging lens needs to be corrected prior to analyzing the data.
The multi-camera is used in visible light and the light from the integrating sphere is white. We don't consider the wave length, while different color channels are take into consideration. L is illuminance of incident light(lux). Luminous power(lm) of sensor received can be expressed as: where A is the sensor area, k i is lens transmittance and it is constant for three channels (i.e., k r , k g , k b ), D f is the aperture. R i represents the channel response (electron/lm), T is the exposure time, so the generated analog signal intensity of pixel can be expressed as follows: After quantization, amplifier and transmission, we can obtain the final digital number (DN) of the image pixel is given by: where G is the gain of the sensor, DC is the number of electrons resulting from dark current, N s is the zero mean Poisson shot noise, and N R is the zero mean read noise generated by the amplifier. Quantization noise N Q is also shown to be zero mean, which is generated by analog-to-digital (ADC) conversion. Zero mean noise can be removed through averaging multiple samples [16][17][18].
The DN of the image pixel after averaging can be write as: where DN and L may not show strict linearity. The obtained DN of image pixels can be fitted to illuminance by a polynomial equation [16]: The dynamic range of a sensor is limited by the sensitivity of the photosensitive material and the bits of photoelectron quantization when shooting a scene with a certain exposure time t and gain G. The dynamic range of the multi-camera system can be extended by shooting the same scene with multiple exposure settings. The pixel number range is from 0 to 255 with 8-bit quantization. The dynamic range can be extended by combining multiple sensors with different exposure time and gain. Suppose EX max and G max are the maximum exposure time and gain. DC max is the dark current when the exposure time and gain are both maximized. DC min is the dark current when the exposure time and gain are both their minimum values, that is, EX is 1-stop, and G isˆ1. According to Equation (6), we can obtain the effective maximum illuminance of incident light L max when both the exposure time gain is the minimum and the digital number DN is the maximum but not saturated. Identically, the effective minimum illuminance of incident light L min can be solved when the both exposure time gain is the maximum and the digital number DN is minimum but above dark current DC max . The dynamic range(DR) can be expressed as: DR " L max {L min (7)

Calibration Process
In this section, we describe the calibration process. A complete calibration of the multi-camera system requires corrections in several domains to be applied to the model equations. According to the photometric model, the quantitative relationship between illuminance of incident light and digital value of image pixel can be determined in the laboratory. An integrating sphere was used as calibration light source for its advantages of large aperture and FOV, as well as uniform intensity [19]. The amount of light illuminance output can be changed by controlling the number of lighted LEDs, which are fixed in the integrating sphere. Wavelength of LEDs output are not considered in our calibration process. The illuminance is read from an illuminometer through an RS232 port at each exposure state. The test method of vignetting is the acquisition of images with an imaging lens under appropriate illuminance and analysis intensity distribution on the image. We also test the bare sensors (no imaging lens) illuminated by 13 LEDs for linearity and non-linearity response. In order to make camera with different point direction face the output port of the integrating sphere, the multi-camera system is mounted on a two-axis platform. Figure 2 shows the photometric calibration diagram. In this section, we describe the calibration process. A complete calibration of the multi-camera system requires corrections in several domains to be applied to the model equations. According to the photometric model, the quantitative relationship between illuminance of incident light and digital value of image pixel can be determined in the laboratory. An integrating sphere was used as calibration light source for its advantages of large aperture and FOV, as well as uniform intensity [19]. The amount of light illuminance output can be changed by controlling the number of lighted LEDs, which are fixed in the integrating sphere. Wavelength of LEDs output are not considered in our calibration process. The illuminance is read from an illuminometer through an RS232 port at each exposure state. The test method of vignetting is the acquisition of images with an imaging lens under appropriate illuminance and analysis intensity distribution on the image. We also test the bare sensors (no imaging lens) illuminated by 13 LEDs for linearity and non-linearity response. In order to make camera with different point direction face the output port of the integrating sphere, the multicamera system is mounted on a two-axis platform. Figure 2 shows the photometric calibration diagram. For the color image sensor, the R, G, and B channels have different response exposures for the same illuminance. Based on the model of radiometric response, that is, Equation (6), the three channels of radiometric response are: We need to collect sufficient images with different exposure state DNk (k = 1,2,3…,n), illuminance Lk (k = 1,2,3…,n), and dark current DCk (k = 1,2,3…,n). By making k ≥ n, the total coefficients can be solved.

Lens Vignetting
We choose one direction of illuminance falloff mode as lens vignetting under the assumption the lens is centrosymmetric. The sensor response distribution is shown in Figure 3a with four lighted LEDs in the integrating sphere (22 lux), and the distribution of the red channel from center to edge is shown in Figure 3b. The vignetting mode of seven cameras is shown in Figure 4. For the color image sensor, the R, G, and B channels have different response exposures for the same illuminance. Based on the model of radiometric response, that is, Equation (6), the three channels of radiometric response are: We need to collect sufficient images with different exposure state DN k (k = 1,2,3 . . . ,n), illuminance L k (k = 1,2,3 . . . ,n), and dark current DC k (k = 1,2,3 . . . ,n). By making k ě n, the total coefficients can be solved.

Lens Vignetting
We choose one direction of illuminance falloff mode as lens vignetting under the assumption the lens is centrosymmetric. The sensor response distribution is shown in Figure 3a with four lighted LEDs in the integrating sphere (22 lux), and the distribution of the red channel from center to edge is shown in Figure 3b. The vignetting mode of seven cameras is shown in Figure 4.

Dark Current
We tested dark current with six exposure time settings (2/25; 4/25; 6/25; 8/25; 10/25; and 12/25 ms) and six gain settings (×1; ×2; ×4; ×8; ×16; and ×32). In order to minimize the influence of thermal effects on the sensor output data, i.e., dark current, we begin to store sample data after the bare sensor worked for 5-10 min in a dark room. The mean of the image channel pixels digital number is used as the final dark current number. The test results are presented in Figure 5. The results show that dark current increases significantly as the gain increases, but not with the exposure time. The dark currents of the R, G, B channels are almost at the same level, so we need not separate these channels for dark current in the subsequent calibration data process. The image data is 8-bit (or gray value between 0-255). The gains ×16 and ×32 are not used in real image data acquisition because these two gains result in uncertain dark current and low signal-to-noise ratio.

Dark Current
We tested dark current with six exposure time settings (2/25; 4/25; 6/25; 8/25; 10/25; and 12/25 ms) and six gain settings (×1; ×2; ×4; ×8; ×16; and ×32). In order to minimize the influence of thermal effects on the sensor output data, i.e., dark current, we begin to store sample data after the bare sensor worked for 5-10 min in a dark room. The mean of the image channel pixels digital number is used as the final dark current number. The test results are presented in Figure 5. The results show that dark current increases significantly as the gain increases, but not with the exposure time. The dark currents of the R, G, B channels are almost at the same level, so we need not separate these channels for dark current in the subsequent calibration data process. The image data is 8-bit (or gray value between 0-255). The gains ×16 and ×32 are not used in real image data acquisition because these two gains result in uncertain dark current and low signal-to-noise ratio.

Dark Current
We tested dark current with six exposure time settings (2/25; 4/25; 6/25; 8/25; 10/25; and 12/25 ms) and six gain settings (ˆ1;ˆ2;ˆ4;ˆ8;ˆ16; andˆ32). In order to minimize the influence of thermal effects on the sensor output data, i.e., dark current, we begin to store sample data after the bare sensor worked for 5-10 min in a dark room. The mean of the image channel pixels digital number is used as the final dark current number. The test results are presented in Figure 5. The results show that dark current increases significantly as the gain increases, but not with the exposure time. The dark currents of the R, G, B channels are almost at the same level, so we need not separate these channels for dark current in the subsequent calibration data process. The image data is 8-bit (or gray value between 0-255). The gainsˆ16 andˆ32 are not used in real image data acquisition because these two gains result in uncertain dark current and low signal-to-noise ratio.
(g) Cam7 Figure 5. Dark current of seven sensors with six exposure times and six gain settings. Figure 5. Dark current of seven sensors with six exposure times and six gain settings.

Radiometric Response
In order to calibrate the radiometric response of the sensors, we have tested the response with all settings which are all configurable exposure times and gains of sensors under different illuminances output by the integrating sphere. The exposure timesettings are 2/25 ms, 4/25 ms, 6/25 ms, 8/25 ms, 10/25 ms, and 12/25 ms. The gain settings areˆ1,ˆ2,ˆ4,ˆ8. The illuminance output by the integrating sphere are between 0-80 lux. The red channel radiometric response results are shown in Figure 6. The figures show that the shorter the exposure time is, the greater the linear range is. The linear range is small with long exposure time, so the long exposure time setting is suitable for homogeneous illuminance. Different sub-cameras have different pixel digital number responses under the same illuminance. The longer the exposure time is, the greater the pixels digital number is. We have used Equation (8) as objective function to fit the sample data with linear fitting (N = 1) and four order polynomial fitting (N = 4). The results of Cam 1 are shown in Table 1.

Radiometric Response
In order to calibrate the radiometric response of the sensors, we have tested the response with all settings which are all configurable exposure times and gains of sensors under different illuminances output by the integrating sphere. The exposure timesettings are 2/25 ms, 4/25 ms, 6/25 ms, 8/25 ms, 10/25 ms, and 12/25 ms. The gain settings are ×1, ×2, ×4, ×8. The illuminance output by the integrating sphere are between 0-80 lux. The red channel radiometric response results are shown in Figure 6. The figures show that the shorter the exposure time is, the greater the linear range is. The linear range is small with long exposure time, so the long exposure time setting is suitable for homogeneous illuminance. Different sub-cameras have different pixel digital number responses under the same illuminance. The longer the exposure time is, the greater the pixels digital number is. We have used Equation (8) as objective function to fit the sample data with linear fitting (N = 1) and four order polynomial fitting (N = 4). The results of Cam 1 are shown in Table 1.
(e) Cam5 (f) Cam6  (g) Cam7  The four-order polynomial fitting results and linear fitting results are almost equal as the sum of squares due to error (SSE) and the coefficent of determination (R-square) results show. The results show that P2, P3, P4 are too small, SSE of linear and four order fitting are very close and R-square almost equals 1. Thus we use the linear model as the radiometric response model. According to the fitting results, we can obtain illuminance range with linear response under a certain exposure setting. The non-linear response are the pixels which are under the level of dark current and above the start of saturation. When an illuminance output by the integrating sphere and a gain setting of sensor are unchanged and fitting the response values of increasing sensor's exposure time step-by-step in the linear response range, the R-square of the linear fitting result is 1. The R-square of the exponential fitting result is also 1 when the gain setting is increased exponentially, i.e., ×1; ×2; ×4; and ×8. Based on these radiometric responses corresponding with exposure time and gain setting, a proper exposure setting (i.e., exposure time and gain) can be obtained quickly in certain circumstances. A reference is provided for the final image display and illuminance of projection from object to sensor is obtained.
DN/Lux represents the ratio of the digital number of pixels with incident light illuminance. The ratio value is from fitting results when the dark current is subtracted and pixel values are not saturated. The photosensitive material and manufacturing technique of the sensor determine the sensitivity level. The result shows that the order of sensitivity from high to low is blue, green, and red channels for the same illuminance. We can further calcutate the luminance of an object based on the sensor exposure setting and the quantitative relationship shown in Table 2. dc N Figure 6. Red channel radiometric response of bear sensor with six exposure times (2/25 ms, 4/25 ms, 6/25 ms, 8/25 ms, 10/25 ms, and 12/25 ms), a digital gainˆ1 and 0-80 lux illuminance. X-axis is exposure time, and the unit is 2/25 ms, Y-axis is illumination and the unit is lux, Z-axis is mean value of red channel digital numbers. The four-order polynomial fitting results and linear fitting results are almost equal as the sum of squares due to error (SSE) and the coefficent of determination (R-square) results show. The results show that P2, P3, P4 are too small, SSE of linear and four order fitting are very close and R-square almost equals 1. Thus we use the linear model as the radiometric response model. According to the fitting results, we can obtain illuminance range with linear response under a certain exposure setting. The non-linear response are the pixels which are under the level of dark current and above the start of saturation. When an illuminance output by the integrating sphere and a gain setting of sensor are unchanged and fitting the response values of increasing sensor's exposure time step-by-step in the linear response range, the R-square of the linear fitting result is 1. The R-square of the exponential fitting result is also 1 when the gain setting is increased exponentially, i.e.,ˆ1;ˆ2;ˆ4; andˆ8. Based on these radiometric responses corresponding with exposure time and gain setting, a proper exposure setting (i.e., exposure time and gain) can be obtained quickly in certain circumstances. A reference is provided for the final image display and illuminance of projection from object to sensor is obtained.
DN/Lux represents the ratio of the digital number of pixels with incident light illuminance. The ratio value is from fitting results when the dark current is subtracted and pixel values are not saturated. The photosensitive material and manufacturing technique of the sensor determine the sensitivity level. The result shows that the order of sensitivity from high to low is blue, green, and red channels for the same illuminance. We can further calcutate the luminance of an object based on the sensor exposure setting and the quantitative relationship shown in Table 2. The luminance quantization is only 8-bit when shooting a scene with a certain exposure setting. A visible image can be obtained shooting in an insufficiently lit environment when using long exposure time and large gain and a visible image can also be obtained shooing in a bright environment when using a short exposure time and little gain. The dynamic range can be extended by combining multiple sensors facing in different directions and with different exposure settings. We can roughly calculate that the minimum illuminance value with exposure setting one (i.e., Ex is 6 stop and G isˆ8) and the maximum illuminance value with exposure setting two (i.e., Ex is 1 stop and G isˆ1). The range of pixel values lies between DC to (255 -DC). We can know the illuminances under these two exposure settings when the pixel value are DC and 255 but not saturated. The dark current DC max (EX is 6 stop, and G isˆ8) is almost twice as the DC min (EX is 1 stop, and G isˆ1) as shown in the dark current calibration results. According to Equation (7), the dynamic range of the multi-camera system is almost 48 times as large as that of single sensor exposure at certain exposure times and gain settings.

Correcting Image and Stitching
We have collected seven images using the multi-camera system. The exposure times of Cam1 to Cam7 were 8/25 ms, 4/25 ms, 4/25 ms, 2/25 ms, 4/25 ms, 6/25 ms, and 12/25 ms, respectively. The gain settings are allˆ1. We have subtracted the dark current and corrected vignetting, and then converted the pixel values to illuminance according to the camera photometric calibration results.
The perspective projection relationship between images can be estimated ahead because the cameras are fixed with a certain geometry and the optical axis angle between the central sensor and outer sensor is 40˝. The multi-band blending method provided by Brown [9] is used to blend the stitched panorama image. Finally, the stitched image is projected on a sphere. The image output by Cam3 is shown in Figure 7a   The luminance quantization is only 8-bit when shooting a scene with a certain exposure setting. A visible image can be obtained shooting in an insufficiently lit environment when using long exposure time and large gain and a visible image can also be obtained shooing in a bright environment when using a short exposure time and little gain. The dynamic range can be extended by combining multiple sensors facing in different directions and with different exposure settings. We can roughly calculate that the minimum illuminance value with exposure setting one (i.e., Ex is 6 stop and G is ×8) and the maximum illuminance value with exposure setting two (i.e., Ex is 1 stop and G is ×1). The range of pixel values lies between DC to (255 -DC). We can know the illuminances under these two exposure settings when the pixel value are DC and 255 but not saturated. The dark current DCmax (EX is 6 stop, and G is ×8) is almost twice as the DCmin (EX is 1 stop, and G is ×1) as shown in the dark current calibration results. According to Equation (7), the dynamic range of the multi-camera system is almost 48 times as large as that of single sensor exposure at certain exposure times and gain settings.

Correcting Image and Stitching
We have collected seven images using the multi-camera system. The exposure times of Cam1 to Cam7 were 8/25 ms, 4/25 ms, 4/25 ms, 2/25 ms, 4/25 ms, 6/25 ms, and 12/25 ms, respectively. The gain settings are all ×1. We have subtracted the dark current and corrected vignetting, and then converted the pixel values to illuminance according to the camera photometric calibration results.
The perspective projection relationship between images can be estimated ahead because the cameras are fixed with a certain geometry and the optical axis angle between the central sensor and outer sensor is 40°. The multi-band blending method provided by Brown [9] is used to blend the stitched panorama image. Finally, the stitched image is projected on a sphere. The image output by Cam3 is shown in Figure 7a   We can obtain large field spherical panorama images as shown in Figure 8. The image is a stitched image with the multi-camera one shot. The distance from the multi-camera system to the building is about 20 m.
We can obtain large field spherical panorama images as shown in Figure 8. The image is a stitched image with the multi-camera one shot. The distance from the multi-camera system to the building is about 20 m.

Conclusions and Future Work
A large FOV multi-camera system is calibrated for each sensor and lens. Vignetting and radiometric response of sensors are obtained. We have corrected factors that affect pixel gray values, which cannot objectively reflect the brightness of the scene. The panorama synthesized by using these corrected brightness values can show more realistic landscapes. and the calibration process is an essential step for photometric stereo applications. Multiple sensor systems have higher dynamic range than traditional single sensors with a wide-angle lens. In future work, we will model the geometric properties of cameras and continue to explore the broad prospects for multiple cameras in 3D applications. High-dynamic range information could also be acquired from the overlapping image regions, and tone-mapped or synthetic exposure images can be rendered. An example of stitched and photometric corrected image obtained with this multi-camera system.

Conclusions and Future Work
A large FOV multi-camera system is calibrated for each sensor and lens. Vignetting and radiometric response of sensors are obtained. We have corrected factors that affect pixel gray values, which cannot objectively reflect the brightness of the scene. The panorama synthesized by using these corrected brightness values can show more realistic landscapes. and the calibration process is an essential step for photometric stereo applications. Multiple sensor systems have higher dynamic range than traditional single sensors with a wide-angle lens. In future work, we will model the geometric properties of cameras and continue to explore the broad prospects for multiple cameras in 3D applications. High-dynamic range information could also be acquired from the overlapping image regions, and tone-mapped or synthetic exposure images can be rendered.