Next Article in Journal
Iterative Blind Deconvolution Algorithm for Deblurring a Single PSP/TSP Image of Rotating Surfaces
Previous Article in Journal
Blood Coagulation Testing Smartphone Platform Using Quartz Crystal Microbalance Dissipation Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accuracy Evaluation of Dense Matching Techniques for Casting Part Dimensional Verification

1
Department of Mechanical Engineering, IK4-Tekniker, 20600 Eibar, Spain
2
Department of Smart and Autonomous Systems, IK4-Tekniker, 20600 Eibar, Spain
3
Department of Graphic Design and Engineering Projects, University of the Basque Country, 48013 Bilbao, Spain
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(9), 3074; https://doi.org/10.3390/s18093074
Submission received: 3 August 2018 / Revised: 4 September 2018 / Accepted: 12 September 2018 / Published: 13 September 2018
(This article belongs to the Section Remote Sensors)

Abstract

:
Product optimization for casting and post-casting manufacturing processes is becoming compulsory to compete in the current global manufacturing scenario. Casting design, simulation and verification tools are becoming crucial for eliminating oversized dimensions without affecting the casting component functionality. Thus, material and production costs decrease to maintain the foundry process profitable on the large-scale component supplier market. New measurement methods, such as dense matching techniques, rely on surface texture of casting parts to enable the 3D dense reconstruction of surface points without the need of an active light source as usually applied with 3D scanning optical sensors. This paper presents the accuracy evaluation of dense matching based approaches for casting part verification. It compares the accuracy obtained by dense matching technique with already certified and validated optical measuring methods. This uncertainty evaluation exercise considers both artificial targets and key natural points to quantify the possibilities and scope of each approximation. Obtained results, for both lab and workshop conditions, show that this image data processing procedure is fit for purpose to fulfill the required measurement tolerances for casting part manufacturing processes.

1. Introduction

The present study was focused on the quality assurance requirements and measurement procedures of the foundry-based manufacturing processes and can also be applied to other similar manufacturing processes such as the forge, where manufacturing tolerances may be similar. Nowadays, large scale raw component suppliers are requesting the optimization of their productive processes to compete with lower cost manufacturers worldwide. This optimization demands more efficient and profitable manufacturing methods to control the stability of the processes through procedures that guarantee the quality of the products. Following this objective, a key stage in the manufacture value chain of wind components is the casting phase. Here, the pre-form of the parts that are subsequently machined is generated to obtain their definitive functionality at the operational level. Therefore, casting parts should ensure a minimum oversize that ensures that material exists in those areas of the piece to be machined. Simulation tools such as Computer Aided Design (CAD) software and Finite Element Analysis (FEA) software are employed targeting this aim. Nevertheless, realistic models of the casting process are hardly established due to the complexity of the manufacturing process and the understanding of the involved parameters. Thus, part inspection becomes crucial to verify that the manufacturing process accuracy is under manufacturing tolerances and these results are also used to adjust simulation models as well as to compensate the pattern’s design and size. In this scenario, casting quality characterization encompasses five main categories: casting finishing, dimensional accuracy, mechanical properties, chemical composition and casting soundness. All aspects are crucial to assure that the part meets customer’s specifications and non-destructive testing methods (visual inspection, measuring machines, magnetic particle inspection, ultrasonic testing, etc.) are usually applied for their characterization.
An accurate dimensional inspection is commonly applied for First Article Inspection (FAI) or even for serial production control, where commonly part-to-CAD analysis and/or geometrical analysis according to ISO 8062-3 Standard is required for manufactured parts conformity assessment.
Either the casting manufacturing processes or the measurement procedures and technologies have evolved to speed up the result obtaining process to feed faster the production quality control stage. Thus, currently, a clear tendency aims to integrate in-situ measurements during the manufacturing process to substitute final closeout inspections. However, total time consumption for part inspection should be tackled to offer a rapid response. Here, two stages can be clearly distinguished: the acquisition and the processing of the measures.
Nowadays, technologies applied in the verification process do not allow a total automation of the inspection stage for several main reasons: the dimensions of the pieces, the multiple stations (positions) of the inspection systems as well as their lack of autonomy. Therefore, there is a latent opportunity in terms of improving existing procedures to reduce inspection times. In this scenario, the present paper highlights dense matching techniques to solve the aforementioned problem.

1.1. Background

The state of the art of 3D measuring technologies and methodologies for casting part verification have substantially been developed in the last few years, offering faster, more complete and more automatic approaches. The evolution of measurement methodologies is closely related to the employed measurement technologies and the output data that they provide. To understand the portfolio of common measuring technologies for medium to large scales components [1,2,3] an overview is presented in Figure 1, where an evolution from contact to non-contact approaches is presented and also a data processing procedures evolution is shown. The difference and the use of each measuring technology depends on the measurement case study and the dimensional analysis required in each case. Therefore, the selection of the most appropriate technology depends on the measurable requirements such as: measuring uncertainty, point resolution, type of processing, measuring scale, part accessibility, etc.
Traditionally, casting part dimensional verification has been executed by handheld manual instruments such as calipers, calibrated fixtures or measuring tapes. However, these devices do not allow verifying the main dimensions of large parts which leads to partial verification with low process quality confidence. To overcome this limitation, large Coordinate Measuring Machines (CMM) and Portable Coordinate Measuring Machines (PCMM) were developed many years ago pushed by large part manufacturers. CMMs are the most accurate approach but they require bringing the part to the machine which is time-consuming and sometimes an unaffordable option due to part dimensions. PCMMs aim to cover this physical limitation. Nowadays, many kinds of systems and brands exist [4,5]. A review of 3D measuring systems is also presented related to manufacturing verification needs in several scales and suitability of measuring devices and techniques is studied for 3D parts.
Laser tracker technology was the pioneer in volumetric eolic component verification (Figure 2). Its parametric modeling is mentioned in [6,7], whereas a general review of its performance and applicability is presented in [8]. It has been widely used for machined component inspection for almost 20 years, thus its application to casting part inspection is imposed by machining companies in the eolic component manufacturing chain. It offers the possibility to quantify and verify the overall dimensions of the part with high accuracy requirements as well as allows verifying the accuracy performance of other measuring or manufacturing means [9,10,11,12]. Recent laser-tracker developments include hand-held accessories to scan the part surface without contact by means of real-time tracking of a laser-triangulation sensor or probe and even highly automated scanning robotic approaches. Compared to initial laser tracker technology that only offers the possibility to measure a discrete point cloud touching the part surface with a calibrated retroreflector, automatic eolic component inspection procedures come closer to the goal of surface scanning.
Conventional shape and dimensional analysis of casting components with tactile coordinate measuring systems has some limitations, such as: discrete point measurement strategy, lack of free-form shape information and time consumption for volumetric component measurement. In the last few years, 3D optical scanners [5] are replacing laser trackers as the most used measuring equipment for the dimensional quality control assessment of machined and casting parts. Modeling and calibration issues are mentioned in [13,14,15,16], whereas some examples of application are studied in [17,18] and even accuracy performance is determined for lighting conditions in [19]. Although their accuracy is still a bit lower (±30 µm/m) than laser-tracker technologies (±15 + 6 µm/m) for high accuracy applications, casting part verification methodologies have been adapted to these technologies as they offer a volumetric view of the part compared to its nominal CAD definition. Optical scanners such as laser-triangulation sensors [20], structured light sensors (Figure 3), or even Time Of Flight (TOF) based laser- scanners [21,22] have the ability to measure dense point clouds that are used to analyze the oversized material distribution or carry out reverse engineering tasks. They capture more detailed and easily interpretable quality information of an object with significantly shorter measuring times and accuracies ranging from 0.05 mm to 2–3 mm for large parts (several meters) which assures post-machining steps. Many technology suppliers already offer this kind of equipment as an alternative to tactile approaches.
The main drawbacks of these optical systems are data acquisition time and part accessibility. As the part is usually bigger than the field of view (FOV) of the measuring system and the required point cloud resolution is quite demanding, the scanning system needs to be located in several positions around the part to scan every surface point. This means that each device location should be accurately aligned to previous locations with common references to reconstruct the 3D point cloud in a common coordinate system. However, these partial alignments add inaccuracy to overall point clouds as a residual error is carried out each time that can suppose a non-continuous point cloud, especially for round parts. Besides, this working procedure forces to see common reference targets among different consecutive scanner locations, which makes the measuring procedure a time consuming and restricted method. Currently, to tackle and enhance the accuracy of device location alignment stage, structured light based scanners are combined with photogrammetric approaches which guarantee higher accuracy referencing points. This method comprises two measuring systems and therefore steps. First, a photogrammetric approach is used to measure some dot targets all around the part and these artificial targets are used as a reference for the subsequent scanning step. The points are identified and used as tie points for each partial scanning, enabling to fuse all 3D data in a common reference system. Moreover, the accuracy is improved as the output of the previous photogrammetric step is more precise than locating the scanners in multiple locations and applying consecutive point cloud fusions which leads to overall 3D point cloud inaccuracies. Besides, using previously calibrated dot targets as a reference enables taking partial scans all around the part without having to assure a consecutive overlapping among each scan and consequently offering a more flexible measuring procedure. Another possibility to improve data acquisition time is to use robotic measuring cells. Nevertheless, this approach is nowadays oriented to other manufacturing sectors where required investment can ensure a fast-economic return.
X-ray computed tomography (CT) is a new technology that offers not only the output of 3D point cloud of the surface but also the inner 3D structure for material integrity checking. However, this technology is currently limited to 1 m range due to required installation costs and X-ray source power. Some references regarding to technique description, modeling, accuracy and applicability are mentioned in [22,23,24,25].
Dense matching techniques are the development of vision-based systems to measure dense 3D point cloud from imagery data. It comprises two main stages (see Figure 4) to reconstruct the dense point cloud: camera network calibration and dense image matching based on previously estimated photogrammetric results. In the literature, network orientation is also referred to as a structure from motion technique [26,27,28,29]. It targets going beyond traditional photogrammetric solutions improving image processing algorithms for textured surface reconstruction. Therefore, it requires texturized surfaces which means that non-homogeneous images are needed to feed and solve feature-based matching problem [30], stereo pair triangulation and multi-view bundle block adjustment [31,32,33,34].
On the one hand, the photogrammetric stage allows sorting out the bundle adjustment of the photogrammetric model the results of which are: 3D point coordinates (XYZ) by means of multi-view triangulation, self-camera calibration (intrinsic parameters) and orientation and position of the images (extrinsic calibration) according to the main coordinate system. These output values are afterwards the input data for dense matching approaches [35]. The photogrammetric stage can be carried out with coded targets [36] as well as based on natural tie points [37].
On the other hand, dense matching techniques [38,39,40] accomplish 3D dense point cloud reconstruction from oriented images and considering known camera intrinsic parameters. Tie points are combined in a bundle block adjustment, resulting in the 3D coordinates of all tie points and, more importantly, the position and orientation of each image. If these parameters are previously and accurately estimated, the expected results for dense matching approaches are supposed to be more precise. An accuracy study of this reconstruction method and algorithms is presented in [41,42]. Matching techniques [43,44] are usually applied in two steps (see Figure 4) to obtain at first sparse matching points (few key points) and then to use them as input for dense matching and triangulation (dense point cloud reconstruction). Matching algorithms are based on known relative orientations among images and image-pair searching schemes. For each image pixel on a reference image, a corresponding point is established in other images and views by means of pixel-based matching algorithm [38,39,40]. Once the dense correspondence is solved, stereo or multi-view triangulation [45] is enabled and determined, which offers 3D colorized dense point clouds from epipolar images and constraints [41,42,43,44]. Most e approaches are based on the minimization of cost functions that consider the degree of the similarity among pixels and includes constraints to consider possible errors in the matching process as well as geometric discontinuity changes.
An increasing number of software solutions for the automatic or semi-automatic generation of textured dense 3D point clouds images have recently appeared in the market [46,47,48]. iWitnessPRO-Agilis© photogrammetric software has evolved to dense reconstruction capabilities. It combines a photogrammetric library (Australis©) and the SURE© tool from the Institute For Photogrammetry at the University of Stuttgart to enable 3D point cloud reconstruction with different camera network calibration approaches (semi-automatic, automatic with coded target, and targetless orientation by means of feature-based matching) [49,50].

1.2. Motivation

Although dense matching photogrammetric based applications are extending more and more into industrial applications, the casting part verification with these novel photogrammetric methods has not been studied before. Most applications target cultural heritage, aerial topography or satellite surveying/mapping applications, but not industrial applications. Thus far, these approaches have not been studied from the point of view of accuracy and traceability, because there is not much information in the current literature [51,52]. It is within this context that this paper aims to provide an approach to assess the accuracy of dense matching techniques from a metrological point of view. Therefore, the approach adopted here is based on established and certified measuring approaches whose accuracy is already known and is used as a reference.
Considering that usually employed certified methods combine photogrammetry with further scanning processes aiming to increase point cloud accuracy and surface continuity, this study established the scope of dense matching techniques as an alternative verification approach employing a unique hand-held industrial camera. If the imagery data from photogrammetric approaches are also suitable to obtain dense point clouds, there is no need to scan the part after the photogrammetric stage has been carried out, which would considerably shorten the overall measuring time and reduce the cost of the measuring solution and technology.
The reduction of the measuring process duration would enable lower cost measuring services and faster feedback for casting part manufacturers to improve the process quality control for high value parts.

1.3. Objectives

The objectives of this research were to understand the accuracy assessment and measuring process requirements for dense matching photogrammetric techniques focused on the verification of medium to large casting parts. One remarkable aspect, in addition to sufficient accuracy to fulfill the manufacturing tolerances, is the time consumption required to carry out the dimensional verification of the parts. To consider this novel method as a suitable procedure, the overall process should reduce time consumption at least during the image data acquisition stage. This point led to carrying out a comparison of automated reconstruction approaches based on coded-targets or target-free schemas. Other factors such as adaptability of the process for different part dimensions and part complexity were also considered in this research.
Targeting the above-mentioned goals, a certified method for 3D verification was considered as a reference to quantify, determine and compare the achieved results between both measuring procedures and therefore to validate the studied approach.
Apart from these objectives, this study also focusede on the adaptability and improvement of dense matching techniques for the case presented, assessing the pros and cons of this method against current state of art approaches.

2. Materials and Methods

2.1. Measurement Method

Dense photogrammetry is an image processing technique based on 2D images to obtain 3D information based on the texturing of the surface to be measured. The multiple images that ensure high overlapping areas are processed with a photogrammetric analysis software to obtain a dense point cloud. The software employed in this research is called iWitnessPRO-Agilis©. This software is a photogrammetric tool (iWitnessPro©) enhanced with dense matching capabilities (SURE© tool) to construct dense point cloud estimation based on image data.
To execute dense photogrammetry, it is necessary to have a part with well-conditioned surface texture such as casting or forged parts. Without this texture, the dense reconstruction of the surface by means of image processing is not possible. Homogeneously texturized parts, such as machined components, cannot be measured with this method.
This research aimed to allow as automatic as possible data processing, thus, in addition to the textured surface, coded targets were placed on the measurand to facilitate this aim. The approach considering only natural key points (feature-based matching-approach) is offered by the software but the processing requires users few controls such as data scaling. The location of coded targets are recommendable to solve with precision the following aspects: the calibration of the camera itself (intrinsic parameters), the external orientation (extrinsic parameter) of each image obtained and the spatial XYZ coordinates of each target whose center is to be determined. The technique also allows sorting the photogrammetric problem out without coded targets if the part offers characteristic geometric elements that can be discerned in several images as reference unequivocal points. However, it is expected that the accuracy is reduced if the use of targets is avoided, thus advantages and disadvantages of each approach should be considered. To automate the processing of the images, it is necessary for the software to independently calculate the external orientation, internal and scaling of the images, in addition to executing the dense reconstruction from these known data. To guarantee this result, the software performs two main stages. First, the photogrammetry is solved and then the dense matching technique is applied to obtain dense point clouds. This dense matching technique employs photogrammetric results to establish the correspondence among images based on image neighborhood data (average pixel values) and define the surface points by means of a combination of multi-stereo pairs and their triangulation for all the images. The overall workflow that comprises both processing stages is shown in Figure 5.
To guarantee accurate results, both photogrammetric and photography requirements have to be considered. If only photogrammetric camera set-ups are used, the surface points are not well focused, and the dense matching is not working. Therefore, a balance between both fields should be considered when image acquisition is carried out. A more detailed definition of the measuring process and the data processing follows these steps and scheme:
  • Preparation of the cast part with targets (if necessary) and a calibrated length bar
  • Adjustment of the camera and lens to measure the measuring scenario
  • Image acquisition process with fixed camera settings
  • Data processing
    (a)
    Photogrammetry with coded targets or targetless approaches
    (b)
    Dense matching
  • Dense point cloud filtering and refinement
In each step, multiple aspects should be analyzed before the measurement to assure that the overall measurement workflow will be successful. The most critical ones are listed below:
  • Camera and lenses adjustment: It defines the resolution, the depth of field, exposition time, lenses aperture, the contrast of surface tie points, etc.
  • Image network design: Minimum overlapping among images both for photogrammetry and dense matching reconstruction is necessary to detect coded targets, key points or unique surface points.
  • Camera self-calibration: Intrinsic parameters (focal length, principal point, optical distortions, etc.) are determined for each measurement as camera internal parameters are not temporally stable.
Thus, the edge-cutting method presented in this research requires a good combination of these influencing factors to obtain suitable results. In the following paragraphs, a detailed description of employed adjustments are defined for the developed measuring method.

2.1.1. Preparation of the Scene

The criteria for the placement of the coded targets (based on dot distribution) are the following:
  • Minimum five common targets between different images
  • Robustness is improved if there are eight or more targets
  • Size of the target should ensure an approximate image size of 10 pixels
  • Image processing module of the software cannot detect another type of coding, which is the reason to use this type of targets. They can be white on black or red on black.
In addition to the targets, it is necessary to locate on the scene to be measured a calibrated length bar (see Figure 6) that establishes the scale of the acquired images. If there is more than one artifact, the scaling is performed with the average value adjusted for both measurements. It is interesting for the scaling to be automatic to use a bar with encoded targets so that the software in question (iWitnessPRO-Agilis©) knows the points to which the scaling corresponds. It is necessary to make this definition in the software at least once.

2.1.2. Camera Adjustment

Once the scene is prepared, it is necessary to adjust the camera parameters to the lighting conditions of the room where the images are taken. Normally, in a photogrammetric process, the camera adjustment is set on the targets to be measured (maximum aperture and a zoom of 10×), but in a dense matching it is also necessary to consider a good contrast on surface points to be scanned. Therefore, when adjusting the camera, a compromise should be found between the coded targets and the surface texture to be measured. The main parameters to be adjusted are:
  • Working distance
  • The aperture
  • Flash intensity (if required)
  • Depth of field
These parameters should be kept constant for the entire measurement to solve the photogrammetric bundle adjustment.
The identification of the camera with the software is automatic and therefore the parameters related to the sensor (resolution and image dimension) are determined automatically. However, it is necessary to verify these parameters and change them if necessary. In this study, a Canon EOS-1Ds Mark III camera was used with 25 mm focal length lenses.

2.1.3. Acquisition of Images

When it comes to image acquisition process, it is necessary to consider what is intended to be measured both with photogrammetry and dense matching techniques. Therefore, in addition to photogrammetric aspects, a proper contrast and triangulation of surface points should be guaranteed. For this, it is necessary to reach a compromise between a good angle of triangulation and an excessive change of perspective between consecutive images. Thus, high angular views should be avoided and angles ranging from 60° to 110° are recommended. Depending on the relative angle between pairs of images, the accuracy of the photogrammetric or dense reconstruction stage will be reinforced.
In addition to triangulation, it is important to take care of the internal calibration aspect, for which it is necessary to acquire rotated images on the image plane. Normally, 4 × 90° rotated images are acquired, respectively, at the beginning of the measurement. The same criteria of acquiring rotated images during the measurement is strongly recommended to enhance camera self-calibration.
In this study, the images were taken around the part following the clockwise criteria and combining different height and view angles for each image acquisition position (see Figure 7).

2.1.4. Image Data Processing

Once the images are acquired, the processing of the photogrammetry and dense matching is carried out in two steps. The bundle adjustment is automatically solved if coded targets are employed for determination of image orientation and scaling, while the processing is more manual when natural key points are determined from imagery. The second step, named dense matching, is supposed to be automatic, but depending on the image data casuistic, few parameters should be fitted, such as: density of point cloud, image-pair combinations, pixel point neighborhood size, etc.
The extension of the generated file format for dense point cloud is “las” extension, therefore CloudCompare© software is used to visualize the point cloud and to convert it to more traditional “ascii” format files as well as to select the region of interest (ROI) of the dense point cloud. These files are afterwards used for the comparison of measured against reference data.

Photogrammetric Processing

Based on coded targets
This is the most common method to solve the bundle adjustment in industrial photogrammetric systems as the automation level is high. Only preliminary image processing parameters need to be adjusted to assure solving intrinsic orientation, extrinsic orientation and coordinates of 3D point of interest at the same time without knowing a priori values of these parameters.
The coded targets enable the automatic referencing between image pairs and among every image rejecting outliers and obtaining more robust image correspondence as the point identification is unequivocal. Moreover, the 3D scene scaling is carried out, automatically identifying the coded targets that correspond to the calibrated length bar (see Figure 8). The main drawback of this approach is the preparation of the scene with the artificial targets and the assuring of a minimum number of common points among images.
Based on Feature Based Matching (FBM)
It is a different photogrammetric processing approach to estimate the extrinsic calibration of images. Instead of using artificial coded targets, natural key features are identified in the images and the correspondence among the multiple stereo pairs is estimated by means of image matching techniques. Approximately 50,000 feature points are used in each image. After establishing the tie points among the images, the relative orientation between image pairs is determined and the absolute orientation considering the first image as a reference. Finally, a bundle adjustment (minimum convergence of three image rays) approach is applied to refine and improve the extrinsic calibration of the camera network.
Once the camera network is defined, 3D scene scaling is established, which is necessary to obtain a metric 3D point cloud. This step is manually executed by choosing an image pair where the calibrated length bar is seen. Based on manual target centroiding identification and epipolar image correspondence, the image points corresponding to the artifact were selected and applied to a known calibrated distance (see Figure 9).

Dense Matching Reconstruction

The dense matching reconstruction is performed by means of SURE© software, which is a tool for photogrammetric surface reconstruction from imagery. It is a software solution for multi-view stereo, which enables the derivation of dense point clouds from a given set of images and known orientations (see Figure 10). Nearly one 3D point can be obtained corresponding to each image pixel with the highest resolution.
The input to SURE is a set of images and the corresponding interior and exterior orientations (structure from motion approach), whereas the output information of the software is dense point clouds or depth images. It is a multi-stereo solution, so single stereo models (using two images) are processed and subsequently fused. It automatically analyzes the block configuration to find suitable stereo models to match in several steps (coarse methods for stereo pair selection and accurate methods for dense stereo 3D reconstruction). During the 3D reconstruction step, the disparity information from multiple images is fused and 3D points or depth images are computed. By exploiting the redundancy of multiple stereo pairs and combinations, blunder and outliers are filtered or rejected. In this research, the full resolution of the camera (21 Megapixels) was used and a point was triangulated as minimum with a stereo pair.

2.2. Validation Procedure for Accuracy Assessment

The validation of the presented methodology was performed by means of a “two-step” measuring procedure combining an optical 3D scanner (ATOS IIe©, Braunschweig, Germany) based on structured light with an optical portable CMM (TRITOP©, Braunschweig, Germany) based on photogrammetry, which assures high accuracy reference data used for the comparisons presented in this research. The fusion of these technologies enables obtaining higher accuracy point cloud data (grown truth) for medium to large dimension parts and a more flexible scanning procedure.
First, a photogrammetric procedure was applied to measure the XYZ coordinates of all the targets placed in the measuring scenario (see Figure 11a). This was accomplished by TRITOP system employing the ring coded targets, a high-tech industrial camera (see Figure 11b), and advanced photogrammetric software to solve the bundle adjustment problem. Dot type coded targets were used as fiducial references for the subsequent scanning stage as the TRITOP system considers these targets as non-coded targets. Each partial scan was referenced by means of these fiducials to a common reference system defined by previously executed photogrammetric stage. The targets based on dot type codification were also used for the automated orientation case study in further analysis presented in this paper. Therefore, both types of coded targets (see Figure 11a) were placed in similar locations to guarantee that the photogrammetric problem can be solved and the reachable accuracy is practically the same.
Once that photogrammetry was solved, a dense 3D scanning measurement of the part was executed with the available structured light 3D scanner (see Figure 11c) using the dot type fiducials coded as references for the partial alignment of each scan. Known fringe patterns were projected onto the surface of the object and recorded by two cameras, based on the stereo camera principle. As the beam paths of both cameras and the projector were calibrated in advance (extrinsic calibration), 3D surface points from three different ray intersections could be calculated. The number of scans depended on the part surface complexity and required detail. Every partial scan was transformed to a common coordinate system employing as a reference the points that were measured before with the photogrammetric method. Once the part was totally scanned, the point cloud was transformed to a continuous mesh by means of Delaunay triangulation approaches [53,54,55,56].

3. Results

The following content presents the results of the comparison of the studied dense matching approaches against photogrammetrically referenced structured light 3D scanner for both lab and shop-floor lighting conditions. Moreover, for each different case study, the photogrammetric orientation problem was solved considering artificial coded targets or natural key features employing the same acquired image dataset.
One of the main differences between using the artificial coded targets or natural key features is the processing time and the automation level for photogrammetric model solving. Whereas the artificial coded target approach is totally automatic for camera network orientation estimation and scaling, the feature based targetless orientation approach requires a specific feature identification image parameter adjustment as well as manual 3D point cloud scaling. The processing time takes twice as long as using coded targets and increases exponentially with the number of images, feature point number and camera resolution. Coded target employment is a method to optimize this step, from the point of view of robustness and computing cost. The highest processing time is for dense matching reconstruction which takes several hours to obtain a full resolution 3D point cloud. The obtained accuracy is shown and discussed below for each case study.
To compare and assess the accuracy of the novel methodology, free 3D inspection software was used (GOM Inspect©, Braunschweig, Germany). The employed method for any comparison was the following one:
  • Import both meshes: the reference data measured by GOM© systems and the one achieved by dense photogrammetry.
  • Perform raw alignment between meshes with six nearly common points on the surface.
  • Perform best-fit alignment for accurate registration considering the reference data as nominal mesh.
  • Compare the 3D color map between aligned meshes and analyze deviations among measured points.
  • Perform statistical evaluation of achieved deviations with histograms (2σ).
In the case of camera network estimation by means of FBM, as this method offers rather high number of 3D points within the photogrammetric stage, the surface comparison is presented both for sparse point cloud and dense point cloud case studies. Although other alignment methods (geometric, datum-based, and 3-2-1) are available, best-fit approach has been chosen as it estimates the minimum average oversize that ensures that material exists in those areas of the piece to be machined. Moreover, it has been considered as a more statistical alignment method for the comparison as it gives the same weight to all surface points instead of fitting some geometric elements and taking them as a reference to define the coordinate system.

3.1. Lab Tests for Medium Size Parts

3.1.1. Orientation Based on Artificial Targets

The first test was performed in lab conditions with stable lighting and environmental conditions. The part used for the experiment was an assembly of three rusted cylinders of Ø 50 mm × 500 mm in height with stochastic surface texture, as shown in Figure 12. After reference data acquisition (32 images), the overall dense photogrammetry was applied following the measuring procedure described in Section 2. The acquired images were initially used to solve the photogrammetry by means of coded-targets and then the dense point cloud reconstruction was established with high resolution.
The relative accuracy of the photogrammetry was 1/80,000 mm (RMS) and the camera self-calibration was accurately established. Regarding the surface data comparison, the deviations fell into ±0.2 mm (2σ), as shown in Figure 12a and its histogram. The obtained dense point cloud contained 10.5 million points (see Figure 12b), and coded targets were placed on lacking areas for determination of orientation. The higher deviation was found in edge points as the triangulation and the definition of the averaged points for dense matching is more challenging in these areas.

3.1.2. Targetless Orientation Based on Feature Based Matching

The second test for lab conditions was performed on feature-based matching for camera network orientation establishment. The same pictures as in Section 3.1.1 were used but coded-targets were not employed to define the absolute orientation of the images. Once the photogrammetric part was solved, the data were scaled (sparse point cloud) and afterwards dense matching was done to obtain dense point cloud.
The relative accuracy of the photogrammetry was 1/27,000 mm (RMS). Camera calibration was considered the same as obtained in the previous analysis due to higher accuracy assurance and convergence difficulties with self-calibration attempts. Regarding the surface data comparison, the deviations fell into ±0.4 mm (2σ), as shown in Figure 13a (sparse point cloud), and ±0.8 mm (2σ), as shown in Figure 13b (dense point cloud). The obtained dense point cloud contained 0.2 million points (see Figure 13b) with empty gaps were coded targets were placed. Again, the higher deviations were found in edge points and a difference of 200% was obtained comparing dense point cloud and sparse one.

3.2. Workshop Tests for Large Parts

3.2.1. Orientation Based on Artificial Targets

The preliminary test executed in workshop conditions did not assure stable lighting or proper accessibility to every part surface. The part used for the experiment was a casting eolic hub of 2500 mm × 2500 mm × 1500 mm with stochastic surface texture, as shown in Figure 14. After reference data acquisition (150 images), the dense photogrammetric was applied following the measuring procedure described in Section 5. The acquired images were initially used to solve the photogrammetry with coded-targets as in lab tests and then the dense point cloud reconstruction was determined.
The relative accuracy of the photogrammetry was 1/157,000 mm (RMS) and the camera self-calibration was accurately established. Regarding the surface data comparison, the deviations fell into ±3 mm (2σ), as shown in Figure 14a and its histogram. The obtained point cloud contained 2 million point with 0.2 mm resolution (see Figure 14b). The highest deviations were found in edge points, as occurred for lab results. The overall surface of the part was not scanned but most of the surfaces were measured, as shown in Figure 14b. To complete the scanning of every surface of the part, another part location is necessary, enabling the non-accessible surfaces to be measured. This requirement is usual for traditional measuring procedures but in this test a unique part location was applied and considered as enough for comparison.

3.2.2. Targetless Orientation Based on Feature Based Matching

The second test for workshop conditions was performed on feature-based matching for camera network orientation establishment. The same pictures as in Section 3.2.1 were used but coded-targets were not employed to define the absolute orientation of the images. Once the photogrammetric part was solved, the data were scaled (sparse point cloud) and afterwards dense matching was carried out to obtain dense point cloud.
The relative accuracy of the photogrammetry was 1/38,500 mm (RMS). Camera calibration was considered the one obtained in the previous analysis due to higher accuracy assurance and convergence difficulties with self-calibration attempts. Regarding the surface data comparison, the deviations fell into ±3 mm (2σ), as shown in Figure 15a (sparse point cloud), and ±6 mm (2σ), as shown in Figure 15b (dense point cloud). The obtained dense point cloud contained 0.5 million points (see Figure 15b) with lacking areas at the upper part of the measurand. Again, the higher deviations were found in edge points and a difference of 200% was obtained comparing dense point cloud and sparse one.

4. Discussion

The trend in casting part dimensional inspection is to use smarter systems and procedures to speed up the measuring process assuring the acceptance criteria of the manufactured parts. Traditionally, measuring devices such as laser trackers or 3D optical scanners are used for FAI inspection or serial quality control. Time of flight based 3D laser scanners are also a recent alternative for 3D information acquisition of casting parts with uncertainties of about ±0.5 mm/m. Nevertheless, these devices do not allow automatic measurements because the measuring system has to be located and adjusted manually in different positions around the part. Partial measurements are placed in a unique coordinate system with the location of common artificial targets around the part.
Apart from this, other objectives such as improvement of the 3D point cloud processing step, automation of the measuring process, reduction of acquisition cost of the measuring devices, and robustness of the measuring process are also desired. This research was focused on studying the capabilities and scope of dense matching techniques and their suitability to be applied to casting part verification, where the manufacturing tolerances are not as tight as with machined parts. A comparison of dense matching approaches with traditional verification approaches is shown in Table 1 as a summary and overall overview of measuring alternatives’ pros and cons. The relative accuracy is compared taking into account only lab conditions and current standards as specifications of currently employed systems are defined in this environment.
The main advantages of studied methods as follows: fast acquisition time, high accessibility to multiple surfaces, high resolution (one point per pixel), high level of automation, and scalability to parts dimensions. The disadvantages are as follows: lower precision than other measuring techniques, the results contain points with noise, data are missing in the obtained point cloud depending on employed resolution and the extent is unknown until overall processing is carried out, high processing time, and expert user is required to ensure an accurate result in non-stable lighting environments.
Another important aspect related to this novel method is the low cost of the whole system comprising both the hardware and the software as well as the flexibility of the system for different scenarios and measuring scales. Moreover, the automation is feasible with a fixed multi-camera network set-up or CNC positioning systems enabling the possibility to develop inspection cells for manufacturing quality control. Currently, one important limitation to completely automatize the solution, is the offline processing step and high computation time for PCs. However, this drawback is day by day improving as the technology is constantly developing and smarter algorithms are being developed. Thus, in few years this limitation will be overcome and even real-time dense matching will be possible. Other aspects to be improved in further developments are the robustness of the technique from the point of view of rejection of outliers and data filtering to extract accurate results as well as development of standards and guidelines based on this technology.
Currently, advanced users are required to apply the measuring process described in this paper, but controlling lighting conditions can reduce this requirement and make more useful this procedure to all photogrammetric users. This way, this procedure could be offered as an alternative service for medium and large casting parts instead of applying other 3D approaches.
In relation to accuracy, dense matching techniques seem to be a suitable measuring procedure for casting part verification as they enable obtaining dense point cloud with accurate enough accuracy (lower than manufacturing tolerance) for part shape and dimension verification. The estimated relative accuracy (RMS) of the measured 3D point coordinates is determined as the ratio of the estimated point standard error over the effective maximum diameter of the object point array along with the corresponding proportional accuracy. Coded-target based approach is the most accurate approach. The grade of tolerances for casting is mentioned in ISO 8062-3:2007 Standard. This part of ISO 8062 defines a system of tolerance grades and machining allowance grades for cast metals and their alloys. It applies to both general dimensional and general geometrical tolerances, although this study was focused mainly on general dimensional verification. As a reference, for a common tolerance grade 12, tolerances of ±10 mm and ±17 mm are required for the dimensions of the parts analyzed in this study. Thus, obtained accuracies are fit to purpose considering these manufacturing tolerances.
A more detailed comparison of analyzed dense matching case studies is presented in Table 2, where obtained accuracy as well as required time for different measuring steps is shown as a summary of the results obtained in this research.
The relative accuracy (proportional RMS) for automated orientation based on coded-targets is 3–4 times higher compared with feature based matching approaches, for both lab and shop-floor case studies, whereas the deviations against traceable measuring methods are approximately twice lower. This difference is mainly due to the referencing step during the photogrammetric problem which is directly influenced by the accuracy of identification of key points for referencing. Whereas the automated orientation method employs the centroiding tool to estimate the center of the high-contrast elliptical targets with an accuracy of 0.03–0.1 pixels, targetless orientation approach uses less accurate (0.1–0.5 pixel) point identification methods, which leads to lower reconstructed point cloud accuracy and point cloud density. The final accuracy of 3D dense point cloud is a direct function of the referencing accuracy, so the more precise the point marking and referencing approach, the better the triangulation and the higher the point quantity of the reconstructed point cloud.
Higher image number and perspectives could also improve the triangulation of each reconstructed point and therefore the relative accuracy for both cases, but it will complicate the data processing phase, thus a compromise between accurate photogrammetry and suitable dense matching needs to be found. The number of reconstructed points totally depends on this balance as a more demanding point triangulation limit gives a lower density point cloud and vice versa. In this research, a minimum intersection of three viewpoints was required to estimate a 3D point. As the camera network orientation for targetless case is less accurate than the one with coded targets, the same parameterization for the point triangulation step gives lower dense point cloud and some missing areas.
The measuring time depends on the preparation of the scene and camera set-up as well as in the size of the part to be measured. However, comparing to other measuring methods, it is fast as only images of the part need to be taken from different perspectives and heights. One of the drawbacks of dense matching techniques is the data processing step once the images have been taken. Whereas photogrammetric problem estimation is affordable from the point of view of time compared to industrial solutions, dense matching step takes much time and PC memory, which makes it more difficult to process the same imagery with different image processing parameterization.

5. Conclusions

Edge-cutting dimensional verification technologies ensure a reliable dimensional assessment of casting parts to check if they fulfill the required conformity assessment. Low enough measuring uncertainties guarantee a proper part inspection, machining process and therefore an overall manufacturing process efficiency.
3D PCMMs are being used more and more nowadays for inspection tasks with advanced probes for data acquisition improvement, but they still require to be moved around the part which takes a long time for data acquisition. Besides, depending on part shape and dimensions, these techniques are hardly applicable to quality control requirements. Dense matching techniques based on imagery data seem to be an affordable alternative to these measuring techniques as they enable faster data acquisition schemas with rich data information. The processing of these data permits reconstructing dense point cloud with high relative accuracy. Therefore, volumetric surface data are estimated which enables surface data comparison against nominal CAD data.
However, this study should be complemented with other accuracy evaluation methods and applications where dense matching techniques are used for verification of textured industrial parts. Currently, few studies are contained in the literature from the metrological point of view that tackle measuring procedure uncertainty or traceability of these measuring techniques. For example, another procedure to establish the measuring accuracy could be to manufacture a calibrated part and use it as a known reference or to characterize each influence error to estimate the measuring uncertainty applying the “error propagation law” as it is carried out for calibration and adjustment of other metrological complex devices. The efficiency of data processing also needs improvement by utilizing parallel processing and hierarchical optimization techniques to quicken result procurement.

Author Contributions

E.G.-A. was in charge of the project and conceived the main objectives; U.M. and G.K. designed and performed the experiments; G.K. analyzed and processed the data; A.T. and R.M. reviewed the overall concept; G.K. wrote the paper; and all co-authors checked it considering their contribution and expertise.

Funding

This research was partially funded by ESTRATEUS project (Reference IE14-396).given are accurate and use the standard spelling of funding agency names at https://search.crossref.org/funding, any errors may affect your future funding.

Acknowledgments

This research was supported by the Basque business development agency and is the output of ESTRATEUS project (Reference IE14-396) where novel manufacturing and measuring methods were studied as strategic lines for Basque industry development.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cuypers, W.; van Gestel, N.; Voet, A.; Kruth, J.; Mingneau, J.; Bleys, P. Optical measurement techniques for mobile and large-scale dimensional metrology. Opt. Lasers Eng. 2009, 47, 292–300. [Google Scholar] [CrossRef] [Green Version]
  2. Schmitt, R.H.; Peterek, M.; Morse, E.; Knapp, W.; Galetto, M.; Härtig, F.; Goch, G.; Hughes, B.; Forbes, A.; Estler, W.T. Advances in Large-Scale Metrology—Review and future trends. CIRP Ann. Manuf. Technol. 2016, 65, 643–665. [Google Scholar] [CrossRef]
  3. Muelaner, J.E.; Maropoulos, P.G. Large volume metrology technologies for the light controlled factory. Procedia CIRP 2014, 25, 169–176. [Google Scholar] [CrossRef] [Green Version]
  4. Goch, G.; Knapp, W.; Ha, F. Precision engineering for wind energy systems. CIRP Ann. Manuf. Technol. 2012, 61, 611–634. [Google Scholar] [CrossRef]
  5. Savio, E.; de Chiffre, L.; Schmitt, R. Metrology of freeform shaped parts. CIRP Ann. Manuf. Technol. 2007, 56, 810–835. [Google Scholar] [CrossRef]
  6. Conte, J.; Santolaria, J.; Majarena, A.C.; Acero, R. Modelling, kinematic parameter identification and sensitivity analysis of a Laser Tracker having the beam source in the rotating head. Meas. J. Int. Meas. Confed. 2016, 89, 261–272. [Google Scholar] [CrossRef]
  7. Conte, J.; Santolaria, J.; Majarena, A.C.; Brau, A.; Aguilar, J.J. Identification and kinematic calculation of Laser Tracker errors. Procedia Eng. 2013, 63, 379–387. [Google Scholar] [CrossRef]
  8. Muralikrishnan, B.; Phillips, S.; Sawyer, D. Laser trackers for large-scale dimensional metrology: A review. Precis. Eng. 2015, 44, 13–28. [Google Scholar] [CrossRef]
  9. Aguado, S.; Santolaria, J.; Aguilar, J.; Samper, D.; Velazquez, J. Improving the Accuracy of a Machine Tool with Three Linear Axes using a Laser Tracker as Measurement System. Procedia Eng. 2015, 132, 756–763. [Google Scholar] [CrossRef]
  10. Nubiola, A.; Bonev, I.A. Absolute calibration of an ABB IRB 1600 robot using a laser tracker. Robot. Comput. Integr. Manuf. 2013, 29, 236–245. [Google Scholar] [CrossRef]
  11. Aguado, S.; Samper, D.; Santolaria, J.; Aguilar, J.J. Machine tool rotary axis compensation trough volumetric verification using laser tracker. Procedia Eng. 2013, 63, 582–590. [Google Scholar] [CrossRef]
  12. Acero, R.; Brau, A.; Santolaria, J.; Pueo, M. Verification of an articulated arm coordinate measuring machine using a laser tracker as reference equipment and an indexed metrology platform. Meas. J. Int. Meas. Confed. 2015, 69, 52–63. [Google Scholar] [CrossRef]
  13. Babu, M.; Franciosa, P.; Ceglarek, D. Adaptive Measurement and Modelling Methodology for In-Line 3D Surface Metrology Scanners. Procedia CIRP 2017, 60, 26–31. [Google Scholar] [CrossRef]
  14. Mahmoud, D.; Khalil, A.; Younes, M. A single scan longitudinal calibration technique for fringe projection profilometry. Optik 2018, 166, 270–277. [Google Scholar] [CrossRef]
  15. Cuesta, E.; Suarez-Mendez, J.M.; Martinez-Pellitero, S.; Barreiro, J.; Alvarez, B.J.; Zapico, P. Metrological evaluation of Structured Light 3D scanning system with an optical feature-based gauge. Procedia Manuf. 2017, 13, 526–533. [Google Scholar] [CrossRef]
  16. Vagovský, J.; Buranský, I.; Görög, A. Evaluation of measuring capability of the optical 3D scanner. Procedia Eng. 2015, 100, 1198–1206. [Google Scholar] [CrossRef]
  17. He, W.; Zhong, K.; Li, Z.; Meng, X.; Cheng, X.; Liu, X. Accurate calibration method for blade 3D shape metrology system integrated by fringe projection profilometry and conoscopic holography. Opt. Lasers Eng. 2018, 110, 253–261. [Google Scholar] [CrossRef]
  18. Garmendia, I.; Leunda, J.; Pujana, J.; Lamikiz, A. In-process height control during laser metal deposition based on structured light 3D scanning. Procedia CIRP 2018, 68, 375–380. [Google Scholar] [CrossRef]
  19. Li, F.; Stoddart, D.; Zwierzak, I. A Performance Test for a Fringe Projection Scanner in Various Ambient Light Conditions. Procedia CIRP 2017, 62, 400–404. [Google Scholar] [CrossRef]
  20. Genta, G.; Minetola, P.; Barbato, G. Calibration procedure for a laser triangulation scanner with uncertainty evaluation. Opt. Lasers Eng. 2016, 86, 11–19. [Google Scholar] [CrossRef]
  21. Isa, M.A.; Lazoglu, I. Design and analysis of a 3D laser scanner. Meas. J. Int. Meas. Confed. 2017, 111, 122–133. [Google Scholar] [CrossRef]
  22. Herráez, J.; Martínez, J.C.; Coll, E.; Martín, M.T.; Rodríguez, J. 3D modeling by means of videogrammetry and laser scanners for reverse engineering. Meas. J. Int. Meas. Confed. 2016, 87, 216–227. [Google Scholar] [CrossRef]
  23. Dosta, M.; Bröckel, U.; Gilson, L.; Kozhar, S.; Auernhammer, G.K.; Heinrich, S. Application of micro computed tomography for adjustment of model parameters for discrete element method. Chem. Eng. Res. Des. 2018, 135, 121–128. [Google Scholar] [CrossRef]
  24. Sarment, D. Cone Beam Computed Tomography. Dent. Clin. NA 2015, 62, 3–6. [Google Scholar] [CrossRef]
  25. Hermanek, P.; Carmignato, S. Porosity measurements by X-ray computed tomography: Accuracy evaluation using a calibrated object. Precis. Eng. 2017, 49, 377–387. [Google Scholar] [CrossRef]
  26. Peterson, E.B.; Klein, M.; Stewart, R.L. Whitepaper on Structure from Motion (SfM) Photogrammetry: Constructing Three Dimensional Models from Photography; Technical Report ST-2015-3835-1; U.S. Bureau of Reclamation: Washington, DC, USA, 2015.
  27. Toldo, R.; Gherardi, R.; Farenzena, M.; Fusiello, A. Hierarchical structure-and-motion recovery from uncalibrated images. Comput. Vis. Image Underst. 2015, 140, 127–143. [Google Scholar] [CrossRef] [Green Version]
  28. Brito, J.H. Autocalibration for Structure from Motion. Comput. Vis. Image Underst. 2017, 157, 240–254. [Google Scholar] [CrossRef]
  29. Westoby, M.J.; Brasington, J.; Glasser, N.F.; Hambrey, M.J.; Reynolds, J.M. “Structure-from-Motion” photogrammetry: A low-cost, effective tool for geoscience applications. Geomorphology 2012, 179, 300–314. [Google Scholar] [CrossRef] [Green Version]
  30. Fraser, C.; Jazayeri, I.; Cronk, S. A feature-based matching strategy for automated 3D model reconstruction in multi-image close-range photogrammetry. In Proceedings of the American Society for Photogrammetry and Remote Sensing Annual Conference 2010: Opportunities for Emerging Geospatial Technologies, San Diego, CA, USA, 26–30 April 2010; Volume 1, pp. 175–183. Available online: https://www.scopus.com/inward/record.uri?eid=2-s2.0-84868587896&partnerID=40&md5=3dc2d82cad7d24d8dce587cd23e2798f (accessed on 5 January 2018).
  31. Moons, T.; Vergauwen, M.; van Gool, L. 3D Reconstruction from Multiple Images. 2008. Available online: https://www.researchgate.net/publication/265190880_3D_reconstruction_from_multiple_images (accessed on 5 January 2018).
  32. Elnima, E.E. A solution for exterior and relative orientation in photogrammetry, a genetic evolution approach. J. King Saud Univ. Eng. Sci. 2015, 27, 108–113. [Google Scholar] [CrossRef]
  33. Jazayeri, I.; Fraser, C.; Cronk, S. Automated 3D object reconstruction via multi-image close-range photogrammetry. In International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, Proceedings of the ISPRS Commission V Mid-Term Symposium ‘Close Range Image Measurement Techniques’, Newcastle upon Tyne, UK, 21–24 June 2010; Volume XXXVIII, pp. 3–8. Available online: http://www.isprs.org/proceedings/XXXVIII/part5/papers/107.pdf (accessed on 5 January 2018).
  34. Zhu, Z.; Stamatopoulos, C.; Fraser, C.S. Accurate and occlusion-robust multi-view stereo. ISPRS J. Photogramm. Remote Sens. 2015, 109, 47–61. [Google Scholar] [CrossRef]
  35. Hernandez, M.; Hassner, T.; Choi, J.; Medioni, G. Accurate 3D face reconstruction via prior constrained structure from motion. Comput. Graph. 2017, 66, 14–22. [Google Scholar] [CrossRef]
  36. Giuseppe, G.; Clement, R. The Use of Self-Identifying Targeting for Feature Based Measurement. 2000. Available online: http://www.geodetic.com/f.ashx/papers/Release-The-Use-of-Self-identifying-Targeting-for-Feature-Based-Measurement.pdf (accessed on 22 January 2018).
  37. Stamatopoulos, C.; Fraser, C.S. Automated Target-Free Network Orienation and Camera Calibration. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. II 2014, 5, 339–346. [Google Scholar] [CrossRef]
  38. Remondino, F.; Spera, M.G.; Nocerino, E.; Menna, F.; Nex, F. State of the art in high density image matching. Photogramm. Rec. 2014, 29, 144–166. [Google Scholar] [CrossRef]
  39. Haala, N. The Landscape of Dense Image Matching Algorithms, Photogramm. Week 2013, 2013, 271–284. Available online: http://www.ifp.uni-stuttgart.de/publications/phowo13/index.en.html (accessed on 22 January 2018).
  40. Haala, N. Multiray Photogrammetry and Dense Image Matching, Photogramm. Week 2011, 11, 185–195. [Google Scholar] [CrossRef]
  41. Re, C.; Robson, S.; Roncella, R.; Hess, M. Metric Accuracy Evaluation of Dense Matching Algorithms in Archeological Applications. Geoinform. FCE CTU 2011, 6, 275–282. [Google Scholar] [CrossRef] [Green Version]
  42. Toschi, I.; Capra, A.; de Luca, L.; Beraldin, J.-A.; Cournoyer, L. On the evaluation of photogrammetric methods for dense 3D surface reconstruction in a metrological context. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. II 2014, 5, 371–378. [Google Scholar] [CrossRef]
  43. Shi, J.; Wang, X. A local feature with multiple line descriptors and its speeded-up matching algorithm. Comput. Vis. Image Underst. 2017, 162, 57–70. [Google Scholar] [CrossRef]
  44. Gesto-Diaz, M.; Tombari, F.; Gonzalez-Aguilera, D.; Lopez-Fernandez, L.; Rodriguez-Gonzalvez, P. Feature matching evaluation for multimodal correspondence. ISPRS J. Photogramm. Remote Sens. 2017, 129, 179–188. [Google Scholar] [CrossRef]
  45. Wenzel, K.; Rothermel, M.; Fritsch, D.; Haala, N. Image Acquisition and Model Selection for Multi-View Stereo. In Proceedings of the 3D-ARCH 2013–3D Virtual Reconstruction and Visualization of Complex Architectures, Trento, Italy, 25–26 February 2013. [Google Scholar]
  46. Robian, A.; Rani, M.R.M.; Ma’Arof, I. Manual and automatic technique needs in producing 3D modelling within PhotoModeler scanner: A preliminary study. In Proceedings of the 2016 7th IEEE Control and System Graduate Research Colloquium (ICSGRC), Shah Alam, Malaysia, 8 August 2016; pp. 181–186. [Google Scholar] [CrossRef]
  47. Jaud, M.; Passot, S.; le Bivic, R.; Delacourt, C.; Grandjean, P.; le Dantec, N. Assessing the accuracy of high resolution digital surface models computed by PhotoScan® and MicMac® in sub-optimal survey conditions. Remote Sens. 2016, 8, 465. [Google Scholar] [CrossRef]
  48. Jebur, A.; Abed, F.; Mohammed, M. Assessing the performance of commercial Agisoft PhotoScan software to deliver reliable data for accurate 3D modelling. In Proceedings of the 3rd International Conference on Buildings, Construction and Environmental Engineering, Sharm el-Sheikh, Egypt, 23–25 October 2017. [Google Scholar] [CrossRef]
  49. Remondino, F.; Nocerino, E.; Toschi, I.; Menna, F. A critical review of automated photogrammetric processing of large datasets. In Proceedings of the 26th International CIPA Symposium–Digital Workflows for Heritage Conservation, Ottawa, ON, Canada, 28 August–1 September 2017; pp. 591–599. [Google Scholar] [CrossRef]
  50. Fraser, C. Advances in Close-Range Photogrammetry; Photogrammetric Week: Belin/Offenbach, German, 2015; pp. 257–268. [Google Scholar]
  51. Remondino, F.; Spera, M.G.; Nocerino, E.; Menna, F.; Nex, F.; Gonizzi-Barsanti, S. Dense image matching: Comparisons and analyses. In Proceedings of the 2013 Digital Heritage International Congress (DigitalHeritage), Marseille, France, 28 October–1 November 2013; pp. 47–54. [Google Scholar] [CrossRef]
  52. Mousavi, V.; Khosravi, M.; Ahmadi, M.; Noori, N.; Haghshenas, S.; Hosseininaveh, A.; Varshosaz, M. The performance evaluation of multi-image 3D reconstruction software with different sensors. Meas. J. Int. Meas. Confed. 2018, 120, 1–10. [Google Scholar] [CrossRef]
  53. Lo, S.H. Parallel Delaunay triangulation in three dimensions. Comput. Methods Appl. Mech. Eng. 2012, 237–240, 88–106. [Google Scholar] [CrossRef]
  54. Lee, D.T.; Schachter, B.J. Two algorithms for constructing a Delaunay triangulation. Int. J. Comput. Inf. Sci. 1980, 9, 219–242. [Google Scholar] [CrossRef]
  55. Aichholzer, O.; Fabila-Monroy, R.; Hackl, T.; van Kreveld, M.; Pilz, A.; Ramos, P.; Vogtenhuber, B. Blocking Delaunay triangulations. Comput. Geom. Theory Appl. 2013, 46, 154–159. [Google Scholar] [CrossRef] [PubMed]
  56. Devillers, O.; Teillaud, M. Perturbations for Delaunay and weighted Delaunay 3D triangulations. Comput. Geom. Theory Appl. 2011, 44, 160–168. [Google Scholar] [CrossRef]
Figure 1. Evolution of measuring technology approaches for casting part dimensional verification: (a) data acquisition technologies; and (b) data processing schemes.
Figure 1. Evolution of measuring technology approaches for casting part dimensional verification: (a) data acquisition technologies; and (b) data processing schemes.
Sensors 18 03074 g001
Figure 2. Laser-tracker technology for machined part dimensional inspection.
Figure 2. Laser-tracker technology for machined part dimensional inspection.
Sensors 18 03074 g002
Figure 3. Scanning of casting part by means of structure light scanner and reference targets.
Figure 3. Scanning of casting part by means of structure light scanner and reference targets.
Sensors 18 03074 g003
Figure 4. Dense reconstruction workflow among multiple views: (a) photogrammetric step; and (b) dense matching approach.
Figure 4. Dense reconstruction workflow among multiple views: (a) photogrammetric step; and (b) dense matching approach.
Sensors 18 03074 g004
Figure 5. Dense matching process detailed workflow for shop-floor case study.
Figure 5. Dense matching process detailed workflow for shop-floor case study.
Sensors 18 03074 g005
Figure 6. Coded targets and calibrated scale bar located around the textured part: (a) lab testing; and (b) shop-floor testing.
Figure 6. Coded targets and calibrated scale bar located around the textured part: (a) lab testing; and (b) shop-floor testing.
Sensors 18 03074 g006
Figure 7. Camera shooting technique around the part.
Figure 7. Camera shooting technique around the part.
Sensors 18 03074 g007
Figure 8. Automated orientation in lab: (a) coded target image identification (in green); and (b) sparse triangulation of coded targets and camera network determination.
Figure 8. Automated orientation in lab: (a) coded target image identification (in green); and (b) sparse triangulation of coded targets and camera network determination.
Sensors 18 03074 g008
Figure 9. Targetless orientation in lab: (a) tie point image identification (in orange) and length bar definition (in red); and (b) sparse triangulation of feature points and camera network determination.
Figure 9. Targetless orientation in lab: (a) tie point image identification (in orange) and length bar definition (in red); and (b) sparse triangulation of feature points and camera network determination.
Sensors 18 03074 g009
Figure 10. Dense matching schema for shop-floor case study: (a) structure from motion; and (b) dense 3D colorized point cloud without filtering.
Figure 10. Dense matching schema for shop-floor case study: (a) structure from motion; and (b) dense 3D colorized point cloud without filtering.
Sensors 18 03074 g010
Figure 11. Measuring method for reference data determination in lab: (a) scene preparation with coded targets (ring and dot type) and scale bar; (b) photogrammetric camera (TRITOP©); and (c) structured light 3D scanner (ATOS IIe©).
Figure 11. Measuring method for reference data determination in lab: (a) scene preparation with coded targets (ring and dot type) and scale bar; (b) photogrammetric camera (TRITOP©); and (c) structured light 3D scanner (ATOS IIe©).
Sensors 18 03074 g011
Figure 12. 3D representation of measured part in lab environment with coded target orientation approach: (a) surface point comparison against reference mesh; and (b) dense point cloud with color and texture information (full resolution).
Figure 12. 3D representation of measured part in lab environment with coded target orientation approach: (a) surface point comparison against reference mesh; and (b) dense point cloud with color and texture information (full resolution).
Sensors 18 03074 g012
Figure 13. 3D representation of measured part in lab environment with targetless orientation approach: (a) surface point comparison against reference mesh with sparse point cloud; and (b) surface point comparison against reference mesh with dense point cloud (quarter resolution).
Figure 13. 3D representation of measured part in lab environment with targetless orientation approach: (a) surface point comparison against reference mesh with sparse point cloud; and (b) surface point comparison against reference mesh with dense point cloud (quarter resolution).
Sensors 18 03074 g013
Figure 14. 3D representation of measured part in shop-floor environment: (a) surface point comparison against reference mesh; and (b) dense point cloud with color and texture information (full resolution).
Figure 14. 3D representation of measured part in shop-floor environment: (a) surface point comparison against reference mesh; and (b) dense point cloud with color and texture information (full resolution).
Sensors 18 03074 g014
Figure 15. 3D representation of measured part in shop-floor environment with targetless orientation approach: (a) surface point comparison against reference mesh with sparse point cloud; and (b) surface point comparison against reference mesh with dense point cloud (quarter resolution).
Figure 15. 3D representation of measured part in shop-floor environment with targetless orientation approach: (a) surface point comparison against reference mesh with sparse point cloud; and (b) surface point comparison against reference mesh with dense point cloud (quarter resolution).
Sensors 18 03074 g015
Table 1. Summary and comparison of measuring techniques for casting part verification.
Table 1. Summary and comparison of measuring techniques for casting part verification.
Measuring Device (PCMM)Relative Accuracy (µm/m)Surface ComparisonAccessibilityScalabilityEnvironmental ConditionsEase of UseDegree of Automation
Laser-tracker (ASME B89.4.19-2006)±15 + 6 (MPE)LowLowLowAir flow stability requiredHighLow
3D scanner (VDI 2634 BLATT 3)±30HighLowLowLaser stabilityHighLow
Dense photogrammetry (Orientation with coded-targets)1/80,000 (RMS)HighHighHighLighting stability requiredLowHigh
Dense photogrammetry (Targetless orientation)1/27,000 (RMS)HighHighHighLighting stability requiredLowMedium to low
Table 2. Summary and comparison of dense matching techniques for lab and shop-floor environments.
Table 2. Summary and comparison of dense matching techniques for lab and shop-floor environments.
Test TypePoint Number (million)Relative Accuracy (mm/m)Measuring Time (min)Processing Time (min)Manufacturing Tolerance Grade 12 (mm)Deviations (2σ)
abSparse (mm)Dense (mm)
Lab test 110.51/80,00055120 to 180±10-±0.2
Lab test 20.21/27,00012±0.4±0.8
Shop-floor test 121/157,0001510>300±17-±3
Shop-floor test 10.51/38,50020±3±6
a: Orientation stage; b: Dense reconstruction.

Share and Cite

MDPI and ACS Style

Kortaberria, G.; Mutilba, U.; Gomez-Acedo, E.; Tellaeche, A.; Minguez, R. Accuracy Evaluation of Dense Matching Techniques for Casting Part Dimensional Verification. Sensors 2018, 18, 3074. https://doi.org/10.3390/s18093074

AMA Style

Kortaberria G, Mutilba U, Gomez-Acedo E, Tellaeche A, Minguez R. Accuracy Evaluation of Dense Matching Techniques for Casting Part Dimensional Verification. Sensors. 2018; 18(9):3074. https://doi.org/10.3390/s18093074

Chicago/Turabian Style

Kortaberria, Gorka, Unai Mutilba, Eneko Gomez-Acedo, Alberto Tellaeche, and Rikardo Minguez. 2018. "Accuracy Evaluation of Dense Matching Techniques for Casting Part Dimensional Verification" Sensors 18, no. 9: 3074. https://doi.org/10.3390/s18093074

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop