the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Nondestructive testing of arbitrarily shaped refractive objects with millimetrewave synthetic aperture radar imaging
Ingrid Ullmann
Julian Adametz
Daniel Oppelt
Andreas Benedikter
Martin Vossiek
Millimetrewave (mmW) imaging is an emerging technique for nondestructive testing. Since many polymers are transparent in this frequency range, mmW imaging is an attractive means in the testing of polymer devices, and images of relatively high resolution are possible. This contribution presents an algorithm for the precise imaging of arbitrarily shaped dielectric objects. The reconstruction algorithm is capable of automatically detecting the object's contour, followed by a materialsensitive reconstruction of the object's interior. As an example we examined a polyethylene device with simulated material defects, which could be depicted precisely.
 Article
(1014 KB)  Fulltext XML
 BibTeX
 EndNote
Throughout the whole process chain, quality management is a fundamental task in industrial production. The monitoring of devices and materials is a core issue in order to guarantee a consistent quality of the products. For the detection of material defects the interior of devices is of essential interest. When the device under test (DUT) is not transparent to the human eye, wavebased imaging, employing electromagnetic (EM) or acoustic waves, can be applied. There is a variety of wavebased techniques utilized for nondestructive testing (NDT). Among them are microwave and terahertz radar, ultrasound, Xray tomography and many more.
In this contribution we present an imaging system employing millimetre waves. Millimetre waves have become an emerging technique in recent years. Due to miniaturization advances in semiconductor technology, leading to considerable cost reduction, they have become attractive for a huge field of applications ranging from NDT to security screening and others (Ahmed et al., 2012; Agarwal et al., 2015). They offer a number of specific advantages compared to the other techniques named above: though they cannot provide the resolution of Xray, they have the advantage of a nonionizing radiation. Furthermore, mmW imaging is less costly than employing EM waves of higher frequencies (e.g. Xray or Terahertz). Sound or ultrasoundbased imaging on the other hand is a costefficient solution for many NDT applications. However, it usually requires the DUT to be immersed in water or another coupling medium – for aircoupled ultrasound typically is not able to properly penetrate into the inside of solids due to the very high difference in the acoustic impedances (Hillger et al., 2015).
From their frequency range and the corresponding wavelengths, millimetre waves offer a good compromise between penetration depth and resolution. The resolution, both laterally and in the range direction, lies in the range of a few millimetres, depending on system parameters like bandwidth or aperture size, but also on the DUT's material (Ahmed et al., 2012).
For the data acquisition a synthetic aperture radar (SAR) is utilized. The SAR technique originates from remote sensing – therefore SAR processing algorithms originally were based on the assumption of a freespace propagation of the electromagnetic wave. This assumption still holds when screening dielectric devices which exhibit a relative electrical permittivity equal (or very close) to one. However, when applying such algorithms to a scenario in which the wave propagates through a material with a refractive index significantly greater than one, the reconstruction is based on false assumptions and the reconstructed image will be of low quality or even faulty. The reasons for this are the change in phase velocity and the resulting refraction of the wave occurring at the material boundary in the case of nonnormal incidence.
The reconstruction algorithm presented in this paper takes into account the effects named above. It can therefore be applied not only to surroundings that exhibit a freespacelike behaviour, but also to subsurface imaging of refractive materials, which includes many polymers, too.
The article is outlined as follows: first a brief review of the theory of electromagnetic wave propagation in heterogeneous media will be given. Then, the image formation concept will be presented. Starting from the automated detection of the DUT's surface, two approaches for the handling of the material inhomogeneity are shown. The concept was evaluated by measurements – the results are presented in Sect. 4. Eventually a conclusion will sum up the main issues of the article.
The propagation of an electromagnetic wave can be described by the Helmholtz equation for the electric field strength in the time domain E(t):
and analogously for the magnetic field. Equation (1) is a homogeneous wave equation, with c denoting the wave's propagation velocity. One possible solution of the partial differential equation Eq. (1) is a plane wave. Assuming the wave to propagate in a direction x, then the plane wave equation reads as
Here, j is the imaginary unit, ω is the angular frequency and k is the wavenumber, related to the wavelength λ by
The phase velocity c is
depending on the material's parameters μ_{r} and ε_{r} (relative magnetic permeability and relative electric permittivity, respectively). The parameters μ_{0} and ε_{0} in Eq. (4) denote the magnetic permeability and electric permittivity of a vacuum. When considering only nonmagnetic materials as is done here, the phase velocity in a material is
where c_{0} is the speed of light in a vacuum and n is the refractive index of the material. When passing through a boundary between two materials of different refractive indices n_{1} and n_{2}, an EM wave will be refracted according to Snell's law,
where α denotes the angle of incidence and β the angle of refraction.
Snell's law can be derived from Fermat's principle (Hecht, 2002). Both Snell's law and Fermat's principle and their notations originate from optics, but can be applied to electromagnetic waves of other frequency ranges, too.
Fermat's principle states that a wave travelling from a point P_{1} to another point P_{2} will follow that optical path whose optical length is shorter than the optical length of any other path in a certain neighbourhood to it (Born and Wolf, 1980). The optical path length (opl) is the geometrical length s multiplied by the refractive index n(x, y, z) of the respective surrounding:
Consequently, the optical path taken between two points in space will minimize the functional Eq. (7). Since the optical length is
it follows that the time it takes the wave to traverse the way between P_{1} and P_{2} will be minimal – or more generally, stationary – also (Born and Wolf, 1980). Therefore, Fermat's principle is also known as the “principle of least time”.
3.1 General aspects on synthetic aperture imaging of heterogeneous media
Since synthetic aperture (SA) data are time and spacedependent, the data processing can operate in the time and space domain or in the temporal and spatial frequency domain, which is obtained from a multidimensional Fourier transform of the measured data. A straightforward way to compress the SA data is by employing matched filtering in the time–space domain (Cumming and Wong, 2005). Here, for each pixel, that is, each possible target ${\mathit{r}}_{\mathrm{T},i}=$ (x_{T,i}, y_{T,i}, z_{T,i})^{T}, a signal hypothesis is stated. This hypothesis equals the signal which the antenna at position ${\mathit{r}}_{\mathrm{A},j}=$ (x_{A,j}, y_{A,j}, z_{A,j})^{T} would receive if there was a point scatterer located in the point r_{T,i}. In a next step, the hypotheses ${s}_{\mathrm{R}x,\mathrm{hyp},ij}\left({\mathit{r}}_{\mathrm{T},i},{\mathit{r}}_{\mathrm{A},j},f,t\right)$ are correlated with the actually measured signal ${s}_{\mathrm{R}x,\mathrm{meas},j}\left({\mathit{r}}_{\mathrm{A},j},f,t\right)$ at the current antenna position.
Here, η denotes the shift of the correlation integral. Those points in which a scattering truly occurs will then exhibit a high value for the correlation. All other points will exhibit a low value. Finally, in each voxel the values of the correlation function ψ_{ij} for all N_{A} antenna positions r_{A,j}, jϵ[1,N_{A}] and, in the case of a multifrequency system, all N_{f} frequency steps f_{m}, mϵ[1,N_{f}], are summed up:
The resulting absolute value of the complex phasor ψ_{sum,i} is a measure of the probability that a point scatterer will be located in the respective point r_{T,i}. Converting the resulting absolute values of all N_{T} voxels r_{T,i}, iϵ[1,N_{T}] to brightness values returns the reconstructed image. That way, due to the pixelwise filtering a precise compression of the targets can be obtained. However, it is obvious that this approach performs poorly in terms of computing time. This is especially true in the field of subsurface imaging: for the matched filtering, the path traversed by the wave needs to be known. In a heterogeneous surrounding, the wave will be refracted according to Eq. (6). However, the two angles α and β are not known a priori due to the unfocused antennas used for SAR. Therefore, the algorithm first has to determine the wave's path and then can proceed to the correlation procedure described above.
There is a number of reconstruction algorithms employed for SAR processing, like rangeDoppler, chirp scaling or ωk, which operate partly or entirely in the frequency domain (Cumming and Wong, 2005). In their original form they were based on the assumption of freespace propagation of the waves, but they have been adapted to inhomogeneous surroundings in past years (Albakhali and Moghaddam, 2009; Skjelvareid et al., 2011). However, in most cases the adapted algorithms can only handle planar objects. Some methods have been proposed which are also suitable for nonplanar objects (Qin et al., 2014), but they require iterative processing steps, which considerably reduces their efficiency.
Therefore, matched filtering, which can be applied to any geometry, is still the most commonly used technique for subsurface imaging of irregularly shaped objects and will be used in this contribution, too.
3.2 Contour detection
If the contour of the object, which is the material boundary between the two media, is not known – or if it is known but its orientation towards the aperture plane is not – then the material boundary must be determined prior to the actual reconstruction. In the following, a way to extract the contour directly from the measurement data is shown.
Therefore, we first reconstruct the space between the aperture and the boundary. Here we can assume a freespace propagation of the EM wave, which means that no refraction has to be taken into account. The resulting image will be defocused, but the boundary will be reconstructed at its true position. Since the boundary will be the strongest reflection in most cases, one way to extract it directly is to search for the brightest pixel in each column. This approach is often used in groundpenetrating radar (Walker and Bell, 2001; Feng et al., 2010). It is convenient because it does not require additional measurements and there are effective algorithms existing for a freespace reconstruction.
Evidently however the error made in estimating the boundary will affect the quality (i.e. the signaltonoise ratio, SNR) of the further reconstruction: due to the imaging principle of interfering single measurements it is essential that the resulting phase error caused by the estimation uncertainty is less than ±90^{∘} in order to avoid destructive interference.
3.3 Refraction ray tracing
This section addresses the matched filtering for heterogeneous surroundings. A model of the scenario with the applied nomenclature is shown in Fig. 1. Here, the indices A, T, and B denote the coordinates of the respective antenna, boundary point and target. For clarity the indices i, j and m utilized in Sect. 3.1 are left out. The lateral (azimuth) direction is x; the range direction is z. For the sake of simplicity, all considerations are derived for a 2D setup.
We use a monostatic synthetic aperture radar, transmitting a signal s_{Tx}(tf). The receiver signal s_{Rx}(tf) is the wave backscattered from the DUT. The antenna positions spanning the synthetic aperture are equidistantly spaced and they are located at z=0. The antenna itself is modelled as an isotropic radiator. It is presumed that the volume to be reconstructed is located in the antenna's farfield. Then, the spherical wave radiated from the antenna will intersect with the object only at a small section of the sphere. This small section can be approximated as a quasiplane phase front. Therefore, we model the incident wave to be a plane wave.
According to the antenna's isotropic directivity pattern, some part of the radiated electromagnetic field will be radiated in such a way that after traversing medium one and being refracted at the boundary it will actually meet the target (r_{1} and r_{2} in Fig. 1). Assuming isotropic scattering, some part of the reflected field will traverse the same way back to the antenna. Then, the received signal's phase will be
which follows from Eqs. (2) to (5).
In Eq. (11), any phase offset due to the scatterer's reflection properties will be neglected, for it is a constant and thus will not influence the reconstruction process. Furthermore, since the reconstruction method will only evaluate the phase, the signal's amplitude is not considered further. Accordingly, the signal hypothesis becomes
when setting the amplitude to a virtual value of one and inserting Eq. (11) for the phase.
Throughout this article we assume the object under test to consist of one nonmagnetic, dielectric, frequencyindependent and lossless material whose electric permittivity is known. For many NDT applications these are valid assumptions.
In the following, two means of finding the optical path in a twomedia system are described. One is based on Snell's law, and the other one is based on Fermat's principle.
3.3.1 Ray tracing based on Fermat's principle
In order to determine the optical path, we search for that point within the boundary which is the true point of transit between the two media. Therefore, we discretize the boundary into a distribution of points. Then, the resulting optical path lengths from the antenna to the assumed target are a function of the boundary distribution z_{B}(x_{B}):
and
see Fig. 1. The overall time it takes the wave to travel from the antenna to the target and back then follows as
In order to determine the true path taken by the wave, the minimum of Eq. (15) has to be found. That boundary point (x_{B}, z_{B}) which minimizes Eq. (15) is the true point of transit between the two materials. The received phase then can be written as
and the signal hypothesis is obtained from Eq. (12).
Note that in order to improve the efficiency of this ray tracing, it is sufficient to consider only those boundary points which lie between the respective antenna and target positions. It is obvious that boundary points lying beyond cannot minimize the optical path.
3.3.2 Ray tracing based on Snell's law
Like before, we discretize the boundary into a distribution of points. From the geometry in Fig. 1 it can be seen that the incident angle is
Herein, δ is the difference angle between the vertical and the normal in the respective point (x_{B}z_{B}).
Likewise, for β we can write
Additionally, α and β also have to fulfill the law of refraction, Eq. (6). That point of the boundary (x_{B}z_{B}) which fits Eqs. (6), (17) and (18) is the true point of refraction. From it the lengths r_{1} and r_{2} are easily found by the Pythagorean theorem:
and
Having determined r_{1} and r_{2}, the signal hypothesis follows from Eqs. (11) and (12).
Note that, of course, this approach is also applicable to a scenario with a planar material boundary. Then, in Eqs. (17) and (18) the difference angle δ can be set to zero.
Again, it is sufficient to consider only those boundary points between the current antenna position and target.
To sum up the concept, a flowchart of the complete image reconstruction procedure is shown in Fig. 2.
4.1 Measurement setup
Measurements were conducted to demonstrate the algorithm's feasibility. Here, we examined a polyethylene (PE) object into which two holes were drilled. Polyethylene displays a relative permittivity of ε_{r} = 2.3 in the relevant frequency range (Elvers, 2016). The drill holes represent air inclusions within the material. Such imperfections can be caused in the production process of the polymer or in the operation of a component.
The object under test is depicted in Fig. 3. It was constructed to be invariant along the vertical direction, thus allowing for a 2D reconstruction. Its surface was chosen to be nonplanar and nonsymmetric in order to demonstrate the algorithm's capacity to reconstruct rather complex objects.
The employed measurement setup is shown in Fig. 4. It consists of one horizontal and two vertical traversing units, on which a pair of antennas is mounted. The traversing units allow for a movement along a distance of approximately 1.1 m in the vertical direction and 0.65 m in the horizontal direction. It is therefore possible to span a synthetic aperture of those dimensions. The two vertical units can be moved separately, which also makes multistatic measurements possible. For the measurements presented in this paper we employed a quasimonostatic setup, i.e. transmitter and receiver antenna were in close proximity; they were mounted with a spacing of 2 mm between them. The antennas were two Hpolarized horn antennas with a physical aperture of 2.45 mm × 4 mm (see Fig. 4). The transmitted signal was generated by a vector network analyser (Agilent PNA E8363B) with frequency extenders (Oleson Microwave Labs V10VNA2T/R). A 201point SFCW signal covering the complete Wband (frequency range: ${f}_{min}=\mathrm{75}$ GHz to ${f}_{max}=\mathrm{110}$ GHz) was employed. Accordingly, the corresponding freespace wavelengths cover a range from λ_{min} = 2.7 mm to λ_{max} = 4 mm and a bandwidth of B = 35 GHz can be used. This bandwidth and wavelengths allow for a resolution in the range of a few millimetres, as will be discussed in Sect. 4.2. The radiated power was P_{Tx} = −17 dBm (20 µW).
In order to generate a 2D image, a line aperture is sufficient. From the spatial sampling theorem the spacing between the antenna array positions must not exceed
(Benetsy et al., 2008), which is 1.35 mm for the Wband. Here, a spacing of 1 mm was chosen. The total length of the line aperture was L_{Ap} = 400 mm. The PE test object was positioned in such a way that its front plateau (see Fig. 3) was orientated parallel to the line aperture at a distance of ${z}_{min}=\mathrm{25.8}$ cm. Thus, the field of view of the reconstructed image needs to reach up to ${z}_{max}=\mathrm{31.8}$ cm in the range direction (see Fig. 3). Table 1 sums up the parameters of the measurements.
The system is calibrated by two calibration measurements (Gumbmann, 2011):

Load standard: an empty space measurement with no reflecting object to eliminate crosstalk between the antennas;

short standard: a measurement with a metal plate placed in front of the aperture at a defined reference distance d_{ref} to eliminate the frequency response of the components (network analyser, extenders, antennas and cables).
The raw data s_{Rx,raw} were then calibrated by
Here, s_{Rx,load} and s_{Rx,short} are the data obtained from the two calibration measurements named above. The notation 〈⋅〉 symbolizes the mean value over a number of measurements (10 in the presented experiments). The factor −1 is the reflection coefficient of the short. The reference phase φ_{ref} is the phase calculated from the short measurement. It is
4.2 Results
As a first step, a freespace reconstruction of the calibrated data was performed. Figure 5 shows the reconstruction of the object assuming freespace propagation throughout the whole domain. As a reference for the reconstructed image, a sketch of the DUT's geometry is depicted also. For the sake of simplicity, the object is depicted in a local coordinate system, starting from z=0 at the DUT's front boundary. In the global coordinate system, the object was located at a range of 25.8 cm from the aperture line, as mentioned above. It can be seen that the target at z ≈ 4.5 cm is not depicted within the object. The lower boundary is not visible either. The image is rather defocused, too.
The reconstruction image is normalized to its maximum intensity value. For clutter reduction, all pixels exhibiting a value below 10 % of the maximum value are set to zero. The object's contour is illustrated by dashed lines.
The boundary contour was estimated by a columnwise maximum search as described in Sect. 3.2. The respective points are depicted in white in Fig. 5. For a comparison the estimated boundary is shown in Fig. 6 together with the true contour. It can be seen that the estimation is a good approximation to the real contour: on the upper plateau the estimated values and the analytical ones match perfectly. The error made by the estimation is depicted in Fig. 7. Its mean value is 0.655 mm; the maximum error is 2.14 mm. The estimation displays an uncertainty that is determined by the system's range resolution. For the Wband, whose bandwidth B is 35 GHz, it is
in free space.
Figure 8 shows the image obtained with the described materialsensitive reconstruction algorithm based on the estimated material boundary. Here, Fermat's principle was used for the ray tracing. The concept based on Snell's law was evaluated in the conference paper (Ullmann et al., 2017). Again, the image is normalized to the maximum intensity and all values below 0.1 are neglected. As before, the reference sketch can be seen on the right side of the figure.
Note that in the reconstructed images only the targets' upper and lower boundaries are visible. They correspond to the material discontinuities at which reflections occur. From Fig. 8 it can be seen that in contrast to the freespace reconstruction, both targets are reconstructed at their true positions. The improvement in the localization is because in the freespace case the propagation velocity is assumed too high. Consequently, since the velocity is proportional to the traversed way, an overly long distance along the range direction is reconstructed. With the adapted algorithm this error is not made. Furthermore, with the developed method the targets and the material boundary are focused more precisely. Since the lateral resolution depends on the wavelength λ, the aperture length L_{Ap} and the distance z between the aperture plane and the depicted position
and the range resolution depends on bandwidth and phase velocity (see Eq. 24), a better resolution can be obtained when taking the material characteristics, which affect the phase velocity and the wavelength, into account. From Eqs. (24) and (25) and Table 1 it can be seen that a resolution with theoretical limits of δ_{r}=2.84 mm in the range direction and δ_{lat}= 1.04 mm in the lateral direction can be obtained with the described setup.
The actual resolution in the reconstructed images can be estimated from Fig. 9. Here, the point that displayed the maximum intensity at the upper left air inclusion (x ≈ 2.5) was picked out from the datasets depicted in Figs. 5 and 8. The range and lateral resolution can then be estimated by finding the distance Δz or Δx, respectively, between the two points of half the maximum intensity. They are depicted by dashed lines in Fig. 9. For the sake of simplicity and because their zvalues do not match, as described previously, the curves are normalized to a dimensionless maximum intensity of 1 and the x/zcoordinate of the maximum point was set to 0. It can be seen from Fig. 9 that the materialsensitive reconstruction (blue curve) displays a smaller (i.e. higher) resolution in both the range and lateral directions compared to the freespace reconstruction (red curve). Note however that in the materialsensitive reconstruction the upper target is reconstructed by two rather than one point of high intensity (see Fig. 8). This is because the hole with a diameter of 4 mm does not represent a point target for the employed system. Consequently, in Fig. 9 there are two peaks rather than one, and they are overlapping, so that a proper determination of the resolution is difficult.
From Fig. 9 the respective resolutions can be estimated to

δ_{r}= 5.99 mm in the freespace reconstruction versus δ_{r}= 3.08 mm in the materialsensitive reconstruction;

δ_{lat}= 6.13 mm in the freespace reconstruction versus δ_{lat}= 4.26 mm in the materialsensitive reconstruction.
Note that, when instead examining the lower right target (x ≈ 7.5), a lateral resolution of approximately 2.2 mm can be estimated for the materialsensitive reconstruction in the same way as before. Noise as well as the error caused by the boundary estimation are possible reasons for the deviation between the theoretically achievable and practically achieved resolutions.
This article presents methods for the detection of subsurface material defects by means of millimetrewave synthetic aperture radar imaging. The proposed reconstruction algorithm first detects the shape of the object's surface automatically. In the actual reconstruction it takes into account the effects occurring at the surface material discontinuity, namely refraction and the change in phase velocity, thus allowing for a precise reconstruction of the object under test. The required ray tracing through the heterogeneous surrounding can be accomplished using different approaches. Here we presented a way based on Snell's law and one based on Fermat's principle. Both methods are feasible (see Fig. 8 in this paper and Fig. 5 in Ullmann et al., 2017); however, compared to the ray tracing based on Fermat's principle, the method based on Snell's law requires the extra step of determining the difference angle δ between the vertical and the normal in the respective boundary point.
An experimental verification was conducted by reconstructing a polyethylene object with air inclusions. Here, the proposed algorithm displayed a higher resolution compared to a conventional freespace reconstruction since the lateral resolution depends on the wavelength and the range resolution depends on the phase velocity.
The underlying measurement data are not publicly available but can be requested from the authors if required.
The authors declare that they have no conflict of interest.
This article is part of the special issue “Sensor/IRS2 2017”. It is a result of the AMA Conferences, Nuremberg, Germany, 30 May–1 June 2017.
The authors would like to thank the European Regional Development Fund (ERDF)
for funding parts of the research activities. This work is part of the
project “Advanced Analytics for Production Optimization” (EASYOpt)
funded by the Bavarian program “Investment for growth and jobs”, 2014–2020
(Bavarian Ministry of Economic Affairs and Media, Energy and
Technology).
Edited by: Bernd
Henning
Reviewed by: two anonymous referees
Agarwal, S., Kumar, B., and Singh, D.: Noninvasive concealed weapon detection and identification using V band millimeter wave imaging radar system, National Conference on Recent Advances in Electronics & Computer Engineering (RAECE), 13–15 February 2015, Roorkee, India, 258–262, https://doi.org/10.1109/RAECE.2015.7510202, 2015.
Ahmed, S. S., Schiessl, A., Gumbmann, F., Tiebout, M., Methfesssel, S., and Schmidt, L.P.: Advanced microwave imaging, IEEE Microw. Mag., 13, 26–43, https://doi.org/10.1109/MMM.2012.2205772, 2012.
Albakhali, M. and Moghaddam, M.: 3D SAR focusing for subsurface point targets, Proc. IGARSS 2009, I60–I63, https://doi.org/10.1109/IGARSS.2009.5416944, 2009.
Benetsy, J., Chen, J., and Huang, Y.: Microphone array signal processing, Springer, Berlin, Germany, https://doi.org/10.1007/9783540786122, 2008.
Born, M. and Wolf, E.: Principles of optics, 6th edn., Pergamon Press, Oxford, UK, 1980.
Cumming, I. G. and Wong, F. H.: Digital processing of synthetic aperture data, Artech House, Norwood, USA, 2005.
Elvers, B.: Ullmann's polymers and Plastics, WileyVCH, Weinheim, Germany, 2016.
Feng, X., Liang, W., Lu, Q., Liu, C., Li, L., Zou, L., and Sato, M.: 3D velocity model and ray tracing of antenna array GPR, Proc. IGARSS 2010, 4204–4207, https://doi.org/10.1109/IGARSS.2010.5651564, 2010.
Gumbmann, F.: Aktive multistatische Nahbereichsabbildung mit ausgedünnten Gruppenantennen, PhD Thesis, FriedrichAlexander University, Erlangen, Germany, 2011.
Hecht, E.: Optics, 5th edn., Pearson, London, UK, 2002.
Hillger, W., Ilse, D., and Bühling, L.: Industrial applications of aircoupled ultrasonic technique, 7th International Symposium on NDT in aerospace, 16–18 November 2015 Bremen, Germany, 1–7, We.2.A.3, 2015.
Qin, K., Yang, C., and Sun, F.: Generalized Frequency Domain Synthetic Aperture Focusing Technique for Ultrsonic Imaging of Irregularly Layered Objects, IEEE T. Ultrason. Ferr., 61, 133–146, https://doi.org/10.1109/TUFFC.2014.6689781, 2014.
Skjelvareid, M. H., Olofson, T., Bikelund, Y., and Larsen, Y.: Synthetic aperture focusing of ultrasonic data from multlayered media using an omegak algorithm, IEEE T. Ultrason. Ferr., 58, 1037–1048, https://doi.org/10.1109/TUFFC.2011.1904, 2011.
Ullmann, I., Adametz, J., Oppelt, D., and Vossiek, M.: Millimeter Wave Radar Imaging for NodDestructive Detection of Material Defects, Proc. SENSOR 2017, 467–472, https://doi.org/10.5162/sensor2017/D3.3, 2017.
Walker, P. D. and Bell, M. R.: Noniterative GPR imaging through a nonplanar airground interface, Proc. IGARSS 2001, 1527–1529, https://doi.org/10.1109/IGARSS.2001.976900, 2001.