Biometric person identification systems identify individuals using personal characteristics such as fingerprints, eyes or facial recognition. However, in some critical situations, such as fires, serious traffic accidents, earthquakes or serious injuries, these features can become ineffective. In certain situations, dental characteristics may become the only valid biometric feature for identification. In these cases, forensic dentists work by examining dental structures to establish a person's identity. Currently, studies are being carried out to develop an automated recognition system based on computer vision to assist forensic dentists. However, due to the difficulties in processing panoramic X-ray images and challenges in accessing the data, person matching studies with these images are limited. This paper presents a novel method for matching people based on panoramic X-ray images. Dental person recognition studies can proceed either by investigating the similarity of teeth or by examining the similarity of jaws. In this work, a new approach that uses keypoint descriptors to perform tooth-jaw matching is proposed. This approach offers a high match rate by allowing to search for dental features on a jaw-by-jaw basis and requires less computational complexity than tooth-to-tooth matching. Unlike jaw-to-jaw approaches, it is possible to match individual teeth. The method presented in this study provides a novel approach with significant matching accuracy and efficiency. By evaluating the effectiveness of these methods on panoramic images, the study contributes to forensic dental identification methods in scenarios where traditional biometric features may fall short.
Notes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
1 Introduction
Biometric systems are designed to identify individuals based on unique personal characteristics, such as fingerprints, face, voice, eyes, and text. These features are highly successful for person identification. However, in cases of severe damage to the body, such as in fires or earthquakes, these biometric features may not be usable for person identification after death. Teeth are considered to be more resilient than other organs used for biometric identification, which makes them a more reliable feature in cases of severe accidents or injuries. Thus, dental features are preferred when other biometric features are not usable. While DNA detection technology is also a method of identification, it can be time-consuming and costly [1]. When other biometric features are not available, forensic dental scientists examine the oral and dental features of the individual to make identifications [2]. Forensic dentists search a specific database that includes information about the person's teeth and jaws. The use of computers in these procedures can aid forensic dentists and accelerate the identification process. It can also allow for searches of larger databases [3]. Panoramic radiography images are X-rays in which the entire mouth is viewed panoramically. These images are used to observe various diseases such as dental and bone anomalies, cysts, tumors and infections [4‐7]. The main stages of person recognition, after preprocessing and segmentation, are feature extraction and matching. In dental image recognition, distinctive differences are identified from the segmented images to facilitate the identification of an individual. These differences are then extracted as features, and the person's information that best matches these features is obtained through matching techniques. These stages are crucial for accurately identifying an individual based on their dental features. During the segmentation, feature extraction, and matching stages of dental image recognition, there are several challenges that need to be addressed. These include variations in dental radiography images taken at different times, inadequate segmentation success, use of images from different types of machines, dental procedures performed on the teeth and jaws of individuals, and images with high noise. These challenges can lead to inaccuracies in identifying an individual based on their dental features, and it is crucial to account for them in the recognition process. In Fig. 1, it can be seen that the noise causes the brightness of the teeth to change. Figure 2 illustrates the changes in a person's teeth after getting braces, and Fig. 3 shows the variations in X-ray images obtained with different machines.
Fig. 1
The effect of noise on images seems to interfere with information retrieval, especially from the central incisor
Fig. 2
Dental procedurals cover dental information because they are shiny and on top, causing the shape information of teeth to change
Fig. 3
The clarity and resolution of images from different devices may be different. This may lead to a poor capture of details
×
×
×
Segmentation is a critical step in the dental X-ray analysis process, as it allows for the extraction of features and information from dental images [8‐10]. Segmentation is usually the first part of the dental human identification process [11‐13]. Despite the fact that dental human identification studies are limited, a number of methods for dental X-ray segmentation have been proposed. The segmentation method to be used can be chosen according to the type of X-ray used, the purpose of the application, and the amount of data available. Different methods established for panoramic X-rays [8, 14‐17] periapical X-rays [18, 19] and bitewing X-rays [20, 21], which are common among dental imaging types. On the other hand, segmentation can be applied not only for dental human identification [17, 22, 33, 34] but also for dental work identification [23, 24], tooth extraction or classification [18, 23] or dental disease detection [25, 26]. Depending on the approach of the dental person recognition process, different methods, such as dental work or tooth segmentation can be applied. Deterministic methods [27, 28] can be used for segmentation, as well as training-based methods [26, 29] such as neural networks. Similarly, the human-matching approach to be applied is important for the choosing the type of segmentation approach to be used. However, another important point here is the amount of available data. Deep learning-based segmentation approaches, such as CNN or Mask R-CNN, require a substantial amount of data, high computing capability, and training time [29, 30].
Advertisement
In the literature, various methods are available for extracting features from dental radiographs to recognize individuals. Studies have utilized techniques such as keypoint-based methods [5, 6], PHOG [33], Hu moments [34], Zernike moments [35, 36], GLCM [37, 38], and Fourier descriptors [13, 37]. When matching source images with a database, distance measures are commonly used. Popular methods include sum of difference squares (SSD) [33], Euclidean distance [34, 39‐41] Hausdorff distance [13, 22, 42], and Mahalanobis distance [38]. Although the proposed algorithms have demonstrated success, their performance varies across different databases due to the lack of a standardized database. Moreover, the desired level of success has not been consistently achieved, and most studies have not been tailored specifically to panoramic X-rays.
This study contributes to the literature by presenting a new method for the identification of individuals with panoramic dental X-ray images. In the literature, the identification of individuals using dental radiography images can be made through two approaches: i) teeth-based [3, 33] and ii) jaw-based [32, 35]. Teeth-based approaches involve segmenting the boundaries of each tooth and extracting features from each tooth, which are then compared to the features of all teeth in the database to identify a match. Meanwhile, jaw-based approaches extract features from the entire jaw in both the target and source X-ray images, which are then compared to the jaw features in the database to identify a match.
The most significant methods developed for dental human identification with panoramic dental X-ray images in recent years are [12, 24, 29, 31, 33, 50]. In [31], the authors used curvature scale space (CSS) transform as the feature extraction method and used Euclidean distance to investigate similarities. In [24], instead of using teeth or jaws as the source of features, they used dental works on the jaw or teeth. In a more recent study, [33] proposed a teeth-based method, using pyramid histogram of oriented gradients (PHOG) on teeth. They used support vector machine (SVM) and sum of squared differences to calculate similarities. In [12], Gurses and Oktay proposed using a deep learning-based segmentation method and Speeded Up Robust Features (SURF) for the features. In both [29] and [50], the authors proposed a deep learning architecture for dental human identification.
When the literature is examined, it is understood that dental person identification studies are still limited today. Especially, it can be said that person recognition systems with panoramic dental X-rays are much more limited [29]. The proposed work contributes to the literature by presenting a novel approach that can be used for person identification with panoramic X-ray images. The highlights of the study are as follows: it can be used even with small datasets without the need for an additional training process, it offers a high matching performance comparable to neural network-based studies, it can be used with different segmentation methods and is less affected by segmentation errors, and it is less impacted by the missing tooth problem due to teeth-jaw matching.
Advertisement
In the article, the X-ray image of the unknown person is referred to as the target image, and the X-ray image database and images of potential matching persons are referred to as the source database and source image, respectively.
Teeth-based approaches have the advantage of being able to examine teeth features independently, which means that even if many of the teeth in the jaw are missing, the person can still be identified in the target database. However, extracting clear tooth boundary information from dental X-ray images can be challenging due to overlapping teeth, machine noise, and dental work-related issues. Additionally, if there are no teeth in the jaw or all existing teeth are prosthetic, these methods may not be effective.
The teeth matching approach is based on measuring the sequential similarity of teeth or tooth regions extracted from two X-ray images. In this perspective:
i.
Teeth are extracted from the X-ray image of the person being searched (target image), and the features of each tooth are extracted.
ii.
X-ray images (source images) are taken from the panoramic X-ray database (source database), respectively.
iii.
Teeth are extracted from each next source image, and the features of each tooth are extracted.
iv.
The similarities of the extracted tooth features are analyzed, and the most similar teeth and X-rays are identified.
An example of such a tooth matching approach is given in [5]. An illustration flow of a teeth-based approach can be seen in Fig. 4.
Fig. 4
Example structure of the teeth-based approach. The teeth in the target jaw are matched one by one with the teeth in the source jaws in the database
×
The jaw-based approach uses information from the entire jaw. Since each tooth is not analyzed separately, the processing time is shorter compared to the tooth matching method. In this approach, the jaw may be identified even if no teeth are present on the jaw. It is also possible to identify a person in images consisting of dental repairs and prosthetic teeth.
The disadvantage of the jaw-based approach is that individual teeth cannot be analyzed separately, which means that if some of the teeth are damaged, the chances of identifying the person in the database would decrease. For instance, in cases of accidental death, if some of the teeth are lost or damaged, it may be more difficult to identify the person. An example structure for jaw-based matching can be seen in Fig. 5.
Fig. 5
Example structure of the jaw-based approach. The target jaw and the source jaws in the database are matched one by one. Individual tooth characteristics are not taken into account
×
In this paper, a novel matching approach for person identification in panoramic X-ray images is proposed. To the best of the author's knowledge, this is the first time that this approach has been presented for person identification in panoramic radiography images. In the proposed approach, the target image is segmented into regions of teeth, and then keypoint features of these teeth are extracted. As the segmentation method, we used the meta-heuristic supported semi-automatic segmentation method [16] and the other segmentation method we proposed, the fully automatic segmentation method [17]. Segmentation was performed using both methods, and the results were analyzed and compared. In the proposed approach, keypoint features are extracted from the segmented regions of teeth in the target image, and similarly, keypoints are extracted from the jaw images in the database. It should be noted that the database images are not divided into tooth regions. The method then matches the tooth features of the target image with the jaw features of the database. An example structure of the teeth-jaw approach is shown in Fig. 6.
Fig. 6
Example structure of teeth-jaw approach. The match between the teeth in the target jaw and the source jaws in the database are examined
×
The advantage of this approach is that it allows searching with a limited number of teeth and reduces the number of searches to be performed. In this way, the proposed method is a solution to the problems of excessive number of searches and insufficient utilization of non-dental regions in the teeth-based approach. A balanced solution is presented in terms of speed and matching performance.
In this study, local keypoint-based algorithms are used for feature extraction. A comparative analysis of local keypoint-based feature extraction methods is presented. The algorithms were evaluated in terms of matching success and execution time. The matching performances of each jaw are considered separately. In this way, the effect of jaw selection on performance is observed. Two different segmentation methods are used, and the effect of segmentation on performance is analyzed. The experimental results show that the proposed approach can be useful in person identification. It is observed that the performance rate of matching the target image in the first 10 images is quite high. The experimental results are promising in this respect.
Section 2 of the paper describes the materials and methods used in the proposed work. The dataset used in the proposed work, i.e., X-ray images, is summarized and illustrated in Sect. 2.1. Section 2.2. describes the method of the proposed work, and the steps of the proposed approach are given graphically. In Sect. 2.2., first, the proposed teeth-jaw-based matching approach (TJMA) is explained. Then, the process of the keypoint-based methods utilized in TJMA is described. In the Method section, the Detection of matching jaw subsection describes the feature matching approach of the proposed method in detail. In addition, a flow diagram summarizes the operation of the system. Section 3 describes the experiments and results of the proposed work in different perspectives. The experimental results and comparisons with similar works in the literature are discussed in Sect. 4. Future directions are explained in Sect. 5. Finally, the conclusion of the study is presented in Sect. 6.
2 Material and method
2.1 Material
For the experimental studies, 250 images were obtained from the Ordu University Department of Dentistry with permission from the Ordu University Clinical Research Ethics Committee. All images were cropped to remove parts of the background other than the jaws. The images featured jaws with varying numbers of teeth and different sizes, but their aspect ratios were normalized as part of the preprocessing stage. Images of jaws without teeth were excluded from the tooth matching experiments, as were mandibular or maxillary jaws without teeth after jaw separation. Sample of X-ray images from the search database (source) is shown in Fig. 7.
Fig. 7
X-ray image samples from the search database
×
2.2 Method
This study aims to identify individuals by analyzing dental radiography images. In the literature, it has been suggested that dental X-ray matching for human identification can mainly be achieved through the following methods.
Comparison of jaw features with jaw features in the database
The advantage of teeth-based studies is that tooth characteristics can be examined independently, allowing for identification even if many of the teeth on the jaw are missing. However, tooth boundary information can be challenging to extract clearly from dental X-ray images due to overlapping teeth, machine noise, and extraction-related issues. Another disadvantage of this approach is that teeth may be absent from the jaw, or all existing teeth may be dentures. In such cases, the distinctive features of this method may not be available for identification purposes.
The jaw-based approach utilizes information from the entire jaw, which reduces the need for analyzing each tooth separately and results in a shorter processing time compared to the tooth matching method. Additionally, this approach allows for identification even if no teeth are present on the jaw. One disadvantage of the jaw-based approach is that individual teeth cannot be examined. Therefore, in cases such as accidental death, damage to some teeth may reduce the likelihood of identifying the person in the database.
In this study, a third approach is proposed, which differs from the two approaches mentioned above. This new approach is based on teeth-jaw matching, and it involves using SIFT, SURF, KAZE, and MSER algorithms for feature extraction. A comparison of these algorithms is presented in terms of matching and speed to determine the most effective approach for this type of matching.
2.2.1 Teeth-jaw-based matching approach
In the teeth-jaw matching approach, the teeth in the target jaw is searched on the jaws in the database, but there is no need to extract tooth boundaries. Instead, the identification of tooth regions is sufficient for matching purposes.
The teeth-jaw-based approach is examined in two main categories based on the segmentation method: matching using teeth obtained by semi-automatic segmentation and matching on teeth obtained by fully automatic segmentation.
In the semi-automated method, jaw separation is performed by examining the jaw in five different regions and identifying the points that best separate the jaw in these regions. Tooth separation is based on the user's selection of tooth midpoints with the help of provided guidelines, and separation lines are automatically generated according to the selected points [16]. In this study, the separation points of the teeth were selected along the dental pulp. From the selected point, lines were created both on the upper separation curve and on the subdental separation curve because the orientation of the crown of some teeth is different from the orientation of the root.
In the fully automatic method, tooth separation points are detected automatically in addition to the semi-automatic method. Tooth separation points are detected by examining the total pixel values of the lines examined at a certain length perpendicular to the jaw separation curve[17]. The general workflow of the approach is given in Fig. 8. As shown in Fig. 8, first, the image to be searched is taken, and preprocessing and segmentation (jaw and teeth separation using [16] or [17]) operations are performed, respectively. In this way, the tooth regions in the jaw to be matched are determined. Feature points are extracted from the tooth regions. Feature points are extracted from the next jaw taken from the database. It is investigated whether the features close to the features in the feature points exist on this jaw. If there are many matches, outlier cleaning is performed by applying RANSAC. It is examined how many common feature points each tooth has with each jaw in the database, that is, with high similarity. Accordingly, the most matching jaw images are determined.
Fig. 8
The general workflow of the approach is provided, including the steps of the approach
×
2.2.1.1 Determination of teeth regions
After segmentation of panoramic X-ray images, tooth regions are detected using the lines obtained during the segmentation step. Each area between two consecutive lines is considered to contain a tooth. However, tooth separation lines can be located at different angles depending on the orientation of the teeth. Therefore, accepting rectangular areas between the closest points or midpoints between two lines as the tooth area may cause a part of the tooth to be missing, or adjacent teeth to be included in the tooth region. To address this issue, the widest region that can be formed by both lines is chosen as the tooth separation region to ensure that the entire tooth is included while avoiding overlap with adjacent teeth.
Figure 9 shows the separation lines of a jaw image. Figure 10 shows an example of teeth obtained separately using the separation lines. The white rectangular region in Fig. 11 shows the region formed by the 2nd and 3rd separation lines. The extracted version of this region is shown in Fig. 10b).
Fig. 9
Example of segmented mandibular jaw image. The segmentation method detects the lines shown in the image between consecutive teeth. The tooth region is considered to be between consecutive lines
Fig. 10
Samples of tooth regions after teeth segmentation
Fig. 11
Variation of the RANSAC method on an example model
×
×
×
2.2.1.2 Extraction of region features
The SIFT, SURF, MSER, and KAZE algorithms, which are keypoint-based methods, were used for feature extraction on the images in both the database and the specified regions. Each of these algorithms is explained separately below.
Scale-Invariant Feature Transform (SIFT).
The scale-invariant feature transform algorithm [43] is a method for extracting unique local features from an image. This algorithm enables the extraction of features that are invariant to scaling, translation, and rotation. The main steps of the algorithm are as follows:
Creation of scale space
LoG convergence
Keypoint detection
Orientation assignment
Creating SIFT features
The SIFT algorithm works on the idea that different objects can be better distinguished at different scales. Therefore, scale spaces are first created. While creating the scale space, the image is blurred using Gaussian algorithm with multiple standard deviation (σ) values. Then the image dimensions are scaled. The group of differently blurred images at each scale forms an octave [43]. Then the extremum points are determined as keypoints. Angles are determined for each keypoint. A histogram with is created to determine the angle. The angle at the position with the highest value in the histogram is selected. Finally, the 16 × 16 region around the feature point is divided into 4 × 4 regions. For each region, an 8-segment histogram is created. Therefore, a feature vector of 4 × 4 × 8 = 128 dimensions is obtained for a keypoint [43]. Once the feature vectors are extracted and the descriptor vectors are defined, the matching between the two images is done by looking at the Euclidean distance between the descriptor vectors. If the Euclidean distance is below a certain threshold, the two points are considered as matched.
Speeded Up Robust Features (SURF).
After the proven capabilities of the SIFT algorithm, the Speeded Up Robust Features (SURF) method was proposed as an algorithm that can produce results comparable to SIFT in terms of performance, while also being faster than SIFT [44]. SURF uses Hessian matrices for LoG convergence, which helps in gaining insight into the variations between regions by analyzing the determinant of the Hessian matrix [44]. Moreover, it employs box filters to compute the Hessian matrix. Instead of operating on the normal image, the SURF method operates on the integral image, allowing for faster box filter calculations. Rather than reducing the image, the SURF method creates image pyramids by applying filters of varying sizes and chooses the maximum-valued consecutive region groups of the image pyramids as keypoints. The angle of the keypoint is determined by analyzing the Haar wavelet responses, and the angle with the highest response sum is selected. The selected area of each keypoint is divided into 4 × 4 subregions, yielding four features per subregion, resulting in a total of 64-dimensional feature data [44].
KAZE Features.
The KAZE method differs from the SIFT and SURF algorithms in that it utilizes a nonlinear diffusion filter [45]. Similar to SIFT, a scale space is generated; however, in the KAZE method, nonlinear filtering is used instead of Gaussian blurring. Gaussian blurring can cause natural boundaries of objects in the image to disappear as it blurs the image [45]. To preserve object edges, an adaptive blurring is applied in the KAZE method with the help of a nonlinear diffusion method.
The KAZE method employs a semi-implicit scheme to define the diffusion equation discretely and generate the scale space [45]. This scheme is based on additive operator splitting (AOS), which is stable in all step sizes [45]. To identify keypoints, Hessian matrices are used. Similarly, to SURF, KAZE calculates the possible directions over the derivatives within a radius of 6σi for each keypoint with σi being the scale of the keypoint. The region of 24σi is divided into 4 × 4 subregions, and the feature vector is obtained from these regions. Finally, the vector is transformed into a unit vector [46].
Maximally Stable Extremal Regions.
The maximally stable extremal regions (MSER) algorithm operates with binary images formed by progressively increasing or decreasing a threshold in the image [47]. These binary images start with a white image with very few black regions and vice versa, and each time a part of an object appears in the binary image, until the object fragments merge to reveal larger objects. The MSER algorithm examines the state of the resulting connected components and identifies the connected components that maintain their state for the longest duration as edge regions [47].
The process begins by sorting the pixels according to their brightness values and placing them in the image in ascending or descending order [47]. The union-find algorithm is then used to keep track of the connected components and their areas [47]. The merging of two connected components results in the disappearance of the smaller component, and all of its pixels are transferred to the larger component. Intensity levels that correspond to local minima of the rate of change of the area function are selected as maximum stable end regions [47]. Each maximally stable edge region is represented by the location of the local brightness minimum or maximum and a threshold value as an output. Therefore, the threshold intervals that produce the maximum stable endpoints are determined by varying the threshold value in the image where the regions are most stable [47]. This is done over the entire sequence of regions to obtain the maximum stationary endpoints.
Random Sample Consensus (RANSAC).
The Random Sample Consensus (RANSAC) algorithm, introduced by Fischler and Bolles [48], is used to detect outliers. RANSAC is a resampling method that generates candidate solutions using a minimum number of data points (observations) to calculate the basic model parameters. RANSAC starts with the smallest possible data set and expands this set with consistent data points.
The algorithm summary is as follows [49]: To determine the model parameters, a minimum number of points are selected at the beginning, and model parameters are created. The solution is performed according to these parameters. The status of all points for this solution is checked. Then, outliers and inliers are identified according to this solution. If the ratio of the number of inliers to the number of all points exceeds a certain threshold value (τ), the model parameters are recalculated using all inliers, and the algorithm terminates. Otherwise, the previous steps are repeated for a certain number of iterations.
An example of the initial and final model orientation of the RANSAC model for the two-dimensional plane is shown in Fig. 11. In the figure, the blue dots represent the data in space, while the orange line represents the model. The orange line symbolizes the initial position chosen for the model and the optimal position to be determined by RANSAC, respectively.
2.2.1.3 Detection of matching jaw
In the proposed method, the first step involves extracting feature points from all jaws and saving them in a database. Next, the target jaw is segmented, and feature points are created on each tooth. These feature points are then matched with those in the database, and the matching point that is most similar to the feature point region on the tooth is selected. This process is repeated for all teeth, and the total number of matching points is calculated. The number of matching points is higher if the jaw similarity is high and lower if it is low. Thus, the jaw with the highest number of matches is identified as the jaw of the matched person.
In tooth-jaw matching, as in the tooth-to-tooth matching approach, all tooth regions are searched in order. Unlike tooth-to-tooth matching, tooth-jaw matching examines the number of matching points of each tooth on the same jaw. This is performed for each image in the panoramic X-ray database, aiming to find the highest similarity.
Methods such as SIFT, SURF MSER and KAZE identify keypoints on the image. Each of these points contains a feature vector. The feature vector size depends on the keypoint-based algorithm. For instance, the SIFT algorithm constructs a 128-dimensional feature vector, while SURF algorithm constructs a 64-dimensional feature vector. Although the number of keypoints searched for matching varies, the feature vector size generated for each keypoint is the same. In the keypoint matching process, each feature is compared with the features of its own type.
Keypoint-based methods detect prominent points in the image. While a larger number of points are detected on the jaw, fewer points are detected on the teeth. The proposed study examines how many of these points are matched. Let M be the number of keypoints in the source jaw image in the queue, and N be the next tooth in the target jaw. Accordingly, it is examined whether N feature vectors match with M feature vectors. Whether the vectors match or not is determined by the distances of the vectors. The sum of squared differences approach is used for this purpose. This process is summarized in Fig. 12.
Fig. 12
Matching feature points obtained from the tooth region with feature points obtained from the jaw region
×
The keypoints obtained from a tooth region and the keypoints obtained from the jaw region are shown as a table in Fig. 12. The keypoint number is given in the first column of the tables. It is seen that N keypoints are detected for the tooth image, and M keypoints are detected for the jaw image. The columns following the first column in the tables show the vector values of the keypoints. The number of keypoint values is denoted by f and is determined by the algorithm used. The f value is the same for both teeth and jaws since the match checks are performed on vectors from the same algorithm. Finally, the arrows show that each keypoint from the tooth region is checked for similarity with each keypoint from the jaw region. The light-colored arrows point to representative matching vectors, i.e., close vectors. A high number of matching keypoints indicate that there is a region on the jaw similar to this tooth region.
An example of a keypoint identified on a jaw image is shown in Fig. 13. Distinctive keypoints can be identified on teeth, implants and even jaw regions within the image. Similarly, keypoints can also be identified on tooth regions. Keypoints for the tooth region are given in Fig. 13.Fig. It is examined whether the identified keypoints have a counterpart on the jaw. As in Fig. 13. and Fig. 13., it is seen that in the regions where the tooth regions are similar on the jaw, distinctive points appear in similar parts.
Fig. 13
Keypoints located in an example jaw region. Each keypoint contains a feature vector
×
Sample matching outputs obtained from X-ray images of the same person are shown in Fig. 14. The tooth regions are sequentially checked to see if there is a matching region on the jaw.
Fig. 14
Keypoints found in sample tooth regions from the same person
×
Figure 14 shows image matching from four different tooth regions. Two of the four regions were selected from healthy teeth and two from implants. Each tooth region has different regions of similarity with the jaw. As can be seen in Fig. 14, the number of matching keypoints is dense between each tooth region and the parts of the tooth region that are similar on the jaw. Since the similarities of all teeth are analyzed, even if some teeth are similar to the incorrect jaw, this can be compensated for by considering the similarities of all teeth. Figure 15 shows the matches between X-ray images of different people. The number of matching points is expected to be very low here as it is between different people. Figure 15 shows the matching of two tooth regions and two implant regions as in Fig. 14 Since these are images between different people, no matching points could be identified between these tooth and implant regions and the jaw X-ray.
Fig. 15
Examples of matching of feature points obtained from tooth regions with feature points obtained from the jaw
×
Many of the keypoints detected on the tooth can also be detected on the jaw. Although false-positive or false-negative points may occur, they usually do not prevent accurate detection. This is because the matching of a multiple number of teeth is examined. This can be thought of as a kind of voting where all teeth participate, and each tooth can contribute to the number of matching points (Fig. 16).
Fig. 16
Examples of matching of feature points obtained from tooth regions of different persons with feature points obtained from the jaw. No matching points was found in this example
×
The flowchart diagram of the process is given in Fig. 17. As can be seen in the diagram, the process starts with the acquisition of the jaw database and tooth regions. In order not to repeat this part in the experimental program, the features of the tooth regions and jaws are extracted once, and these extracted features are saved. Afterward, there is no need to repeat the image feature extraction process. In the flow, Tx refers to the target X-ray image from which the tooth regions will be extracted, abbreviated as target X-ray. Similarly, for the images in the database, it is called source X-ray and abbreviated as Sx. After the teeth are separated in the segmentation process, these tooth regions are called target regions and abbreviated as Tr. The keypoints extracted from these regions are denoted as kpTr. Similarly, the keypoints extracted from the database images are named kpSx. To denote the matching points, "m_" was added to the beginning of these names, and they were named as m_kpTr, m_kpSx,, respectively. Outlier points were cleaned with RANSAC algorithm and the values in the RANSAC output are shown as c_m_kpTr and c_m_kpSx. Finally, the number of matches of each tooth with each jaw is recorded and stored in mk_list. This list is created and kept separately for all teeth. In this way, the number of matches can be compared, and the jaws with the most matching points can be identified.
Fig. 17
Flowchart of the proposed approach
×
3 Experimental results
In the tooth-jaw matching approach, various keypoint-based features are utilized for person detection. Feature extraction algorithms are applied to the regions obtained in the segmentation phase. Next, feature extraction is performed on the images in the database. Since the features of the entire jaw are extracted during database feature extraction, segmentation is not necessary. The extracted feature points are then matched, and it is expected that the number of matches will be higher in images with a high similarity ratio. Based on this, the images with the highest number of matches are determined to be the matching images. SIFT, SURF, MSER, and KAZE features are all used for feature extraction.
The matching results for each of these features are presented in Table 1 (SIFT), Table 2 (SURF), Table 3 (MSER), and Table 4 (KAZE), respectively. The rows of the tables contain the matching results for different segmentation methods and jaw regions, while the columns show the ratios of matching images found in the first n images (where n = 1, 2, 3, 4, 5, 10, 15).
For each method, two different segmentation methods were tested. These are Semi Auto Segmentation and Auto Segmentation, which we have previously presented in [16] and [17], respectively.
Comparisons of the matching results obtained are given in Fig. 17, Fig. 18, Fig. 19, and Fig. 20. Figure 17 and Fig. 19 show the comparison of SIFT, SURF, MSER and KAZE features for mandibular jaw.
Fig. 18
Comparison of matching results on mandibular jaw using semi-automated segmentation
Fig. 19
Comparison of matching results on maxillary jaw using semi-automated segmentation
Fig. 20
Comparison of matching results on mandibular jaw using automated segmentation
×
×
×
Figure 18 and Fig. 20 show the comparison of SIFT, SURF, MSER and KAZE features for maxillary jaw (Fig. 21).
Fig. 21
Comparison of matching results on maxillary jaw using automated segmentation
×
In addition to the matching success, the proposed feature extraction methods are compared in terms of time. Figure 22 shows the average time needed to identify a person in the given database.
Fig. 22
Average matching times
×
4 Discussion
The proposed study evaluates two methods, one semi-automatic and one fully automatic, for identifying tooth regions. In the semi-automatic method, separation curves are determined after jaw separation, and the user selects the potential positions between consecutive teeth on the upper separation curve. Separation lines are then created perpendicular to the curve from the points selected by the user, and these lines are automatically oriented in accordance with the orientation of the teeth. This method offers the advantage of generating no missing or excessive separation lines. However, it also has a disadvantage in that it is semi-automatic, and the user's involvement in the segmentation process increases the user's workload and sensitivity to errors.
The fully automated method, on the other hand, automatically detects the separation lines with high accuracy. However, there is often an overestimation of accuracy, which is addressed by checking the rejects and automatically eliminating unnecessary ones. This method is preferred because it is fully automatic, reduces user workload, and is resistant to user mistakes.
In the next stage of the study, the person matching process is applied using a novel approach. Various experimental studies were conducted to evaluate this approach, and the resulting performances were analyzed.
When examining teeth-person matching studies in the literature, it becomes clear that the studies generally attempt to extract individual tooth boundaries and then detect the most similar tooth in the database. To the best of the author's knowledge, the teeth-jaw matching approach used in this study has not been previously used for person detection from panoramic dental x-ray images. This method offers several advantages over individual tooth matching.
The first advantage is that tooth boundary segmentation can be challenging, and incomplete or inaccurate determination of tooth boundaries can decrease the performance of the algorithm. In the proposed approach, it is not necessary to determine the tooth boundaries; instead, it is sufficient to determine the tooth region. Additionally, even if there are excess or missing segmentation regions, matching can still be performed using the obtained regions.
Another advantage of the teeth-jaw matching approach is the reduced number of searches required. Since the tooth region is matched with an entire jaw, the number of searches is significantly reduced. Let m denote the number of jaws and n denote the number of teeth. In this case, the search complexity is reduced from O(m2n2) to O(m2n).
Teeth-based approach, teeth-jaw-based approach and jaw-based approach are compared in Table 5. In Table 5, the pseudocodes for each approach are given and the complexity analysis is expressed.
Table 5
Complexity analysis of different matching approaches
For instance, suppose it is needed to search for 100 images with an average of 32 teeth within a database of 100 images. This means that it is needed to search for 320 target images over 320 database images, requiring 10,240,000 (3200 × 3200) searches. However, in the proposed method, since teeth and jaws are matched, only 320,000 (3200 × 100) searches are needed. The workload is significantly reduced, particularly since the number of jaws (m) is usually much lower than the number of teeth (n).
Jaw-to-jaw matching is the lowest approach in terms of computational load. Its complexity corresponds to O(m2). According to the example above, the number of searches required for jaw-jaw matching is 10,000 (100 × 100). On the other hand, in the jaw-to-jaw matching approach, the target jaw must be very intact. Also, individual teeth cannot be evaluated.
It is challenging to conduct a fair comparison since there is no publicly available dental panoramic dataset for comparing the human identification techniques. [33]. To the best of the author's knowledge, the most notable methods developed for panoramic dental images in recent years are [12, 24, 29, 31, 33, 50]. The method described in reference [31] was evaluated on a dataset consisting of 60 images, each belonging to one of 30 individuals. The results showed that the method achieved a rank-1 accuracy of 67%. However, this study does not present rank-5 or rank-10 accuracy rates. The relatively low rank-1 accuracy performance is a weakness of this study. The method described in reference [24] was evaluated on a dataset of 30 images of 20 different individuals. It achieved a success rate of 90%. However, it is important to note that the effectiveness of this approach directly depends on the presence of dental work in the images. Therefore, in datasets with a limited number of dental work images or in cases where dental work is absent, the performance of the method can be significantly degraded. Furthermore, as Oktay emphasized in [33], the performance of the method was also observed to degrade when dealing with different types of dental work.
In [33], it was found that about 81% of the images were successfully matched in the first order, indicating a high level of accuracy. Furthermore, 89% of the images were matched in the first or second rank, further emphasizing the efficiency of the system. It is worth noting that the 5th rank accuracy is 92%. This shows that a significant portion of the images is correctly matched within the first five ranks. Furthermore, the rank-10 accuracy reaches 94%. It is understood that the changes between rank-1, 2, 5 and 10 do not change very sharply. This implies a more stable performance progression. The downside of this study is that it employs a teeth-based technique, which results in increased computing complexity.
A novel deep learning approach for dental human identification was proposed in [29], where the authors developed and trained a CNN-based model using dental images. This was the first attempt to apply CNNs to this problem domain, and the results were promising. The model achieved 85.16% accuracy for rank-1 identification and 97.74% for rank-5 identification. However, a limitation of this method is that it is requires a training process and needs a large and diverse dataset for training, as CNN performance is highly influenced by the quality and quantity of data [29, 30].
A similar work is proposed in [50]. In the study, a novel deep neural network is proposed. It is provided 81.90% rank-1 and 91.22 rank-5 performance. As it seen in [29], this study also investigates jaw similarity.
Another study utilizing deep learning was presented in [12]. In this study, deep learning was used for tooth segmentation, and SURF features were used for matching. Rank-1 performance is 80.39% and rank-5 performance is 96.08%. This method and [50] have similar downside with [29] and [33]. First, it is a teeth-based method and has higher computer complexity. Second, it is a deep learning-based method and requires many dental X-ray images to train.
These deep learning methods focus on jaw-to-jaw similarity. This has the downside that individual tooth matching is difficult. In post-mortem x-rays, some of the teeth may be deteriorated or missing. In this case, the person needs to be matched with the remaining teeth. While this deterioration or loss complicates all approaches, it becomes a more significant problem for methods that primarily search for similarity between jaws. However, in tooth-to-tooth matching or tooth-jaw matching approaches, as in the proposed study, it is possible to examine the matching pattern of individual teeth.
Characteristics of other person matching studies in the literature are summarized in Table 6.
Table 6
Comparison the dental human identification methods
In the Table 6, the tooth-jaw matching approach proposed in this study is abbreviated as TJMA.
In the proposed work, deterministic approaches are used for both tooth segmentation and tooth matching that do not require a training process. Still, it is seen that high performances can be achieved comparable to the approaches utilizing deep learning.
The proposed approach is evaluated with multiple methods, and it is observed how well the methods provide suitable results for dental matching. In this respect, it gives an insight into the performance of different keypoint-based approaches for dental human identification. Another contribution of the proposed study, different from other studies, is that the stability of the method is tested by using different segmentation methods. This study examines whether the proposed method is robust to segmentation errors. As seen in Tables 1, 2, 3 and 4, the method provides consistent results for both semi-automatic and automatic segmentation.
5 Future directions
In the proposed work, a teeth-jaw-based matching approach is introduced as a novel method for dental person recognition. The study focuses on keypoint matching, necessitating the evolution and enhancement of various keypoint-based feature extractors. Significant progress can be achieved in this field by creating or improving keypoint approaches tailored for dental X-ray images. Preprocessing, i.e., image enhancement, of dental images is also important in this respect. Image enhancement methods focused on dental images will help to better detect keypoints. Finally, it is planned to advance studies on matching without certain parts of dental images.
6 Conclusion
This study presents a novel matching approach to use teeth as a reliable biometric feature to identify individuals following serious injuries caused by accidents or disasters. It discusses the performance and potential of this method. The current process of identification through dental records performed by forensic dentists is known to be time-consuming and laborious. Therefore, this research is aimed to present a computer-aided methodology to streamline and improve the identification process performed by forensic dentists.
Identification of individuals using dental features can be achieved by comparing each tooth of the individual with records in the database or by comparing the features of the source jaw with those in the database. Previous research has predominantly focused on jaw-to-jaw (jaw-based) or tooth-to-tooth (tooth-based) matching approaches. However, this paper proposes a new tooth-jaw-based approach. This method involves extracting keypoint-based features for each tooth and then matching these features with jaw keypoints in the database. Notably, this study marks the first time that such an approach has been applied for human identification using panoramic X-ray images. The proposed methodology combines the advantages of using jaw keypoints for matching in the database while simultaneously allowing the extraction of dental features.
Declarations
Conflict of interest
The authors declare no competing interests
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.