• The development of automated palmprint identification using major flexion creases

      Sutton, Raul Dr; Cook, Thomas (University of Wolverhampton, 2012-01)
      Palmar flexion crease matching is a method for verifying or establishing identity. New methods of palmprint identification, that complement existing identification strategies, or reduce analysis and comparison times, will benefit palmprint identification communities worldwide. To this end, this thesis describes new methods of manual and automated palmar flexion crease identification, that can be used to identify palmar flexion creases in online palmprint images. In the first instance, a manual palmar flexion crease identification and matching method is described, which was used to compare palmar flexion creases from 100 palms, each modified 10 times to mimic some of the types of alterations that can be found in crime scene palmar marks. From these comparisons, using manual palmar flexion crease identification, results showed that when labelled within 10 pixels, or 3.5 mm, of the palmar flexion crease, a palmprint image can be identified with a 99.2% genuine acceptance rate and a 0% false acceptance rate. Furthermore, in the second instance, a new method of automated palmar flexion crease recognition, that can be used to identify palmar flexion creases in online palmprint images, is described. A modified internal image seams algorithm was used to extract the flexion creases, and a matching algorithm, based on kd-tree nearest neighbour searching, was used to calculate the similarity between them. Results showed that in 1000 palmprint images from 100 palms, when compared to manually identified palmar flexion creases, a 100% genuine acceptance rate was achieved with a 0.0045% false acceptance rate. Finally, to determine if automated palmar flexion crease recognition can be used as an effective method of palmprint identification, palmar flexion creases from two online palmprint image data sets, containing images from 100 palms and 386 palms respectively, were automatically extracted and compared. In the first data set, that is, for images from 100 palms, an equal error rate of 0.3% was achieved. In the second data set, that is, for images from 386 palms, an equal error rate of 0.415% was achieved.