Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/93696
Full metadata record
DC FieldValueLanguage
dc.contributor.authorVella, Joseph G.-
dc.contributor.authorFarrugia, Neil-
dc.date.accessioned2022-04-13T10:01:18Z-
dc.date.available2022-04-13T10:01:18Z-
dc.date.issued2021-
dc.identifier.citationVella, J. G., & Farrugia, N. (2021). Footwear impressions retrieval through textures and local features. In T. Tagarev, K. T. Atanassov, V. Kharchenko & J. Kacprzyk (Eds.), Digital transformation, cyber security and resilience of modern societies (pp. 343-360). Switzerland: Springer.en_GB
dc.identifier.urihttps://www.um.edu.mt/library/oar/handle/123456789/93696-
dc.description.abstractThe proposed artefact applies pre-processing filters, extracts key features, and retrieves the relevant matches from a shoeprint impression repository. Two functions were utilized for matching impressions. One function is texture based and creates an MPEG-1 movie out of two input images and employs the size of the output movie as a similarity measure. The other function is local feature based and uses SURF feature extraction and MSAC for matching. For pre-processing of the prints, a set of well-known techniques were employed. Also, we implemented a technique to facilitate better matching through splitting the input prints into smaller prints and then matching on these. FID 300 is a publicly available dataset of footwear impressions in greyscale. It comes with 1175 reference prints (e.g. sole images from tip to heel), and 300 prints lifted from real crime scenes, the latter being incomplete and with low image quality. The evaluation was done over various options and always against all reference prints in the FID 300. Clearly the evaluation results are affected by the quality of the lifted images. Evaluations were done in three batches (each having different pre-processing): first, all crime scene prints with the texture function got an average accuracy of 61%; second, a sample of 43 lifted prints with the texture function got 65% average accuracy; third, all crime scene prints and the local feature function applied got 50% average accuracy.en_GB
dc.language.isoenen_GB
dc.publisherSpringeren_GB
dc.rightsinfo:eu-repo/semantics/restrictedAccessen_GB
dc.subjectDigital forensic scienceen_GB
dc.subjectImage processing -- Digital techniquesen_GB
dc.subjectVisual texture recognitionen_GB
dc.subjectPattern perceptionen_GB
dc.titleFootwear impressions retrieval through textures and local featuresen_GB
dc.title.alternativeDigital transformation, cyber security and resilience of modern societiesen_GB
dc.typebookParten_GB
dc.rights.holderThe copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holder.en_GB
dc.description.reviewedpeer-revieweden_GB
dc.identifier.doi10.1007/978-3-030-65722-2-
Appears in Collections:Scholarly Works - FacICTCIS

Files in This Item:
File Description SizeFormat 
Footwear_impressions_retrieval_through_textures_and_local_features(2021).pdf
  Restricted Access
772.8 kBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.