Please use this identifier to cite or link to this item: https://www.um.edu.mt/library/oar/handle/123456789/93881
Full metadata record
DC FieldValueLanguage
dc.date.accessioned2022-04-18T07:59:00Z-
dc.date.available2022-04-18T07:59:00Z-
dc.date.issued2010-
dc.identifier.citationGaldies, K. (2010). Automatic avatar creator (Bachelor’s dissertation).en_GB
dc.identifier.urihttps://www.um.edu.mt/library/oar/handle/123456789/93881-
dc.descriptionB.Sc. IT (Hons)(Melit.)en_GB
dc.description.abstractThe proposed system aims to recreate a user digitally as an avatar, taking a live webcam feed as input. This digital caricature will mainly focus on the head and translate features such as skin colour, hair style, hair and eye colour together with mouth style and any eye wear detected. It aims to offer a less time consuming approach to avatar construction and provide a solid starting point for further customizability. All that is required to run the program is a windows pc and a standard webcam. After the installation, the program launches automatically and a webcam feed may be seen on the left whilst an avatar editor is seen on the right. If the user is positioned at an adequate distance from the webcam and the room lighting is not of a hindrance, the face will be detected instantly through the use of appearance-based detection methods employed in the Open CV library. The same method is then employed to detect eyes, nose and mouth and in this way feature extraction may take place. Extracted readings may be classified as either 'hard variables' or 'soft variables'. The former are only detected once and do not change throughout the whole session. On the other hand 'soft variables' such as glasses and 'mouth style' detection are constantly being refreshed and should reflect real-time changes. In other words once the hair colour is detected as "black" it will not change throughout the whole session. However, putting on a pair of glasses after the avatar has been constructed will result in a change. The outcome of this process can be seen in the avatar window where the detected features will be reflected in the digital model. The editor is also open to manual changes, giving the user the ability to correct wrongly detected features or else change other characteristics which for some reason or another were not automatically detected. From the results obtained, the proposed solution seems to be fairly successful in obtaining the correct facial information and delivers on its promise to link features extraction to avatar construction.en_GB
dc.language.isoenen_GB
dc.rightsinfo:eu-repo/semantics/restrictedAccessen_GB
dc.subjectAvatars (Virtual reality)en_GB
dc.subjectIcons (Computer graphics)en_GB
dc.subjectGraphical user interfaces (Computer systems)en_GB
dc.subjectComputer graphicsen_GB
dc.subjectImage processing -- Digital techniquesen_GB
dc.titleAutomatic avatar creatoren_GB
dc.typebachelorThesisen_GB
dc.rights.holderThe copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holder.en_GB
dc.publisher.institutionUniversity of Maltaen_GB
dc.publisher.departmentFaculty of Information and Communication Technology. Department of Computer Scienceen_GB
dc.description.reviewedN/Aen_GB
dc.contributor.creatorGaldies, Keith (2010)-
Appears in Collections:Dissertations - FacICT - 2010
Dissertations - FacICTCS - 2010-2015

Files in This Item:
File Description SizeFormat 
B.SC.(HONS)ICT_Galdies_Keith_2010.pdf
  Restricted Access
18.35 MBAdobe PDFView/Open Request a copy


Items in OAR@UM are protected by copyright, with all rights reserved, unless otherwise indicated.