Please use this identifier to cite or link to this item: http://repository.kln.ac.lk/handle/123456789/11281
Full metadata record
DC FieldValueLanguage
dc.contributor.authorDissanayaka, D.M.M.T.-
dc.contributor.authorLiyanage, S.R.-
dc.date.accessioned2016-01-22T03:37:01Z-
dc.date.available2016-01-22T03:37:01Z-
dc.date.issued2015-
dc.identifier.citationDissanayaka, D.M.M.T. and Liyanage, S.R. 2015. Real Time Emotion Based Music Player for Android, p. 221, In: Proceedings of the International Postgraduate Research Conference 2015 University of Kelaniya, Kelaniya, Sri Lanka, (Abstract), 339 pp.en_US
dc.identifier.urihttp://repository.kln.ac.lk/handle/123456789/11281-
dc.description.abstractListening to music has been found to affect the human brain activities. Emotion based music players with automated playlists can help users to maintain a selected emotional state. This research proposes an emotion based music player that create playlists based on real time photos of the user. Two emotional statuses, happy and not-happy were considered in this study. User‘s images were captured in real-time using an android device camera. Grey scaled images were used to compress the image files. Eye and lip areas were cropped and sent to the MATLAB backend via client server-socket connections. Gaussian filtering was applied to reduce noise. Canny Edge Detection algorithm was used for edge detection. Eigen face-based pattern recognition was used for emotion recognition. PCA eigenvectors were learnt from the dataset via unsupervised training to learn the Eigen face models. The dissimilarity between pairs of face images projected to the Eigen space were measured using the Euclidean distance. The matched image was the one with the lowest dissimilarity. The identified label, happy/not-happy was transmitted back to the Android music player via a client server socket connection. Songs that are pre-categorised as happy/ not-happy are stored in the android application. When emotional label of the perceived face image is received, songs relevant to the received emotional label are loaded to the android music player. 120 face images were collected at the Department of Statistics & Computer Science, University of Kelaniya for validation. Another 100 happy and 100 not-happy images were collected for testing. Out of the 100 test cases with happy faces 75 were detected as happy and out of the 100 not-happy faces 66 were classified as not-happy. The overall accuracy of the developed system for the 200 test cases was 70.5%. This concept can be extended from a single face to multiple faces and the system has to be made more robust to noises, different poses, and structural components. The system can be extended to include other emotions that are recognizable via facial expressions.en_US
dc.language.isoenen_US
dc.publisherFaculty of Graduate Studies, University of Kelaniyaen_US
dc.subjectGaussian Filteren_US
dc.subjectCanny Edge Detectoren_US
dc.subjectEigen Faceen_US
dc.subjectPrinciple Component Analysisen_US
dc.titleReal Time Emotion Based Music Player for Androiden_US
dc.typeArticleen_US
Appears in Collections:IPRC - 2015

Files in This Item:
File Description SizeFormat 
221.pdf127.36 kBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.