Enhancing child safety online: A multi-modal not safe for work content detection system

dc.contributor.authorErandaka, E. R.
dc.contributor.authorSandunwala, S. T.
dc.contributor.authorMadurapperuma, M. T. S. K.
dc.contributor.authorRathnayaka, R. M. H. I.
dc.date.accessioned2025-11-13T08:00:24Z
dc.date.issued2025
dc.description.abstractBackground: The proliferation of digital platforms has significantly increased children's exposure to Not Safe for Work (NSFW) content, encompassing adult and violent material, posing substantial risks globally. Current video detection systems have limitations in automation, real-time processing, and comprehensive detection of violent content. Furthermore, existing models do not concurrently identify multiple NSFW content types (including text and imagery), highlighting a critical research gap. This study was conducted to propose a multi-modal NSFW detection system designed to surpass existing accuracy benchmarks to safeguard children in online environments. Method: The system employed a fine-tuned DistilBERT model for NSFW text detection and a YOLOv1ln model for image and video analysis. An automated OpenCV pipeline extracts video frames at 15 FPS, facilitating real-time processing. The training process utilizes a custom dataset alongside publicly available repositories from Hugging Face and Kaggle. The detection models were integrated into a web-based browser application through a streamlined data pipeline. Results: The system achieved 91% accuracy in text detection with DistilBERT and 80% accuracy in image and video analysis using YOLOv11n, trained on a dataset comprising over 15,000 annotated samples. Integration into a functional web-based application demonstrated effective NSFW content moderation capabilities. Conclusions: This research introduces a novel multi-modal NSFW detection framework capable of real-time content moderation, significantly contributing to enhanced child safety in digital spaces within Sri Lanka. Future recommendations include expanding dataset diversity, optimizing model architectures, and implementing the system in policy-driven environments, such as educational institutions and national safety networks.
dc.identifier.citationErandaka, E. R., Sandunwala, S. T., Madurapperuma, M. T. S. K., & Rathnayaka, R. M. H. I. (2025). Enhancing child safety online: A multi-modal not safe for work content detection system. International Conference on Child Protection 2025, University of Kelaniya, Sri Lanka. (p. 178).
dc.identifier.urihttp://repository.kln.ac.lk/handle/123456789/30258
dc.publisherInternational Conference on Child Protection 2025, University of Kelaniya, Sri Lanka.
dc.subjectNot safe for work
dc.subjectchild protection
dc.subjectmachine learning
dc.subjectDistilBERT
dc.subjectYOLO11n
dc.titleEnhancing child safety online: A multi-modal not safe for work content detection system
dc.typeArticle

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
ICCP'25_Proceedings Final-2-179.pdf
Size:
214.03 KB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: