Cross-Platform Multi-Modal Transfer Learning Framework for Cyberbullying Detection

  • Weiqi ZHANG
  • , Chengzu DONG*
  • , Aiting YAO
  • , Asef NAZARI*
  • , Anuroop GADDAM
  • *Corresponding author for this work

Research output: Journal PublicationsJournal Article (refereed)peer-review

Abstract

Cyberbullying and hate speech increasingly appear in multi-modal social media posts, where images and text are combined in diverse and fast changing ways across platforms. These posts differ in style, vocabulary and layout, and labeled data are sparse and noisy, which makes it difficult to train detectors that are both reliable and deployable under tight computational budgets. Many high performing systems rely on large vision language backbones, full parameter fine tuning, online retrieval or model ensembles, which raises training and inference costs. We present a parameter efficient cross-platform multi-modal transfer learning framework for cyberbullying and hateful content detection. Our framework has three components. First, we perform domain adaptive pretraining of a compact ViLT backbone on in domain image-text corpora. Second, we apply parameter efficient fine tuning that updates only bias terms, a small subset of LayerNorm parameters and the classification head, leaving the inference computation graph unchanged. Third, we use noise aware knowledge distillation from a stronger teacher built from pretrained text and CLIP based image-text encoders, where only high confidence, temperature scaled predictions are used as soft labels during training, and teacher models and any retrieval components are used only offline. We evaluate primarily on Hateful Memes and use IMDB as an auxiliary text only benchmark to show that the deployment aware PEFT + offline-KD recipe can still be applied when other modalities are unavailable. On Hateful Memes, our student updates only 0.11% of parameters and retain about 96% of the AUROC of full fine-tuning.
Original languageEnglish
Article number442
JournalElectronics
Volume15
Issue number2
Early online date20 Jan 2026
DOIs
Publication statusPublished - Jan 2026

Bibliographical note

Publisher Copyright:
© 2026 by the authors.

Keywords

  • multi-modal cyberbullying detection
  • transfer learning
  • domain adaptive pretraining
  • knowledge distillation
  • parameter efficient fine tuning
  • lightweight deployment

Fingerprint

Dive into the research topics of 'Cross-Platform Multi-Modal Transfer Learning Framework for Cyberbullying Detection'. Together they form a unique fingerprint.

Cite this