Picture for Dinesh Kumar Vishwakarma

Dinesh Kumar Vishwakarma

MHS-STMA: Multimodal Hate Speech Detection via Scalable Transformer-Based Multilevel Attention Framework

Add code
Sep 08, 2024
Figure 1 for MHS-STMA: Multimodal Hate Speech Detection via Scalable Transformer-Based Multilevel Attention Framework
Figure 2 for MHS-STMA: Multimodal Hate Speech Detection via Scalable Transformer-Based Multilevel Attention Framework
Figure 3 for MHS-STMA: Multimodal Hate Speech Detection via Scalable Transformer-Based Multilevel Attention Framework
Figure 4 for MHS-STMA: Multimodal Hate Speech Detection via Scalable Transformer-Based Multilevel Attention Framework
Viaarxiv icon

Hate Content Detection via Novel Pre-Processing Sequencing and Ensemble Methods

Add code
Sep 08, 2024
Viaarxiv icon

A Noise and Edge extraction-based dual-branch method for Shallowfake and Deepfake Localization

Add code
Sep 02, 2024
Viaarxiv icon

Tex-ViT: A Generalizable, Robust, Texture-based dual-branch cross-attention deepfake detector

Add code
Aug 29, 2024
Viaarxiv icon

Modelling Visual Semantics via Image Captioning to extract Enhanced Multi-Level Cross-Modal Semantic Incongruity Representation with Attention for Multimodal Sarcasm Detection

Add code
Aug 05, 2024
Figure 1 for Modelling Visual Semantics via Image Captioning to extract Enhanced Multi-Level Cross-Modal Semantic Incongruity Representation with Attention for Multimodal Sarcasm Detection
Figure 2 for Modelling Visual Semantics via Image Captioning to extract Enhanced Multi-Level Cross-Modal Semantic Incongruity Representation with Attention for Multimodal Sarcasm Detection
Figure 3 for Modelling Visual Semantics via Image Captioning to extract Enhanced Multi-Level Cross-Modal Semantic Incongruity Representation with Attention for Multimodal Sarcasm Detection
Figure 4 for Modelling Visual Semantics via Image Captioning to extract Enhanced Multi-Level Cross-Modal Semantic Incongruity Representation with Attention for Multimodal Sarcasm Detection
Viaarxiv icon

VyAnG-Net: A Novel Multi-Modal Sarcasm Recognition Model by Uncovering Visual, Acoustic and Glossary Features

Add code
Aug 05, 2024
Viaarxiv icon

Target-Dependent Multimodal Sentiment Analysis Via Employing Visual-to Emotional-Caption Translation Network using Visual-Caption Pairs

Add code
Aug 05, 2024
Figure 1 for Target-Dependent Multimodal Sentiment Analysis Via Employing Visual-to Emotional-Caption Translation Network using Visual-Caption Pairs
Figure 2 for Target-Dependent Multimodal Sentiment Analysis Via Employing Visual-to Emotional-Caption Translation Network using Visual-Caption Pairs
Figure 3 for Target-Dependent Multimodal Sentiment Analysis Via Employing Visual-to Emotional-Caption Translation Network using Visual-Caption Pairs
Figure 4 for Target-Dependent Multimodal Sentiment Analysis Via Employing Visual-to Emotional-Caption Translation Network using Visual-Caption Pairs
Viaarxiv icon

Contrastive Learning-based Multi Modal Architecture for Emoticon Prediction by Employing Image-Text Pairs

Add code
Aug 05, 2024
Viaarxiv icon

Towards Effective Image Forensics via A Novel Computationally Efficient Framework and A New Image Splice Dataset

Add code
Jan 13, 2024
Viaarxiv icon

Datasets, Clues and State-of-the-Arts for Multimedia Forensics: An Extensive Review

Add code
Jan 13, 2024
Viaarxiv icon