Lists (1)
Sort Name ascending (A-Z)
Stars
Code and some materials from the papers "Selection of Source Images Heavily Influences the Effectiveness of Adversarial Attacks" (BMVC 2021) and "Evaluating Adversarial Attacks on ImageNet:A Realit…
A multimodal approach on emotion recognition using audio and text.
[EMNLP2023] Conversation Understanding using Relational Temporal Graph Neural Networks with Auxiliary Cross-Modality Interaction
Pretrained Pytorch face detection (MTCNN) and facial recognition (InceptionResnet) models
IEEE T-BIOM : "Audio-Visual Fusion for Emotion Recognition in the Valence-Arousal Space Using Joint Cross-Attention"
This is an example code for audio-visual emotion recognition via a hybrid deep model
The FG2019 paper submitted as Audio-Visual Emotion Forecasting. This repository contains the codes and other necessary information for that.
This repository contains various models targetting multimodal representation learning, multimodal fusion for downstream tasks such as multimodal sentiment analysis.
ABAW3 (CVPRW): A Joint Cross-Attention Model for Audio-Visual Fusion in Dimensional Emotion Recognition
Repository for collecting and categorizing papers outlined in our survey paper: "Large Language Models on Tabular Data -- A Survey".
We achieved the 2nd and 3rd places in ABAW3 and ABAW5, respectively.
ABAW6 (CVPR-W) We achieved second place in the valence arousal challenge of ABAW6
Explore my diverse collection of projects showcasing machine learning, data analysis, and more. Organized by project, each directory contains code, datasets, documentation, and resources. Dive in, …
NeuralTalk is a Python+numpy project for learning Multimodal Recurrent Neural Networks that describe images with sentences.
Code supporting the paper "Multimodal Autoencoder: A Deep Learning Approach to Filling In Missing Sensor Data and Enabling Better Mood Prediction"
Lightweight and Interpretable ML Model for Speech Emotion Recognition and Ambiguity Resolution (trained on IEMOCAP dataset)
TensorFlow implementation of "Multimodal Speech Emotion Recognition using Audio and Text," IEEE SLT-18
This repository contains the code for the paper `End-to-End Multimodal Emotion Recognition using Deep Neural Networks`.
A real time Multimodal Emotion Recognition web app for text, sound and video inputs
This repository provides implementation for the paper "Self-attention fusion for audiovisual emotion recognition with incomplete data".
A python package that includes many methods for decoding neural activity
A publicly-editable collection of open computational neuroscience resources
Python ETL framework for stream processing, real-time analytics, LLM pipelines, and RAG.