This repository contains various models targetting multimodal representation learning, multimodal fusion for downstream tasks such as multimodal sentiment analysis.
Hi
how can we fit our own dataset .
Are models classifying the sentiment of speaker -
based on tone of speech, what they speech, with what facial expression do they speak ?
Hello,
Can you share the way you extract audio features in the work "Multi-level Multiple Attentions for Contextual Multimodal Sentiment Analysis"? I have no idea that how to extract 100 dimensions sentence-level audio features.
Thank you !
File "hfusion.py", line 111, in load_bimodal_activations
with open('./bimodal.pickle', 'rb') as handle:
FileNotFoundError: [Errno 2] No such file or directory: './bimodal.pickle'
the google drive link you have provided is not working. its showing 404 error. kindly check it out and is there any other source available to download the CMU-MOSEI dataset
I have trained the Context-Dependent Sentiment Analysis in User Generated Videos , and it went well! But how can I use the model to infer new data! Thanks! @soujanyaporia
I am confused about how to add distributed training so that I can define gpu for training. I don't know which file should I modify and add some codes in? Does anybody can help me?