论文标题
通过关节微调进行多模式分析的转移学习
Transfer Learning with Joint Fine-Tuning for Multimodal Sentiment Analysis
论文作者
论文摘要
大多数现有方法都集中在文本数据的情感分析上。但是,最近在社交平台上大量使用了图像和视频,激发了其他模式的情感分析。当前的研究表明,探索其他方式(例如,图像)会增加情感分析表现。最先进的多模型(例如剪辑和Visualbert)已在数据集中预先训练,并与文本与图像配对。尽管这些模型获得的结果是有希望的,但是这些模型的预训练和情感分析微调任务在计算上却很昂贵。本文使用联合微调进行情感分析介绍了一种转移学习方法。我们的建议使用更直接的替代微调策略实现了竞争结果,该策略利用了不同的预训练的单峰模型,并有效地将它们结合在多模式空间中。此外,我们的建议可以在关节微调阶段合并任何预训练的文本和图像的预培训模型时灵活,这对于低资源场景中的情感分类尤其有趣。
Most existing methods focus on sentiment analysis of textual data. However, recently there has been a massive use of images and videos on social platforms, motivating sentiment analysis from other modalities. Current studies show that exploring other modalities (e.g., images) increases sentiment analysis performance. State-of-the-art multimodal models, such as CLIP and VisualBERT, are pre-trained on datasets with the text paired with images. Although the results obtained by these models are promising, pre-training and sentiment analysis fine-tuning tasks of these models are computationally expensive. This paper introduces a transfer learning approach using joint fine-tuning for sentiment analysis. Our proposal achieved competitive results using a more straightforward alternative fine-tuning strategy that leverages different pre-trained unimodal models and efficiently combines them in a multimodal space. Moreover, our proposal allows flexibility when incorporating any pre-trained model for texts and images during the joint fine-tuning stage, being especially interesting for sentiment classification in low-resource scenarios.