Multimodal analysis and prediction of latent user dimensions

Wendlandt, Laura and Mihalcea, Rada and Boyd, Ryan L. and Pennebaker, James W. (2017) Multimodal analysis and prediction of latent user dimensions. In: Social Informatics - 9th International Conference, SocInfo 2017, Proceedings :. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) . Springer-Verlag, GBR, pp. 323-340. ISBN 9783319672168

Full text not available from this repository.


Humans upload over 1.8 billion digital images to the internet each day, yet the relationship between the images that a person shares with others and his/her psychological characteristics remains poorly understood. In the current research, we analyze the relationship between images, captions, and the latent demographic/psychological dimensions of personality and gender. We consider a wide range of automatically extracted visual and textual features of images/captions that are shared by a large sample of individuals Using correlational methods, we identify several visual and textual properties that show strong relationships with individual differences between participants. Additionally, we explore the task of predicting user attributes using a multimodal approach that simultaneously leverages images and their captions. Results from these experiments suggest that images alone have significant predictive power and, additionally, multimodal methods outperform both visual features and textual features in isolation when attempting to predict individual differences.

Item Type:
Contribution in Book/Report/Proceedings
Uncontrolled Keywords:
?? vision modelsmultimodal predictiontheoretical computer sciencegeneral computer science ??
ID Code:
Deposited By:
Deposited On:
22 Jun 2019 01:04
Last Modified:
16 Jul 2024 04:38