One is not enough: Multimodal learning for richer information.

Unimodal deep learning focuses on processing, analyzing, and generating data from a single modality. For example, this could involve training neural nets on images, text, or audio to classify, regress, or generate on only images, text, or audio. Multimodal deep learning, on the other hand, involves integrating and jointly processing data from multiple modalities. This … Continue reading One is not enough: Multimodal learning for richer information.

SOCIAL ROBOTS AND FACIAL EMOTIONS​

Humans express emotion through various media. Humans use their faces to present internal affective states. Humans use speech to express emotion as well. Among various types of expressions of emotions, facial expressions are predominant. Various expressions displayed on humans' faces represent different cognitive processes. In communicative or interactive conditions, humans can recognize the emotion being … Continue reading SOCIAL ROBOTS AND FACIAL EMOTIONS​