Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

Unimodal statistical learning produces multimodal object-like representations

View through CrossRef
The concept of objects is fundamental to cognition and is defined by a consistent set of sensory properties and physical affordances. Although it is unknown how the abstract concept of an object emerges, most accounts assume that visual or haptic boundaries are crucial in this process. Here, we tested an alternative hypothesis that boundaries are not essential but simply reflect a more fundamental principle: consistent visual or haptic statistical properties. Using a novel visuo-haptic statistical learning paradigm, we familiarised participants with objects defined solely by across-scene statistics provided either visually or through physical interactions. We then tested them on both a visual familiarity and a haptic pulling task, thus measuring both within-modality learning and across-modality generalisation. Participants showed strong within-modality learning and ‘zero-shot’ across-modality generalisation which were highly correlated. Our results demonstrate that humans can segment scenes into objects, without any explicit boundary cues, using purely statistical information.
Title: Unimodal statistical learning produces multimodal object-like representations
Description:
The concept of objects is fundamental to cognition and is defined by a consistent set of sensory properties and physical affordances.
Although it is unknown how the abstract concept of an object emerges, most accounts assume that visual or haptic boundaries are crucial in this process.
Here, we tested an alternative hypothesis that boundaries are not essential but simply reflect a more fundamental principle: consistent visual or haptic statistical properties.
Using a novel visuo-haptic statistical learning paradigm, we familiarised participants with objects defined solely by across-scene statistics provided either visually or through physical interactions.
We then tested them on both a visual familiarity and a haptic pulling task, thus measuring both within-modality learning and across-modality generalisation.
Participants showed strong within-modality learning and ‘zero-shot’ across-modality generalisation which were highly correlated.
Our results demonstrate that humans can segment scenes into objects, without any explicit boundary cues, using purely statistical information.

Related Results

AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model
AFR-BERT: Attention-based mechanism feature relevance fusion multimodal sentiment analysis model
Multimodal sentiment analysis is an essential task in natural language processing which refers to the fact that machines can analyze and recognize emotions through logical reasonin...
Investigating the difficulty level of multimodal representations used by science teachers of gifted students
Investigating the difficulty level of multimodal representations used by science teachers of gifted students
In view of the tendency for gifted learners to experience disinterest in commonplace concepts and oversimplified explanations, coupled with their inclination towards intellectually...
Meta-Representations as Representations of Processes
Meta-Representations as Representations of Processes
In this study, we explore how the notion of meta-representations in Higher-Order Theories (HOT) of consciousness can be implemented in computational models. HOT suggests that consc...
Automatic Modulation Recognition Method Basedon Multimodal I/Q-FRFT Fusion
Automatic Modulation Recognition Method Basedon Multimodal I/Q-FRFT Fusion
Abstract Automatic modulation recognition (AMR) is a key technology in the domain of cognitive radio communications. Accurately identifying the modulation schemes of signal...
DESIGNING A MULTIMODAL TRANSPORT NETWORK
DESIGNING A MULTIMODAL TRANSPORT NETWORK
Objective: To create a methodology for designing a multimodal transport network under various scenarios of socioeconomic development of the Russian Federation and its regions which...
Sparse Fusion for Multimodal Transformers
Sparse Fusion for Multimodal Transformers
Multimodal classification is a core task in human-centric machine learning.We observe that information is highly complementary across modalities, thus unimodal information can be d...

Back to Top