Mateo Ve´ lez-Fort,1,3 Lee Cossell,1,3 Laura Porta,1 Claudia Clopath,1,2 and Troy W. Margrie1,4,*
1 Sainsbury Wellcome Centre, University College London, London, UK
2 Bioengineering Department, Imperial College London, London, UK
3 These authors contributed equally
4 Lead contact
*Correspondence: t.margrie@ucl.ac.uk
https://doi.org/10.1016/j.cell.2025.01.032
SUMMARY
Knowing whether we are moving or something in the world is moving around us is possibly the most critical sensory discrimination we need to perform. How the brain and, in particular, the visual system solves this motion-source separation problem is not known. Here, we find that motor, vestibular, and visual motion signals are used by the mouse primary visual cortex (VISp) to differentially represent the same visual flow information according to whether the head is stationary or experiencing passive versus active translation. During locomotion, we find that running suppresses running-congruent translation input and that translation signals dominate VISp activity when running and translation speed become incongruent. This cross-modal interaction between the motor and vestibular systems was found throughout the cortex, indicating that running and translation signals provide a brain-wide egocentric reference frame for computing the internally generated and actual speed of self when moving through and sensing the external world.
Ugne Klibaite,1,4,* Tianqing Li,2,4 Diego Aldarondo,1,3 Jumana F. Akoad,1 Bence P. O¨ lveczky,1,* and Timothy W. Dunn2,5,*
1 Department of Organismic and Evolutionary Biology, Harvard University, Cambridge, MA 02138, USA
2 Department of Biomedical Engineering, Duke University, Durham, NC 27708, USA
3 Present address: Fauna Robotics, New York, NY 10003, USA
4 These authors contributed equally
5 Lead contact
*Correspondence: klibaite@fas.harvard.edu (U.K.), olveczky@fas.harvard.edu (B.P.O¨ .), timothy.dunn@duke.edu (T.W.D.)
https://doi.org/10.1016/j.cell.2025.01.044
Social interaction is integral to animal behavior. However, lacking tools to describe it in quantitative and rigorous ways has limited our understanding of its structure, underlying principles, and the neuropsychiatric disorders, like autism, that perturb it. Here, we present a technique for high-resolution 3D tracking of postural dynamics and social touch in freely interacting animals, solving the challenging subject occlusion and partassignment problems using 3D geometric reasoning, graph neural networks, and semi-supervised learning. We collected over 110 million 3D pose samples in interacting rats and mice, including seven monogenic autism rat lines. Using a multi-scale embedding approach, we identified a rich landscape of stereotyped actions, interactions, synchrony, and body contacts. This high-resolution phenotyping revealed a spectrum of changes in autism models and in response to amphetamine not resolved by conventional measurements. Our framework and large library of interactions will facilitate studies of social behaviors and their neurobiological underpinnings.
Authors
Igor Zilberman
Vitaliya Boyar
Keywords
Nancy A. Faller
Elizabeth A. LaGro
Victoria E. Elliott
伤口世界平台生态圈,以“关爱人间所有伤口患者”为愿景,连接、整合和拓展线上和线下的管理慢性伤口的资源,倡导远程、就近和居家管理慢性伤口,解决伤口专家的碎片化时间的价值创造、诊疗经验的裂变复制、和患者的就近、居家和低成本管理慢性伤口的问题。
2019广东省医疗行业协会伤口管理分会年会
扫一扫了解详情:
任何关于疾病的建议都不能替代执业医师的面对面诊断。所有门诊时间仅供参考,最终以医院当日公布为准。
网友、医生言论仅代表其个人观点,不代表本站同意其说法,请谨慎参阅,本站不承担由此引起的法律责任。