Lawrence D. Shriberg, PhD -Slide of the Week

The goals of this research were to obtain initial estimates of the prevalence of each of four types of motor speech disorders in children with idiopathic Speech Delay (SD) and to use findings to estimate the population-based prevalence of each disorder. Analyses were completed on audio-recorded conversational speech samples from 415 children recruited for research in idiopathic SD in six USA cities during the past three decades.

Karl S. Rosengren, PhD – Slide of the Week

Children’s drawings have long been used to assess aspects of general cognitive functioning, intelligence, perceptual motor development, and even socio-emotional development.  The goal of the current study was to examine the structure of children’s drawings using crowd-sourced human similarity judgments and machine vision approaches. 

Seth Pollak, PhD – Slide of the Week

Although the configurations of facial muscles that humans perceive vary continuously, we often represent emotions as categories. This suggests that, as in other domains of categorical perception such as speech and color perception, humans become attuned to features of emotion cues that map onto meaningful thresholds for these signals given their environments.

Robert Pearce, MD, PhD – Slide of the Week

Previous experiments using genetic and pharmacological manipulations have provided strong evidence that etomidate impairs synaptic plasticity and memory by modulating a5-subunit containing GABAARs (a5-GABAARs). Since a5-GABAARs mediate tonic inhibition (TI) in hippocampal CA1 pyramidal cells, and etomidate enhances TI, etomidate enhancement of TI in pyramidal cells has been proposed as the underlying mechanism (Martin et al., 2009).

Ben Parrell, PhD – Slide of the Week

We present a new computational model of speech motor control: the Feedback-Aware Control of Tasks in Speech or FACTS model. This model is based on a state feedback control architecture, which is widely accepted in non-speech motor domains. The FACTS model employs a hierarchical observer-based architecture, with a distinct higher-level controller of speech tasks and a lower-level controller of speech articulators.