Melody recognition is an online process of evaluating incoming information and comparing this information to an existing internal corpus, thereby reducing prediction error. The predictive-coding model postulates top-down control on sensory processing accompanying reduction in prediction error. To investigate the relevancy of this model to melody processing, the current study examined early magnetoencephalogram (MEG) auditory responses to familiar and unfamiliar melodies in 25 participants. The familiar melodies followed and primed an octave-scrambled version of the same melody...

Listeners usually understand without difficulty even speech that sounds atypical. When they encounter noncanonical realizations of speech sounds, listeners can make short-term adjustments of their long-term representations of those sounds. Previous research, focusing mostly on adaptation in consonants, has suggested that for perceptual adaptation to take place some local cues (lexical, phonotactic, or visual) have to guide listeners' interpretation of the atypical sounds. In the present experiment we investigated perceptual adaptation in vowels...

Three experiments examine the role of previously read text in sentence comprehension and the control of eye movements during spontaneous rereading. Spontaneous rereading begins with a regressive saccade and involves reinspection of previously read text. All 3 experiments employed the gaze-contingent change technique to modulate the availability of previously read text. In Experiment 1, previously read text was permanently masked either immediately to the left of the fixated word (beyond wordn) or more than 1 word to the left (beyond wordn-1)...

There is ample evidence that motor actions are stored in terms of, and controlled by, the sensory effects that these actions produce. At present it is unclear, though, whether action control is governed by intended sensory changes (e.g., the transition from darkness to brightness when switching on a light) or only by intended sensory end states (e.g., the light being on). The present study explored the role of sensory changes for action control. To address this issue, participants engaged in a spatial tracking task...

The size congruity effect refers to the interaction between numerical magnitude and physical digit size in a symbolic comparison task. Though this effect is well established in the typical 2-item scenario, the mechanisms at the root of the interference remain unclear. Two competing explanations have emerged in the literature: an early interaction model and a late interaction model. In the present study, we used visual conjunction search to test competing predictions from these 2 models. Participants searched for targets that were defined by a conjunction of physical and numerical size...

The misallocation of driver visual attention has been suggested as a major contributing factor to vehicle accidents. One possible reason is that the relatively high cognitive demands of driving limit the ability to efficiently allocate gaze. We present an experiment that explores the relationship between attentional function and visual performance when driving. Drivers performed 2 variations of a multiple-object tracking task targeting aspects of cognition including sustained attention, dual-tasking, covert attention, and visuomotor skill...

It has been proposed that humans possess an automatic system to represent mental states ('implicit mentalizing'). The existence of an implicit mentalizing system has generated considerable debate however, centered on the ability of various experimental paradigms to demonstrate unambiguously such mentalizing. Evidence for implicit mentalizing has previously been provided by the 'dot perspective task,' where participants are slower to verify the number of dots they can see when an avatar can see a different number of dots...

Speech segmentation is supported by multiple sources of information that may either inform language processing specifically, or serve learning more broadly. The Iambic/Trochaic Law (ITL), where increased duration indicates the end of a group and increased emphasis indicates the beginning of a group, has been proposed as a domain-general mechanism that also applies to language. However, language background has been suggested to modulate use of the ITL, meaning that these perceptual grouping preferences may instead be a consequence of language exposure...

Hearing is confronted by a similar problem to vision when the observer moves. The image motion that is created remains ambiguous until the observer knows the velocity of eye and/or head. One way the visual system solves this problem is to use motor commands, proprioception, and vestibular information. These "extraretinal signals" compensate for self-movement, converting image motion into head-centered coordinates, although not always perfectly. We investigated whether the auditory system also transforms coordinates by examining the degree of compensation for head rotation when judging a moving sound...

To gain deeper insight into respective contributions of perceptual and motor experience of an observed action to anticipating its result, we examined the perceptual anticipation of players with different action roles in striking sports. Baseball pitchers and batters at both advanced and intermediate levels were asked to make a decision about whether to swing the bat when viewing a series of videos showing incomplete sequences of a model pitcher throwing a strike or a ball. The results revealed that first 100 ms of ball flight could discriminate advanced batters from intermediate pitchers and batters (with no difference between intermediate pitchers and batters)...

Viewed objects have been shown to afford suitable actions, even in the absence of any intention to act. However, little is known as to whether gaze behavior (i.e., the way we simply look at objects) is sensitive to action afforded by the seen object and how our actual motor possibilities affect this behavior. We recorded participants' eye movements during the observation of tools, graspable and ungraspable objects, while their hands were either freely resting on the table or tied behind their back. The effects of the observed object and hand posture on gaze behavior were measured by comparing the actual fixation distribution with that predicted by 2 widely supported models of visual attention, namely the Graph-Based Visual Saliency and the Adaptive Whitening Salience models...

Research on stimulus-response (S-R) associations as the basis of behavioral automaticity has a long history. Traditionally, it was assumed that S-R associations are formed as a consequence of the (repeated) co-occurrence of stimulus and response, that is, when participants act upon stimuli. Here, we demonstrate that S-R associations can also be established in the absence of action. In an item-specific priming paradigm, participants either classified everyday objects by performing a left or right key press (task-set execution) or they were verbally presented with information regarding an object's class and associated action while they passively viewed the object (verbal coding)...

Residual switch costs are deficits in task-switching performance that occur despite considerable time to prepare for a task switch. In the present study, the author investigated whether increased phasic alertness modulates residual switch costs. In 2 experiments involving the task-cuing procedure, subjects performed numerical categorization tasks on target digits, with and without an alerting stimulus presented shortly before the target (alert and no-alert trials, respectively). Switch costs were obtained that decreased with a longer cue-target interval, indicating subjects engaged in preparation, but large residual switch costs remained...

Sounds offer a rich source of information about events taking place in our physical and social environment. However, outside the domains of speech and music, little is known about whether humans can recognize and act upon the intentions of another agent's actions detected through auditory information alone. In this study we assessed whether intention can be inferred from the sound an action makes, and in turn, whether this information can be used to prospectively guide movement. In 2 experiments experienced and novice basketball players had to virtually intercept an attacker by listening to audio recordings of that player's movements...

Previous research has shown that attentional sets can be tuned to implicitly prioritize awareness of universally aversive or rewarding stimuli. But can mere ownership modulate implicit attentional prioritization as well? In Experiments 1 and 2, participants learned whether everyday objects belonged to them (self-owned) or the experimenter (other-owned) and completed a temporal order judgment task in which pairs of stimuli appeared onscreen with staggered timing. Results revealed a prior-entry effect, in which participants were more likely to report seeing a self-owned object first when 2 objects appeared simultaneously...

The literature on perceptual learning for speech shows that listeners use lexical information to disambiguate phonetically ambiguous speech sounds and that they maintain this new mapping for later recognition of ambiguous sounds for a given talker. Evidence for this kind of perceptual reorganization has focused on phonetic category boundary shifts. Here, we asked whether listeners adjust both category boundaries and internal category structure in rapid adaptation to foreign accents. We investigated the perceptual learning of Mandarin-accented productions of word-final voiced stops in English...

Many studies have revealed cultural differences in the way Easterners and Westerners attend to their visual world. It has been proposed that these cultural differences reflect the utilization of different processes, namely holistic processes by Easterners and analytical processes by Westerners. In the face processing literature, eye movement studies have revealed different fixation patterns for Easterners and Westerners that are congruent with a broader spread of attention by Easterners: compared with Westerners, Easterners tend to fixate more toward the center of the face even if they need the information provided by the eyes and mouth...

We tested an information foraging framework to characterize the mechanisms that drive active (visual) sampling behavior in decision problems that involve multiple sources of information. Experiments 1 through 3 involved participants making an absolute judgment about the direction of motion of a single random dot motion pattern. In Experiment 4, participants made a relative comparison between 2 motion patterns that could only be sampled sequentially. Our results show that: (a) Information (about noisy motion information) grows to an asymptotic level that depends on the quality of the information source; (b) The limited growth is attributable to unequal weighting of the incoming sensory evidence, with early samples being weighted more heavily; (c) Little information is lost once a new source of information is being sampled; and (d) The point at which the observer switches from 1 source to another is governed by online monitoring of his or her degree of (un)certainty about the sampled source...

Reading comes with a clear binocular advantage, expressed in shorter fixation times and fewer regressions in binocular relative to monocular visual presentations. Little is known, however, about whether the cost associated with monocular viewing derives primarily from the encoding of foveal information or in obtaining a preview benefit from upcoming parafoveal text. In the present sentence reading eye tracking experiment, the authors used a novel dichoptic binocular gaze-contingent moving window technique to selectively manipulate the amount of text made available to the reader both binocularly and monocularly in the fovea and parafovea on a fixation-by-fixation basis...

We can discriminate and recognize many faces, despite their visual similarity. Individual differences in this ability have been linked to 2 face coding mechanisms: adaptive norm-based coding of identity and holistic coding. However, it is not yet known whether these mechanisms are distinct. Nor is it known whether they make unique contributions to face recognition ability because no studies have measured the operation of both these mechanisms in the same individuals. We measured individual differences in both the strength of adaptive norm-based coding (with a face identity aftereffect task) and holistic coding (with a composite face task)...