AJSLP
Tutorial
Brain–Computer Interfaces for
Augmentative and Alternative
Communication: A Tutorial
Jonathan S. Brumberg,a Kevin M. Pitt,b Alana Mantie-Kozlowski,c and Jeremy D. Burnisond
Purpose: Brain–computer interfaces (BCIs) have the potential
to improve communication for people who require but are
unable to use traditional augmentative and alternative
communication (AAC) devices. As BCIs move toward clinical
practice, speech-language pathologists (SLPs) will need to
consider their appropriateness for AAC intervention.
Method: This tutorial provides a background on BCI
approaches to provide AAC specialists foundational knowledge
necessary for clinical application of BCI. Tutorial descriptions
were generated based on a literature review of BCIs for
restoring communication.
Results: The tutorial responses directly address 4 major
areas of interest for SLPs who specialize in AAC: (a) the
current state of BCI with emphasis on SLP scope of
practice (including the subareas: the way in which
individuals access AAC with BCI, the efficacy of BCI
for AAC, and the effects of fatigue), (b) populations for
whom BCI is best suited, (c) the future of BCI as an
addition to AAC access strategies, and (d) limitations
of BCI.
Conclusion: Current BCIs have been designed as
access methods for AAC rather than a replacement;
therefore, SLPs can use existing knowledge in AAC as
a starting point for clinical application. Additional training
is recommended to stay updated with rapid advances
in BCI.
I
& Baker, 2012). In the most serious cases of total paralysis
with loss of speech (e.g., locked-in syndrome; Plum &
Posner, 1972), even these advanced methods are not sufficient to provide access to language and literacy (Oken
et al., 2014). Access to communication is critical for maintaining social interactions and autonomy of decision-making
in this population (Beukelman & Mirenda, 2013); therefore,
individuals with paralysis and akinetic mutism have been
identified as potential candidates for brain–computer interface (BCI) access to AAC (Fager et al., 2012).
BCIs for communication take AAC and access technology to the next level and provide a method for selecting
and constructing messages by detecting changes in brain
activity for controlling communication software (Wolpaw,
Birbaumer, McFarland, Pfurtscheller, & Vaughan, 2002).
In particular, they are devices that provide a direct link between an individual and a computer device through brain
activity alone, without requiring any overt movement or
behavior. As an access technique, BCIs have the potential
to reduce or eliminate some physical barriers to successful
AAC intervention for individuals with severe speech and
physical impairments. Similar to AAC and associated access
techniques, current BCI technology can take a variety of
forms on the basis of the neural signal targeted and the method
used for individuals to interact with the communication
ndividuals with severe speech and physical impairments often rely on augmentative and alternative
communication (AAC) and specialized access technologies to facilitate communication on the basis of the
nature and severity of their speech, motor, and cognitive
impairments. In some cases, people who use AAC are able
to use specially modified computer peripherals (e.g., mouse,
joystick, stylus, or button box) to access AAC devices,
whereas in other, more severe cases, sophisticated methods
are needed to detect the most subtle of movements (e.g.,
eye gaze tracking; Fager, Beukelman, Fried-Oken, Jakobs,
a
Department of Speech-Language-Hearing: Sciences and Disorders,
Neuroscience Graduate Program, The University of Kansas,
Lawrence
b
Department of Speech-Language-Hearing: Sciences and Disorders,
The University of Kansas, Lawrence
c
Communication Sciences and Disorders Department, Missouri State
University, Springfield
d
Neuroscience Graduate Program, The University of Kansas,
Lawrence
Correspondence to Jonathan S. Brumberg: brumberg@ku.edu
Editor-in-Chief: Krista Wilkinson
Editor: Erinn Finke
Received December 31, 2016
Revision received April 6, 2017
Accepted August 14, 2017
https://doi.org/10.1044/2017_AJSLP-16-0244
Disclosure: The authors have declared that no competing interests existed at the time
of publication.
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018 • Copyright © 2018 American Speech-Language-Hearing Association
1
interface. Each of these factors may impose different demands on the cognitive and motor abilities of individuals
who use BCI (Brumberg & Guenther, 2010). Although the
field of BCI has grown over the past decade, many stakeholders including speech-language pathologists (SLPs), other
practitioners, individuals who use AAC and potentially
BCI, and caretakers are unfamiliar with the technology.
SLPs are a particularly important stakeholder given their
role as the primary service providers who assist clients with
communicative challenges secondary to motor limitations
through assessment and implementation of AAC interventions and strategies. A lack of core knowledge on the potential use of BCI for clinical application may limit future
intervention with BCI for AAC according to established
best practices. This tutorial will offer some basic explanations regarding BCI, including the benefits and limitations
of this access technique, and the different varieties of BCI.
It will also provide a description of individuals who may be
potentially best suited for using BCI to access AAC. An
understanding of this information is especially important
for SLPs specializing in AAC who are most likely to interact with BCI as they move from research labs into real-world
situations (e.g., classrooms, home, work).
Tutorial Descriptions by Topic Area
Topic 1: How Do People Who Use BCI Interact
With the Computer?
BCIs are designed to allow individuals to control
computers and communication systems using brain activity alone and are separated according to whether signals
are recorded noninvasively from/through the scalp or invasively through implantation of electrodes in or on the brain.
Noninvasive BCIs, those that are based on brain recordings made through the intact skull without requiring a surgical procedure (e.g., electroencephalography or EEG,
magnetoencephalography, functional magnetic resonance
imaging, functional near-infrared spectroscopy), often use
an indirect technique to map brain signals unrelated to communication onto controls for a communication interface
(Brumberg, Burnison, & Guenther, 2016). Though there
are many signal acquisition modalities for noninvasive
recordings of brain activity, noninvasive BCIs typically
use EEG, which is recorded through electrodes placed on
the scalp according to a standard pattern (Oostenveld &
Praamstra, 2001) and record voltage changes that result
from the simultaneous activation of millions of neurons. EEG
can be analyzed for its spontaneous activity, or in response
to a stimulus (e.g., event-related potentials), and both
have been examined for indirect access BCI applications.
In contrast, another class of BCIs attempts to directly output speech from imagined/attempted productions (Blakely,
Miller, Rao, Holmes, & Ojemann, 2008; Brumberg, Wright,
Andreasen, Guenther, & Kennedy, 2011; Herff et al., 2015;
Kellis et al., 2010; Leuthardt et al., 2011; Martin et al., 2014;
Mugler et al., 2014; Pei, Barbour, Leuthardt, & Schalk,
2011; Tankus, Fried, & Shoham, 2012); however, these
2
techniques typically rely on invasively recorded brain signals (via implanted microelectrodes or subdural electrodes)
related to speech motor preparation and production. Though
in their infancy, direct BCIs for communication have the
potential to completely replace the human vocal tract for
individuals with severe speech and physical impairments
(Brumberg, Burnison, & Guenther, 2016; Chakrabarti,
Sandberg, Brumberg, & Krusienski, 2015); however, the
technology does not yet provide a method to “read thoughts.”
For the remainder of this tutorial, we focus on noninvasive,
indirect methods for accessing AAC with BCIs, and we refer
readers to other sources for descriptions of direct BCIs for
speech (Brumberg, Burnison, & Guenther, 2016; Chakrabarti
et al., 2015).
Indirect methods for BCI parallel other access methods
for AAC devices, where nonspeech actions (e.g., button
press, direct touch, eye gaze) are translated to a selection on
a communication interface. The main difference between
the two access methods is that BCIs rely on neurophysiological signals related to sensory stimulation, preparatory
motor behaviors, and/or covert motor behaviors (e.g., imagined or attempted limb movements), rather than overt motor
behavior used for conventional access. The way in which
individuals control a BCI greatly depends on the neurological signal used by the device to make selections on the
communication interface. For instance, in the case of an
eye-tracking AAC device, one is required to gaze at a communication icon, and the system makes a selection on the
basis of the screen coordinates of the eye gaze location. For
a BCI, individuals may be required to (a) attend to visual
stimuli to generate an appropriate visual–sensory neural
response to select the intended communication icon (e.g.,
Donchin, Spencer, & Wijesinghe, 2000), (b) take part in
an operant conditioning paradigm using biofeedback of
EEG (e.g., Kübler et al., 1999), (c) listen to auditory stimuli
to generate auditory–sensory neural responses related to the
intended communication output (e.g., Halder et al., 2010),
or (d) imagine movements of the limbs to alter the sensorimotor rhythm (SMR) to select communication items (e.g.,
Pfurtscheller & Neuper, 2001).
At present, indirect BCIs are more mature as a technology, and many have already begun user trials (Holz,
Botrel, Kaufmann, & Kübler, 2015; Sellers, Vaughan, &
Wolpaw, 2010). Therefore, SLPs are most likely to be involved with indirect BCIs first as they move from the research lab to the real world. Indirect BCI techniques are very
similar to current access technologies for high-tech AAC;
for example, the output of the BCI system can act as an
input method for conventional AAC devices. Below, we
review indirect BCI techniques and highlight their possible
future in AAC.
The P300-Based BCI
The visual P300 grid speller (Donchin et al., 2000) is
the most well-known and most mature technology with ongoing at-home user trials (Holz et al., 2015; Sellers et al.,
2010). Visual P300 BCIs for communication use the P300
event-related potential, a neural response to novel, rare
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018
visual stimuli in the presence of many other visual stimuli,
to select items on a communication interface. The traditional graphical layout for a visual P300 speller is a 6 ×
6 grid that includes the 26 letters of the alphabet, space,
backspace, and numbers (see Figure 1). Each row and
column1 on the spelling grid are highlighted in a random
order, and a systematic variation in the EEG waveform is
generated when one attends to a target item for selection,
the “oddball stimulus,” which occurs infrequently compared with the remaining items (Donchin et al., 2000). The
event-related potential in response to the target item will
contain a positive voltage fluctuation approximately 300 ms
after the item is highlighted (Farwell & Donchin, 1988).
The BCI decoding algorithm then selects items associated
with detected occurrences of the P300 for message creation
(Donchin et al., 2000). The P300 grid speller has been
operated by individuals with amyotrophic lateral sclerosis
(ALS; Nijboer, Sellers, et al., 2008; Sellers & Donchin, 2006)
and has been examined as part of at-home trials by individuals with neuromotor impairments (Holz et al., 2015;
Sellers & Donchin, 2006), making it a likely candidate for
future BCI-based access for AAC.
In addition to the cognitive requirements for operating the P300 speller, successful operation depends somewhat on the degree of oculomotor control (Brunner et al.,
2010). Past findings have shown that the P300 amplitude
can be reduced if individuals are unable to use an overt
attention strategy (gazing directly at the target) and, instead,
must use a covert strategy (attentional change without
ocular shifting), which can degrade BCI performance (Brunner
et al., 2010). An alternative P300 interface displays a single
item at a time on the screen (typically to the center as in
Figure 1, second from left) to alleviate concerns for individuals with poor oculomotor control. This interface, known
as the rapid serial visual presentation speller, has been successfully controlled by a cohort of individuals across the
continuum of locked-in syndrome severity (Oken et al., 2014).
All BCIs that use spelling interfaces require sufficient
levels of literacy, though many can be adapted to use icon
or symbol-based communication (e.g., Figure 2).
Auditory stimuli can also be used to elicit P300 responses for interaction with BCI devices for individuals
with poor visual capability (McCane et al., 2014), such as
severe visual impairment, impaired oculomotor control,
and cortical blindness. Auditory interfaces can also be used
in poor viewing environments, such as outdoors or in the
presence of excessive lighting glare. Like its visual counterpart, the auditory P300 is elicited via an oddball paradigm,
and has been typically limited to binary (yes/no) selection
by attending to one of two different auditory tones presented monaurally to each ear (Halder et al., 2010), or linguistic stimuli (e.g., attending to a “yep” target among
“yes” presentations in the right ear vs. “nope” and “no” in
the left; Hill et al., 2014). The binary control achieved
1
Each individual item may also be highlighted, rather than rows and
columns.
using the auditory P300 interface has the potential to be
used to navigate a spelling grid similar to conventional auditory scanning techniques for accessing AAC systems, by
attending to specific tones that correspond to rows and columns (Käthner et al., 2013; Kübler et al., 2009). There is
evidence that auditory grid systems may require greater attention than their visual analogues (Klobassa et al., 2009;
Kübler et al., 2009), which should be considered when
matching clients to the most appropriate communication
device.
Steady State Evoked Potentials
BCIs can be controlled using attention-modulated
steady state brain rhythms, as opposed to event-related
potentials, in both visual (steady state visually evoked potential [SSVEP]) and auditory (auditory steady state response
[ASSR]) domains. Both the SSVEP and ASSR are physiological responses to a driving input stimulus that are amplified
when an individual focuses his or her attention on the stimulus (Regan, 1989). Strobe stimuli are commonly used for
SSVEP, whereas amplitude-modulated tones are often
used for ASSR (Regan, 1989).
BCIs using SSVEP exploit the attention-modulated
response to strobe stimuli by simultaneously presenting
multiple communication items for selection, each flickering
at a different frequency (Cheng, Gao, Gao, & Xu, 2002;
Friman, Luth, Volosyak, & Graser, 2007; Müller-Putz,
Scherer, Brauneis, & Pfurtscheller, 2005).2 As a result, all
item flicker rates will be observed in the EEG recordings,
but the frequency of the attended stimulus will contain the
largest amplitude (Lotte, Congedo, Lécuyer, Lamarche,
& Arnaldi, 2007; Müller-Putz et al., 2005; Regan, 1989)
and greatest temporal correlation to the strobe stimulus (Chen,
Wang, Gao, Jung, & Gao, 2015; Lin, Zhang, Wu, & Gao,
2007). The stimulus with the greatest neurophysiological response will then be selected by the BCI to construct a message,
typically via an alphanumeric keyboard (shown in Figure 1),
though icons can be adapted for different uses and levels of
literacy (e.g., Figure 2). Major advantages of this type of
interface are the following: (a) high accuracy rates, often
reported above 90% with very little training (e.g., Cheng
et al., 2002; Friman et al., 2007); (b) overlapping, centrally
located stimuli could be used for individuals with impaired
oculomotor control (Allison et al., 2008). A major concern with this technique, however, is an increased risk for
seizures (Volosyak, Valbuena, Lüth, Malechka, & Gräser,
2011).
BCIs that use the ASSR require one to shift his or
her attention to a sound stream that contains a modulated
stimulus (e.g., a right monoaural 38-Hz amplitude modulation, 1000-Hz carrier tone presented with a left monoaural 42-Hz modulated, 2500-Hz carrier; Lopez, Pomares,
Pelayo, Urquiza, & Perez, 2009). As with the SSVEP, the
modulation frequency of the attended sound stream is
2
There are other variants that use a single flicker rate with a specific
strobe pattern that is beyond the scope of this tutorial.
Brumberg et al.: AAC-BCI Tutorial
3
Figure 1. From left to right, example visual displays for the following BCIs: P300 grid speller, RSVP P300, SSVEP, and motor-based (SMR
with keyboard). For the P300 grid, each row and column are highlighted until a letter is selected. In the RSVP, each letter is displayed
randomly, sequentially in the center of the screen. For the SSVEP, this example uses four flickering stimuli (at different frequencies) to
represent the cardinal directions, which are used to select individual grid items. This can also be done with individual flicker frequencies
for all 36 items with certain technical considerations. For the motor-based BCI, this is an example of a binary-selection virtual keyboard;
imagined right hand movements select the right set of letters. RSVP = rapid serial visual presentation; SSVEP = steady state visually evoked
potential; SMR = sensorimotor rhythm; BCI = brain–computer interfaces. Copyright © Tobii Dynavox. Reprinted with permission.
observable in the recorded EEG signal and will be amplified relative to the other competing stream. Therefore,
in this example, if the BCI detects the greatest EEG amplitude at 38 Hz, it will perform a binary action associated
with the right-ear tone (e.g., yes or “select”), whereas detection of the greatest EEG amplitude at 42 Hz will generate a left-ear tone action (e.g., no or “advance”).
Motor-Based BCIs
Another class of BCIs provides access to communication interfaces using changes in the SMR, a neurological
signal related to motor production and motor imagery
(Pfurtscheller & Neuper, 2001; Wolpaw et al., 2002), for
individuals with and without neuromotor impairments
(Neuper, Müller, Kübler, Birbaumer, & Pfurtscheller,
2003; Vaughan et al., 2006). The SMR is characterized
by the μ (8–12 Hz) and β (18–25 Hz) band spontaneous
EEG oscillations that are known to desynchronize, or reduce in amplitude, during covert and overt movement
attempts (Pfurtscheller & Neuper, 2001; Wolpaw et al.,
2002). Many motor-based BCIs use left and right limb
movement imagery because the SMR desynchronization
will occur on the contralateral side, and are most often
used to control spelling interfaces (e.g., virtual keyboard:
Scherer, Müller, Neuper, Graimann, & Pfurtscheller, 2004;
DASHER: Wills & MacKay, 2006; hex-o-spell: Blankertz
et al., 2006; see Figure 1, right, for an example), though
they can be used as inputs to commercial AAC devices
as well (Brumberg, Burnison, & Pitt, 2016).
Two major varieties of motor-based BCIs have been
developed for controlling computers: those that provide
continuous cursor control (analogous to mouse/joystick and
eye gaze) and others that use discrete selection (analogous
to button presses). An example layout of keyboard-based
and symbol-based motor-BCI interfaces are shown in Figures 1 and 2. Cursor-style BCIs transform changes in the
SMR continuously over time into computer control signals
(Wolpaw & McFarland, 2004). One example of a continuous, SMR-based BCI uses imagined movements of the hands
and feet to move a cursor to select progressively refined
Figure 2. From left to right, examples of how existing BCI paradigms can be applied to page sets from current AAC devices: P300 grid, SSVEP,
motor based (with icon grid). For the P300 grid interface, a row or column is highlighted until a symbol is selected (here, it is yogurt). For the
SSVEP, either directional (as shown here) or individual icons flicker at specified strobe rates to either move a cursor or directly select an item. For
motor based, the example shown here uses attempted or imagined left hand movements to advance the cursor and right hand movements to
choose the currently selected item. SSVEP = steady state visually evoked potential; SMR = sensorimotor rhythm; BCI = brain–computer
interfaces; AAC = augmentative and alternative communication. Copyright © Tobii Dynavox. Reprinted with permission.
4
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018
groups of letters organized at different locations around a
computer screen (Miner, McFarland, & Wolpaw, 1998;
Vaughan et al., 2006). Another continuous-style BCI is
used to control the “hex-o-spell” interface in which imagined movements of the right hand rotate an arrow to point
at one of six groups of letters, and imagined foot movements extend the arrow to select the current letter group
(Blankertz et al., 2006).
Discrete-style motor BCIs perform this transformation using the event-related desynchronization (Pfurtscheller
& Neuper, 2001), a change to the SMR in response to
some external stimulus, like an automatically highlighted
row or column via scanning interface. One example of a
discrete-style motor BCI uses the event-related desynchronization to control a virtual keyboard consisting of a
binary tree representation of letters, in which individuals
choose between two blocks of letters, selected by (imagined) right or left hand movements until a single letter or
item remains (Scherer et al., 2004). Most motor-based
BCIs require many weeks or months for successful operation and report accuracies greater than 75% for individuals without neuromotor impairments and, in one study,
69% accuracy for individuals with severe neuromotor
impairments (Neuper et al., 2003).
Motor-based BCIs are inherently independent from
interface feedback modality because they rely only on an
individual’s ability to imagine his or her limbs moving,
though users are often given audio or visual feedback of
BCI choices (e.g., Nijboer, Furdea, et al., 2008). A recent,
continuous motor BCI has been used to produce vowel
sounds with instantaneous auditory feedback by using limb
motor imagery to control a two-dimensional formant frequency speech synthesizer (Brumberg, Burnison, & Pitt,
2016). Other recent discrete motor BCIs have been developed for row–column scanning interfaces (Brumberg,
Burnison, & Pitt, 2016; Scherer et al., 2015).
2003). Nearly all BCIs require some amount of cognitive
effort or selective attention, though the amount of each
depends greatly on the style and modality of the specific
device. Individuals with other neuromotor disorders, such
as cerebral palsy, muscular dystrophies, multiple sclerosis,
Parkinson’s disease, and brain tumors, may require AAC
(Fried-Oken, Mooney, Peters, & Oken, 2013; Wolpaw et al.,
2002) but are not yet commonly considered for BCI studies and interventions (cf. Neuper et al., 2003; Scherer et al.,
2015), due to concomitant impairments in cognition, attention, and memory. In other instances, elevated muscle
tone and uncontrolled movements (e.g., spastic dysarthria,
dystonia) limit the utility of BCI due to the introduction of
physical and electromyographic movement artifacts (i.e.,
muscle-based signals that are much stronger than EEG
and can distort recordings of brain activity). BCI research
is now beginning to consider important human factors
involved in appropriate use of BCI for individuals (FriedOken et al., 2013) and for coping with difficulties in brain
signal acquisition due to muscular (Scherer et al., 2015)
and environmental sources of artifacts. Developing BCI
protocols to help identify the BCI technique most appropriate for each individual must be considered as BCI
development moves closer to integration with existing
AAC techniques.
Topic 2: Who May Best Benefit From a BCI?
BCI Summary
BCIs use a wide range of techniques for mapping
brain activity to communication device control through a
combination of signals related to sensory, motor, and/or
cognitive processes (see Table 1 for a summary of BCI
types). The choice of BCI protocol and feedback methods
trade off with cognitive abilities needed for successful device operation (e.g., Geronimo, Simmons, & Schiff, 2016;
Kleih & Kübler, 2015; Kübler et al., 2009). Many BCIs
require individuals to follow complex, multistep procedures
and require potentially high levels of attentional capacity
that are often a function of the sensory or motor process
used for BCI operation. For example, the P300 speller BCI
(Donchin et al., 2000) requires that individuals have an
ability to attend to visual stimuli and make decisions about
them (e.g., recognize the intended visual stimulus among
many other stimuli). BCIs that use SSVEPs depend on the
neurological response to flickering visual stimuli (Cheng
et al., 2002) that is modulated by attention rather than
other cognitive tasks. These two systems both use visual
stimuli to elicit neural activity for controlling a BCI but differ in their demands on cognitive and attention processing.
In contrast, motor-based BCI systems (e.g., Pfurtscheller
& Neuper, 2001; Wolpaw et al., 2002) require individuals
to have sufficient motivation and volition, as well as an
ability to learn how changing mental tasks can control a
communication device.
At present, BCIs are best suited for individuals with
acquired neurological and neuromotor impairments leading to paralysis and loss of speech with minimal cognitive
involvement (Wolpaw et al., 2002), for example, brainstem
stroke and traumatic brain injury (Mussa-Ivaldi & Miller,
Sensory, Motor, and Cognitive Factors
Alignment of the sensory, motor, and cognitive requirements for using BCI to access AAC devices with individuals’ unique profile will help identify and narrow down
Operant Conditioning BCIs
This interface operates by detecting a stimulusindependent change in brain activity, which is used to select
options on a communication interface. The neural signals
used for controlling the BCI are not directly related to
motor function or sensation. Rather, it uses EEG biofeedback for operant conditioning to teach individuals to voluntarily change the amplitude and polarity of the slow cortical
potential, a slow-wave (< 1 Hz) neurological rhythm that
is related to movements of a one-dimensional cursor. In BCI
applications, cursor vertical position is used to make binary
selections for communication interface control (Birbaumer
et al., 2000; Kübler et al., 1999).
Brumberg et al.: AAC-BCI Tutorial
5
Table 1. Summary of BCI varieties and their feedback modality.
EEG signal type
Event-related potentials
Sensory/Motor modality
User requirements
Visual P300 (grid)
Visual P300 (RSVP)
Auditory P300
Steady state evoked potentials
Steady state visually evoked potential
Motor-based
Auditory steady state response
Continuous sensorimotor rhythm
Discrete event-related desynchronization
Operant conditioning
Motor preparatory signals, for example,
contingent negative variation
Slow cortical potentials
Visual oddball paradigm, requires selective attention
around the screen
Visual oddball paradigm, requires selective attention
to the center of the screen only (poor oculomotor
control)
Auditory oddball paradigm, requires selective
auditory attention, no vision requirement
Attention to frequency tagged visual stimuli, may
increase seizure risk
Attention to frequency modulated audio stimuli
Continuous, smooth control of interface (e.g., cursors)
using motor imagery (first person)
Binary (or multichoice) selection of interface items
(# choices = # of imagined movements), requires
motor imagery ability
Binary selection of communication interface items
using imagined movements
Binary selection of communication interface items
after biofeedback-based learning protocol
Note. BCI = brain–computer interface; EEG = electroencephalography; RSVP = rapid serial visual presentation.
the number of candidate BCI variants (e.g., feature matching; Beukelman & Mirenda, 2013; Light & McNaughton,
2013), which is important for improving user outcomes with
the chosen device (Thistle & Wilkinson, 2015). Matching
possible BCIs should also include overt and involuntary
motor considerations, specifically the presence of spasticity
or variable muscle tone/dystonia, which may produce
electromyographic artifacts that interfere with proper BCI
function (Goncharova, McFarland, Vaughan, & Wolpaw,
2003). In addition, there may be a decline in brain signals
used for BCI decoding as symptoms of progressive neuromotor diseases become more severe (Kübler, Holz, Sellers,
& Vaughan, 2015; Silvoni et al., 2013) that may result in
decreased BCI performance. The wide range in sensory,
motor, and cognitive components of BCI designs point to
a need for user-centered design frameworks (e.g., Lynn,
Armstrong, & Martin, 2016) and feature matching/screening
protocols (e.g., Fried-Oken et al., 2013; Kübler et al.,
2015), like those used for current practices in AAC intervention (Light & McNaughton, 2013; Thistle & Wilkinson,
2015).
Topic 3: Are BCIs Faster Than Other
Access Methods for AAC?
Current AAC devices yield a range of communication
rates that depend on access modality (e.g., direct selection,
scanning), level of literacy, and information represented by
each communication item (e.g., single-meaning icons or
images, letters, icons representing complex phrases; Hill &
Romich, 2002; Roark, Fried-Oken, & Gibbons, 2015), as
well as word prediction software (Trnka, McCaw, Yarrington,
McCoy, & Pennington, 2008). Communication rates using
AAC are often less than 15 words per minute (Beukelman
& Mirenda, 2013; Foulds, 1980), and slower speeds (two to
6
five words per minute; Patel, 2011) are observed for letter
spelling due to the need for multiple selections for spelling
words (Hill & Romich, 2002). Word prediction and language modeling can increase both speed and typing efficiency (Koester & Levine, 1996; Roark et al., 2015; Trnka
et al., 2008), but the benefits may be limited due to additional cognitive demands (Koester & Levine, 1996). Scan
rate in auto-advancing row–column scanning access methods
also affects communication rate, and though faster scan
rates should lead to faster communication rates, slower
scan rates can reduce selection errors (Roark et al., 2015).
BCIs are similarly affected by scan rate (Sellers & Donchin,
2006); for example, a P300 speller can only operate as fast
as each item is flashed. Increases in flash rate may also
increase cognitive demands for locating desired grid items
while ignoring others, similar to effects observed using
commercial AAC visual displays (Thistle & Wilkinson,
2013).
Current BCIs for communication generally yield
selection rates that are slower than existing AAC methods,
even with incorporation of language prediction models
(Oken et al., 2014). Table 2 provides a summary of selection rates from recent applications of conventional access
techniques and BCI to communication interfaces. Both
individuals with and without neuromotor impairments
using motor-based BCIs have achieved selection rates under 10 selections (letters, numbers, symbols) per minute
(Blankertz et al., 2006; Neuper et al., 2003; Scherer et al.,
2004), and those using P300 methods commonly operate
below five selections per minute (Acqualagna & Blankertz,
2013; Donchin et al., 2000; Nijboer, Sellers, et al., 2008;
Oken et al., 2014). A recent P300 study using a novel presentation technique has obtained significantly higher communication rates of 19.4 characters per minute, though
the method has not been studied in detail with participants
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018
Table 2. Communication rates from recent BCI and conventional access to communication interfaces.
BCI method
Population
Selection rate
Berlin BCI (motor imagery)
Graz BCI (motor imagery)
Graz BCI (motor imagery)
P300 speller (visual)
Healthy
Healthy
Impaired
Healthy
P300 speller (visual)
RSVP P300
RSVP P300
ALS
ALS
LIS
Healthy
2.3–7.6 letters/min
2.0 letters/min
0.2–2.5 letter/min
4.3 letters/min
19.4 char/min (120.0 bits/min)
1.5–4.1 char/min (4.8–19.2 bits/min)
3–7.5 char/min
0.4–2.3 char/min
1.2–2.5 letter/min
SSVEP
Healthy
AAC (row–column)
Healthy
LIS
Healthy
AAC (direct selection)
33.3 char/min
10.6 selections/min (27.2 bits/min)
18–22 letters/min
6.0 letters/min
5.2 words/min
Source
Blankertz et al. (2006)
Scherer et al. (2004)
Neuper et al. (2003)
Donchin et al. (2000)
Townsend and Platsko (2016)
Nijboer, Sellers, et al. (2008)
Mainsah et al. (2015)
Oken et al. (2014)
Acqualagna and Blankertz (2013),
Oken et al. (2014)
Chen et al. (2015)
Friman et al. (2007)
Roark et al. (2015)
Trnka et al. (2008)
Note. BCI = brain–computer interface; ALS = amyotrophic lateral sclerosis; RSVP = rapid serial visual presentation; LIS = locked-in syndrome;
SSVEP = steady state visually evoked potential; AAC = augmentative and alternative communication; char = character.
with neuromotor impairments (Townsend & Platsko, 2016).
BCIs, on the basis of the SSVEP, have emerged as a promising technique often yielding both high accuracy (> 90%)
and communication rates as high as 33 characters per minute (Chen et al., 2015).
From these reports, BCI performance has started to
approach levels associated with AAC devices using direct
selection, and the differences in communication rates for
scanning AAC devices and BCIs (shown in Table 2) are
reduced when making comparisons between individuals
with neuromotor impairments rather than individuals without impairments (e.g., AAC: six characters per minute;
Roark et al., 2015; BCI: one to eight characters per minute; Table 2). Differences in communication rate can also
be reduced based on the type of BCI method (e.g., 3–7.5
characters per minute; Mainsah et al., 2015). These results
suggest that BCI has become another clinical option for
AAC intervention that should be considered during the
clinical decision-making process. BCIs have particular utility when considered for the most severe cases; the communication rates described in the literature are sufficient
to provide access to language and communication for those
who are currently without both. Recent improvements in
BCI designs have shown promising results (e.g., Chen et al.,
2015; Townsend & Platsko, 2016), which may start to
push BCI communication efficacy past current benchmarks
for AAC. Importantly, few BCIs have been evaluated over
extended periods of time (Holz et al., 2015; Sellers et al.,
2010); therefore, it is possible that BCI selection may improve over time with training.
Topic 4: Fatigue and Its Effects
BCIs, like conventional AAC access techniques, require
various levels of attention, working memory, and cognitive
load that all affect the amount of effort (and fatigue) needed
Table 3. Take-home points collated from the interdisciplinary research team that highlight the major considerations for BCI as possible access
methods for AAC.
BCIs do not yet have the ability to translate thoughts or speech plans into fluent speech productions. Direct BCIs, usually involving a surgery
for implantation of recording electrodes, are currently being developed as speech neural prostheses.
Noninvasive BCIs are most often designed as an indirect method for accessing AAC, whether custom developed or commercial.
There are a variety of noninvasive BCIs that can support clients with a range of sensory, motor, and cognitive abilities—and selecting the
most appropriate BCI technique requires individualized assessment and feature matching procedures.
The potential population of individuals who may use BCIs is heterogeneous, though current work is focused on individuals with acquired
neurological and neuromotor disorders (e.g., locked-in syndrome due to stroke, traumatic brain injury, and ALS); limited study has involved
individuals with congenital disorders such as CP.
BCIs are currently not as efficient as existing AAC access methods for individuals with some form of movement, though the technology is
progressing. For these individuals, BCIs provide an opportunity to augment or complement existing approaches. For individuals with
progressive neurodegenerative diseases, learning to use BCI before speech and motor function worsen beyond the aid of existing access
technologies may help maintain continuity of communication. For those who are unable to use current access methods, BCIs may provide
the only form of access to communication.
Long-term BCI use is only just beginning; BCI performance may improve as the technology matures and as individuals who use BCI gain
greater proficiency and familiarity with the device.
Note. BCI = brain–computer interface; AAC = augmentative and alternative communication; ALS = amyotrophic lateral sclerosis; CP =
cerebral palsy.
Brumberg et al.: AAC-BCI Tutorial
7
to operate the device (Kaethner, Kübler, & Halder, 2015;
Pasqualotto et al., 2015). There is evidence that scanningtype AAC devices are not overly tiring (Gibbons & Beneteau,
2010; Roark, Beckley, Gibbons, & Fried-Oken, 2013), but
prolonged AAC use can have a cumulative effect and reduce
communication effectiveness (Trnka et al., 2008). In these
cases, language modeling and word prediction can reduce
fatigue and maintain high communication performance
using an AAC device (Trnka et al., 2008). Within BCI, reports
of fatigue, effort, and cognitive load are mixed. Individuals with ALS have reported that visual P300 BCIs required more effort and time compared with eye gaze access
(Pasqualotto et al., 2015), whereas others reported that a visual
P300 speller was easier to use, and not overly exhausting compared with eye gaze, because it does not require precise eye
movements (Holz et al., 2015; Kaethner et al., 2015). Other
findings from these studies indicate that the visual P300
speller incurred increased cognitive load and fatigue for
some (Kaethner et al., 2015), whereas for others, there is
less strain compared to eye-tracking systems (Holz et al., 2015).
The application of many conventional and BCI-based
AAC access techniques with the same individual may permit an adaptive strategy to rely on certain modes of access
based on each individual’s level of fatigue. This will allow
one to change his or her method of AAC access to suit his
or her fatigue level throughout the day.
Topic 5: BCI as an Addition to Conventional AAC
Access Technology
At their current stage of development, BCIs are
mainly the primary choice for individuals with either absent, severely impaired, or highly unreliable speech and
motor control. As BCIs advance as an access modality for
AAC, it is important that the goal of intervention remains
on selecting an AAC method that is most appropriate
versus selecting the most technologically advanced access
method (Light & McNaughton, 2013). Each of the BCI
devices discussed has unique sensory, motor, and cognitive requirements that may best match specific profiles of
individuals who may require BCI, as well as the training
required for device proficiency. The question then of BCIs
replacing any form of AAC must be determined according
to the needs, wants, and abilities of the individual. These
factors play a crucial role on motivation, which has direct
impact on BCI effectiveness (Nijboer, Birbaumer, & Kübler,
2010). Other assessment considerations include comorbid
conditions, such as a history of seizures, which is a contraindication for some visual BCIs due to the rapidly flashing
icons (Volosyak et al., 2011). Cognitive factors, such as differing levels of working memory (Sprague, McBee, & Sellers,
2015) and an ability to focus one’s attention (Geronimo
et al., 2016; Riccio et al., 2013), are also important considerations because they have been correlated to successful
BCI operation.
There are additional considerations for motor-based
BCIs, including (a) a well-known observation that the SMR,
which is necessary for device control, cannot be adequately
8
estimated in approximately 15%–30% of all individuals with
or without impairment (Vidaurre & Blankertz, 2010) and
(b) the possibility of performance decline or instability as a
result of progressive neuromotor disorders, such as ALS
(Silvoni et al., 2013). These concerns are currently being
addressed using assessment techniques to predict motor-based
BCI performance, including a questionnaire to estimate
kinesthetic motor imagery (e.g., first person imagery or imagining performing and experiencing the sensations associated
with motor imagery) performance (Vuckovic & Osuagwu,
2013), which is known to lead to better BCI performance
compared with a third person motor imagery (e.g., watching
yourself from across the room; Neuper, Scherer, Reiner, &
Pfurtscheller, 2005). Overall, there is limited research available on the inter- and intraindividual considerations for BCI
intervention that may affect BCI performance (Kleih &
Kübler, 2015); therefore, clinical assessment tools and guidelines must be developed to help determine the most appropriate method of accessing AAC (that includes both
traditional or BCI-based technologies) for each individual.
These efforts have already begun (e.g., Fried-Oken et al.,
2013; Kübler et al., 2015), and more work is needed to ensure that existing AAC practices are well incorporated with
BCI-based assessment tools.
In summary, the ultimate purpose of BCI access
techniques should not be seen as a competition or a replacement for existing AAC methods that have a history
of success. Rather, the purpose of BCI-based communication is to provide a feature-matched alternate or complementary method for accessing AAC for individuals with
suitability, preference, and motivation for BCI or for
those who are unable to utilize current communicative
methods.
Topic 6: Limitations of BCI and Future Directions
Future applications of noninvasive BCIs will continue to focus on increasing accuracy and communication
rate for use either as standalone AAC options or to access
existing AAC devices. One major area of future work is
to improve the techniques for noninvasively recording brain
activity needed for BCI operation. Though a large majority
of people who may potentially use BCI have reported that
they are willing to wear an EEG cap (84%; Huggins, Wren,
& Gruis, 2011), the application of EEG sensors and their
stability over time are still obstacles needed to be overcome
for practical use. Most EEG-based BCI systems require the
application of electrolytic gel to bridge the contact between
electrodes and the scalp for good signal acquisition. Unfortunately, this type of application has been reported to be
inconvenient and cumbersome by individuals who currently
use BCI and may also be difficult to set up and maintain
by a trained facilitator (Blain-Moraes, Schaff, Gruis, Huggins,
& Wren, 2012). Further, electrolytic gels dry out over time,
gradually degrading EEG signal acquisition. Recent advances
in dry electrode technology may help overcome this limitation (Blain-Moraes et al., 2012) by allowing for recording of
EEG without electrolytic solutions and may lead to easier
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018
application of EEG sensors and prolonged stability of EEG
signal acquisition.
In order to be used in all environments, EEG must
be portable and robust to external sources of noise and
artifacts. EEG is highly susceptible to electrical artifacts
from the muscles, environment, and other medical equipment (e.g., mechanical ventilation). Therefore, an assessment is needed for likely environments of use, as are
guidelines for minimizing the effect of these artifacts.
Simultaneous efforts should be made toward improving
the tolerance of EEG recording equipment to these outsize sources of electrical noise (Kübler et al., 2015).
The ultimate potential of BCI technology is the development of a system that can directly decode brain activity
into communication (e.g., written text or spoken), rather
than indirectly operate a communication device. This type
of neural decoding is primarily under investigation using
invasive methods using electrocorticography and intracortical microelectrodes and has focused on decoding phonemes (Blakely et al., 2008; Brumberg et al., 2011; Herff
et al., 2015; Mugler et al., 2014; Tankus et al., 2012), words
(Kellis et al., 2010; Leuthardt et al., 2011; Pei et al., 2011),
and time-frequency representations (Martin et al., 2014).
Invasive methods have the advantage of increased signal
quality and resistance to sources of external noise but require
a surgical intervention to implant recording electrodes
either in or on the brain (Chakrabarti et al., 2015). The goal
of these decoding studies and other invasive electrophysiological investigations of speech processing is to develop a
neural prosthesis for fluent-like speech production (Brumberg,
Burnison, & Guenther, 2016). Although invasive techniques
come at a surgical cost, one study reported that 72% of individuals with ALS indicated they were willing to undergo
outpatient surgery, and 41% were willing to have a surgical
intervention with a short hospital stay to access invasive
BCI methods (Huggins et al., 2011). That said, very few
invasive BCIs are available for clinical research or long-term
at-home use (e.g., Vansteensel et al., 2016); therefore, noninvasive methods will likely be first adopted for use in AAC
interventions.
Conclusions
This tutorial has focused on a few important considerations for the future of BCIs as AAC: (a) Despite broad
speech-language pathology expertise in AAC, there are
few clinical guidelines and recommendations for the use
of BCI as an AAC access technique; (b) the most mature
BCI technologies have been designed as methods to access
communication interfaces rather than directly accessing
thoughts, utterances, and speech motor plans from the brain;
and (c) BCI is an umbrella term for a variety of brain-tocomputer techniques that require comprehensive assessment
for matching people who may potentially use BCI with the
most appropriate device. The purpose of this tutorial was
to bridge the gaps in knowledge between AAC and BCI
practices, describe BCIs in the context of current AAC conventions, and motivate interdisciplinary collaborations to
pursue rigorous clinical research to adapt AAC feature
matching protocols to include intervention with BCIs.
A summary of take-home messages to help bridge the gap
between knowledge of AAC and BCI was compiled from
our interdisciplinary team and summarized in Table 3.
Additional training and hands-on experience will improve
acceptance of BCI approaches for interventionists targeted by
this tutorial, as well as people who may use BCI in the future.
Key to the clinical acceptance of BCI are necessary
improvements in communication rate and accuracy via
BCI access methods (Kageyama et al., 2014). However,
many people who may use BCIs understand the current
limitations, yet they recognize the potential positive benefits of BCI, reporting that the technology offers “freedom,”
“hope,” “connection,” and unlocking from their speech
and motor impairments (Blain-Moraes et al., 2012). A significant component of future BCI research will focus on
meeting the priorities of people who use BCIs. A recent
study assessed the opinions and priorities of individuals
with ALS in regard to BCI design and reported that individuals with ALS prioritized performance accuracy of at
least 90% and a rate of at least 15 to 19 letters per minute
(Huggins et al., 2011). From our review, most BCI technologies have not yet reached these specifications, though
some recent efforts have made considerable progress (e.g.,
Chen et al., 2015; Townsend & Platsko, 2016). A renewed
emphasis on user-centered design and development is
helping to move this technology forward by best matching the wants and needs of individuals who may use BCI
with realistic expectations of BCI function. It is imperative
to include clinicians, individuals who use AAC and BCI,
and other stakeholders into the BCI design process to
improve usability and performance and to help find the
optimal translation from the laboratory to the real world.
Acknowledgments
This work was supported in part by the National Institutes
of Health (National Institute on Deafness and Other Communication Disorders R03-DC011304), the University of Kansas
New Faculty Research Fund, and the American Speech-LanguageHearing Foundation New Century Scholars Research Grant, all
awarded to J. Brumberg.
References
Acqualagna, L., & Blankertz, B. (2013). Gaze-independent BCIspelling using rapid serial visual presentation (RSVP). Clinical
Neurophysiology, 124(5), 901–908.
Allison, B. Z., McFarland, D. J., Schalk, G., Zheng, S. D., Jackson,
M. M., & Wolpaw, J. R. (2008). Towards an independent brain–
computer interface using steady state visual evoked potentials.
Clinical Neurophysiology, 119(2), 399–408.
Beukelman, D., & Mirenda, P. (2013). Augmentative and alternative communication: Supporting children and adults with complex communication needs (4th ed.). Baltimore, MD: Brookes.
Birbaumer, N., Kübler, A., Ghanayim, N., Hinterberger, T.,
Perelmouter, J., Kaiser, J., . . . Flor, H. (2000). The thought
translation device (TTD) for completely paralyzed patients. IEEE
Transactions on Rehabilitation Engineering, 8(2), 190–193.
Brumberg et al.: AAC-BCI Tutorial
9
Blain-Moraes, S., Schaff, R., Gruis, K. L., Huggins, J. E., & Wren,
P. A. (2012). Barriers to and mediators of brain–computer
interface user acceptance: Focus group findings. Ergonomics,
55(5), 516–525.
Blakely, T., Miller, K. J., Rao, R. P. N., Holmes, M. D., & Ojemann,
J. G. (2008). Localization and classification of phonemes using
high spatial resolution electrocorticography (ECoG) grids. In
2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (pp. 4964–4967).
https://doi.org/10.1109/IEMBS.2008.4650328
Blankertz, B., Dornhege, G., Krauledat, M., Müller, K.-R., Kunzmann,
V., Losch, F., & Curio, G. (2006). The Berlin brain–computer
interface: EEG-based communication without subject training.
IEEE Transactions on Neural Systems and Rehabilitation Engineering, 14(2), 147–152.
Brumberg, J. S., Burnison, J. D., & Guenther, F. H. (2016). Brainmachine interfaces for speech restoration. In P. Van Lieshout,
B. Maassen, & H. Terband (Eds.), Speech motor control in
normal and disordered speech: Future developments in theory
and methodology (pp. 275–304). Rockville, MD: ASHA Press.
Brumberg, J. S., Burnison, J. D., & Pitt, K. M. (2016). Using
motor imagery to control brain–computer interfaces for communication. In Foundations of Augmented Cognition: Neuroergonomics and Operational Neuroscience. Cham, Switzerland:
Spring International Publishing.
Brumberg, J. S., & Guenther, F. H. (2010). Development of speech
prostheses: Current status and recent advances. Expert Review
of Medical Devices, 7(5), 667–679.
Brumberg, J. S., Wright, E. J., Andreasen, D. S., Guenther, F. H.,
& Kennedy, P. R. (2011). Classification of intended phoneme production from chronic intracortical microelectrode
recordings in speech-motor cortex. Frontiers in Neuroscience,
5, 65.
Brunner, P., Joshi, S., Briskin, S., Wolpaw, J. R., Bischof, H., &
Schalk, G. (2010). Does the “P300” speller depend on eye gaze?
Journal of Neural Engineering, 7(5), 056013.
Chakrabarti, S., Sandberg, H. M., Brumberg, J. S., & Krusienski,
D. J. (2015). Progress in speech decoding from the electrocorticogram. Biomedical Engineering Letters, 5(1), 10–21.
Chen, X., Wang, Y., Gao, S., Jung, T.-P., & Gao, X. (2015). Filter
bank canonical correlation analysis for implementing a highspeed SSVEP-based brain–computer interface. Journal of Neural
Engineering, 12(4), 46008.
Cheng, M., Gao, X., Gao, S., & Xu, D. (2002). Design and implementation of a brain–computer interface with high transfer
rates. IEEE Transactions on Biomedical Engineering, 49(10),
1181–1186.
Donchin, E., Spencer, K. M., & Wijesinghe, R. (2000). The mental
prosthesis: Assessing the speed of a P300-based brain–computer
interface. IEEE Transactions on Rehabilitation Engineering,
8(2), 174–179.
Fager, S., Beukelman, D., Fried-Oken, M., Jakobs, T., & Baker, J.
(2012). Access interface strategies. Assistive Technology, 24(1),
25–33.
Farwell, L., & Donchin, E. (1988). Talking off the top of your
head: Toward a mental prosthesis utilizing event-related brain
potentials. Electroencephalography and Clinical Neurophysiology, 70(6), 510–523.
Foulds, R. (1980). Communication rates of nonspeech expression as a function in manual tasks and linguistic constraints.
In International Conference on Rehabilitation Engineering
(pp. 83–87).
Fried-Oken, M., Mooney, A., Peters, B., & Oken, B. (2013). A clinical screening protocol for the RSVP keyboard brain–computer
10
interface. Disability and Rehabilitation. Assistive Technology,
10(1), 11–18.
Friman, O., Luth, T., Volosyak, I., & Graser, A. (2007). Spelling
with steady-state visual evoked potentials. In 2007 3rd International
IEEE/EMBS Conference on Neural Engineering (pp. 354–357).
Kohala Coast, HI: IEEE.
Geronimo, A., Simmons, Z., & Schiff, S. J. (2016). Performance
predictors of brain–computer interfaces in patients with
amyotrophic lateral sclerosis. Journal of Neural Engineering,
13(2), 026002.
Gibbons, C., & Beneteau, E. (2010). Functional performance using
eye control and single switch scanning by people with ALS.
Perspectives on Augmentative and Alternative Communication,
19(3), 64–69.
Goncharova, I. I., McFarland, D. J., Vaughan, T. M., & Wolpaw,
J. R. (2003). EMG contamination of EEG: Spectral and
topographical characteristics. Clinical Neurophysiology, 114(9),
1580–1593.
Halder, S., Rea, M., Andreoni, R., Nijboer, F., Hammer, E. M.,
Kleih, S. C., . . . Kübler, A. (2010). An auditory oddball brain–
computer interface for binary choices. Clinical Neurophysiology, 121(4), 516–523.
Herff, C., Heger, D., de Pesters, A., Telaar, D., Brunner, P., Schalk,
G., & Schultz, T. (2015). Brain-to-text: Decoding spoken phrases
from phone representations in the brain. Frontiers in Neuroscience, 9, 217.
Hill, K., & Romich, B. (2002). A rate index for augmentative and
alternative communication. International Journal of Speech
Technology, 5(1), 57–64.
Hill, N. J., Ricci, E., Haider, S., McCane, L. M., Heckman, S.,
Wolpaw, J. R., & Vaughan, T. M. (2014). A practical, intuitive
brain–computer interface for communicating “yes” or “no”
by listening. Journal of Neural Engineering, 11(3), 035003.
Holz, E. M., Botrel, L., Kaufmann, T., & Kübler, A. (2015).
Long-term independent brain–computer interface home use
improves quality of life of a patient in the locked-in state: A
case study. Archives of Physical Medicine and Rehabilitation,
96(3), S16–S26.
Huggins, J. E., Wren, P. A., & Gruis, K. L. (2011). What would
brain–computer interface users want? Opinions and priorities
of potential users with amyotrophic lateral sclerosis. Amyotrophic Lateral Sclerosis, 12(5), 318–324.
Kaethner, I., Kübler, A., & Halder, S. (2015). Comparison of eye
tracking, electrooculography and an auditory brain–computer
interface for binary communication: A case study with a participant in the locked-in state. Journal of Neuroengineering and
Rehabilitation, 12(1), 76.
Kageyama, Y., Hirata, M., Yanagisawa, T., Shimokawa, T.,
Sawada, J., Morris, S., . . . Yoshimine, T. (2014). Severely affected ALS patients have broad and high expectations for
brain-machine interfaces. Amyotrophic Lateral Sclerosis and
Frontotemporal Degeneration, 15(7–8), 513–519.
Käthner, I., Ruf, C. A., Pasqualotto, E., Braun, C., Birbaumer, N.,
& Halder, S. (2013). A portable auditory P300 brain–computer
interface with directional cues. Clinical Neurophysiology, 124(2),
327–338.
Kellis, S., Miller, K., Thomson, K., Brown, R., House, P., &
Greger, B. (2010). Decoding spoken words using local field
potentials recorded from the cortical surface. Journal of
Neural Engineering, 7(5), 056007.
Kleih, S. C., & Kübler, A. (2015). Psychological factors influencing brain–computer interface (BCI) performance. 2015 IEEE
International Conference on Systems, Man, and Cybernetics,
3192–3196.
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018
Klobassa, D. S., Vaughan, T. M., Brunner, P., Schwartz, N. E.,
Wolpaw, J. R., Neuper, C., & Sellers, E. W. (2009). Toward a
high-throughput auditory P300-based brain–computer interface. Clinical Neurophysiology, 120(7), 1252–1261.
Koester, H. H., & Levine, S. (1996). Effect of a word prediction
feature on user performance. Augmentative and Alternative
Communication, 12(3), 155–168.
Kübler, A., Furdea, A., Halder, S., Hammer, E. M., Nijboer, F., &
Kotchoubey, B. (2009). A brain–computer interface controlled
auditory event-related potential (p300) spelling system for
locked-in patients. Annals of the New York Academy of Sciences, 1157, 90–100.
Kübler, A., Holz, E. M., Sellers, E. W., & Vaughan, T. M. (2015).
Toward independent home use of brain–computer interfaces:
A decision algorithm for selection of potential end-users.
Archives of Physical Medicine and Rehabilitation, 96(3),
S27–S32.
Kübler, A., Kotchoubey, B., Hinterberger, T., Ghanayim, N.,
Perelmouter, J., Schauer, M., . . . Birbaumer, N. (1999). The
thought translation device: A neurophysiological approach
to communication in total motor paralysis. Experimental
Brain Research, 124(2), 223–232.
Leuthardt, E. C., Gaona, C., Sharma, M., Szrama, N., Roland, J.,
Freudenberg, Z., . . . Schalk, G. (2011). Using the electrocorticographic speech network to control a brain–computer interface in humans. Journal of Neural Engineering, 8(3), 036004.
Light, J., & McNaughton, D. (2013). Putting people first: Re-thinking
the role of technology in augmentative and alternative communication intervention. Augmentative and Alternative Communication,
29(4), 299–309.
Lin, Z., Zhang, C., Wu, W., & Gao, X. (2007). Frequency recognition based on canonical correlation analysis for SSVEPBased BCIs. IEEE Transactions on Biomedical Engineering,
54(6), 1172–1176.
Lopez, M.-A., Pomares, H., Pelayo, F., Urquiza, J., & Perez, J.
(2009). Evidences of cognitive effects over auditory steadystate responses by means of artificial neural networks and its
use in brain–computer interfaces. Neurocomputing, 72(16–18),
3617–3623.
Lotte, F., Congedo, M., Lécuyer, A., Lamarche, F., & Arnaldi, B.
(2007). A review of classification algorithms for EEG-based
brain–computer interfaces. Journal of Neural Engineering,
4(2), R1–R13.
Lynn, J. M. D., Armstrong, E., & Martin, S. (2016). User centered design and validation during the development of domestic
brain computer interface applications for people with acquired
brain injury and therapists: A multi-stakeholder approach.
Journal of Assistive Technologies, 10(2), 67–78.
Mainsah, B. O., Collins, L. M., Colwell, K. A., Sellers, E. W.,
Ryan, D. B., Caves, K., & Throckmorton, C. S. (2015). Increasing BCI communication rates with dynamic stopping towards
more practical use: An ALS study. Journal of Neural Engineering, 12(1), 016013.
Martin, S., Brunner, P., Holdgraf, C., Heinze, H.-J., Crone, N. E.,
Rieger, J., . . . Pasley, B. N. (2014). Decoding spectrotemporal
features of overt and covert speech from the human cortex.
Frontiers in Neuroengineering, 7, 14.
McCane, L. M., Sellers, E. W., McFarland, D. J., Mak, J. N.,
Carmack, C. S., Zeitlin, D., . . . Vaughan, T. M. (2014). Brain–
computer interface (BCI) evaluation in people with amyotrophic
lateral sclerosis. Amyotrophic Lateral Sclerosis and Frontotemporal Degeneration, 15(3–4), 207–215.
Miner, L. A., McFarland, D. J., & Wolpaw, J. R. (1998). Answering
questions with an electroencephalogram-based brain–computer
interface. Archives of Physical Medicine and Rehabilitation, 79(9),
1029–1033.
Mugler, E. M., Patton, J. L., Flint, R. D., Wright, Z. A., Schuele,
S. U., Rosenow, J., . . . Slutzky, M. W. (2014). Direct classification of all American English phonemes using signals from
functional speech motor cortex. Journal of Neural Engineering,
11(3), 035015.
Müller-Putz, G. R., Scherer, R., Brauneis, C., & Pfurtscheller, G.
(2005). Steady-state visual evoked potential (SSVEP)-based
communication: Impact of harmonic frequency components.
Journal of Neural Engineering, 2(4), 123–130.
Mussa-Ivaldi, F. A., & Miller, L. E. (2003). Brain-machine interfaces: Computational demands and clinical needs meet basic
neuroscience. Trends in Neurosciences, 26(6), 329–334.
Neuper, C., Müller, G. R., Kübler, A., Birbaumer, N., & Pfurtscheller,
G. (2003). Clinical application of an EEG-based brain–computer
interface: A case study in a patient with severe motor impairment. Clinical Neurophysiology, 114(3), 399–409.
Neuper, C., Scherer, R., Reiner, M., & Pfurtscheller, G. (2005).
Imagery of motor actions: Differential effects of kinesthetic
and visual-motor mode of imagery in single-trial EEG. Cognitive Brain Research, 25(3), 668–677.
Nijboer, F., Birbaumer, N., & Kübler, A. (2010). The influence of
psychological state and motivation on brain–computer interface performance in patients with amyotrophic lateral sclerosis
—A longitudinal study. Frontiers in Neuroscience, 4, 1–13.
Nijboer, F., Furdea, A., Gunst, I., Mellinger, J., McFarland, D. J.,
Birbaumer, N., & Kübler, A. (2008). An auditory brain–computer
interface (BCI). Journal of Neuroscience Methods, 167(1),
43–50.
Nijboer, F., Sellers, E. W., Mellinger, J., Jordan, M. A., Matuz, T.,
Furdea, A., . . . Kübler, A. (2008). A P300-based brain–computer
interface for people with amyotrophic lateral sclerosis. Clinical Neurophysiology, 119(8), 1909–1916.
Oken, B. S., Orhan, U., Roark, B., Erdogmus, D., Fowler, A.,
Mooney, A., . . . Fried-Oken, M. B. (2014). Brain–computer
interface with language model-electroencephalography fusion
for locked-in syndrome. Neurorehabilitation and Neural Repair,
28(4), 387–394.
Oostenveld, R., & Praamstra, P. (2001). The five percent electrode
system for high-resolution EEG and ERP measurements.
Clinical Neurophysiology, 112(4), 713–719.
Pasqualotto, E., Matuz, T., Federici, S., Ruf, C. A., Bartl, M.,
Olivetti Belardinelli, M., . . . Halder, S. (2015). Usability and workload of access technology for people with severe motor impairment: A comparison of brain–computer interfacing and eye
tracking. Neurorehabilitation and Neural Repair, 29(10), 950–957.
Patel, R. (2011). Message formulation, organization, and navigation schemes for icon-based communication aids. In Proceedings of the 33rd Annual International Conference of the IEEE
Engineering in Medicine and Biology Society (EMBC ’11)
(pp. 5364–5367). Boston, MA: IEEE.
Pei, X., Barbour, D. L., Leuthardt, E. C., & Schalk, G. (2011).
Decoding vowels and consonants in spoken and imagined
words using electrocorticographic signals in humans. Journal
of Neural Engineering, 8(4), 046028.
Pfurtscheller, G., & Neuper, C. (2001). Motor imagery and direct
brain–computer communication. Proceedings of the IEEE,
89(7), 1123–1134.
Plum, F., & Posner, J. B. (1972). The diagnosis of stupor and
coma. Contemporary Neurology Series, 10, 1–286.
Regan, D. (1989). Human brain electrophysiology: Evoked potentials and evoked magnetic fields in science and medicine.
New York, NY: Elsevier.
Brumberg et al.: AAC-BCI Tutorial
11
Riccio, A., Simione, L., Schettini, F., Pizzimenti, A., Inghilleri, M.,
Belardinelli, M. O., . . . Cincotti, F. (2013). Attention and P300based BCI performance in people with amyotrophic lateral
sclerosis. Frontiers in Human Neuroscience, 7, 732.
Roark, B., Beckley, R., Gibbons, C., & Fried-Oken, M. (2013).
Huffman scanning: Using language models within fixed-grid
keyboard emulation. Computer Speech and Language, 27(6),
1212–1234.
Roark, B., Fried-Oken, M., & Gibbons, C. (2015). Huffman and
linear scanning methods with statistical language models.
Augmentative and Alternative Communication, 31(1), 37–50.
Scherer, R., Billinger, M., Wagner, J., Schwarz, A., Hettich, D. T.,
Bolinger, E., . . . Müller-Putz, G. (2015). Thought-based row–
column scanning communication board for individuals with
cerebral palsy. Annals of Physical and Rehabilitation Medicine,
58(1), 14–22.
Scherer, R., Müller, G. R., Neuper, C., Graimann, B., & Pfurtscheller, G.
(2004). An asynchronously controlled EEG-based virtual keyboard: Improvement of the spelling rate. IEEE Transactions
on Biomedical Engineering, 51(6), 979–984.
Sellers, E. W., & Donchin, E. (2006). A P300-based brain–computer
interface: Initial tests by ALS patients. Clinical Neurophysiology,
117(3), 538–548.
Sellers, E. W., Vaughan, T. M., & Wolpaw, J. R. (2010). A brain–
computer interface for long-term independent home use. Amyotrophic Lateral Sclerosis, 11(5), 449–455.
Silvoni, S., Cavinato, M., Volpato, C., Ruf, C. A., Birbaumer, N.,
& Piccione, F. (2013). Amyotrophic lateral sclerosis progression and stability of brain–computer interface communication.
Amyotrophic Lateral Sclerosis and Frontotemporal Degeneration, 14(5–6), 390–396.
Sprague, S. A., McBee, M. T., & Sellers, E. W. (2015). The effects
of working memory on brain–computer interface performance.
Clinical Neurophysiology, 127(2), 1331–1341.
Tankus, A., Fried, I., & Shoham, S. (2012). Structured neuronal
encoding and decoding of human speech features. Nature
Communications, 3, 1015.
Thistle, J. J., & Wilkinson, K. M. (2013). Working memory demands of aided augmentative and alternative communication
for individuals with developmental disabilities. Augmentative
and Alternative Communication, 29(3), 235–245.
Thistle, J. J., & Wilkinson, K. M. (2015). Building evidence-based
practice in AAC display design for young children: Current
12
practices and future directions. Augmentative and Alternative
Communication, 31(2), 124–136.
Townsend, G., & Platsko, V. (2016). Pushing the P300-based
brain–computer interface beyond 100 bpm: Extending performance guided constraints into the temporal domain. Journal of
Neural Engineering, 13(2), 026024.
Trnka, K., McCaw, J., Yarrington, D., McCoy, K. F., & Pennington, C.
(2008). Word prediction and communication rate in AAC,
In Proceedings of the IASTED International Conference on
Telehealth/Assistive Technologies (Telehealth/AT ’08) (pp. 19–24).
Baltimore, MD: ACTA Press Anaheim, CA, USA.
Vansteensel, M. J., Pels, E. G. M., Bleichner, M. G., Branco, M. P.,
Denison, T., Freudenburg, Z. V., . . . Ramsey, N. F. (2016). Fully
implanted brain–computer interface in a locked-in patient with
ALS. New England Journal of Medicine, 375(21), 2060–2066.
Vaughan, T. M., McFarland, D. J., Schalk, G., Sarnacki, W. A.,
Krusienski, D. J., Sellers, E. W., & Wolpaw, J. R. (2006). The
Wadsworth BCI Research and Development Program: At
home with BCI. IEEE Transactions on Neural Systems and
Rehabilitation Engineering, 14(2), 229–233.
Vidaurre, C., & Blankertz, B. (2010). Towards a cure for BCI
illiteracy. Brain Topography, 23(2), 194–198.
Volosyak, I., Valbuena, D., Lüth, T., Malechka, T., & Gräser, A.
(2011). BCI demographics II: How many (and what kinds of )
people can use a high-frequency SSVEP BCI? IEEE Transactions on Neural Systems and Rehabilitation Engineering, 19(3),
232–239.
Vuckovic, A., & Osuagwu, B. A. (2013). Using a motor imagery
questionnaire to estimate the performance of a brain–computer
interface based on object oriented motor imagery. Clinical
Neurophysiology, 124(8), 1586–1595.
Wills, S. A., & MacKay, D. J. C. (2006). DASHER—An efficient
writing system for brain–computer interfaces? IEEE Transactions on Neural Systems and Rehabilitation Engineering, 14(2),
244–246.
Wolpaw, J. R., Birbaumer, N., McFarland, D. J., Pfurtscheller, G.,
& Vaughan, T. M. (2002). Brain–computer interfaces for communication and control. Clinical Neurophysiology, 113(6),
767–791.
Wolpaw, J. R., & McFarland, D. J. (2004). Control of a twodimensional movement signal by a noninvasive brain–computer
interface in humans. Proceedings of the National Academy of
Sciences of the United States of America, 101(51), 17849–17854.
American Journal of Speech-Language Pathology • Vol. 27 • 1–12 • February 2018
Copyright of American Journal of Speech-Language Pathology is the property of American
Speech-Language-Hearing Association and its content may not be copied or emailed to
multiple sites or posted to a listserv without the copyright holder's express written permission.
However, users may print, download, or email articles for individual use.
Augmentative and Alternative Communication, 2015; 31(2): 124–136
© 2015 International Society for Augmentative and Alternative Communication
ISSN 0743-4618 print/ISSN 1477-3848 online
DOI: 10.3109/07434618.2015.1035798
RESEARCH ARTICLE
Building Evidence-based Practice in AAC Display Design for Young
Children: Current Practices and Future Directions
JENNIFER J. THISTLE & KRISTA M. WILKINSON
Department of Communication Sciences and Disorders, The Pennsylvania State University, University Park, PA, USA
Abstract
Each time a practitioner creates or modifies an augmentative and alternative communication (AAC) display for a client, that practitioner must make a series of decisions about which vocabulary concepts to include, as well as physical and organizational features
of the display. Yet, little is known about what factors influence the actual decisions and their outcomes. This research examined
the design factors identified as priorities by speech-language pathologists (SLPs) when creating AAC displays for young children
(age 10 years and under), and their rationale for the selection of these priorities. An online survey gathered ratings and comments
from 112 SLPs with experience in AAC concerning the importance of a variety of factors related to designing an aided AAC
display. Results indicated that some decisions were supported by existing research evidence, such as choosing vocabulary, collaborating with key stakeholders, and supporting partner modeling. Other decisions highlight areas for future research, including use
of visual scene display layouts, symbol background color, and supports for motor planning.
Keywords: Clinical practices; Display design; Survey; Augmentative and alternative communication
Introduction
of an individual, and using these characteristics of the
individual to drive the selection and design of the system
(Light & McNaughton, 2013a). For example, depending on the individual’s visual or motor access abilities,
the size of the symbols may or may not be an important
feature to manipulate (Kovach & Kenyon, 2003).
A 2006 survey examined SLPs’ perceptions on
what contributes to success and abandonment of AAC
technology (Johnson, Inglebret, Jones, & Ray, 2006).
SLPs reported that an appropriate match between the
individual and the system is one factor that promotes
greater success with the device. Intrinsic abilities such
as motor, cognitive/linguistic, literacy skills, and sensory
perceptual skills must be assessed and compared to external features of systems to determine the best match.
With the great variety of aided AAC technologies
available, matching external features to intrinsic abilities is no small task. Ideally, practitioners are combining
their practical knowledge and experiences with available evidence to inform a trial-based, feature-matching
approach. However, such an approach may increase
the time it takes an individual to reach competence
with a system. Rather, if there are design decisions that
follow specific patterns, these could potentially reduce
the number of trials needed to identify the best fit for
Individuals with complex communication needs often
rely on augmentative and alternative communication
(AAC) to participate in communication interactions.
An AAC system encompasses a variety of methods to
support communication, such as gestures, sign language,
communication boards, and speech generating devices
(Beukelman & Mirenda, 2013). Techniques that utilize
tools outside of the body, such as a communication
board with graphic symbols or a computer programmed
with voice output, are called aided AAC. Substantial
evidence suggests the use of AAC interventions increases
language development with individuals with a variety of
communication disabilities (e.g., Binger & Light, 2007;
Drager et al., 2006; Romski & Sevcik, 1996).
Once a system is selected, AAC intervention requires
more than taking the device out of the box and handing it over to the individual. One of the challenges facing practitioners such as speech-language pathologists
(SLPs), special education teachers, and occupational
therapists is creating an aided AAC system that maintains an appropriate balance between the benefits of the
communication afforded by the system, and the costs
of learning how to use it (Beukelman, 1991). Achieving
this balance requires determining the needs and abilities
Correspondence: Jennifer Thistle, Department of Communication Sciences and Disorders, HSS 112, University of Wisconsin-Eau Claire, Eau Claire,
WI 54702, USA. Tel: ⫹ 1 715 836 6015. E-mail: thistljj@uwec.edu
(Received 28 March 2014; revised 23 March 2015; accepted 24 March 2015)
124
AAC Display Design Decisions
the individual. The first step in identifying decision
processes that are more likely to result in positive
outcomes is to better understand the factors considered
by practitioners in designing AAC displays.
There has been limited research specifically exploring
the kinds of decisions SLPs and other practitioners
make related to display design. One of the only studies
to date that examined this topic was a preliminary and
qualitative study using a hypothetical case study format.
Specifically, McFadd and Wilkinson (2010) provided
six clinicians who specialize in the field of AAC with a
partial list of vocabulary to include on an aided AAC
display. Clinicians selected additional vocabulary and
created a low-tech communication board for a hypothetical student to use during snack time. Clinicians
were asked to narrate their thought processes while making decisions regarding the selection and arrangement
of vocabulary on the display.
The clinicians in McFadd and Wilkinson (2010)
applied research-based recommendations by incorporating vocabulary that supported a variety of communicative functions (Adamson, Romski, Deffebach,
& Sevcik, 1992; Light & Drager, 2005). For example,
five of the clinicians included verbs and social terms
to support a variety of interactions. Another clinician
only included object labels, but described her rationale
that those labels would support social communication by allowing the child and peers to talk about the
foods they were eating. Five of the six clinicians used
Boardmaker Picture Communication Symbols (PCS;
Mayer-Johnson, 1992) to represent the content, based
on the instruction from the researchers; the sixth pulled
similar types of images from the Internet because her
school did not have access to the Boardmaker program.
Choices related to arrangement of the symbols were
less consistent across clinicians. All clinicians organized
the vocabulary in some fashion, and those that included
different types of words (e.g., object, verbs, socialregulatory) created subsets based on those types. However, the placement of the subsets varied. For instance,
some clinicians placed social-regulatory symbols along
the top row while others placed the same symbols along
the left-hand column. Still another clinician used spacing within the page to separate types of words. Finally,
some clinicians used background color to distinguish
different symbol word classes while another placed the
symbols on a colored page to support navigation across
multiple pages.
One challenge when examining clinical practices lies
in the heterogeneity of individuals who can benefit from
AAC systems. Children with developmental disorders,
for example, may have very different needs and abilities
compared to adults with acquired disorders. The current study focused on practitioners, specifically SLPs,
working with young children (aged 10 years and under)
in an attempt to constrain some of the variability seen
in AAC decision-making. Even when limited to elementary school children, however, the caseloads of SLPs will
influence the experiences upon which they draw when
© 2015 International Society for Augmentative and Alternative Communication
125
designing AAC displays. In the 2014 American SpeechLanguage-Hearing Association (ASHA) Schools Survey
(ASHA, 2014), the 55% of SLPs who regularly provided
AAC-related services reported serving an average of five
students. This represents 10% of an elementary school
SLPs’ average monthly caseload, and 20% of a residential/special day school SLPs’ average monthly caseload (ASHA, 2014). Furthermore SLPs who work in
residential/special day schools reported that 71% of their
caseloads consisted of students with severe communication impairments. It is likely, then, that SLPs will have
had different experiences designing AAC displays.
Professional preparation in AAC also may influence SLPs’ comfort level with providing AAC services, thereby affecting the decisions they make when
designing AAC displays. In a survey of 71 SLPs, 72%
rated their competence in providing AAC services as
fair to poor (Marvin, Montano, Fusco, & Gould, 2003).
Similar results emerged in surveys conducted in Egypt
(Wormnaes & Abdel Malek, 2004) and New Zealand
(Sutherland, Gillon, & Yoder, 2005). Such low levels of
competence may be precipitated by education and training provided by SLP programs. In a survey of SLP training programs, 33% of respondents felt that the majority
of their students were prepared to work with individuals who use AAC (Ratcliff, Koul, & Lloyd, 2008).
Just under half (47%) of the respondents reported that
only up to one quarter of their students receive clinical practice in AAC. In a review of research conducted
from 1985–2009, Costigan and Light (2010) examined
surveys of pre-service AAC training programs in the
US, and reported that the majority of SLPs received
minimal to no pre-service training in AAC.
Thus, it is possible that the variability noted between
SLPs in McFadd and Wilkinson’s (2010) study reflected
the SLPs’ educational background and experiences
with individualizing displays for the wide variety of
children who use AAC. If there are some practices that
professionals have found to be more successful than
other practices, it is important to identify the successful
approaches in order to reduce the number of trials of
different features that an individual who uses AAC must
go through.
Research Questions
This research addressed the following question: What
design factors do SLPs identify as priorities when
they create aided AAC displays for young school-aged
children, and what are their rationales for the selection of these factors as priorities? Through an online
survey, participants answered questions related to the
decisions they make regarding vocabulary selection,
symbol types and arrangement, and manipulation of a
variety of visual features (e.g., size, color, etc.) of aided
AAC displays. The responses were analyzed not only to
gain a broad view of the general clinical practices but
also to understand the factors that might influence the
decision-making process.
126
J. J. Thistle & K. M.Wilkinson
Method
Survey Development
Survey questions were developed and refined through
initial pilot testing and reviews by experts in survey design. Initial questions were developed to target
decisions related to vocabulary selection, symbol types
and arrangement, and manipulation of a variety of
visual features of aided AAC displays. The initial pilot
survey from which the final version was developed was
completed by three SLPs, with an average of 8 years
(range: 7–10 years) experience providing AAC services
to children. Feedback from the pilot participants ensured
that the focus of the questions centered on the goals
of the study. The university’s Survey Research Center
then reviewed the survey for structure and adherence
to survey design principles. As a result, demographic
questions were moved from the beginning to the end
of the survey, based on the rationale that it may be perceived as less intrusive to answer personal demographic
questions at the end of the survey (Groves et al., 2009).
The final version of the survey consisted of 42 questions intended to solicit information about two aspects
of display design: the principles guiding aided AAC
display design in general, and child-specific decisions
driven by a given case study. Participants advanced
through the survey sequentially (answering the general
questions first, then the child-specific questions) and
received an error message if they attempted to advance
without completing a question. Therefore, as the survey
progressed, no questions could be skipped. Because
the survey could be abandoned prior to completion, a
greater number of responses were provided to questions
asked earlier in the survey than those that were asked
later.
The current study reports on responses related to
the first section because the goal was to outline general
principles guiding aided AAC display design decisions.
Appendix A (to be found online at http://informahealthcare.com/doi/abs/10.3109/07434618.2015.1035798)
presents these survey questions. The results of the
answers related to the specific case will be reported in a
separate study.
Participants
Target participants were practicing SLPs who (a) maintained a current certificate of clinical competence from
ASHA, (b) had at least 1 year of experience supporting individuals who use AAC, and (c) provided AAC
services to school-age children aged 10 and under. The
online survey was available for 12 weeks and participants were recruited through multiple contact points to
allow for adequate opportunity for responding and to
increase sample size (Dillman, Smyth, & Christian,
2009). Qualtrics1 online survey software hosted the
web-based survey. Participants completed the survey at
a computer of their choosing and were able to take the
survey over multiple sessions if they chose to do so.
The University’s Office for Research Protections
provided human subjects approval for this research
project. An implied consent form was embedded as the
first page of the online survey and participants were
advised that continuing the survey indicated consent.
Participants had the option of downloading the implied
consent form if desired.
Survey Distribution
Members of two list serves were contacted at three time
points. The list serves were the ASHA Special Interest
Group 12-Augmentative and Alternative Communication
(SIG-12) and Quality Indicators for Assistive Technology (QIAT). A general recruitment notice describing the
study and soliciting participation was posted to each list
serve at the initial time point, 3 weeks later, and again
7 weeks from the initial posting. Throughout this data
collection period, in-person recruitment also occurred
during the ISAAC biennial convention. Finally, appeals
to personal contacts and postings on social media
websites provided additional advertising regarding
availability of the survey.
Data Analysis
The survey consisted of a mix of open-ended and
closed-ended questions. Descriptive methods of data
analysis were utilized due to the exploratory nature of the
questions and the goal of the survey to identify trends
to inform future research directions. Descriptive data in
the form of frequency tables were used to examine the
closed-ended questions.
The open-ended questions were coded for common
themes using scrutiny techniques (Ryan & Bernard,
2003). Like Ryan and Bernard, three research assistants and the first author initially identified themes
and subthemes by reading each response and listing
commonly repeated terms and identifying similarities
and differences in responses. Refinement of the themes
and subthemes occurred during a cycle of consensus
coding. The research team formally defined the codes
in a codebook that contained the definition as well
as examples and non-examples. A summary page
of the codes is presented in Appendix B (to be found
online at http://informahealthcare.com/doi/abs/10.3109/
07434618.2015.1035798). Two of the primary codes
were each further refined into five secondary codes, resulting in 16 possible codes. The primary codes were used
to identify responses that (a) were unclear to the coder,
(b) noted features the participant did not think were
important, (c) related to the child’s skills and abilities,
(d) related to the communication demands, (e) related
to the AAC device, and (f) related to key stakeholders
(e.g., clinicians, teachers, communication partners). The
secondary codes provided detail related to the child’s
abilities (e.g., vision abilities) and the communication
demands (e.g., functional vocabulary for the setting).
The first author divided the responses into individual
thought units consisting of the smallest meaningful piece
Augmentative and Alternative Communication
AAC Display Design Decisions
of information contained in the response (Fraenkel,
2006). Typically, the thought units corresponded with
a participant’s sentence. However, when the participant
included a variety of ideas in one sentence, the resulting thought units were individual words or phrases.
Research assistants then assigned one code per thought
unit. Inter-observer agreement was assessed on the final
coding of the thought units. Two research assistants
each independently coded thought units. After a period
of training, the coders reached a minimum Kappa
coefficient of .7 (Fleiss, Levin, & Paik, 2003). Coders
then individually coded all questions and subsequent
reliability was recalculated on 25% of the responses.
Agreement on each question was on average .76
(range .71–.78). Kappa values of .4–.75 are considered
good and values of .75 or greater signify excellent agreement (Fleiss et al., 2003).
Results
Responses
In total, 192 individuals accessed the survey. Of those,
24 dropped out during the initial screening section,
17 were excluded because they did not meet the selection criteria, and two reported living outside the United
States, in Canada and South Africa. Due to the small
number of international participants, these responses
were excluded from the final analysis.
Of the 149 eligible participants, 112 completed the
broad design questions2 but provided only some demographic data, 77 completed the entire survey (including
all demographic data), and 37 did not complete the
primary questions. The presentation of the results follows the sequence of the survey, although demographics
and initial design decisions for the 77 participants who
provided that information will be described first. A discussion of the clinical implications and future directions
follow the survey results.
Demographics
Of the 77 participants who provided complete demographic data, 60 (78%) were members of ASHA
SIG-12, and nearly half (48%) reported living in the
Northeast. Table I presents a summary of the demographic information, including distribution by geographical region, participant gender, and race/ethnicity.
One of the screening questions asked participants’
years of experience supporting children who use AAC.
Thus, although only 77 participants completed the
demographics section of the survey, all 112 participants
provided years of experience. Table II presents the
proportion of participants who completed the broad
design questions but did not complete the demographic
questions and those who completed both sections by
their level of experience. The following results address
similarities and differences observed in responses across
the different levels of experience.
© 2015 International Society for Augmentative and Alternative Communication
127
Table I. Percentage of Participants by Geographical Region, Gender,
and Race/Ethnicity.
Characteristic
Participants (%)
Geographic region
Northeast
Southeast
North Central
South Central
West/Mountain
Gender
Female
Male
Race/ethnicity
White/Caucasian
African American
Hispanic
Asian
n
%
37
12
7
6
15
48.0
7.8
15.6
9.1
19.5
75
2
97.4
2.6
72
1
3
1
93.5
1.3
3.9
1.3
Initial Design Decisions
One of the first decisions a clinician must make when
creating a new display for young children is whether to
modify the page set provided by the manufacturer. Of
the 77 participants answering this question, 60 (78%)
reported often or always making changes to the page
set provided by the AAC manufacturer, and 8 (10%)
reported rarely or never making changes to the page
set provided by the manufacturer. An examination of
the responses by level of experience did not reveal a
distinctive pattern related to experience level. Across
most experience levels, only 9% (5 out of 53) of the
participants reported rarely or never making changes.
However, 21% (3 out of 14) of the participants with
13–20 years of experience reported rarely or never
making changes. This difference in responding by participants with this level of experience recurs throughout the
survey and will be explored in the Discussion section.
Decisions Related to Vocabulary Selection
SLPs often play a key role in choosing the display
content, including what concepts and communication
functions the content supports. Several themes emerged
in terms of decisions made by SLPs with regard to the
importance of child preferences, other stakeholders, the
role of core vocabulary, and the range of word classes
to include.
Table II. Number and Percentage of Participants Completing the
Broad Design and the Demographic Questions Sections.
Years of
experience
1–3
4–7
8–12
13–20
21 or more
Completed
demographic
questions (n ⫽ 77)
n
%
8
23
13
14
19
10.4
29.9
16.9
18.2
24.6
Completed only broad
design questions
(n ⫽ 35)
n
%
3
7
5
11
9
8.6
20.0
14.3
31.4
25.7
128
J. J. Thistle & K. M.Wilkinson
Child Preferences. The child’s preferences were noted to
be extremely or fairly important in vocabulary selection by 87% (97 out of 112) of respondents. Figure 1
illustrates the level of importance participants placed
upon the child’s preferences in their vocabulary selection process, for each level of clinician experience. All
of the participants with 1–3 years of experience felt the
child’s preferences were extremely or fairly important.
On the other hand, 28% (7 of 25) of respondents with
13–20 years of experience indicated that the child’s
preferences were only somewhat or not very important.
This accounted for half of the 13% (15 out of 112) of
all participants who felt that child’s preferences were
somewhat or not very important.
Additional Priorities. An open-ended question provided
some insight into what additional priorities respondents
felt were important when choosing vocabulary. Figure 2
illustrates the main categories of priorities identified by
participants by years of experience.
Key Stakeholders. As a whole, 46 of 112 (41%) participants mentioned collaborating with key stakeholders.
Once again, however, there was a somewhat unusual
pattern in the group with 13–20 years’ experience, as
only 20% (5 out of 25) mentioned key stakeholders
compared to 41% (41) of the remaining 87 participants. Some participants mentioned specific instances
that would influence vocabulary selection. One wrote,
“I add vocabulary based on family preferences too –
parents often like some of the politeness terms, for
instance.” Other participants were more general in their
description and rationale of including key stakeholders:
“I take into account what the family and classroom find
important.”
Role of Core Vocabulary. Clinicians with more than 13
years of experience reported choosing core vocabulary
based on frequency of words more often (15%, 8 out
of 53) than those with less than 13 years of experience
(8%, 5 out of 59). Some respondents prioritized core
vocabulary above the child’s or key stakeholders’ preferences. For instance, one participant stated, “I would
ONLY consider the child’s vocabulary preferences
Figure 2. Percentage of participants mentioning additional vocabulary
selection considerations beyond a child’s preferences, by years of
experience.
when I am including personal words that are part of
the child’s extended vocabulary set. These personal
extended vocabulary words are second on my list after
CORE vocabulary” (emphasis provided by participant).
This quote also illustrates the challenge inherent in
design decision – many decisions influence other decisions, and trade-offs must be made. In this case, it seems
the participant was making a choice between providing
core vocabulary or personalized vocabulary.
Range of Word Classes. Most participants indicated that
they frequently used a variety of word forms in support
of language acquisition and use. Specifically, participants incorporated subjects (82%, 92 of 112), actions
(97%, 109 of 112), objects (83%, 93 of 112), and emotion words (84%, 94 of 112) most or all of the time.
Figure 3 shows the frequency with which participants
incorporate each of these types of words by their level
of experience. In all, 93% (26 out of 28) of participants
with the most experience reported incorporating emotion words most or all of the time, whereas on average
81% (68 out of 84) of the clinicians with other levels of
experience incorporated emotion words most or all of
the time.
Decisions Related to Symbol Type
Following identification of appropriate content, SLPs
consider options regarding how best to represent that
content. The type of symbol representation was rated as
fairly or extremely important by 100% of participants
with 1–3 years’ experience (n ⫽ 11), but only 76% (77
out of 101) of all other participants. When asked what
factors influence their choice of symbol type, 90% (10
out of 11) of the participants with 1–3 years of experiences cited the child’s cognitive abilities as an important consideration. Across all other experience levels,
just under half (45%, 45 out of 101) reported that the
child’s cognitive level should be considered.
Decisions Related to Visual Features of the Display
Figure 1. Percentage of participants who rated child’s preference in
vocabulary selection as either fairly/extremely important or not very/
somewhat important, by years of experience.
The visual nature of an aided AAC display allows for
manipulation of such features as symbol arrangement
Augmentative and Alternative Communication
AAC Display Design Decisions
129
Figure 3. Percent of participants indicating the frequency with which they include subjects, action words, descriptors, and emotion words in
displays, by years of experience.
and display layout, symbol size, and use of color.
Three themes that emerged from this survey concerned
choices related to (a) the type of display layout, (b) the
use of black and white versus colored symbols, and
(c) the use of background color cuing.
Type of Display Layout. When asked to estimate the
percentage of hybrid or visual scene displays (VSDs)
participants design, 83% (64 out of 77) reported using
these displays less than 25% of the time, suggesting grid-based displays were used most of the time.
Only 3% (2 out of 77) of participants reported using
VSD/hybrid displays more than 50% of the time.
Use of Symbol Internal Color. Of the 112 participants,
94 (84%) reported utilizing symbols that have internal color most or all of the time; only one participant
indicated very rarely using color symbols, but did not
provide a reason in the following open-ended question.
Despite the widespread use of symbol color, 78 (69%)
reported using black and white symbols some of the
time. Participants reported that black and white symbols were used to highlight new symbols, when color
would not contribute to the meaning (e.g., prepositional
words), or when team members did not have access to
color printers.
Use of Background Color. Color can also be featured in
the background of symbols. All participants reported
using symbol background color at least sometimes, and
49 (43%) reported that they used it most of the time,
a trend that was consistent across all experience levels.
The top two reasons provided regarding use of background color were (a) to support development of grammatical skills through color coding parts of speech, and
© 2015 International Society for Augmentative and Alternative Communication
(b) to draw attention to specific symbols. Using background color as a cue to word class category reflects a
common clinical recommendation (Goossens’, Crain, &
Elder, 1999; Kent-Walsh & Binger, 2009); however, to
date there has been no research that specifically examines if the dimension of color aids in learning appropriate sentence structure.
Additional Decisions
When given the opportunity to discuss any additional
general factors not previously mentioned, 32% (36 out
of 112) of participants supported the use of consistency
in display design to support motor planning and automaticity. In this approach, the location of previously
learned symbols on the display does not change as new
symbols are added to the display. Finally, an additional
feature participants consider was mentioned in response
to several different, unrelated, open-ended questions:
designing the display in a way that supports partner
modeling.
Discussion
The goal of AAC intervention for a child is to provide
support for participation and language development
across all environments, facilitating early communication (Light & Drager, 2005; Romski & Sevcik, 1996),
advancing linguistic growth and functional communication (Binger & Light, 2007; Drager et al., 2006;
Johnston, McDonnell, Nelson, & Magnavito, 2003), and
providing early literacy experiences (Koppenhaver &
Erickson, 2003; Light & McNaughton, 2013b). Designing an appropriate AAC display is one part of AAC
intervention that may contribute to this goal. There are
130
J. J. Thistle & K. M.Wilkinson
many factors to consider when designing an AAC display, including but not limited to appeal of the display
(Light, Drager, & Nemser, 2004; Light, Page, Curran, &
Pitkin, 2007), ease of use (Drager, Light, Speltz, Fallon
& Jeffries, 2003; Drager et al., 2004; Fallon, Light, &
Achenbach, 2003), and communicative functions supported by the display (Adamson et al., 1992; Romski &
Sevcik, 1996). Furthermore, these considerations must
be weighed against the needs and abilities of the child
who will be using the display (Light & McNaughton,
2013a). Certainly, as a field, we are at times successful as
we strive toward this goal; at other times, however, we do
not succeed (Johnson et al., 2006; Snell et al., 2010).
In this study, SLPs reported considering a number of
factors in the development of AAC displays, suggesting
at least some awareness of the need for AAC systems to
be responsive to the needs of the child and team. This is
in line with the multifaceted nature of AAC interventio...
Purchase answer to see full
attachment