Subido por claudia ovalle

KeepingItRealwithAuthAssessment

Anuncio
See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/254307494
Keeping It “REAL” with Authentic Assessment
Article in NHSA Dialog A Research-to-Practice Journal for the Early Intervention Field · January 2010
DOI: 10.1080/15240750903458105
CITATIONS
READS
33
1,257
2 authors:
Marisa Macy
Stephen Bagnato
University of Nebraska at Kearney
University of Pittsburgh
58 PUBLICATIONS 508 CITATIONS
108 PUBLICATIONS 1,910 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Infant Mental Health--Early Regulatory Disorders Assessment View project
Innovative Disabilities Interventions for Rural & Urban Contexts View project
All content following this page was uploaded by Stephen Bagnato on 12 August 2014.
The user has requested enhancement of the downloaded file.
SEE PROFILE
This article was downloaded by: [Macy, Marisa]
On: 15 January 2010
Access details: Access Details: [subscription number 918633748]
Publisher Routledge
Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office: Mortimer House, 3741 Mortimer Street, London W1T 3JH, UK
NHSA Dialog
Publication details, including instructions for authors and subscription information:
http://www.informaworld.com/smpp/title~content=t775653686
Keeping It “R-E-A-L” with Authentic Assessment
Marisa Macy a; Stephen J. Bagnato b
a
Department of Education, Lycoming College, b Departments of Pediatrics and Psychology-inEducation, University of Pittsburgh Medical Center (UPMC),
Online publication date: 15 January 2010
To cite this Article Macy, Marisa and Bagnato, Stephen J.(2010) 'Keeping It “R-E-A-L” with Authentic Assessment', NHSA
Dialog, 13: 1, 1 — 20
To link to this Article: DOI: 10.1080/15240750903458105
URL: http://dx.doi.org/10.1080/15240750903458105
PLEASE SCROLL DOWN FOR ARTICLE
Full terms and conditions of use: http://www.informaworld.com/terms-and-conditions-of-access.pdf
This article may be used for research, teaching and private study purposes. Any substantial or
systematic reproduction, re-distribution, re-selling, loan or sub-licensing, systematic supply or
distribution in any form to anyone is expressly forbidden.
The publisher does not give any warranty express or implied or make any representation that the contents
will be complete or accurate or up to date. The accuracy of any instructions, formulae and drug doses
should be independently verified with primary sources. The publisher shall not be liable for any loss,
actions, claims, proceedings, demand or costs or damages whatsoever or howsoever caused arising directly
or indirectly in connection with or arising out of the use of this material.
NHSA DIALOG, 13(1), 1–20
C 2010, National Head Start Association
Copyright ISSN: 1524-0754 print / 1930-9325 online
DOI: 10.1080/15240750903458105
RESEARCH ARTICLES
Keeping It “R-E-A-L” with Authentic Assessment
Marisa Macy
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Lycoming College, Department of Education
Stephen J. Bagnato
University of Pittsburgh Medical Center (UPMC), Departments of Pediatrics
and Psychology-in-Education
The inclusion of young children with disabilities has remained a function of the Head Start program
since its inception in the 1960s when the United States Congress mandated that children with
disabilities comprise 10% of the Head Start enrollment (Zigler & Styfco, 2000). Standardized, normreferenced tests used to identify children with delays are problematic because (a) they have low
treatment validity, (b) they are not universally designed or adaptable, (c) it is difficult to capture the
real life behaviors/skills of young children, (d) they do little to facilitate collaboration with parents,
(e) they lack sensitivity to changes in the child’s development and learning, and (f) children with
disabilities are often excluded from group data. This study examined the use of an alternative approach
that uses early childhood authentic assessment to determine a young child’s eligibility for special
services. Results of this study have implications for adopting authentic assessment practices.
Keywords: developmental delay, systems integration
The inclusion of young children with disabilities has remained a function of the Head Start program since its inception in the 1960s. The United States Congress mandated that children with
disabilities comprise 10% of the Head Start enrollment (Zigler & Styfco, 2000). Informal and formal types of assessment practices are used to determine children who are eligible for specialized
services (Bagnato, 2007; McLean, Bailey, & Wolery, 2004; National Association of School Psychologists, 2005; National Research Council, 2008). Informally, Head Start professionals observe
children on an ongoing basis as children participate in their early childhood programs. Professionals note when children’s behavior and skills deviate from developmental expectations. Parents and
caregivers also contribute to understanding a young child’s development. Home visits and other
opportunities for collaboration allow Head Start programs to gather a holistic picture of children.
Correspondence should be addressed to Marisa Macy, Lycoming College, Department of Education, 700 College
Place, Williamsport, PA 17701. E-mail: macy@lycoming.edu
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
2
MACY AND BAGNATO
Formally, professionals may incorporate the use of tools to screen children for delays in development and learning. An individual assessment may occur if a child’s academic or developmental
performance is suspicious to a professional and parent, or a group of children may be screened.
The Individuals with Disabilities Education Improvement Act of 2004 (IDEA) requires the implementation of child-find efforts in order to locate young children who are eligible for services due
to delay or risk conditions. Screening tools offer a snapshot and are not comprehensive enough
to determine eligibility for IDEA services.
Eligibility assessments are used when enough data are collected to warrant an in-depth look
at a child’s development. Standardized, norm-referenced tests are the most widely used tools
to determine whether a child is eligible for IDEA services. This practice is a result of the fact
that standardized, norm-referenced tests often contain standard scores that are required under
many state guidelines (Danaher, 2005; Shakelford, 2006). Standard deviation and percentage
delay are frequently used in order to identify the eligible population of young children; however, several states make use of informed clinical opinion, which promotes more flexibility for
children, their families, and professionals (Bagnato, McKeating-Esterle, Fevola, Bartalomasi, &
Neisworth, 2008; Bagnato, Smith-Jones, Matesa, & McKeating-Esterle, 2006; Dunst & Hamby,
2004; Shakelford, 2002). Informed clinical opinion, or clinical judgment, is often used by professionals to understand a child’s special needs without having to administer formal tests to a child
(Bagnato et al., 2006).
It is not unreasonable to assume that standardized, norm-referenced tests have been proven to
be reliable and valid. These tests have been in existence for several years and have widespread use
among professionals who believe them to have strong psychometric properties. A recent research
synthesis examined the use of standardized, norm-referenced tests and established that few if any
of these tests are effective in determining eligibility for IDEA special services for young children
with disabilities (Bagnato, Macy, Salaway, & Lehman, 2007).
In addition to weak evidence supporting use of standardized, norm-referenced tests to identify
eligible young children with delays, the study (Bagnato et al., 2007) also pointed out several
flaws with using these tests for eligibility determination. First, most of the standardized, normreferenced tests examined in Bagnato et al.’s (2007) study did not include children with disabilities
in the standardization process. In addition, all of the standardized, norm-referenced tests lacked
item density and procedural flexibility, which are salient elements to assessing young children
with delays. The standardized, norm-referenced tests were also deficient in offering graduated
scoring options.
Another concern with using standardized, norm-referenced tests for eligibility determination
is that they do not inform treatment efforts or instruction. These tests can show the extent a child’s
behavior or skill is different from the norm, but they do not lay a foundation for the next steps in
helping the child reach his or her potential or supports she or he will need to be fully included
in a Head Start classroom. Differentiating children’s performances on a test is the main focus of
standardized, norm-referenced tests, not necessarily what will occur after the test is over.
One of the basic tenets of IDEA is that assessment practices must be fair and nonbiased.
Court cases have challenged the use of inappropriate assessments. Professional organizations
like the Division for Early Childhood (DEC) and the National Association for the Education
of Young Children (NAEYC) propose using multiple methods for collecting information about
children. A promising alternative to standardized, norm-referenced testing is the use of authentic
assessment, which promotes a natural context to best understand what a child can do. It is a
form of assessment that favors events, materials, and individuals who are familiar to the child.
AUTHENTIC ASSESSMENT
3
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Real life conditions become the backdrop for children to apply what they know to a given
situation.
Head Start programs would benefit by having authentic eligibility assessments that link to
curriculum and instruction because children will enter their programs with meaningful information (Bagnato et al., 2007; Grisham-Brown, Hallam, & Pretti-Frontczak, 2008). Results from the
authentic assessment can be used to directly inform program planning, curriculum, instruction,
and lesson plans. To implement authentic assessment, practitioners should pay close attention to
the R-E-A-L framework: roles, equipment, assessment tools, and location.
Role of data collector. Authentic assessment relies upon a team of people that consists of informed
caregivers such as parents, grandparents, and other family members as well as teachers, speech
therapists, and other professionals who are familiar with and have knowledge of the child’s
skills and abilities (Bagnato & Neisworth, 1999; Guralnick, 2006). Effective teams have
mutual respect for one another’s roles and expertise, ability to communicate with others,
and openness to share assessment role responsibilities. Assessment responsibilities are shared
when (a) parents are considered central members of the team with valuable observations
and information to contribute (Meisels, Xue, Bickel, Nicholson, & Atkins-Burnett, 2001);
(b) teachers and practitioners provide input (Dunst, 2002; Keilty, LaRocco, & Casell, 2009);
and (c) the team relies upon the observations and evaluations of trained professionals such as
occupational and speech therapists, depending on the child’s need (Meisels, Bickel, Nicholson,
Xue, & Atkins-Burnett, 2001).
Equipment and materials. Familiar equipment and materials are used to assess children using an
authentic assessment framework. Common toys or household items are examples of materials
that children will recognize from their natural environments. When assessment includes the
actual or authentic activity with companion materials, the child is operating under more usual
conditions and has experience performing similar tasks. For example, assessing a child’s
adaptive skills during snack with all the familiar accompanying utensils, food items, and
furniture can help to obtain an accurate assessment of the child’s true ability.
Assessment tools. Select authentic assessment tools that bring together interdisciplinary and interagency teams (Losardo & Notari-Syverson, 2001; Slentz & Hyatt, 2008). Curriculum-based
tools connect assessment to programming and intervention planning (Macy & Bricker, 2006).
Curriculum-based assessments allow teams to gather information from various sources, including parents and teachers. Another benefit is that they can be used to monitor individual and
group progress. Bagnato, Neisworth, and Pretti-Frontczak (in press) recommend eight quality
standards for selecting assessment tools to include (a) acceptability—the social aspects of using
a tool; (b) authenticity—contextual factors (e.g., everyday situations); (c) collaboration—the
tool can be easily incorporated into interdisciplinary teamwork; (d) evidence—the tool has
been found through research and practice to be valid, reliable, and useful; (e) multiple factors—
used to gather information about a child (e.g., various methods, individuals, situations, and
time points); (f) sensitivity—the tool is sensitive to child performance; (g) universality—
children with special needs can be accommodated; and (h) utility—the extent the tool is useful
to parents and professionals. These eight standards are illustrated in Figure 1. Approximately
100 different authentic and standard, norm-referenced assessments were rated using the eight
standards by over 1,000 assessment users in an online study. Results from the online survey
indicated that users identified authentic assessment as meeting more of the eight standards
than the standardized, norm-referenced tests.
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
4
MACY AND BAGNATO
FIGURE 1
Standards for judging assessment tools (from Bagnato et al., in press).
Location. Authentic assessment approaches reflect the ongoing experiences children may encounter in their home, school, community, and other places where young children spend
time. Authentic assessment takes place during the real life conditions under which the target
behaviors/skills are needed for the child to function. Young children are often more comfortable and relaxed in familiar locations, and this will result in a more accurate assessment
(Neisworth & Bagnato, 2004).
The R-E-A-L framework can be used by Head Start professionals to facilitate implementation
of an authentic assessment approach. The foundation for assessment should be to measure skills
that reflect what a child is capable of doing in genuine situations. Authentic assessment is used
to understand what children can do in naturalistic settings by using typical early childhood
experiences to assess children. This is different from standardized, norm-referenced tests that
are often administered in a clinical setting, with structured and adult-directed procedures, by
people who are unfamiliar to the child, and test discrete isolated tasks that are unrelated to the
child’s daily life. Given the questionable practice of using standardized, norm-referenced tests to
determine a child is eligible for IDEA special services, further research and analysis on assessment
AUTHENTIC ASSESSMENT
5
alternatives is necessary. The purpose of this study was to examine the technical adequacy of
authentic assessment measures.
METHOD
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Nine authentic measures were chosen through a review of literature and surveys of practitioners
investigating the most commonly used measures in preschool and early intervention settings
(Bagnato, Neisworth, & Munson, 1997; Bagnato et al., in press; Pretti-Frontczak, Kowalski, &
Brown, 2002). To be included in our review, the authentic measure needed to have published
research available. Other authentic measures were considered for inclusion (like the Developmental Continuum from Teaching Strategies); however, we did not find published research on
these measures. The following nine authentic measures are included in the study:
r
r
r
r
r
r
r
r
r
Adaptive Behavior Assessment System (ABAS),
Assessment Evaluation and Programming System (AEPS),
Carolina Curriculum for Preschoolers with Special Needs (Carolina),
Child Observation Record (COR),
Developmental Observation Checklist System (DOCS),
Hawaii Early Learning Profile (HELP),
Pediatric Evaluation of Disability Inventory (PEDI),
Transdisciplinary Play-Based Assessment (TPBA), and
Work Sampling System (WSS)/Ounce
Some measures had multiple editions (i.e., AEPS, Carolina, ABAS, COR, and TPBA).
Table 1 offers information about these measures.
Research studies on the identified authentic assessments (i.e., ABAS, AEPS, Carolina, COR,
DOCS, HELP, PEDI, TPBA, and WSS/Ounce) were reviewed in this synthesis. Research characteristics fell into two categories: accuracy (reliability) and effectiveness (validity, utility).
Accuracy refers to the extent to which a tool identifies young children with disabilities. This
includes reliability of the measure (e.g., consistency across test items and the use of cutoff scores
in order for the tool to precisely or accurately measure a skill or behavior). Examples of accuracy
include test-retest reliability, interrater reliability, intrarater reliability, and interitem consistency.
Effectiveness refers to the extent to which a tool successfully identifies young children with
disabilities. This includes the validity of the measure (i.e., including to what extent does the tool
measure what it was designed to measure) and how it relates significantly to similar measures.
Examples of effectiveness include predictive validity, concurrent validity, construct validity, test
floors, and item gradients. The next section describes our strategy for searching the literature base
for research studies on the accuracy and effectiveness of authentic assessment measures.
Search Strategy
Search Terms. Relevant published research studies were identified using the following
search terms: authentic assessment, testing, early intervention, preschool, early childhood, eligibility, pediatrics, disabilities, handicap identification, and referral. More general terms of special
schools, state programs resource, centers, and evaluations were also used.
6
Birth to
89 years
5 rating
forms
Birth to 6
Birth to
preschool
ABAS
AEPS
Carolina
Age Range
Parent/Primary Caregiver
form is designed to be
completed by parents or
other primary caregivers.
Two forms are available:
Ages birth to 5 and Ages
5–21.
4-point rating scale:
3 = Always when needed
2 = Sometimes when needed
1 = Never when needed
0 = Is not able
Includes a box to check if
rater guessed. Includes
section for rater to make
comments regarding a
specific item.
3-point rating scale:
2 = Consistently meets
criterion
1 = Inconsistently meets
criterion
0 = Does not meet criterion
Six qualifying notes
3-point rating scale:
(+) Mastery
(+/−) Inconsistent/emerging
skill
(−) Unable to perform skill
Qualifying notes
Parent Form
(B-5): 241
Teacher Form
(2–5): 216
B-3: 249
3–6: 217
B-2: 359
2–5: 400
Communication, Community
Use, Functional
Pre-Academics,
School/Home Living,
Health and Safety, Leisure,
Self-Care, Self-Direction,
Social, Motor (10)
Adaptive, Cognitive, Fine
Motor, Gross Motor,
Social-Communication, &
Social (6)
Cognition,
Cognition/Communication,
Communication,
Personal-Social, Fine
Motor, Gross Motor (6)
Families encouraged to be
involved throughout the
assessment and instruction
process.
Family Report allows parents
to be involved in collecting
information and
list/prioritize areas of
interest.
Family Involvement
Scoring Features
# of Items
Domains
TABLE 1
Nine Authentic Measures and Their Characteristics
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
n/a
Provides cutoff
scores by age
intervals.
Provides normreferenced
scores based on
age.
Eligibility
Features
7
21/2 to 6
Birth to 6
COR
DOCS
Developmental Checklist
(DC): Language, Social,
Motor, Cognition (4)
Adjustment Behavior
Checklist (ABC)
Parental Stress and Support
Checklist (PSSC)
Initiative, Social Relations,
Creative Representation,
Music and Movement,
Language and Literacy,
Mathematics and Science
(6)
Provides normreferenced
scores based on
age.
Parents are viewed as primary
informant on the DC
questionnaire. PSSC
assesses parental stress,
parental support, child
adaptability, parent-child
interaction, and
environmental impact.
DC: Raters check Yes or No
(Yes = 1; No =0)
ABC: 4-point rating scale:
Very much like, Somewhat
like, Not much like, Not at
all like
PSSC: 4-point rating scale:
Highly agree, Somewhat
agree, Sometimes agree,
Do not agree
DC: 475
ABC: 25
PSSC: 40
(Continued on next page)
n/a
Family Report is available to
create reports for parents
about their child that can
be discussed at parent
conferences or home visits.
Parents are able to record
notes about the child based
on the parents’
observations of the child’s
behavior at home on the
report. Parent Guide
available to explain the
COR and for parents to
record anecdotes based on
the COR.
5-point rating scale:
Five descriptive statements
that represent a range of
functioning from very poor
to very superior
30 items, each
with 5 levels
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
8
Birth to 6
6 months
to 71/2
years
HELP
PEDI
Age Range
Self-Care
Mobility
Social Function (3)
Cognition, Fine Motor, Gross
Motor, Language, Social
Emotional, Self Help (6)
Domains
Functional
Skills: 197
Caregiver
Assistance:
20
Modifications:
20
685
# of Items
Materials available to
increase parent
participation in the
assessment process;
guidelines for inclusion of
parent input are spread
throughout.
Options for administration
include parent interview.
Functional Skills: 0 =
Unable, or limited in
capability to perform item
in most situations; 1 =
Capable of performing
item in most situations
Caregiver Assistance:
Independent,
Supervise/Prompt/Monitor,
Minimal assistance,
Moderate assistance,
Maximal assistance, Total
assistance
Modifications: No
modifications,
Child-oriented,
Rehabilitation, Extensive
modifications
Family Involvement
4-point rating scale:
(+) Skill or behavior is
present
(−) Skill is not present
(+/−) Skill appears to be
emerging
(A) Skill or behavior is
atypical or dysfunctional
Qualifying notes
Scoring Features
TABLE 1
Nine Authentic Measures and Their Characteristics (Continued)
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Provides normreferenced
scores.
Provides age
ranges for
skills.
Eligibility
Features
WSS: PreGrade
5
WSS/Ounce
WSS: Personal and Social
Development, Language
and Literacy, Mathematical
Thinking, Scientific
Thinking, Social Studies,
The Arts, Physical
Development & Health (7)
Ounce: Personal
Connections, Feelings
About Self, Relationships
with Other Children,
Understanding and
Communication,
Exploration and Problem
Solving, Movement and
Coordination (6)
Cognitive, Social-Emotional,
Communication and
Language, Sensorimotor
Development
n/a
A list of developmental
skills
observed
during play
3 types of ratings:
Not Yet
In Process
Proficient
Scoring system:
(+) Skill at age level and
his/her skills are
qualitatively strong
(−) Skill is below age level
and team has qualitative
concerns
√
( ) Need for further
observation and/or testing
(NO) No opportunity
(NA) Not applicable due to
age or disability
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Ounce: Contains a Family
Album element that is used
by families to collect
observations, photos, and
mementos of their child’s
growth and development.
Includes play with parent in
the assessment sequence.
n/a
n/a
Note. ABAS = Adaptive Behavior Assessment System; AEPS = Assessment Evaluation & Programming System; Carolina = Carolina Curriculum for Infants/
Toddlers/Preschoolers with Special Needs; COR = Child Observation Record; DOCS = Developmental Observation Checklist System; HELP = Hawaii Early Learning
Profile; PEDI = Pediatric Evaluation of Disability Inventory; TPBA = Transdisciplinary Play-Based Assessment; WSS/Ounce = Work Sampling System.
Ounce:
Birth to
31/2
Birth to 6
TPBA
9
10
MACY AND BAGNATO
The nine authentic assessments (i.e., ABAS, AEPS, Carolina, COR, DOCS, HELP, PEDI,
TPBA, and WSS/Ounce) were also included within the search. The search was done broadly in
the fields of psychology, developmental disabilities, special education, allied health fields (speech
and language therapy, physical therapy, occupational therapy), and as early intervention.
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Sources. The primary databases included the following sources: CINAHL, Cochrane Library, Digital Dissertations, Ebsco Host, Education Resource Information Center (ERIC), Google
Scholar, Health Source, Illumina, Medline, Ovid/Mental Measurements Yearbook Buros, Psychological Abstracts (PsycINFO), and Social Sciences Citation Index. Additionally, we conducted
selective searches of unpublished masters’ theses and doctoral dissertations. Hand searches of
select journals and ancestral searches were also conducted.
Selection Criteria. The study had to meet the following criteria for inclusion: (a) researched
one or more of the selected authentic assessment measures; (b) involved the evaluation of young
children with disabilities or at risk for developing a disability due to environmental or biological
risk conditions; (c) examined the accuracy or effectiveness of the measure at testing infants,
toddlers, and preschool children with disabilities; and (d) disseminated in a scientific and scholarly
publication, which included dissertation and thesis studies. This research synthesis was conducted
as part of literature reviews and syntheses conducted at the Tracking, Referral and Assessment
Center for Excellence (Dunst, Trivette, & Cutspe, 2002).
RESULTS
A total of 27 studies on authentic assessment were identified from the fields of child development,
early intervention, psychology, special education, physical therapy, pediatrics, and behavioral
development. The most studies were conducted on the AEPS. The following information is
presented in Table 2: total number of studies that met the inclusion criteria, years articles were
published, age range included in the studies, and the total number of participants in study samples.
Participants
There were over 10,000 young children who participated in these research studies on authentic
assessment. Children’s ages ranged from birth to 224 months. Children were identified with
various disabilities, and there were several studies that included children without disabilities
and children who were at risk for developing a disability. Table 3 shows child characteristics
that include total sample size, mean age in months, age range in months, and child ability
characteristics.
Types of Studies
Each study reported in this synthesis examined some aspect of accuracy and/or effectiveness
related to one or more of the authentic measures. We found the following types of studies:
AUTHENTIC ASSESSMENT
11
TABLE 2
Research Studies on the Nine Measures
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Tool
# of Studies
Publication Years
ABAS
AEPS
Carolina
2
9
1
2006–2007
1986–2008
2006
COR
DOCS
HELP
PEDI
TPBA
WSS/Ounce
3
5
2
2
4
1
1993–2005
1997–2005
1995–1996
1993–1998
1994–2003
2008
Children’s Age Range (in Months)
33 to 216
0 to 72
Mean age for the treatment
group was 4.5 months
48 to 68
1 to 72
22 to 34
36 to 224
6 to 46
45 to 60
Sample Size (Children)
151
2,897
47
4,902
2,000+
29
50
74
112
Note. ABAS = Adaptive Behavior Assessment System; AEPS = Assessment Evaluation & Programming System;
Carolina = Carolina Curriculum for Infants/Toddlers/Preschoolers with Special Needs; COR = Child Observation
Record; DOCS = Developmental Observation Checklist System; HELP = Hawaii Early Learning Profile; PEDI =
Pediatric Evaluation of Disability Inventory; TPBA = Transdisciplinary Play-Based Assessment; WSS/Ounce = Work
Sampling System.
13 interitem/interrater reliability, 5 test–retest reliability, 5 sensitivity/specificity, 15 concurrent
validity, 3 predictive validity, and 6 construct/criterion validity. Accuracy (reliability) and effectiveness (validity) of the research studies are identified in Table 4.
Reported Results
A total of 16 studies examined the accuracy of authentic measures. There were 20 studies that
examined the effectiveness of authentic measures. The number of studies exceeds 27 because
some studies examined accuracy and effectiveness. Table 5 incorporates results on the accuracy
and effectiveness of authentic assessment measures.
DISCUSSION
Using authentic assessment to determine children eligible for IDEA services has the potential to
improve Head Start services for children (Bagnato, 2007; Grisham-Brown et al., 2008; Gulikers,
Bastiaens, & Kirschner, 2004; Layton & Lock, 2007; Neisworth & Bagnato, 2004). In order to
establish whether or not a child was eligible for early intervention services, many of the studies
in our review had standardized, norm-referenced tests as a comparison from which to judge the
merit of the authentic measures. This was the case for the majority of concurrent and constructs
validity studies reviewed in this synthesis. McLean et al. (2004) suggest that one way to examine
construct validity is to establish convergent validity by examining high positive correlations with
other tests that measure the same constructs. Authentic assessments do not measure constructs in
the same way as standardized, norm-referenced tests. Instead of comparing authentic measures
12
MACY AND BAGNATO
TABLE 3
Research Studies with Participant Demographic Information
Author(s) and Year
(N = 27)
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Anthony (2003)
Bailey & Bricker (1986)
Baird, Campbell, Ingram, &
Gomez (2001)
Bricker, Bailey, & Slentz
(1990)
Bricker, Yovanoff, Capt, &
Allen (2003)
Bricker et al. (2008)
Calhoon (1997)
Cody (1995)
Sample
Size
Age Range
in Months
10
32
13
6 to 46
n/a
11 to 47
Developmental delay
Children with and without developmental delay
Cri-du-Chat syndrome
335
2 to 72
861
1 to 72
1,381
4
25
1 to 72
22 to 35
22 to 34
Children with (mild, moderate, and severe) and without developmental delay,
and at risk
Children eligible and not eligible for early intervention/early childhood
special education
Same as above
Language delay
Previously identified as delayed in the areas of behavior, cognition, and
language
Down syndrome
DelGiudice, Brogna,
47
n/a
Romano, Paludetto, &
Toscano (2006)
Di Pinto (2006)
60
60 to 216
Fantuzzo, Grim, & Montes
733/1,427
n/a
(2002)
Friedli (1994)
20
n/a
Gilbert (1997)
100
1 to 72
Hsia (1993)
82
36 to 72
Knox & Usen (1998)
10
45 to 224
Macy, Bricker, & Squires
68
6 to 36
(2005)
McKeating-Esterle, Bagnato,
91
33 to 71
Fevola, & Hawthorne
(2007)
Meisels, Xue, & Shamblott
112
45.24 to 59.76
(2008)
Morgan (2005)
32
4 to 60
Myers, McBride, & Peterson
40
7 to 36
(1996)
Noh (2005)
65
36 to 64
Sayers, Cowden, Newton,
Warren, & Eason (1996)
Schweinhart, McNair,
Barnes, & Larner (1993)
Sekina & Fantuzzo (2005)
Sher (2000)
Slentz (1986)
Wright & Boschen (1993)
Child
Characteristics
ADHD (ADHD/PI; ADHD/C)
Urban and low income
Children with and without developmental delay
Children with and without developmental delay
Children with and without developmental delay
Cerebral palsy
Children eligible and not eligible for early intervention/early childhood
special education
Developmental delay, autism, hearing impairment/deafness, Down
syndrome, MR, CP/muscular dystrophy, speech/language impairment,
visual impairment/blindness, other health impairment, multiple disabilities
At risk; children with special needs whose IEPs indicated that they were in
the mild to moderate range (speech or physical impairment)
Reactive attachment disorder
Developmental delay
4
n/a
Children eligible and not eligible for early intervention/early childhood
special education
Down syndrome
2,500
n/a
Low income
242
20
55 to 68
36 to 67
53
40
36 to 72
36 to 84
Urban
Children eligible and not eligible for early intervention/early childhood
special education
Children with and without developmental delay
Cerebral palsy
Note. ABAS = Adaptive Behavior Assessment System; ADHD = Attention Deficit Hyperactivity Disorder; AEPS =
Assessment Evaluation & Programming System; Carolina = Carolina Curriculum for Infants/Toddlers/Preschoolers
with Special Needs; COR = Child Observation Record; CP = Cerebral Palsy; DOCS = Developmental Observation
Checklist System; HELP = Hawaii Early Learning Profile; IEP = Individualized Education Plan; MR = Mental
Retardation; PEDI = Pediatric Evaluation of Disability Inventory; TPBA = Transdisciplinary Play-Based Assessment;
WSS/Ounce = Work Sampling System.
13
AUTHENTIC ASSESSMENT
TABLE 4
Research Study Characteristics
Accuracy (Reliability)
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Author(s) and Year (N = 27)
Test
Anthony (2003)
Bailey & Bricker (1986)
Baird, Campbell, Ingram, & Gomez
(2001)
Bricker, Bailey, & Slentz (1990)
Bricker, Yovanoff, Capt, & Allen
(2003)
TPBA
AEPS
DOCS
Bricker et al. (2008)
AEPS
Calhoon (1997)
Cody (1995)
DelGiudice, Brogna, Romano,
Paludetto, & Toscano (2006)
Di Pinto (2006)
Fantuzzo, Grim, & Montes (2002)
Friedli (1994)
Gilbert (1997)
Hsia (1993)
TPBA
AEPS
AEPS
ABAS
COR
TPBA
DOCS
AEPS
Knox & Usen (1998)
Macy, Bricker, & Squires (2005)
PEDI
AEPS
Meisels, Xue, & Shamblott, (2008)
McKeating-Esterle, Bagnato, Fevola,
& Hawthorne (2007)
Morgan (2005)
Myers, McBride, & Peterson (1996)
Noh (2005)
WSS
ABAS
AEPS
AEPS
DOCS
TPBA
AEPS
Effectiveness (Validity)
Interitema
Sensitivitya
Constructa
Interraterb Test-Retest Specificityb Concurrent Criterionb Predictive
xb
x
x
x
xa
xb
x
x
xa
xb
xa
xb
x
x
x
x
xa
xb
xb
xa
xb
x
x
x
xa
xa
xa
xb
xb
x
x
x
x
x
xa
xa
xb
Sayers, Cowden, Newton, Warren, &
Eason (1996)
Schweinhart, McNair, Barnes, &
Larner (1993)
Sekina, & Fantuzzo (2005)
Sher (2000)
Slentz (1986)
COR
xb
COR
AEPS
AEPS
xb
Wright & Boschen (1993)
PEDI
x
AEPS
x
xa
xb
xa
xb
x
x
x
x
xa
xa
xa
Note. ABAS = Adaptive Behavior Assessment System; AEPS = Assessment Evaluation & Programming System;
Carolina = Carolina Curriculum for Infants/Toddlers/Preschoolers with Special Needs; COR = Child Observation
Record; DOCS = Developmental Observation Checklist System; HELP = Hawaii Early Learning Profile; PEDI =
Pediatric Evaluation of Disability Inventory; TPBA = Transdisciplinary Play-Based Assessment; WSSS/Ounce =
Work Sampling System.
14
MACY AND BAGNATO
TABLE 5
Reported Research Results
Author(s) and Year (N = 27)
Anthony (2003)
Bailey & Bricker (1986)
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Baird, Campbell, Ingram, & Gomez
(2001)
Bricker, Bailey, & Slentz (1990)
Bricker, Yovanoff, Capt, & Allen
(2003)
Bricker et al. (2008)
Calhoon (1997)
Cody (1995)
DelGiudice, Brogna, Romano,
Paludetto, & Toscano (2006)
Di Pinto (2006)
Fantuzzo, Grim, & Montes (2002)
Friedli (1994)
Gilbert (1997)
Hsia (1993)
Knox & Usen (1998)
Macy, Bricker, & Squires (2005)
Meisels, Xue, & Shamblott (2008)
Reported Results
TPBA visual development guidelines were used by raters at Denver’s PLAY
clinic and had positive interrater agreement results.
AEPS correlation with the Gesell Developmental Schedule (Knobloch et al.,
1980) was strong for the whole test but not individual areas. AEPS could be
successfully administered in a reasonable amount of time.
DOCS may lack sensitivity in detecting variations in development.
AEPS correlation across areas was r = .88 (p < .001).
AEPS newly established cutoff scores in the 2nd edition identified eligible
children accurately most of the time.
AEPS cutoff scores performed similarly to the Bricker et al. (2003) study. The
measure accurately identified the majority of children correctly.
Children performed better (i.e., higher scores) on the TPBA than the
conventional test, and the play-based assessment provided a richer
description of children’s emerging skills.
In the HELP study, the play age obtained from the authentic assessment was
highly correlated with the Developmental Age Equivalent of the
conventional assessment (i.e., BSID).
After 1 year, children in the Carolina condition made progress and had higher
DQ than children in the comparison condition who made slight progress but
improvements were not statistically significant.
ABAS accurately documents poor social adaptive outcomes for children with
ADHD.
The study supports the use of the COR assessment method for low-income
urban preschool children; however, a three-factor model should replace the
proposed six-factor model.
TPBA had favorable test retest results, interrater agreement, and concurrent
validity.
Significant difference among raters was found: mothers rated the child’s skills
highest, the fathers next, and the teachers last. Differences among raters on
the DOCS may influence eligibility decisions.
The AEPS has strong interrater agreement at both domain (ranging from .87
social to .94 adaptive) and total test (.90) levels. Strong relationship between
individual domain scores (.64 to .96) and total test (.98) when internal
consistency was examined. Findings also showed that the AEPS was
sensitive to performance differences of children with delays.
The PEDI is a useful tool for describing the area of functional delay in children
with cerebral palsy. It also appears to be sensitive to changes that were
observed clinically.
The AEPS accurately classified all eligible children and over 94% (n = 64/68)
of the noneligible children. The overall sensitivity was 100%; specificity
was 89%. The AEPS used to determine eligibility was positively and
significantly correlated with conventional eligibility measures. Finally, the
observers who scored the AEPS had strong agreement on observations made
about child performance.
Study found evidence for validity and reliability of WSS, suggesting that
WSHS accurately assesses language development, literacy, and mathematics
skills in young children.
(Continued on next page)
AUTHENTIC ASSESSMENT
15
TABLE 5
Reported Research Results (Continued)
Author(s) and Year (N = 27)
McKeating-Esterle, Bagnato, Fevola,
& Hawthorne (2007)
Morgan (2005)
Myers, McBride, & Peterson (1996)
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Noh (2005)
Sayers, Cowden, Newton, Warren, &
Eason (1996)
Schweinhart, McNair, Barnes, &
Larner (1993)
Sekina, & Fantuzzo (2005)
Sher (2000)
Slentz (1986)
Wright & Boschen (1993)
Reported Results
ABAS-II is correlated with ratings of informed opinion when assessing
children for early intervention.
Evidence supports the predictive validity of the DOCS-II in detecting RAD in
a randomized sample.
The overall mean number of days to complete the eligibility assessment
process took the group using the TPBA 22 days less than it took the group
using a conventional test.
The AEPS has satisfactory interrater reliability agreement in the cognitive and
social domains. Strong relationship between individual domain scores and
items in the domains.
When children’s scores increased on the gross motor domain of the HELP,
they did the same for the PSI.
The COR was found to be a psychometrically promising tool for the
assessment of children’s development in developmentally appropriate early
childhood programs. Also, the COR helped staff to understand early
childhood development and curriculum and to prepare individualized
education programs for their children.
Univariate and multivariate results provide support for convergent and
divergent validity of the COR dimensions. Fifteen of the 18 variables
differentiated the three COR dimensions, particularly the COR Cognitive
and Social Engagement dimensions.
Professionals using the AEPS were able to identify eligible children. Moderate
interrater reliability for the communication domain and high reliability for
other domains.
Positive results of this study support the technical properties of the AEPS.
Interrater agreement was very high at .94 for the entire test and ranged from
.84 to .94 for the six domains. Results for two administrations of the AEPS
(N = 18) revealed strong level of stability across time for the total test (.91),
domain scores varied between high (fine motor .86, cognitive .91, social
communication .77) to moderate (social .50) to low stability (gross motor
.07 and self-care .13). Internal consistency was strong for all domains except
self-care. Concurrent validity was examined by comparing the AEPS with
the McCarthy (1972) and the Uniform Performance Assessment System
(Haring, White, Edgar, Affleck, & Hayden, 1980) with mixed results
ranging from very weak to strong relationships between domains and scales.
Satisfactory information is provided to confirm the PEDI’s usefulness for
clinical and research purposes with children with cerebral palsy.
Note. ABAS = Adaptive Behavior Assessment System; AEPS = Assessment Evaluation & Programming System;
Carolina = Carolina Curriculum for Infants/Toddlers/Preschoolers with Special Needs; COR = Child Observation
Record; DOCS = Developmental Observation Checklist System; HELP = Hawaii Early Learning Profile; PEDI =
Pediatric Evaluation of Disability Inventory; TPBA = Transdisciplinary Play-Based Assessment; WSS/Ounce = Work
Sampling System.
with other good examples of authentic and linked eligibility assessment, many of the studies in
our review made comparisons with nonlinked eligibility tests. For example, in the Macy, Bricker,
and Squires (2005) study, the AEPS (curriculum linked measure) was compared with the Battelle
Developmental Inventory (not linked to curriculum).
16
MACY AND BAGNATO
Future research should continue not only to examine the accuracy and effectiveness of authentic measures but also to make comparisons using an external standard. Some examples of
external standards are (a) correct identification rates based on expert consensus, (b) the need for
services—service-based eligibility, and (c) probability of succeeding/progressing in Head Start
or typical setting with typical peers without support services. Another area of research should
study the effects of the initial eligibility assessment using authentic measures on child outcomes
using a longitudinal design. Cost–benefit studies need to be conducted on the use of authentic
assessment practices used to determine children eligible for IDEA services. This type of evidence
could be helpful to policymakers when reauthorizing policies and updating eligibility assessment
guidelines.
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Limitations
A number of authentic assessment measures are commercially available; however, we chose to
examine only these nine measures because they appeared most often in the professional literature
and by accounts from practitioners. Other authentic measures, like the Developmental Continuum
from Teaching Strategies, were not included in this review because they did not meet our inclusion
criteria. For example, tools like the Developmental Continuum have conducted research studies;
however, we were unable to locate these in journals and/or the databases we searched. Sometimes
publishers maintain the results from studies for proprietary reasons and they are not available to
the public. The body of literature contained other publications related to authentic assessment
and we included only research studies that involved young children who were at risk or had a
disability in the sample.
Authentic assessment can be used for designing a program for a child, creating interventions,
and to evaluate the efficacy of a child’s individualized program. Not only does authentic assessment have potential to accurately identify children in need of services, it also has important
implications beyond eligibility determination (Neisworth & Bagnato, 2004). Findings of this
study will help professionals to critically identify characteristics of authentic assessment research
findings that influence the accurate and representative documentation of a young child’s early
intervention eligibility assessment experience.
Head Start programs would benefit from an assessment approach that provides useful information linked to curriculum and instruction in order to serve children more efficiently. Additionally,
an approach is needed that can monitor ongoing child performance (Downs & Strand, 2006;
Grisham-Brown, Hallam, & Brookshire, 2006; Grisham-Brown et al., 2008; Hebbeler, Barton,
& Mallik, 2008; Meisels, Liaw, Dorfman, & Nelson, 1995). Authentic assessment is a viable alternative to eligibility assessments that use standardized, norm-referenced tests (Bagnato, 2005,
2007; Bagnato & Neisworth, 1992; Macy & Hoyt-Gonzales, 2007; McLean, 2005; Neisworth &
Bagnato, 2004). An authentic assessment approach has growing support from early childhood
professional organizations (i.e., DEC and NAEYC) and from the literature base (Bagnato &
Neisworth, 2005). The authentic assessment studies reported here in this research study are the
first phase of research supporting the use of an authentic assessment approach for eligibility
determination. Although this foundation is a good start, more research is needed to continue to
explore authentic measures and approaches.
AUTHENTIC ASSESSMENT
17
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
REFERENCES
*Indicates studies used in the synthesis.
**Indicates measures used in the synthesis.
*Anthony, T. L. (2003). The inclusion of vision-development guidelines in the Transdisciplinary Play-Based Assessment:
A study of reliability and validity. Unpublished doctoral dissertation, University of Denver, Denver, CO.
Bagnato, S. J. (2005). The authentic alternative for assessment in early intervention: An emerging evidence-based practice.
Journal of Early Intervention, 28, 17–22.
Bagnato, S. J. (2007). Authentic assessment for early childhood intervention best practices: The Guilford school practitioner series. New York: Guilford.
Bagnato, S. J., Macy, M., Salaway, J., & Lehman, C. (2007). Research foundations for conventional tests and testing
to ensure accurate and representative early intervention eligibility. Pittsburgh, PA: TRACE Center for Excellence
in Early Childhood Assessment, Early Childhood Partnerships, Children’s Hospital/University of Pittsburgh, U.S.
Department of Education Office of Special Education Programs, and Orelena Hawks Puckett Institute.
Bagnato, S. J., McKeating-Esterle, E., Fevola, A. F., Bartalomasi, M., & Neisworth, J. T. (2008). Valid use of clinical
judgment (informed opinion) for early intervention eligibility: Evidence-base and practice characteristics. Infants and
Young Children, 21, 334–348.
Bagnato, S. J., & Neisworth, J. T. (1992). The case against intelligence testing in early intervention. Topics in Early
Childhood Special Education, 12, 1–20.
Bagnato, S. J., & Neisworth, J. T. (1999). Collaboration and teamwork in assessment for early intervention. Comprehensive
Psychiatric Assessment of Young Children, 8, 347–363.
Bagnato, S. J., & Neisworth J. T. (2005). DEC recommended practices: Assessment. In S. Sandall, M. L. Hemmeter, B.
J. Smith, & M. McLean (Eds.), DEC recommended practices (pp. 45–50). Longmont, CO: Sopris West.
Bagnato, S. J., Neisworth, J. T., & Munson, S. (1997). LINKing assessment and early intervention. Baltimore, MD:
Brookes.
Bagnato, S. J., Neisworth, J. T., & Pretti-Frontczak, K. L. (in press). LINKing assessment and early intervention (4th ed.).
Baltimore, MD: Brookes.
Bagnato, S. J., Smith-Jones, J., Matesa, M., & McKeating-Esterle, E. (2006). Research foundations for using clinical
judgment (informed opinion) for early intervention eligibility determination. Cornerstones, 2(3), 1–14.
*Bailey, E., & Bricker, D. (1986). A psychometric study of a criterion-referenced assessment designed for infants and
young children. Journal of the Division of Early Childhood, 10, 124–134.
*Baird, S. M., Campbell, D., Ingram, R., & Gomez, C. (2001). Young children with Cri-du-Chat: Genetic, developmental,
and behavioral profiles. Infant-Toddler Intervention, 11, 1–14.
**Bricker, D. (Ed.). (1993). Assessment, evaluation, and programming system for infants and children (1st ed.). Baltimore,
MD: Brookes.
**Bricker, D. (Ed.). (2002). Assessment, evaluation, and programming system for infants and children (2nd ed.).
Baltimore, MD: Brookes.
*Bricker, D., Bailey, E., & Slentz, K. (1990). Reliability, validity, and utility of the Evaluation and Programming System:
For infants and young children (EPS-I). Journal of Early Intervention, 14, 147–160.
*Bricker, D., Clifford, J., Yovanoff, P., Pretti-Frontczak, K., Waddell, M., Allen, D., et al. (2008). Eligibility determination
using a curriculum-based assessment: A further examination. Journal of Early Intervention, 31, 3–21.
*Bricker, D., Yovanoff, P., Capt, B., & Allen, D. (2003). Use of a curriculum-based measure to corroborate eligibility
decisions. Journal of Early Intervention, 26, 20–30.
*Calhoon, J. M. (1997). Comparison of assessment results between a formal standardized measure and a play-based
format. Infant-Toddler Intervention: The Transdisciplinary Journal, 7, 201–214.
*Cody, D. A. (1995). Play assessment as a measure of cognitive ability in developmentally delayed preschoolers.
Unpublished doctoral dissertation, George Mason University, Fairfax, VA.
Danaher, J. (2005). Eligibility policies and practices for young children under Part B of IDEA (NECTAC Notes No.15).
Chapel Hill, NC: The University of North Carolina, FPG Child Development Institute, National Early Childhood
Technical Assistance Center.
*DelGiudice, E., Brogna, G., Romano, A., Paludetto, R., & Toscano, E. (2006). Early intervention for children with Down
syndrome in southern Italy: The role of parent-implemented developmental training. Infants & Young Children, 19,
50–58.
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
18
MACY AND BAGNATO
*Di Pinto, M. (2006). The ecological validity of the Behavior Rating Inventory of Executive Functioning (BRIEF) in
attention deficit hyperactivity disorder: Predicting academic achievement and social adaptive behavior in the subtypes
of ADHD. Dissertation Abstracts International, 67(03), 1696B. (UMI No. 3209793)
Downs, A., & Strand, P. S. (2006). Using assessment to improve the effectiveness of early childhood education. Journal
of Child and Family Studies, 15, 671–680.
Dunst, C. J. (2002). Family-centered practices: Birth through high school. Journal of Special Education, 36(3), 139–147.
Dunst, C. J., Trivette, C. M., & Cutspe, P. A. (2002). An evidence-based approach to documenting the characteristics
and consequences of early intervention practices. Centerscope, 1(2), 1–6. Retrieved July 11, 2009, from http://www.
evidencebasepractices.org/centerscope/centerscopevol1no2.pdf
Dunst, C. J., & Hamby, D. W. (2004). States’ Part C eligibility definitions account for differences in the percentage of children participating in early intervention programs. Snapshots, 1(4). Retrieved July 12, 2009, from
http://www.tracecenter.info/products.php
*Fantuzzo, D., Grim, S., & Montes, G. (2002). Generalization of the Child Observation Record: A validity study for
diverse samples of urban, low-income preschool children. Early Childhood Research Quarterly, 17, 106–125.
*Friedli, C. (1994). Transdisciplinary play-based assessment: A study of reliability and validity. Unpublished doctoral
dissertation, University of Colorado, Denver, CO.
*Gilbert, S. L. (1997). Parent and teacher congruency on variations of a screening assessment: An examination (Report
No. H023B50009). Auburn, AL: Auburn University. (ERIC Document Reproduction Service No. ED413726)
Grisham-Brown, J., Hallam, R., & Brookshire, R. (2006). Using authentic assessment to evidence children’s progress
toward early learning standards. Early Childhood Education Journal, 34, 45–51.
Grisham-Brown, J., Hallam, R., & Pretti-Frontczak, K. L. (2008). Preparing Head Start personnel to use a curriculumbased assessment: An innovative practice in the “age of accountability.” Journal of Early Intervention, 30, 271–281.
Gulikers, T. M., Bastiaens, T. J., & Kirschner, P. A. (2004). A five-dimensional framework for authentic assessment.
Educational Technology Research and Development, 52(3), 67–86.
Guralnick, M. J. (2006). Family influences on early development: Integrating the science of normative development,
risk and disability, and intervention. In K. McCartney & D. Phillips (Eds.), Blackwell handbook of early childhood
development (pp. 44–61). Malden, MA: Blackwell.
**Haley, S. M., Coster, W. J., Ludlow, L. H., Haltiwanger, J. T., & Andrellos, P. J. (1992). Pediatric Evaluation of
Disability Inventory (PEDI). Boston: New England Medical Center Hospitals, Inc.
**Harrison, P. L., & Oakland, T. (2003). Adaptive Behavior Assessment System-Second Edition. San Antonio, TX: The
Psychological Corporation.
Hebbeler, K., Barton, L. R., & Mallik, S. (2008). Assessment and accountability for programs serving young children
with disabilities. Exceptionality, 16, 48–63.
**High/Scope Educational Research Foundation. (2003). Preschool Child Observation Record–Second Edition. Ypsilanti,
MI: High/Scope Press.
**Hresko, W. P., Miguel, S. A., Sherbenou, R. J., & Burton, S. D. (1994). Developmental Observation Checklist System.
Austin, TX: PRO-ED.
*Hsia, T. (1993). Evaluating the psychometric properties of the Assessment, Evaluation, and Programming System for
Three to Six Years: AEPS Test. Unpublished doctoral dissertation, University of Oregon, Eugene, OR.
Individuals with Disabilities Education Improvement Act, Pub. L. No. 108-446, 20 U.S.C. (2004).
**Johnson-Martin, N. M., Attermeier, S. M., & Hacker, B. J. (2004). The Carolina curriculum for infants and toddlers
with special needs (3rd ed.). Baltimore, MD: Brookes.
**Johnson-Martin, N. M., Hacker, B. J., & Attermeier, S. M. (2004). The Carolina curriculum for preschoolers with
special needs (2nd ed.). Baltimore, MD: Brookes.
Keilty, B., LaRocco, D. J., & Casell, F. B. (2009). Early interventionists’ reports of authentic assessment methods through
focus group research. Topics in Early Childhood Special Education, 28, 244–256.
*Knox, V., & Usen, Y. (1998). Clinical review of the Pediatric Evaluation of Disability Inventory. The British Journal of
Occupational Therapy, 63, 29–32.
Layton, C. A., & Lock, R. H. (2007). Use authentic assessment techniques to fulfill the promise of No Child Left Behind.
Intervention in School & Clinic, 42, 169–173.
**Linder, T. W. (1993). Transdisciplinary play-based assessment: A functional approach to working with young children.
Baltimore, MD: Brookes.
Losardo, A., & Notari-Syverson, A. (2001). Alternative approaches to assessing young children. Baltimore, MD: Brookes.
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
AUTHENTIC ASSESSMENT
19
Macy, M., & Bricker, D. (2006). Practical applications for using curriculum-based assessment to create embedded learning
opportunities for young children. Young Exceptional Children, 9(4), 12–21.
*Macy, M. G., Bricker, D. D., & Squires, J. K. (2005). Validity and reliability of a curriculum-based assessment approach
to determine eligibility for part C services. Journal of Early Intervention, 28, 1–16.
Macy, M., & Hoyt-Gonzales, K. (2007). A linked system approach to early childhood special education eligibility
assessment. TEACHING Exceptional Children, 39(3), 40–44.
*McKeating-Esterle, E., Bagnato, S. J., Fevola, A., & Hawthorne, C. (2007). A comparison of clinical judgment (informed
opinion) with performance-based assessment to document early intervention status. Unpublished master’s thesis,
University of Pittsburgh, Pittsburgh, PA.
McLean, M. (2005). Using curriculum-based assessment to determine eligibility: Time for a paradigm shift? Journal of
Early Intervention, 28, 23–27.
McLean, M., Bailey, D. B., & Wolery, M. (2004). Assessing infants and preschoolers with special needs (3rd ed.). Upper
Saddle River, NJ: Pearson.
Meisels, S. J., Bickel, D. D., Nicholson, J., Xue, Y., & Atkins-Burnett, S. (2001). Trusting teachers’ judgments: A validity
study of a curriculum-embedded performance assessment in kindergarten to grade 3. American Educational Research
Journal, 38, 73–95.
**Meisels, S. J., Dombro, A. L., Marsden, D. B., Weston, D., & Jewkes, A. (2003). The Ounce Scale: An observational
assessment for infants, toddlers, and families. New York: Pearson Early Learning.
**Meisels, S. J., Jablon, J. R., Marsden, D. B., Dichtelmiller, M. L., & Dorfman, A. B. (2001). The Work Sampling System
(4th ed.). New York: Pearson Early Learning.
Meisels, S. J., Liaw, F., Dorfman, A., & Nelson, R. F. (1995). The Work Sampling System: Reliability and validity of a
performance assessment for young children. Early Childhood Research Quarterly, 10, 277–296.
Meisels, S. J., Xue, Y., Bickel, D. D., Nicholson, J., & Atkins-Burnett, S. (2001). Parental reactions to authentic
performance assessment. Educational Assessment, 7, 61–85.
*Meisels, S. J., Xue, Y., & Shamblott, M. (2008). Assessing language, literacy, and mathematics skills with Work
Sampling Head Start. Early Education and Development, 19, 963–981.
Morgan, F. T. (2005). Development of a measure to screen for reactive attachment disorder in children 0–5 years.
Dissertation Abstracts International, 66(04), 2312B (UMI No. 3173576).
*Myers, C. L., McBride, S. L., & Peterson, C. A. (1996). Transdisciplinary, play-based assessment in early childhood
special education: An examination of social validity. Topics in Early Childhood Special Education, 16, 102–126.
National Association of School Psychologists. (2005). NASP position statement on early childhood assessment. Bethesda,
MD: Author.
National Research Council. (2008). Early childhood assessment: Why, what, and how. Washington, DC: National
Academies Press.
Neisworth, J. T., & Bagnato, S. J. (2004). The mismeasure of young children: The authentic assessment alternative.
Infants & Young Children, 17, 198–212.
*Noh, J. (2005). Examining the psychometric properties of the second edition of the Assessment, Evaluation, and
Programming System for three to six years: AEPS Test–Second Edition (3–6). Unpublished doctoral dissertation,
University of Oregon, Eugene, OR.
**Parks, S., Furono, S., O’Reilly, K., Inatsuka, T., Hoska, C. M., & Zeisloft-Falbey, B. (1994). Hawaii early learning
profile (HELP). Palo Alto, CA: Vort.
Pretti-Frontczak, K., Kowalski, K., & Brown, R. D. (2002). Preschool teachers’ use of assessments and curricula: A
statewide examination. Exceptional Children, 69, 109–123.
*Sayers, L. K., Cowden, J. E., Newton, M., Warren, B., & Eason, B. (1996). Qualitative analysis of a pediatric strength
intervention on the developmental stepping movements of infants with Down syndrome. Adapted Physical Activity
Quarterly, 13, 247–268.
*Schweinhart, L. J., McNair, S., Barnes, H., & Larner, M. (1993). Observing young children in action to assess their
development: The High/Scope Child Observation Record study. Educational and Psychological Measurement, 53,
445–454.
*Sekina, Y., & Fantuzzo, J. (2005). Validity of the Child Observation Record: An investigation of the relationship between
COR dimensions and social-emotional and cognitive outcomes for Head Start children. Journal of Psychoeducational
Assessment, 23, 242–261.
20
MACY AND BAGNATO
Downloaded By: [Macy, Marisa] At: 23:06 15 January 2010
Shakelford, J. (2002). Informed clinical opinion (NECTAC Note No. 10). Chapel Hill, NC: The University of North
Carolina, Frank Porter Graham Child Development Institute, National Early Childhood Technical Assistance Center.
Shakelford, J. (2006). State and jurisdictional eligibility definitions for infants and toddlers with disabilities under IDEA
(NECTAC Note No. 21). Chapel Hill, NC: The University of North Carolina, Frank Porter Graham Child Development
Institute, National Early Childhood Technical Assistance Center.
*Sher, N. (2000). Activity-based assessment: Facilitating curriculum linkage between eligibility evaluation and intervention. Unpublished doctoral dissertation, University of Oregon, Eugene, OR.
*Slentz, K. (1986). Evaluating the instructional needs of young children with handicaps: Psychometric adequacy of the
Evaluation and Programming System-Assessment Level II. Dissertation Abstracts International, 47 (11), 4072A.
Slentz, K. L., & Hyatt, K. J. (2008). Best practices in applying curriculum-based assessment in early childhood. In A.
Thomas & J. Grimes (Eds.), Best practices in school psychology V (Vol. 2, pp. 519–534). Bethesda, MD: National
Association of School Psychologists.
*Wright, F. V., & Boschen, K. (1993). The Pediatric Evaluation of Disability Inventory (PEDI) validation of a new
functional assessment outcome instrument. Canadian Journal of Rehabilitation, 7, 41–42.
Zigler, E., & Styfco, S. J. (2000). Pioneering steps (and fumbles) in developing a federal preschool intervention. Topics
in Early Childhood Special Education, 20, 67–70.
View publication stats
Descargar