April 30, 2018
This is the fourth post in a five-part blog series, DLL Data Gaps. Click here to learn more about this project and access the other blogs in the series.
Inadequate data on young dual language learners (DLLs) hampers efforts to support these children in early care and education (ECE). This blog series has explored gaps that exist in tracking the enrollment of DLLs and rating the quality of services for these learners. Finally, as this post will address, there is also a lack of meaningful assessment data to validly capture the full range of DLLs’ development in ECE.
Age-appropriate testing of students’ proficiencies can serve many purposes in ECE, including formative assessment for instruction, screening for special needs, or program-wide research or evaluations. State policy leaders are increasingly focused on student outcomes through more standardized assessment data, collected and aggregated at the systems level, to inform decision-making and the allocation of ECE resources.
Kindergarten readiness assessments (KRAs), in particular, have gained traction as a strategy to provide educators, families, and district and state leaders with more standardized data on the status of children's abilities when they enter kindergarten—a “snapshot on development,” according to the BUILD Initiative. KRAs are intended to both support instruction in the early elementary years and provide information that can help policy leaders support school readiness, not to prevent children from enrolling in school.
More than 40 states are currently developing or implementing KRAs, up from just seven in 2010. The Obama administration promoted states’ adoption of KRAs through the federal Race to the Top – Early Learning Challenge, a discretionary grant program launched in 2011. In part, the grant program encouraged states to measure children’s outcomes in a range of developmental domains in tandem with “implementing comprehensive data systems and using data to improve instruction, practices, services, and policies.” At least 25 states have now mandated KRA use by state law.
States are rolling out their KRA systems in a variety of ways. Some use a commercial assessment tool, such as Teaching Strategies GOLD®, while others created their own state tool or participated in one of three interstate consortia supported by federal grants. The tools can involve direct assessment (requiring a direct interaction between the test administrator and the child), observation of the child in authentic classroom activities, or a combination of these two approaches.
KRAs have weathered a fair amount of concerns and pushback in their roll-out. Some teachers experienced the testing as an extra burden. Researchers cautioned against drawing inferences from a one-time assessment of young children when development is in great flux. As the National Education Goals Panel asserted, “the younger the child, the more difficult it is to obtain reliable and valid assessment data. It is particularly difficult to assess children's cognitive abilities accurately before age 6.” Due to questions of validity and reliability, policy experts, including my colleagues at New America, have also stressed that KRAs should not be used punitively as an accountability measure for ECE providers.
These broader concerns over KRAs have implications for all children—including DLLs. But states also need to specifically consider how to incorporate DLLs in KRAs as a special population. The development of bilingual children looks fundamentally different than their monolingual peers given that DLLs’ knowledge and skills are spread across two languages. For example, research suggests that DLLs have smaller vocabularies in English and their home language when taken separately, but their total vocabulary sizes—the sum of what they know in both languages—are similar to monolingual peers. As such, it is critical that young DLLs have an opportunity to show what they know and can do in their home language.
And yet, at present, almost all state KRAs assess children only in English. “If we want equitable assessments for DLLs, we have to assess across both languages,” said DLL expert Linda Espinosa, who works with states and districts across the country and served on the National Technical Advisory Committee for KRAs. She said nearly all states are failing “to take on the challenge of creating truly equivalent [bilingual] forms of these test items,” a complicated, costly endeavor in the context of tight state budgets. According to the Migration Policy Institute, New Jersey, Oregon, and Texas have developed KRAs entirely in Spanish while Illinois and Washington state allow DLLs to be assessed in their home language for some test items. Bilingual KRA testing also goes hand-in-hand with a need for more bilingual, bicultural assessors, which requires additional investments.
Moreover, while most states have not even attempted to do so, even those pursuing bilingual KRA assessments tend to use direct translation of the English version, Espinosa said. This method can produce tests that are psychometrically unsound (e.g. not normed or validated on DLLs) and/or irrelevant in content for linguistically and culturally diverse children (e.g. asking a DLL in Alaska to recognize a beach umbrella on a vocabulary test). For these reasons, ‘truly’ bilingual KRA testing for DLLs will likely remain a long-term challenge.
Still, KRA implementation needs major improvements even in English-only testing contexts to collect higher-quality DLL data. The U.S. Department of Education highlighted this reality in a 2016 case study of four states’ initial implementation of KRAs. Teachers reported that they did not fully understand guidelines for assessing DLLs, and a majority were unsure about testing procedures for DLLs. For example, some states allow certain testing accommodations for DLLs, such as accepting correct answers in non-verbal forms like pointing or gesturing. Teachers also voiced a desire for greater support, such as more explicit training on administering KRAs with DLLs and on-site assistance from bilingual staff.
In addition to serving DLLs more equitably at the stage of assessment, state leaders should consider how to share and disseminate data on DLLs’ KRA results. Decision-making around how to publish KRA results often reflects the different ways states view KRAs and their purpose. “Whether to publicly report [KRA] data and what should be included...is often a complicated discussion, involving many stakeholders,” concluded ECE researcher GG Weisenfeld in a 2017 report.
Some states, like Maryland and Oregon, publish KRA data publicly on state websites. Washington state also decided to use the KRA data on its state report card, but officials “recognized the paradox of reporting formative assessment data in a summative presentation, and landed on multiple ways to accurately portray the data...opt[ing] not to use a single, composite ‘readiness’ score.” Others states have resisted aggregating or reporting out results. New Jersey, for example, emphasizes that its voluntary KRA is a formative tool to inform instruction and professional development—not to publicize achievement gaps and trends. Michigan similarly states that it finds such summative use “inappropriate.”
For DLLs in states that do decide to publicly report, state leaders must also decide whether to disaggregate the data results by DLL status for public users. In the K-12 context, federal law requires the disaggregation of academic data by English learner (EL) status in grades 3-8, as well as by race and ethnicity, family income, and disability status. Civil rights groups stress the importance of these mandates for ensuring that disadvantaged students do not get ignored or masked in data systems. At the same time, for ELs at lower proficiencies, language barriers will definitionally interfere with their academic performance and thereby drive down the subgroups’ results.
Similarly, KRA systems leaders can use subgroups to expose achievement gaps in ECE, but they also must consider if data is reliable and valid enough to do so. As New America stressed in a 2017 report on K-12 EL data, below a certain threshold of English proficiency, it is impossible to make valid claims about academic proficiencies in English. In the case of English-only KRAs, DLLs’ scores in literacy and math development may reflect English proficiencies rather than true knowledge of concepts and skills. If states publicly report KRA data without this context, and rely on “native English speakers as the norm against which all students are compared, the unique characteristics of DLLs are likely to be misinterpreted, or worse, determined to be delays,” according to DLL expert Espinosa. Again, this is why testing young DLLs bilingually is so critical—to capture an accurate, complete picture of their development.
As shown above, Maryland disaggregates by DLL/EL status in its public reporting of KRA data. The gaps between EL and English proficient kindergarteners are significantly greater in literacy and math than in physical and social domains (which are less language-dependent), perhaps indicating the extent to which language interferes with DLL data’s validity. As the report notes, "Because the KRA is not given in the student’s home language, the knowledge and skills of ELs may not be fully captured.”
Recommendations for State Leaders:
- Assess DLLs/ELs in their home language on kindergarten readiness assessments (KRAs).
- Invest in the development of valid bilingual assessment tools in home languages.
- Invest in expanding access to bilingual assessors.
- Improve and increase professional development and guidance for teachers on administering KRAs with DLLs/ELs.
- If publicly reporting data by DLL/EL status for KRAs, provide guidance and explain limitations of these data to users.