New Tool to Provide Tech Inventory for Common Core
A new readiness tool will provide a national snapshot of school technology in preparation for common-core online assessments in 2014-15
A national inventory of educational technology is evolving as school districts try to determine what digital tools they have—and what they'll need—to deploy online testing for all students on common academic standards just a few years from now.
A new tool released by the two coalitions helping to develop those online assessments is intended to aid states and districts in taking a snapshot of their current rosters of laptops, netbooks, and other mobile devices, as well as their overall technological bandwidth. It then will highlight where districts are lacking in their capability to assess students under the Common Core State Standards by 2014-15, when such testing is set to be introduced.
The free, Web-based Technology Readiness Tool is kicking up myriad concerns among educators, who worry that there's little new money to bring their technology capabilities up to the level needed, that such testing could overwhelm district infrastructure, and that assessments could end up evaluating students' technology skills more than their mastery of common-core material.
But the readiness tool is the first step toward addressing those concerns, and some ed-tech leaders hope it can provide the leverage needed to encourage state lawmakers to add funding to bring lagging districts up to speed, says Douglas Levin, the executive director of the State Educational Technology Directors Association. The Glen Burnie, Md.-based group is helping state education agencies deploy the tool.
The publishing and educational technology company Pearson, based in London, which developed the readiness tool, is seeking to develop assessments for the common core.
"We haven't done an inventory like this [on a nationwide scale] ever," says Levin. "People are viewing this as an assessment issue, but it's also a large-scale technology project. At the end of the day, the test can't work if the technology doesn't work."
What Districts Need
Forty-five states and the District of Columbia have adopted the common standards in both English/language arts and math, and a 46th state—Minnesota—has adopted just the English/language arts standards. The standards were unveiled in 2010 under an initiative led by the nation's governors and state schools chiefs and are now moving into the implementation phase. However, the challenge of implementing the common core has some districts and education experts concerned about finances and logistics.
The intention is to use "next generation" assessments to determine how well students have grasped instruction based on those standards.
Most states are choosing to back assessments being developed by one of two nonprofit coalitions, the Smarter Balanced Assessment Consortium or the Partnership for Assessment of Readiness for College and Careers (PARCC), although some states have joined both. Assessments from both consortia will be administered using technology, and both will make use of new testing options such as simulations, video, and audio.
The main difference between the consortia is that the assessments created by Smarter Balanced will be adaptive, meaning the level of difficulty changes based on how well students are answering questions. That could create a situation in which students being tested on the same material could end up taking exams that are significantly different from those their classmates take.
Some states are already doing many or all of their student assessments online. They include Delaware, Indiana, and Oregon, which have adopted the new standards, as well as Virginia, which hasn't. But most states haven't moved in that direction in a big way, and the idea of doing common-core assessments online by 2014-15 is daunting.
"There's a big concern that school districts won't have the capacity to do that," says Daniel A. Domenech, the executive director of the American Association of School Administrators, based in Alexandria, Va. "The tool is a great idea to give us a factual definition of where school districts are and sound the alert that resources are going to be needed."
The readiness tool, which was released to states in March and is just starting to reach school districts, allows schools and district technology leaders to log in and register how many and what types of computers and other devices they have. There will be at least two rounds of data collection, to be analyzed by the assessment coalitions. The first window of data collection ran from March 20 through June 14, says Chad Colby, a spokesman for PARCC.
The coalitions are seeking information on school district operating systems, the types of technological devices they have, the ratio of students to those devices, available bandwidth, wireless access, network speed, and other categories.
Raymond Reitz, the chief technology officer for the 12,000-student Chapel Hill-Carrboro school system in North Carolina says he's just starting to work with the readiness tool and is concerned about his district's ability to be ready for online assessments by the 2014-15 deadline.
He says his district at the very least will have to increase its number of mobile devices and its wireless network capacity. And that will require additional funding, he says, but with no extra money in sight.
Reitz hopes the data gathered by the readiness tool will "somehow paint a picture and communicate to the legislature and get them concerned about what it's going to take to make this possible."
Other states have made it work.
Michael Stetter, the director of accountability resources for the Delaware Department of Education, says that over the past two years, Delaware has implemented computer-based assessments to track student growth in reading and math several times a year.
To get school districts ready, the state had them do their own technology inventories, and lawmakers allocated money for 10,000 additional netbooks, Stetter says.
"States are worried right now because they're doing paper-and-pencil [tests] and can't imagine having all the computers to get this done," he says.
But the Delaware inventory showed that districts had more resources, in some cases, than they initially thought they had. For example, Stetter says, different departments in a school might buy computers for one dedicated purpose, but wouldn't share them, even though they might be used infrequently.
Domenech, of the AASA, says he, too, hopes the information gathered by the tool can be used as leverage. "There's no question it will definitely flag the need for greater investment in technology," he says. "But because we're still seeing states cutting back on educational dollars, we're wondering where that money will come from."
Testing Content Knowledge
The information gathered by the technology-readiness tool will have an added benefit, says Colby, of PARCC. It will help the two coalitions creating the assessments ensure the tests, at least in part, work with the technology districts already have, rather than what they might acquire.
"We want to know what devices are already being used, and the assessments should follow using that infrastructure," says Colby. "We don't want to create a scenario where the assessment is driving the purchasing."
But the information collected by the readiness tool will likely drive some purchasing decisions, says Wes Bruce, the chief assessment officer for the Indiana Department of Education and the chairman of PARCC's technology operational working group.
For example, even though Microsoft has said it's going to terminate support for its Windows XP operating system by the 2014-15 school year, when the online assessments are launched, if enough schools and districts say they're still using it, the consortia must make sure that the assessments will work with Windows XP.
In addition, says Levin, of the State Educational Technology Directors Association, the hope is that once schools and districts focus on their technology needs, they'll get up to speed enough to give students a chance to try out the technology before the real assessments take place.
"We're critically aware that the test itself should not be the first time the student is exposed to this technology," he says. "We want to assess their content knowledge, not their technology skills."
Vol. 05, Issue 03, Pages 4, 43-45Published in Print: June 13, 2012, as Ready or Not
- Resident Teacher - Master's Degree Programs
- Academy for Urban School Leadership, Chicago, Illinois
- Head of School
- The Sharon Academy, Sharon, Vermont
- University Guidance Counselor. IB School. Dubai.
- Galvin Education, Dubai (City) (AE)
- School Bus Driver
- Madison Highland Prep, Phoenix, Arizona
- Educational Consultant
- Teaching Matters, New York City, New York