Computer Tutors Prod Students to Ask for Help
Ask a friend to meet you at a restaurant in a city he or she has never visited, and it’s entirely possible the friend will get lost along the way. You might expect the friend to pull out a map, turn on the car’s navigation system, or even ask a gas station attendant for directions. You wouldn’t expect your lost friend to wander around town, knocking on random restaurant doors until eventually giving up and hopping in a taxi.
Yet research suggests this random searching and reluctance to seek basic help is exactly how high school students often approach problem-solving.
“Students often misuse help,” said Ido Roll, a postdoctoral researcher at the Carl Wieman Science Education Initiative at the University of British Columbia and a member of the Pittsburgh Science of Learning Center. “Either they don’t ask for help at all, or they ask for all the help there is.”
In a series of studies presented at the American Educational Research Association’s annual meeting, held here April 8-12, researchers from the Vancouver-based university and Carnegie Mellon University in Pittsburgh found that students typically will go to extreme lengths to avoid asking for help when working on computer-based tutoring programs. Yet if they learn to think about when and how to ask for help, they are more likely to avoid simply cheating to get answers.
In the classroom, it can be difficult to determine why a student does or doesn’t ask for help. Yet when students use an online program, the computer can record how fast and how often a student tries to solve a problem, uses a dictionary, or asks for help. In several studies since 2006, Mr. Roll and his research partners at Carnegie Mellon’s Human-Computer Interaction Institute—assistant professor Vincent Aleven, senior systems scientist Bruce M. McLaren, and professor Kenneth R. Koedinger—mined data from math-tutoring programs to gauge how 10th and 11th grade students used a help button, which offers progressively more in-depth hints and eventually gives the answer to the question.
“If you get one error, 25 percent of students will ask for help; 75 percent of them will try again. And the pattern persists after any consecutive number of errors; after five, six attempts, I am still more likely to try again than to ask for help,” Mr. Roll said.
“Across hundreds of thousands of actions in multiple studies, we never get more than 30 percent of students who ask for help; they always try to do it again themselves,” he said. “When we asked students [why they didn’t ask for help], they said their parents told them [essentially] ‘real men don’t ask for help.’ ”
By the time students finally did ask for help, the data showed they had given up trying to solve the problem and were aiming to cheat; Mr. Roll said 82 percent of students who used the hint tool did not stop to read it, but instead clicked through multiple hints to get to the answer.
Thinking About Help
In a 2008 study in the same series of studies, Mr. Roll and other researchers found students often “gamed the system” through guessing and looking for the answer when they felt frustrated or disliked the subject.
Research on metacognition—the study of how students think about what they know or learn—suggests that encouraging students to reflect on how they are learning can lead them to use better strategies, such as asking for appropriate help or studying more effectively.
“Simply giving students the tools may not be enough if the extent to which the student can use a strategy depends on metacognitive skills,” said Katherine A. Rawson, an assistant professor of psychology at Kent State University, in Ohio. She studies the role of metacognition in study skills but was not part of the study on students’ digital help-seeking behaviors.
With that in mind, in Mr. Roll’s most recent experiments, the researchers changed an adaptive, computer-based geometry tutor so that the help tool would encourage students to reflect on their problem-solving strategies.
For example, if a student had answered incorrectly several times in quick succession, which would suggest he or she was guessing, a help window could pop up noting that the student seemed to be struggling and point out the glossary and hint buttons. Or if the student hit the hint button repeatedly without enough time to read the hints, suggesting he or she had given up and was just looking for the answer, a help window would encourage the student to slow down and think about the hint.
The help tutor based its responses on how quickly and correctly a student answered problems, as well as how often he or she made “help mistakes,” such as not asking for help when it was needed or asking for help just to get to the answer.
“For unskilled students, just a few errors is enough to trigger a warning; for skilled students, we let them grapple with it a bit more,” Mr. Roll said.
He tested the new system first with 58 10th and 11th graders in two urban classes with high concentrations of students from minority groups and two suburban, low-minority classes, and then in a second experiment with 67 students in a rural high school. In each trial, half the students used the updated tutoring program, while the other two classes used the standard program.
In the first, three-week trial, students who used the hint tool prompting students to be more reflective about solving a problem used hints inappropriately 10 percent less often than those who did not receive feedback. In the control group, 70 percent of the time that students clicked on the help button, it was to “game the system” by clicking straight through to the final answer; by contrast, only 48 percent of the students given the help tutor did so.
The second, four-month trial had similar findings: Students in the control group used the hint button to cut straight to the final answer 60 percent of the time, while students who received feedback from the help tutor tried to game the system only 45 percent of the time. The students using the adaptive help tutor also sought help more productively in other ways: They took more time considering one hint before requesting additional hints and solved the problem with lower-level hints than students using the standard program did.
The researchers will seek to replicate these initial findings with more students.
Vol. 30, Issue 28
Get 10 free stories, e-newsletters, and more!
- Cristo Rey Jesuit College Preparatory School of Houston, Houston, TX
- Assistant Principal - Drake High School
- Tamalpais Union High School District, San Anselmo, CA
- The New School Montessori, Cincinnati, OH
- Senior Education Policy Analyst, North Carolina
- The Hunt Institute, Durham, NC
- Part-time AP class instructors - remote
- Ascendant Institute, US