The Need for a Series of How-to Articles
In a field as complex as educational technology, conflict is common; consensus is not. Yet there is widespread agreement on at least one point: We need a more organized and persuasive body of evidence on technology’s benefits to classroom practice. Criticism of educational methods has accelerated in recent years, and the cost and complexity of implementing technology-based teaching strategies make them a favorite target. In the last decade, a diverse and growing collection of critics (e.g., Cordes & Miller, 1999; Cuban, 2001; Oppenheimer, 1997, 2003; Stoll, 1995) has been steadily making the case for reducing or eliminating the use of electronic technologies in teaching. At the same time, technology’s advocates have acknowledged an increasing need for research that could address these concerns and make a strong case for technology’s pedagogical contributions.
The current climate calling for accountability through evidence-based research has been building for some time and seems unlikely to diminish. Yet many authors confirm that there are weaknesses in research designs and reports that have rendered past educational studies less than helpful (Berliner, 2002; Burkhardt & Schoenfled, 2003; Kaestle, 1993; Lederman, 2003). Other writers have focused on the special problems reflected in studies of educational technology topics (Clark, 1983, 1985, 1991, 1994; Cradler, 2003; Roblyer & Knezek, 2003; Roblyer, 2004). These weaknesses include fragmented and uncoordinated approaches to studying technology resources and strategies, methods that lack rigor or are ill-matched to the research questions at hand, and poorly written reports that render problematic subsequent attempts at replication and follow-up. We in the field of educational technology have a clear and imminent challenge. We must design and carry out research that will both address past concerns about methods and findings and clarify the directions we should take in the future.
With this article, CITE Journal introduces a series of examples designed to address one of the central problems of providing useful educational technology research: quality assurance. The entries in this series will serve as exemplars of each of several types of technology research study. Each will be an article re-published from a research journal, accompanied by a discussion that deconstructs the characteristics that make it a model for other researchers. This introductory article describes the challenges that must be met in order to do high quality research, discusses the criteria that will be used to select published models, and outlines the types of research studies represented in the series.
Problems and Challenges of Educational Technology Research
In one of several reaction pieces published in the Educational Researcher shortly after the passage of the No Child Left Behind Act of 2001 (NCLB), Berliner (2002) made an interesting distinction between what he termed “easy-to-do science” and “hard-to-do science.” He said,
Easy-to-do science is what those in physics, chemistry, geology, and some other fields do. Hard-to-do science is what the social scientists do and, in particular, it is what we educational researchers do. In my estimation, we have the hardest to do science of them all….We face particular problems and must deal with local conditions that limit generalizations and theory-building—problems that are different from those faced by the easier-to-do sciences. (p. 18)
Berliner and others like him (Lederman, 2003) offer these descriptions not as excuses for ill-conceived research, but rather by way of explanation for why even good research has had so little impact on shaping teaching methods. He cites in particular the problem of “ubiquitous interactions,” or the sheer number of variables in any given education study that increases the difficulty of both isolating effects and of combining results across studies. For example, in any study of teaching behaviors, variables of student IQ, socioeconomic status, motivation to learn, and other factors come into play. A set of practical and logistical problems such as Institutional Review Board (IRB) approvals and permissions from schools and parents, and long-term access to busy classrooms have added yet another layer of difficulty.
As Roblyer (2004) observed, several conditions common to educational technology research add to the already considerable problems inherent in all behavioral research. Technologies change so quickly that it is difficult to build a body of findings over time on any given application. The flaws in the educational technology research paradigm described so convincingly by Clark (1983, 1985, 1991, 1994) have never been satisfactorily resolved. Clark insisted that comparisons of technology-based strategies with non-technology-based ones were confounded by variables such as instructional design and teacher effects. Despite Kozma’s (1991) insightful proposal that research should focus on technology-enhanced instructional designs, rather than the technologies themselves, research of the kind he proposed has been almost nonexistent.
It is tempting to conclude that researchers in educational technology have become convinced that “challenging” means “impossible” and have eschewed studies of technology’s impact. Yet as Burkhardt and Schoenfeld (2003) noted, despite the acknowledged barriers, there are precedents for even more difficult areas of study that have found the means to carry out meaningful research. They note that prior to the Flexner Report in 1910, funding for medical research was almost nonexistent and the field itself was unproven. Yet “…once balkanized and disputatious, medicine began slowly to cohere into a discipline,” thanks to the work of a few pioneering institutions. As research became more coherent and reliable, funding followed. Though some educators object that the study of classroom learning is different from studies of physiological systems (Lederman, 2003), Burkhardt and Schoenfeld and others feel that education should use medical education research as its model and begin as it did, with the establishment of standards for quality.
Instructional Purposes the Series Will Serve
There are indications that the field of educational technology is gearing up to begin foundation work of the kind Burkhardt and Schoenfeld say is so essential. Recent articles (Roblyer, 2004; Roblyer & Knezek, 2003) and editorials (Bull, Knezek, Roblyer, Schrum, & Thompson, 2005; Cradler, 2003; Thompson, Bull, & Bell, 2005) serve both as an acknowledgement of past failures and a rallying cry for new attacks on the problem of providing a sound research base for technology in teaching. The field is beginning to resound with the call for a new educational technology research agenda—one that focuses on capturing the unique impact of technology-enhanced instructional designs, rather than of the digital technologies, themselves.
This article introduces a series of educational technology research studies that can serve as models for others who wish to do meaningful research. This series proposes to build on and extend the foundation work begun in the aforementioned articles and editorials. By modeling and explicating some of the criteria required for good research, the articles in this series will show that useful research, while not easy, is doable. While a “Flexner Report” for educational technology remains a goal for the future, perhaps an essential first step is the recognition that it is an achievable goal. The series also should prove useful as a set of “how to” directions for those just beginning their research efforts (e.g., dissertation studies). To set the stage for these examples, we present major categories of criteria that future research must reflect.
The Pillars of Good Educational Research
The current demands for “scientifically-based research” (read “randomized experiments”) in education notwithstanding, studies that can move the field forward have much less to do with the type of method used in a particular study than with certain features that lend structural support to its findings. Studies that will be helpful are those that meet the following five criteria or “pillars” of good research.
Pillar 1: The Significance Criterion
Every educational research study should make a clear and compelling case for its existence. While significance is a subjective criterion, it seems especially important in technology research that would-be researchers recognize what makes a study significant enough to undertake in today’s educational climate.
In technology research, studies frequently focus on issues such as low use or non-adoption of a technology or perceptions of its current or potential usefulness. However, compared to traditional teaching methods, electronic resources are nearly always more expensive to purchase and maintain over time and technology-based methods more complex to learn and to implement. According to diffusion of innovations expert Everett Rogers (2003), expense and complexity are two of the key concerns people have about adopting innovations, so even more justification is needed for recommending them.
Education is replete with recognized problems related to the “bottom-line indicators” of quality in educational experiences: achievement, school attendance, course and degree completion, and learning time. The fact that the NCLB Act ties funding to scientific evidence of benefits (i.e., quality indicators such as achievement) has given even greater impetus for research on the impact of technology-enhanced strategies on these quality indicators. The legislation itself may change with the administration; the clamor for such evidence seems unlikely to subside. Articles that report research studies with technology-based teaching strategies should begin by making it clear that they address a significant educational problem, as opposed to a proposed technology solution.
Pillar 2: The Rationale Criterion
Burkhardt and Schoenfeld said that many barriers to progress in educational research are, in large part, “self-inflicted wounds” (p. 13). The most damaging of these, they say, is the failure to attend to the need for a solid theory base in research.
Lack of attention to coherent theory building leaves us looking balkanized and incoherent….It also leaves us vulnerable to attack from outside—powerful politicians, and some academics, who understand little of what educational research is all about feel empowered to tell us how to go about our business. (p. 13)
Researchers who wish to counter this problem have not only reviewed the past research in the area, but also analyzed the findings in light of an underlying theory of why we might expect certain results.
For example, studies of interactive/electronic storybook uses in reading might propose, as did Matthew (1997) that auditory cues provided by electronic storybooks reduce the cognitive load required for comprehension by assisting with decoding words. This leads to higher comprehension, especially among students who have decoding difficulties to begin with, since it scaffolds their acquisition of better language skills. Evidence of the impact of cognitive load and its relationship to language learning has been established with past studies.
If we are to make progress in this field or any other, new research must carefully consider previous lines of research, and each study must be built on a foundation of theory about expected effects derived from past work. Many researchers seem to use “theory” in the popular sense. However, as Kerlinger (1973) said, “While the man in the street uses ‘theories’ and concepts…in a loose fashion….The scientist…systematically builds his theoretical structures, tests them for internal consistency, and subjects them to empirical tests” (p. 3).
After reading and scrutinizing past work that has been done on a topic, researchers ask questions like, “Why might a technology-based strategy have impact?” “What evidence do we have that a strategy has potential for impact?” and “What remains to be done to explicate past findings?” This kind of analysis leads to sound research questions on predicted impact. The literature review part of the research report should show that the current study has a solid theory base and builds on and adds important information to past findings.
Pillar 3: The Design Criterion
After establishing research questions, researchers must decide on a research approach and methods that are well-suited to capturing and measuring impact on the variables of interest. The most challenging of the five criteria, this area reflects ongoing debates about what constitutes “evidence-based” approaches and effective design.
Although the NCLB Act equates “good design” with “randomized experiments,” the research community has responded with uncharacteristic unanimity that true experimental1 or quasi-experimental methods are neither always required nor possible. The method selected should depend on the type of problem being studied and the type of information desired. Researchers seem in agreement that selecting the research approach (i.e., experimental and quasi-experimental designs) before identifying the problem is poor practice that will not advance our understanding of what works. What is essential is a design that is a logical choice for the questions under study.
Choice of design falls into two broad areas: (a) “objective” (sometimes referred to as “scientific”), which is usually associated with experimental, quasi-experimental, or other quantitative methods; and (b) “naturalistic inquiry,” which is usually associated with qualitative methods such as narratives, phenomenologies, ethnographies, grounded theory studies, or case studies.
In experimental studies the researcher usually seeks to generalize results to a population, and this means ensuring as much internal and external validity as is possible by including adequate numbers of subjects, subjects selected for good representation of a larger population, methods that attempt to control or describe alternative explanations of impact, data collection that clearly measures the variables under consideration, instrumentation with reported reliability and validity, and statistical analyses that are clearly suited for analyzing the data and that are correctly applied.
Usually, qualitative studies aim to study impact at a given site, rather than generalize findings to other similar sites. While the focus in these studies is on thorough observation and analysis rather than controlled measurement, it is especially important that descriptions of methods in qualitative studies are detailed enough so that they may be understood and interpreted. As with all research, readers should be able to ascertain from the descriptions that methods of gathering and analyzing data were appropriate to the questions being explored and that conclusions were reasonable in light of the findings.
With research on educational technology, some design problems seem to be more frequent than others. Cradler (2003) found, for example, that there is often an overreliance on self-report data in methods that purport to be quantitative and measures of impact that are not objective or validated.
Although studies that employ experimental or qualitative approaches, as well as mixed-methods ones (Chatterji, 2004), can help researchers understand the effects of technology on student learning, an article reporting technology research should have a methods section that shows a good match between the questions being addressed in the study and the designs and methods used to carry it out.
Pillar 4: The Comprehensive Reporting Criterion
Burkhardt and Schoenfeld (2003) emphasized that educational research is in desperate need of “…cumulativity through studies that build on previous work” (p. 8). However, due to what Mosteller, Nave, and Miech (2004) called the “sprawling nature of education research” (p. 29), this objective is difficult. They said that cumulative studies are possible only to the extent that each one reports enough sufficiently detailed information to allow others to analyze and build on previous work. To help meet this need, Mosteller, Nave, and Miech made a good case for a “structured abstract” to be included with every research report. They stated that a structured abstract will be one way of assuring that completed research is useful to those who can use and build on it. The abstract they described and provided an example for includes an original APA-style abstract, background on the study, purpose, setting, subjects, intervention, research design, data collection and analysis, findings, and conclusion.
Naturally, all these components should be well articulated, and conclusions should flow from them logically in light of the study findings. An article reporting technology research should include all of this information in as detailed a form as possible.
Pillar 5: The Cumulativity Criterion
Studies may be well-designed and beautifully reported, meeting criteria 1-4 in all respects, and still not be as helpful as they might to moving the field forward. Even in the so-called “hard sciences,” no single study can address all relevant issues. Therefore, the final quality each study should reflect is a clear indication of where the current study fits in providing the required information, what kinds of studies must follow, and why.
Technology research, especially, suffers from what might be called the single study syndrome. Perhaps the goal was to get the doctoral degree or to achieve tenure. The study is done; case closed. The best, most well-bred research has a lineage that looks to the future as well as builds on the past. In studies of human behavior, where variables are numerous and complex, it takes many studies over time to build a case for causation and impact. Consequently, articles reporting research ideally should make it clear that the study is part of a current or proposed line of research, along with proposed next steps in the line.
Types of Studies That Move the Field Forward
As Clark (1983) argued, there is a need for a new paradigm for what he called “media research.” Although Kozma (1991) answered Clark by proposing that subsequent research should focus on technology-enhanced instructional designs, rather than the technologies themselves, this kind of research has not happened to any great degree. The current climate for accountability asks educators to produce a smoking gun: clear, irrefutable, scientific evidence to justify that their choice of given technology-based designs work better than any other choice in any situation. However, it may be more realistic to seek evidence that technology-based designs lead to improved learning. One way to do this is to establish potential for relative advantage.
Everett Rogers (2003) said that people are more likely to adopt an innovation if they see the advantage of the new strategy relative to what they currently use. When there is a clear need for a better instructional method than those used in the past, researchers can propose that a given technology-based method is the best choice because it offers the combination of relevant symbol systems, processing capabilities, and logistical feasibility to address the need—and then do research to support that it has this relative advantage and clarify the conditions under which it works best. Roblyer (2004) proposed four kinds of impact research to support this paradigm: establishing relative advantage, improving implementation strategies, monitoring impact on societal goals, and reporting on common practices in order to measure sociological impact and shape directions accordingly.
Study Type 1: Research to Establish Relative Advantage
Some supposed benefits of technology-based strategies have been talked about and discussed so frequently and for so long that educational technology professionals often assume their benefits to be well-established, yet descriptions of impact are often anecdotal and few collections of empirical studies (scientific or otherwise) exist to support them. Research lines are urgently needed to confirm and clarify oft-discussed benefits, such as prevention of inert knowledge (Brown, Collins, & Duguid, 1989) with visual technologies like simulations and video-based scenarios; increasing cultural awareness and acceptance (e.g., with distance collaboration projects); increased reading comprehension with interactive technologies (e.g., interactive/electronic storybooks) and increased comprehension of abstract concepts (e.g., with spreadsheets, geometry software) to clarify concepts that students traditionally find complex and difficult to understand.
Study Type 2: Research to Improve Implementation Strategies
Many technology-based strategies are already in common use (e.g., online distance learning, word processing in writing instruction) and only seem likely to expand in the future. Yet, implementations of these technology-based strategies vary widely, and relatively little is known about what has the greatest impact on educational outcomes and why it works well in given situations. For example, word processing is often used in writing assessments, and online chats and conferences are much used in distance courses. Since turning back to precomputer ways of doing the same thing does not seem a good option, much more needs to be known about how to implement the current technology-based strategies to achieve the desired impact.
Sometimes what is not known about implementation strategies can make all the difference, as when we found that students tended to get higher scores on handwritten compositions than on word-processed ones (Roblyer, 1997). Raters seemed to have higher expectations for typed essays and were able to spot errors more easily. Once raters were trained to guard against this unintended bias, the rating differences disappeared. If educators know what causes effects like this, they can adjust their methods. If implementations are not studied over time, effects like these may never be known.
Study Type 3: Research to Monitor Impact on Important Societal Goals
Some technologies have been much promoted as ways of supporting and providing more equitable access to learning opportunities. Educators need to know whether or not students are gaining access as expected to the benefits that technology makes possible or if technologies are instead widening still further the Digital Divide. For example, are underserved students benefiting as much from access to Virtual High Schools as are their privileged counterparts? Are computer-based remedial strategies that are often viewed as the strategies of choice for struggling math students having the desired impact on improved achievement?
Study Type 4: Studies That Monitor and Report on Common Uses and Shape Desired Directions
Many new technologies (e.g., distance learning, handheld devices) are already in such common use that what we need now is clear evidence about what sociological impact they are having on school life and whether they are meeting their own ostensible goals. Do we have theories from sociology, psychology, or elsewhere that can help predict negative side effects and shape technology uses to make their impact more positive? For example:
- In distance education: Are course evaluations lower than with traditional face-to-face learning? If so, what should we do about that?
- With handheld devices: Do they promote cheating, as well as sharing data quickly? What should we do to implement them differently?
- Regarding use of the Internet: Does it encourage students to plagiarize? Do they get undesirable information, as well as relevant information from online sources?
Invitation to Nominate Exemplary Studies
This introductory article ends with an invitation to all educators in the field of educational technology and in the content areas to nominate studies to serve as exemplars of the criteria described here. We would like to include examples of the four types of studies as reflected in content-area research. Nominations may be submitted to CITE Journal editors for inclusion in this series.
As Roblyer (2004) noted,
There is much to be said for starting small and building results. If we are able to offer strong, consistent evidence of benefits for even a handful of technology uses, it will make it more likely that they would become methods of choice for given instructional situations.…If we can make a strong case for even a few technology-based methods as best practices, we can justify the costs of technology resources and, consequently, even more applications will come into common use. Positive results will drive future practice and research.
The need for this research foundation can hardly be overstated. If technology is to be viewed as having a clear and essential role to play in education, it must have a clearly articulated research agenda and high quality studies that both document and shape its impact. The evidence these studies yield will help drive effective classroom practice, preparation of educators of the future, and funding for technology-based instructional methods. The criteria described in this article and the model studies that will appear in later CITE Journal issues constitute one small but critically important step on the road to providing educational research that makes a difference.
1Note that some equate the term “empirical” with “experimental.” This author uses “empirical” according to the definition in the MCREL Policymakers Primer on Education Research, which calls any systematic study “empirical,” regardless of its approach to observation. Thus, scientific, qualitative, and mixed methods approaches are all empirical.
Berliner, D. (2002). Educational research: The hardest science of all. Educational Researcher, 31(8), 18-20.
Brown, J. S., Collins, A., & Duguid, P. (1989). Situated cognition and the culture of learning. Educational Researcher, 18(1), 32–41.
Bull, G., Knezek, G., Roblyer, M. D., Schrum, L., & Thompson, A. (2005). A proactive approach to a research agenda for educational technology. Journal of Research on Technology in Education, 37(3), 217-220.
Burkhart, H., & Schoenfeld, A. (2003). Improving educational research: Toward a more useful, more influential, and better-funded enterprise. Educational Researcher, 32(9), 3-14.
Chatterji, M. (2004). Evidence on “what works”: An argument for extended-term mixed-method (ETMM) evaluation designs. Educational Researcher, 33(9), 3-13.
Clark, R. (1983). Reconsidering research on learning from media. Review of Educational Research, 53(4), 445-459.
Clark, R. (1985). Evidence for confounding in computer-based instruction studies: Analyzing the meta-analyses. Educational Communications and Technology Journal, 33(4), 249-262.
Clark, R. (1991). When researchers swim upstream: Reflections on an unpopular argument about learning from media. Educational Technology, 31(2), 34-40.
Clark, R. E. (1994). Media will never influence learning. Educational Technology Research and Development, 42(2), 21-29.
Cordes, C., & Miller, E. (1999). Fools gold: A critical look at computers and childhood. College Park, MD: Alliance for Childhood. Retrieved January 8, 2004, from http://www.allianceforchildhood.net/projects/computers/computers_reports.htm
Cradler, J. (2003). Technology’s impact on teaching and learning. Learning and Leading with Technology, 30(7), 54-57.
Cuban, L., Kirkpatrick, H., & Peck, C. (2001). High access and low use of technologies in high school classrooms: Explaining an apparent paradox. American Educational Research Journal, 38(4), 813-834.
Cuban, L. (2001). Oversold and underused: Computers in the classroom. Cambridge, MA: Harvard University Press.
Kaestle, C. (1993). The awful reputation of educational research. Educational Researcher, 22(1), 23-30.
Kerlinger, F. (1973). Foundations of behavioral research (2nd ed.). New York: Holt, Rinehart, and Winston.
Kozma, R. (1991). Learning with media. Review of Educational Research, 61(2), 179-211.
Lederman, N. (2003). What works: A commentary on the nature of scientific research. Contemporary Issues in Technology and Teacher Education, 3(1). Retrieved July 14, 2005, from http://www.citejournal.org/vol3/iss1/editorial/article1.cfm
Matthew, K. (1997). A comparison of the influence of interactive CD-ROM storybooks and traditional print storybooks on reading comprehension. Journal of Research on Computing in Education, 29(3), 263-275.
Mosteller, F., Nave, B., & Miech, E. (2004). Why we need a structured abstract in education research. Educational Researcher, 33(1), 29–34.
Oppenheimer, T. (1997). The computer delusion. Atlantic Monthly, 280(1), 45–62.
Oppenheimer, T. (2003). The flickering mind: The false promise of technology in the classroom and how learning can be saved. New York: Random House.
Papert, S. (1987). Computer criticism vs. technocentric thinking. Educational Researcher, 16(1), 22–30.
Rogers, E. (2003). Diffusion of innovations (5th ed.). New York: The Free Press.
Roblyer, M. D. (1997). Technology and the oops! effect: Finding a bias against word processing. Learning and Leading With Technology, 24(7), 14-16.
Roblyer, M.D. (2004). If technology is the answer, what’s the question? Research to help make the case for why we use technology in teaching. Technology and Teacher Education Annual, 2004. Charlottesville, VA: Association for the Advancement of Computing.
Roblyer, M. D., & Knezek, G. (2003). New millennium research for educational technology: A call for a national research agenda. Journal of Research on Technology in Education, 36(1), 60-71.
Stoll, C. (1995). Silicon snake oil: Second thoughts on the information highway. New York: Doubleday.
Thompson, A., Bull, G., & Bell, L. (2005). A proactive approach to a research agenda: A call to action. Journal of Computing in Teacher Education, 21(3), 74, 76.
37 total views, 6 views today