Quantitative research in education : Background information

  • Background information
  • SAGE researchmethods SAGE Research Methods is a tool created to help researchers, faculty and students with their research projects. Users can explore methods concepts to help them design research projects, understand particular methods or identify a new method, conduct their research, and write up their findings. Since SAGE Research Methods focuses on methodology rather than disciplines, it can be used across the social sciences, health sciences, and other areas of research.

Cover Art

  • The American freshman, national norms for ... From the Higher Education Research Institute, University of California, Los Angeles
  • Education at a glance : OECD indicators
  • Global education digest From UNESCO
  • Next: Recent e-books >>
  • Recent e-books
  • Recent print books
  • Connect to Stanford e-resources

Profile Photo

  • Last Updated: Jan 23, 2024 12:46 PM
  • URL: https://guides.library.stanford.edu/quantitative_research_in_ed
  • Our Mission

How to Read and Interpret Research to Benefit Your Teaching Practice

Teachers can find helpful ideas in research articles and take a strategic approach to get the most out of what they’re reading.

Photo of teacher working at home

Have you read any education blogs, attended a conference session this summer, or gone to a back-to-school meeting so far where information on PowerPoint slides was supported with research like this: “Holland et al., 2023”? Perhaps, like me, you’ve wondered what to do with these citations or how to find and read the work cited. We want to improve our teaching practice and keep learning amid our busy schedules and responsibilities. When we find a sliver of time to look for the research article(s) being cited, how are we supposed to read, interpret, implement, and reflect on it in our practice? 

There has been much research over the past decade building on research-practice partnerships . Teachers and researchers should work collaboratively to improve student learning. Though researchers in higher education typically conduct formal research and publish their work in journal articles, it’s important for teachers to also see themselves as researchers. They engage in qualitative analysis while circulating the room to examine and interpret student work and demonstrate quantitative analysis when making predictions around student achievement data.

There are different sources of knowledge and timely questions to consider that education researchers can learn and take from teachers. So, what if teachers were better equipped to translate research findings from a journal article into improved practice relevant to their classroom’s immediate needs? I’ll offer some suggestions on how to answer this question.

Removing Barriers to New Information

For starters, research is crucial for education. It helps us learn and create new knowledge. Teachers learning how to translate research into practice can help contribute toward continuous improvement in schools. However, not all research is beneficial or easily applicable. While personal interests may lead researchers in a different direction, your classroom experience holds valuable expertise. Researchers should be viewed as allies, not sole authorities.

Additionally, paywalls prevent teachers from accessing valuable research articles that are often referenced in professional development. However, some sites, like Sage and JSTOR , offer open access journals where you can find research relevant to your classroom needs. Google Scholar is another helpful resource where you can plug in keywords like elementary math , achievement , small-group instruction , or diverse learners to find articles freely available as PDFs. Alternatively, you can use Elicit and get answers to specific questions. It can provide a list of relevant articles and summaries of their findings.

Approach research articles differently than other types of writing, as they aren’t intended for our specific audience but rather for academic researchers. Keep this in mind when selecting articles that align with your teaching vision, student demographic, and school environment.

Using behavioral and brain science research, I implemented the spacing effect . I used this strategy to include spaced fluency, partner practices, and spiral reviews (e.g., “do nows”) with an intentional selection of questions and tasks based on student work samples and formative/summative assessment data. It improved my students’ memory, long-term retention, and proficiency, so I didn’t take it too personally when some of them forgot procedures or symbols.

What You’ll Find in a Research Article

Certain elements are always included in a research article. The abstract gives a brief overview. Following that, the introduction typically explains the purpose and significance of the research—often through a theoretical framework and literature review. Other common sections of a research article may include methodology, results or findings, and discussion or conclusion.

The methodology section explains how the researchers answered their research question(s) to understand the topic. The results/findings section provides the answer(s) to the research question(s), while the discussion/conclusion section explains the importance and meaning of the results/findings and why it matters to readers and the field of education at large.

How to Process Information to Find What You’re Looking For

To avoid getting overwhelmed while reading research, take notes. Many articles are lengthy and filled with complex terminology and citations. Choose one relevant article at a time, and jot down important points or questions.

You could apply many strategies to read research, but here’s an idea that takes our time constraints and bandwidth as teachers into account:

  • First, read the title and full abstract, then scan and skim the introduction. You’ll be able to see if it’s relevant to your interests, needs, and whether you need to continue reading. 
  • After you’ve decided if the research is relevant to your classroom and professional development, jump straight to the discussion/conclusion section to see the “so what” about the research findings and how they could apply to your classroom. Review the findings/results section after for more details if needed.

Decipher the Details in the Data 

As a math, science, or English language arts teacher, you might come across figures, tables, or graphs that could spark ideas for your lessons. Some of these visuals and data may seem complex and difficult to understand. To make sense of them, take it slow and read through the notes and descriptions carefully.             

For example, researchers C. Kirabo Jackson and Alexey Makarin created a graph to show that middle school math teachers who had online access and support to use high-quality materials saw a positive impact on math test scores, especially when they used the materials for multiple lessons. The notes below the graph explain how the data was collected and which school districts were involved in the study.

Lastly, after reading the findings/results section, you’ll understand the gist of the research and if it’s applicable to your needs. Reading beyond these sections depends on your schedule and interests. It’s perfectly normal if it takes additional time to digest these sections.

When it comes to reading research, teachers don’t have to go it alone. School and district leaders can involve us in discussions about research findings and their practical implications for our school during professional learning community meetings or professional development sessions before the start of the school year. Even if only a few teachers participate in this process, sharing the main points with peers and the principal can have a significantly positive impact on improving direct instruction for students.

  • Find My Rep

You are here

What are the benefits of educational research for teachers.

Ask an Expert Rebecca Austin Researching Primary Education

Cultivating a research-based approach to developing your practice provides evidence to effect change in your teaching, your classroom, your school, and beyond. Rebecca Austin, author of Researching Primary Education  and Senior Lecturer at the School of Teacher Education and Development at Canterbury Christchurch University, highlights what the benefits are of research to your practice…

In the context of the debate about what works and why, there is a wide range of benefits to researching your own practice, whether directly feeding into improvement through action research or, more broadly, gaining understanding and knowledge on themes of interest and relevance. This is why research is embedded into initial teacher education. As research becomes embedded in your practice you can gain a range of benefits. Research can:

  • clarify purposes, processes and priorities when introducing change – for example, to  curriculum, pedagogy or assessment  
  • develop your agency, influence, self-efficacy and voice within your own school and  more widely within the profession.

Each of these can involve investigation using evidence from your own setting, along with wider research evidence. 

Chapter Icon

  • Site search

CBT Supervision

The ABC of CBT

CBT for Beginners

CBT Values and Ethics

Reflection in CBT

CBT for Older People

Overcoming Obstacles in CBT

The CBT Handbook

CBT for Personality Disorders

CBMCS Multicultural Training Program

CBMCS Multicultural Reader

CBDNA Journal: Research & Review

An Introduction to CBT Research

CBT for Common Trauma Responses

Person-centred Therapy and CBT

Low-intensity CBT Skills and Interventions

CBT for Depression: An Integrated Approach

CBT with Children, Young People and Families

CBT for Worry and Generalised Anxiety Disorder

Action Research

Journal of Research in Nursing

Product Type plus Created with Sketch. minus Created with Sketch.

  • Textbook (32) Apply Textbook filter
  • Journal (13) Apply Journal filter
  • Academic Book (5) Apply Academic Book filter
  • Professional Book (4) Apply Professional Book filter
  • Reference Book (4) Apply Reference Book filter

Disciplines plus Created with Sketch. minus Created with Sketch.

  • Education (31) Apply Education filter
  • Counselling and Psychotherapy (General) (18) Apply Counselling and Psychotherapy (General) filter
  • Research Methods & Evaluation (General) (18) Apply Research Methods & Evaluation (General) filter
  • Nursing (6) Apply Nursing filter
  • Public Health (4) Apply Public Health filter
  • Psychology (General) (3) Apply Psychology (General) filter
  • Social Work & Social Policy (General) (3) Apply Social Work & Social Policy (General) filter
  • Clinical Medicine (3) Apply Clinical Medicine filter
  • Anthropology & Archaeology (General) (1) Apply Anthropology & Archaeology (General) filter
  • Arts & Humanities (General) (1) Apply Arts & Humanities (General) filter
  • History (General) (1) Apply History (General) filter
  • Business & Management (General) (1) Apply Business & Management (General) filter
  • Communication and Media Studies (General) (1) Apply Communication and Media Studies (General) filter
  • Cultural Studies (General) (1) Apply Cultural Studies (General) filter
  • Economics & Development Studies (General) (1) Apply Economics & Development Studies (General) filter
  • Life & Biomedical Sciences (1) Apply Life & Biomedical Sciences filter
  • Politics & International Relations (1) Apply Politics & International Relations filter
  • Study Skills (General) (1) Apply Study Skills (General) filter
  • Other Health Specialties (1) Apply Other Health Specialties filter

Status plus Created with Sketch. minus Created with Sketch.

  • Published (44) Apply Published filter
  • Forthcoming (1) Apply Forthcoming filter
  • Copy/Paste Link Link Copied

Using Research and Reason in Education: How Teachers Can Use Scientifically Based Research to Make Curricular & Instructional Decisions

Paula J. Stanovich and Keith E. Stanovich University of Toronto

Produced by RMC Research Corporation, Portsmouth, New Hampshire

This publication was produced under National Institute for Literacy Contract No. ED-00CO-0093 with RMC Research Corporation. Sandra Baxter served as the contracting officer's technical representative. The views expressed herein do not necessarily represent the policies of the National Institute for Literacy. No official endorsement by the National Institute for Literacy or any product, commodity, service, or enterprise is intended or should be inferred.

The National Institute for Literacy

Sandra Baxter, Interim Executive Director Lynn Reddy, Communications Director

To order copies of this booklet, contact the National Institute for Literacy at EdPubs, PO Box 1398, Jessup, MD 20794-1398. Call 800-228-8813 or email [email protected] .

The National Institute for Literacy, an independent federal organization, supports the development of high quality state, regional, and national literacy services so that all Americans can develop the literacy skills they need to succeed at work, at home, and in the community.

The Partnership for Reading, a project administered by the National Institute for Literacy, is a collaborative effort of the National Institute for Literacy, the National Institute of Child Health and Human Development, the U.S. Department of Education, and the U.S. Department of Health and Human Services to make evidence-based reading research available to educators, parents, policy makers, and others with an interest in helping all people learn to read well.

Editorial support provided by C. Ralph Adler and Elizabeth Goldman, and design/production support provided by Diane Draper and Bob Kozman, all of RMC Research Corporation.

Introduction

In the recent move toward standards-based reform in public education, many educational reform efforts require schools to demonstrate that they are achieving educational outcomes with students performing at a required level of achievement. Federal and state legislation, in particular, has codified this standards-based movement and tied funding and other incentives to student achievement.

At first, demonstrating student learning may seem like a simple task, but reflection reveals that it is a complex challenge requiring educators to use specific knowledge and skills. Standards-based reform has many curricular and instructional prerequisites. The curriculum must represent the most important knowledge, skills, and attributes that schools want their students to acquire because these learning outcomes will serve as the basis of assessment instruments. Likewise, instructional methods should be appropriate for the designed curriculum. Teaching methods should lead to students learning the outcomes that are the focus of the assessment standards.

Standards- and assessment-based educational reforms seek to obligate schools and teachers to supply evidence that their instructional methods are effective. But testing is only one of three ways to gather evidence about the effectiveness of instructional methods. Evidence of instructional effectiveness can come from any of the following sources:

  • Demonstrated student achievement in formal testing situations implemented by the teacher, school district, or state;
  • Published findings of research-based evidence that the instructional methods being used by teachers lead to student achievement; or
  • Proof of reason-based practice that converges with a research-based consensus in the scientific literature. This type of justification of educational practice becomes important when direct evidence may be lacking (a direct test of the instructional efficacy of a particular method is absent), but there is a theoretical link to research-based evidence that can be traced.

Each of these methods has its pluses and minuses. While testing seems the most straightforward, it is not necessarily the clear indicator of good educational practice that the public seems to think it is. The meaning of test results is often not immediately clear. For example, comparing averages or other indicators of overall performance from tests across classrooms, schools, or school districts takes no account of the resources and support provided to a school, school district, or individual professional. Poor outcomes do not necessarily indict the efforts of physicians in Third World countries who work with substandard equipment and supplies. Likewise, objective evidence of below-grade or below-standard mean performance of a group of students should not necessarily indict their teachers if essential resources and supports (e.g., curriculum materials, institutional aid, parental cooperation) to support teaching efforts were lacking. However, the extent to which children could learn effectively even in under-equipped schools is not known because evidence-based practices are, by and large, not implemented. That is, there is evidence that children experiencing academic difficulties can achieve more educationally if they are taught with effective methods; sadly, scientific research about what works does not usually find its way into most classrooms.

Testing provides a useful professional calibrator, but it requires great contextual sensitivity in interpretation. It is not the entire solution for assessing the quality of instructional efforts. This is why research-based and reason-based educational practice are also crucial for determining the quality and impact of programs. Teachers thus have the responsibility to be effective users and interpreters of research. Providing a survey and synthesis of the most effective practices for a variety of key curriculum goals (such as literacy and numeracy) would seem to be a helpful idea, but no document could provide all of that information. (Many excellent research syntheses exist, such as the National Reading Panel, 2000; Snow, Burns, & Griffin, 1998; Swanson, 1999, but the knowledge base about effective educational practices is constantly being updated, and many issues remain to be settled.)

As professionals, teachers can become more effective and powerful by developing the skills to recognize scientifically based practice and, when the evidence is not available, use some basic research concepts to draw conclusions on their own. This paper offers a primer for those skills that will allow teachers to become independent evaluators of educational research.

The Formal Scientific Method and Scientific Thinking in Educational Practice

When you go to your family physician with a medical complaint, you expect that the recommended treatment has proven to be effective with many other patients who have had the same symptoms. You may even ask why a particular medication is being recommended for you. The doctor may summarize the background knowledge that led to that recommendation and very likely will cite summary evidence from the drug's many clinical trials and perhaps even give you an overview of the theory behind the drug's success in treating symptoms like yours.

All of this discussion will probably occur in rather simple terms, but that does not obscure the fact that the doctor has provided you with data to support a theory about your complaint and its treatment. The doctor has shared knowledge of medical science with you. And while everyone would agree that the practice of medicine has its "artful" components (for example, the creation of a healing relationship between doctor and patient), we have come to expect and depend upon the scientific foundation that underpins even the artful aspects of medical treatment. Even when we do not ask our doctors specifically for the data, we assume it is there, supporting our course of treatment.

Actually, Vaughn and Dammann (2001) have argued that the correct analogy is to say that teaching is in part a craft, rather than an art. They point out that craft knowledge is superior to alternative forms of knowledge such as superstition and folklore because, among other things, craft knowledge is compatible with scientific knowledge and can be more easily integrated with it. One could argue that in this age of education reform and accountability, educators are being asked to demonstrate that their craft has been integrated with science--that their instructional models, methods, and materials can be likened to the evidence a physician should be able to produce showing that a specific treatment will be effective. As with medicine, constructing teaching practice on a firm scientific foundation does not mean denying the craft aspects of teaching.

Architecture is another professional practice that, like medicine and education, grew from being purely a craft to a craft based firmly on a scientific foundation. Architects wish to design beautiful buildings and environments, but they must also apply many foundational principles of engineering and adhere to structural principles. If they do not, their buildings, however beautiful they may be, will not stand. Similarly, a teacher seeks to design lessons that stimulate students and entice them to learn--lessons that are sometimes a beauty to behold. But if the lessons are not based in the science of pedagogy, they, like poorly constructed buildings, will fail.

Education is informed by formal scientific research through the use of archival research-based knowledge such as that found in peer-reviewed educational journals. Preservice teachers are first exposed to the formal scientific research in their university teacher preparation courses (it is hoped), through the instruction received from their professors, and in their course readings (e.g., textbooks, journal articles). Practicing teachers continue their exposure to the results of formal scientific research by subscribing to and reading professional journals, by enrolling in graduate programs, and by becoming lifelong learners.

Scientific thinking in practice is what characterizes reflective teachers--those who inquire into their own practice and who examine their own classrooms to find out what works best for them and their students. What follows in this document is, first, a "short course" on how to become an effective consumer of the archival literature that results from the conduct of formal scientific research in education and, second, a section describing how teachers can think scientifically in their ongoing reflection about their classroom practice.

Being able to access mechanisms that evaluate claims about teaching methods and to recognize scientific research and its findings is especially important for teachers because they are often confronted with the view that "anything goes" in the field of education--that there is no such thing as best practice in education, that there are no ways to verify what works best, that teachers should base their practice on intuition, or that the latest fad must be the best way to teach, please a principal, or address local school reform. The "anything goes" mentality actually represents a threat to teachers' professional autonomy. It provides a fertile environment for gurus to sell untested educational "remedies" that are not supported by an established research base.

Teachers as independent evaluators of research evidence

One factor that has impeded teachers from being active and effective consumers of educational science has been a lack of orientation and training in how to understand the scientific process and how that process results in the cumulative growth of knowledge that leads to validated educational practice. Educators have only recently attempted to resolve educational disputes scientifically, and teachers have not yet been armed with the skills to evaluate disputes on their own.

Educational practice has suffered greatly because its dominant model for resolving or adjudicating disputes has been more political (with its corresponding factions and interest groups) than scientific. The field's failure to ground practice in the attitudes and values of science has made educators susceptible to the "authority syndrome" as well as fads and gimmicks that ignore evidence-based practice.

When our ancestors needed information about how to act, they would ask their elders and other wise people. Contemporary society and culture are much more complex. Mass communication allows virtually anyone (on the Internet, through self-help books) to proffer advice, to appear to be a "wise elder." The current problem is how to sift through the avalanche of misguided and uninformed advice to find genuine knowledge. Our problem is not information; we have tons of information. What we need are quality control mechanisms.

Peer-reviewed research journals in various disciplines provide those mechanisms. However, even with mechanisms like these in behavioral science and education, it is all too easy to do an "end run" around the quality control they provide. Powerful information dissemination outlets such as publishing houses and mass media frequently do not discriminate between good and bad information. This provides a fertile environment for gurus to sell untested educational "remedies" that are not supported by an established research base and, often, to discredit science, scientific evidence, and the notion of research-based best practice in education. As Gersten (2001) notes, both seasoned and novice teachers are "deluged with misinformation" (p. 45).

We need tools for evaluating the credibility of these many and varied sources of information; the ability to recognize research-based conclusions is especially important. Acquiring those tools means understanding scientific values and learning methods for making inferences from the research evidence that arises through the scientific process. These values and methods were recently summarized by a panel of the National Academy of Sciences convened on scientific inquiry in education (Shavelson & Towne, 2002), and our discussion here will be completely consistent with the conclusions of that NAS panel.

The scientific criteria for evaluating knowledge claims are not complicated and could easily be included in initial teacher preparation programs, but they usually are not (which deprives teachers from an opportunity to become more efficient and autonomous in their work right at the beginning of their careers). These criteria include:

  • the publication of findings in refereed journals (scientific publications that employ a process of peer review),
  • the duplication of the results by other investigators, and
  • a consensus within a particular research community on whether there is a critical mass of studies that point toward a particular conclusion.

In their discussion of the evolution of the American Educational Research Association (AERA) conference and the importance of separating research evidence from opinion when making decisions about instructional practice, Levin and O'Donnell (2000) highlight the importance of enabling teachers to become independent evaluators of research evidence. Being aware of the importance of research published in peer-reviewed scientific journals is only the first step because this represents only the most minimal of criteria. Following is a review of some of the principles of research-based evaluation that teachers will find useful in their work.

Publicly verifiable research conclusions: Replication and Peer Review

Source credibility: the consumer protection of peer reviewed journals..

The front line of defense for teachers against incorrect information in education is the existence of peer-reviewed journals in education, psychology, and other related social sciences. These journals publish empirical research on topics relevant to classroom practice and human cognition and learning. They are the first place that teachers should look for evidence of validated instructional practices.

As a general quality control mechanism, peer review journals provide a "first pass" filter that teachers can use to evaluate the plausibility of educational claims. To put it more concretely, one ironclad criterion that will always work for teachers when presented with claims of uncertain validity is the question: Have findings supporting this method been published in recognized scientific journals that use some type of peer review procedure? The answer to this question will almost always separate pseudoscientific claims from the real thing.

In a peer review, authors submit a paper to a journal for publication, where it is critiqued by several scientists. The critiques are reviewed by an editor (usually a scientist with an extensive history of work in the specialty area covered by the journal). The editor then decides whether the weight of opinion warrants immediate publication, publication after further experimentation and statistical analysis, or rejection because the research is flawed or does not add to the knowledge base. Most journals carry a statement of editorial policy outlining their exact procedures for publication, so it is easy to check whether a journal is in fact, peer-reviewed.

Peer review is a minimal criterion, not a stringent one. Not all information in peer-reviewed scientific journals is necessarily correct, but it has at the very least undergone a cycle of peer criticism and scrutiny. However, it is because the presence of peer-reviewed research is such a minimal criterion that its absence becomes so diagnostic. The failure of an idea, a theory, an educational practice, behavioral therapy, or a remediation technique to have adequate documentation in the peer-reviewed literature of a scientific discipline is a very strong indication to be wary of the practice.

The mechanisms of peer review vary somewhat from discipline to discipline, but the underlying rationale is the same. Peer review is one way (replication of a research finding is another) that science institutionalizes the attitudes of objectivity and public criticism. Ideas and experimentation undergo a honing process in which they are submitted to other critical minds for evaluation. Ideas that survive this critical process have begun to meet the criterion of public verifiability. The peer review process is far from perfect, but it really is the only external consumer protection that teachers have.

The history of reading instruction illustrates the high cost that is paid when the peer-reviewed literature is ignored, when the normal processes of scientific adjudication are replaced with political debates and rhetorical posturing. A vast literature has been generated on best practices that foster children's reading acquisition (Adams, 1990; Anderson, Hiebert, Scott, & Wilkinson, 1985; Chard & Osborn, 1999; Cunningham & Allington, 1994; Ehri, Nunes, Stahl, & Willows, 2001; Moats, 1999; National Reading Panel, 2000; Pearson, 1993; Pressley, 1998; Pressley, Rankin, & Yokol, 1996; Rayner, Foorman, Perfetti, Pesetsky, & Seidenberg, 2002; Reading Coherence Initiative, 1999; Snow, Burns, & Griffin, 1998; Spear-Swerling & Sternberg, 2001). Yet much of this literature remains unknown to many teachers, contributing to the frustrating lack of clarity about accepted, scientifically validated findings and conclusions on reading acquisition.

Teachers should also be forewarned about the difference between professional education journals that are magazines of opinion in contrast to journals where primary reports of research, or reviews of research, are peer reviewed. For example, the magazines Phi Delta Kappan and Educational Leadership both contain stimulating discussions of educational issues, but neither is a peer-reviewed journal of original research. In contrast, the American Educational Research Journal (a flagship journal of the AERA) and the Journal of Educational Psychology (a flagship journal of the American Psychological Association) are both peer-reviewed journals of original research. Both are main sources for evidence on validated techniques of reading instruction and for research on aspects of the reading process that are relevant to a teacher's instructional decisions.

This is true, too, of presentations at conferences of educational organizations. Some are data-based presentations of original research. Others are speeches reflecting personal opinion about educational problems. While these talks can be stimulating and informative, they are not a substitute for empirical research on educational effectiveness.

Replication and the importance of public verifiability.

Research-based conclusions about educational practice are public in an important sense: they do not exist solely in the mind of a particular individual but have been submitted to the scientific community for criticism and empirical testing by others. Knowledge considered "special"--the province of the thought of an individual and immune from scrutiny and criticism by others--can never have the status of scientific knowledge. Research-based conclusions, when published in a peer reviewed journal, become part of the public realm, available to all, in a way that claims of "special expertise" are not.

Replication is the second way that science uses to make research-based conclusions concrete and "public." In order to be considered scientific, a research finding must be presented to other researchers in the scientific community in a way that enables them to attempt the same experiment and obtain the same results. When the same results occur, the finding has been replicated . This process ensures that a finding is not the result of the errors or biases of a particular investigator. Replicable findings become part of the converging evidence that forms the basis of a research-based conclusion about educational practice.

John Donne told us that "no man is an island." Similarly, in science, no researcher is an island. Each investigator is connected to the research community and its knowledge base. This interconnection enables science to grow cumulatively and for research-based educational practice to be built on a convergence of knowledge from a variety of sources. Researchers constantly build on previous knowledge in order to go beyond what is currently known. This process is possible only if research findings are presented in such a way that any investigator can use them to build on.

Philosopher Daniel Dennett (1995) has said that science is "making mistakes in public. Making mistakes for all to see, in the hopes of getting the others to help with the corrections" (p. 380). We might ask those proposing an educational innovation for the evidence that they have in fact "made some mistakes in public." Legitimate scientific disciplines can easily provide such evidence. For example, scientists studying the psychology of reading once thought that reading difficulties were caused by faulty eye movements. This hypothesis has been shown to be in error, as has another that followed it, that so-called visual reversal errors were a major cause of reading difficulty. Both hypotheses were found not to square with the empirical evidence (Rayner, 1998; Share & Stanovich, 1995). The hypothesis that reading difficulties can be related to language difficulties at the phonological level has received much more support (Liberman, 1999; National Reading Panel, 2000; Rayner, Foorman, Perfetti, Pesetsky, & Seidenberg, 2002; Shankweiler, 1999; Stanovich, 2000).

After making a few such "errors" in public, reading scientists have begun, in the last 20 years, to get it right. But the only reason teachers can have confidence that researchers are now "getting it right" is that researchers made it open, public knowledge when they got things wrong. Proponents of untested and pseudoscientific educational practices will never point to cases where they "got it wrong" because they are not committed to public knowledge in the way that actual science is. These proponents do not need, as Dennett says, "to get others to help in making the corrections" because they have no intention of correcting their beliefs and prescriptions based on empirical evidence.

Education is so susceptible to fads and unproven practices because of its tacit endorsement of a personalistic view of knowledge acquisition--one that is antithetical to the scientific value of the public verifiability of knowledge claims. Many educators believe that knowledge resides within particular individuals--with particularly elite insights--who then must be called upon to dispense this knowledge to others. Indeed, some educators reject public, depersonalized knowledge in social science because they believe it dehumanizes people. Science, however, with its conception of publicly verifiable knowledge, actually democratizes knowledge. It frees practitioners and researchers from slavish dependence on authority.

Subjective, personalized views of knowledge degrade the human intellect by creating conditions that subjugate it to an elite whose "personal" knowledge is not accessible to all (Bronowski, 1956, 1977; Dawkins, 1998; Gross, Levitt, & Lewis, 1997; Medawar, 1982, 1984, 1990; Popper, 1972; Wilson, 1998). Empirical science, by generating knowledge and moving it into the public domain, is a liberating force. Teachers can consult the research and decide for themselves whether the state of the literature is as the expert portrays it. All teachers can benefit from some rudimentary grounding in the most fundamental principles of scientific inference. With knowledge of a few uncomplicated research principles, such as control, manipulation, and randomization, anyone can enter the open, public discourse about empirical findings. In fact, with the exception of a few select areas such as the eye movement research mentioned previously, much of the work described in noted summaries of reading research (e.g., Adams, 1990; Snow, Burns, & Griffin, 1998) could easily be replicated by teachers themselves.

There are many ways that the criteria of replication and peer review can be utilized in education to base practitioner training on research-based best practice. Take continuing teacher education in the form of inservice sessions, for example. Teachers and principals who select speakers for professional development activities should ask speakers for the sources of their conclusions in the form of research evidence in peer-reviewed journals. They should ask speakers for bibliographies of the research evidence published on the practices recommended in their presentations.

The science behind research-based practice relies on systematic empiricism

Empiricism is the practice of relying on observation. Scientists find out about the world by examining it. The refusal by some scientists to look into Galileo's telescope is an example of how empiricism has been ignored at certain points in history. It was long believed that knowledge was best obtained through pure thought or by appealing to authority. Galileo claimed to have seen moons around the planet Jupiter. Another scholar, Francesco Sizi, attempted to refute Galileo, not with observations, but with the following argument:

There are seven windows in the head, two nostrils, two ears, two eyes and a mouth; so in the heavens there are two favorable stars, two unpropitious, two luminaries, and Mercury alone undecided and indifferent. From which and many other similar phenomena of nature such as the seven metals, etc., which it were tedious to enumerate, we gather that the number of planets is necessarily seven...ancient nations, as well as modern Europeans, have adopted the division of the week into seven days, and have named them from the seven planets; now if we increase the number of planets, this whole system falls to the ground...moreover, the satellites are invisible to the naked eye and therefore can have no influence on the earth and therefore would be useless and therefore do not exist. (Holton & Roller, 1958, p. 160)

Three centuries of the demonstrated power of the empirical approach give us an edge on poor Sizi. Take away those years of empiricism, and many of us might have been there nodding our heads and urging him on. In fact, the empirical approach is not necessarily obvious, which is why we often have to teach it, even in a society that is dominated by science.

Empiricism pure and simple is not enough, however. Observation itself is fine and necessary, but pure, unstructured observation of the natural world will not lead to scientific knowledge. Write down every observation you make from the time you get up in the morning to the time you go to bed on a given day. When you finish, you will have a great number of facts, but you will not have a greater understanding of the world. Scientific observation is termed systematic because it is structured so that the results of the observation reveal something about the underlying causal structure of events in the world. Observations are structured so that, depending upon the outcome of the observation, some theories of the causes of the outcome are supported and others rejected.

Teachers can benefit by understanding two things about research and causal inferences. The first is the simple (but sometimes obscured) fact that statements about best instructional practices are statements that contain a causal claim. These statements claim that one type of method or practice causes superior educational outcomes. Second, teachers must understand how the logic of the experimental method provides the critical support for making causal inferences.

Science addresses testable questions

Science advances by positing theories to account for particular phenomena in the world, by deriving predictions from these theories, by testing the predictions empirically, and by modifying the theories based on the tests (the sequence is typically theory -> prediction -> test -> theory modification). What makes a theory testable? A theory must have specific implications for observable events in the natural world.

Science deals only with a certain class of problem: the kind that is empirically solvable. That does not mean that different classes of problems are inherently solvable or unsolvable and that this division is fixed forever. Quite the contrary: some problems that are currently unsolvable may become solvable as theory and empirical techniques become more sophisticated. For example, decades ago historians would not have believed that the controversial issue of whether Thomas Jefferson had a child with his slave Sally Hemings was an empirically solvable question. Yet, by 1998, this problem had become solvable through advances in genetic technology, and a paper was published in the journal Nature (Foster, Jobling, Taylor, Donnelly, Deknijeff, Renemieremet, Zerjal, & Tyler-Smith, 1998) on the question.

The criterion of whether a problem is "testable" is called the falsifiability criterion: a scientific theory must always be stated in such a way that the predictions derived from it can potentially be shown to be false. The falsifiability criterion states that, for a theory to be useful, the predictions drawn from it must be specific. The theory must go out on a limb, so to speak, because in telling us what should happen, the theory must also imply that certain things will not happen. If these latter things do happen, it is a clear signal that something is wrong with the theory. It may need to be modified, or we may need to look for an entirely new theory. Either way, we will end up with a theory that is closer to the truth.

In contrast, if a theory does not rule out any possible observations, then the theory can never be changed, and we are frozen into our current way of thinking with no possibility of progress. A successful theory cannot posit or account for every possible happening. Such a theory robs itself of any predictive power.

What we are talking about here is a certain type of intellectual honesty. In science, the proponent of a theory is always asked to address this question before the data are collected: "What data pattern would cause you to give up, or at least to alter, this theory?" In the same way, the falsifiability criterion is a useful consumer protection for the teacher when evaluating claims of educational effectiveness. Proponents of an educational practice should be asked for evidence; they should also be willing to admit that contrary data will lead them to abandon the practice. True scientific knowledge is held tentatively and is subject to change based on contrary evidence. Educational remedies not based on scientific evidence will often fail to put themselves at risk by specifying what data patterns would prove them false.

Objectivity and intellectual honesty

Objectivity, another form of intellectual honesty in research, means that we let nature "speak for itself" without imposing our wishes on it--that we report the results of experimentation as accurately as we can and that we interpret them as fairly as possible. (The fact that this goal is unattainable for any single human being should not dissuade us from holding objectivity as a value.)

In the language of the general public, open-mindedness means being open to possible theories and explanations for a particular phenomenon. But in science it means that and something more. Philosopher Jonathan Adler (1998) teaches us that science values another aspect of open-mindedness even more highly: "What truly marks an open-minded person is the willingness to follow where evidence leads. The open-minded person is willing to defer to impartial investigations rather than to his own predilections...Scientific method is attunement to the world, not to ourselves" (p. 44).

Objectivity is critical to the process of science, but it does not mean that such attitudes must characterize each and every scientist for science as a whole to work. Jacob Bronowski (1973, 1977) often argued that the unique power of science to reveal knowledge about the world does not arise because scientists are uniquely virtuous (that they are completely objective or that they are never biased in interpreting findings, for example). It arises because fallible scientists are immersed in a process of checks and balances --a process in which scientists are always there to criticize and to root out errors. Philosopher Daniel Dennett (1999/2000) points out that "scientists take themselves to be just as weak and fallible as anybody else, but recognizing those very sources of error in themselvesÉthey have devised elaborate systems to tie their own hands, forcibly preventing their frailties and prejudices from infecting their results" (p. 42). More humorously, psychologist Ray Nickerson (1998) makes the related point that the vanities of scientists are actually put to use by the scientific process, by noting that it is "not so much the critical attitude that individual scientists have taken with respect to their own ideas that has given science its success...but more the fact that individual scientists have been highly motivated to demonstrate that hypotheses that are held by some other scientists are false" (p. 32). These authors suggest that the strength of scientific knowledge comes not because scientists are virtuous, but from the social process where scientists constantly cross-check each others' knowledge and conclusions.

The public criteria of peer review and replication of findings exist in part to keep checks on the objectivity of individual scientists. Individuals cannot hide bias and nonobjectivity by personalizing their claims and keeping them from public scrutiny. Science does not accept findings that have failed the tests of replication and peer review precisely because it wants to ensure that all findings in science are in the public domain, as defined above. Purveyors of pseudoscientific educational practices fail the test of objectivity and are often identifiable by their attempts to do an "end run" around the public mechanisms of science by avoiding established peer review mechanisms and the information-sharing mechanisms that make replication possible. Instead, they attempt to promulgate their findings directly to consumers, such as teachers.

The principle of converging evidence

The principle of converging evidence has been well illustrated in the controversies surrounding the teaching of reading. The methods of systematic empiricism employed in the study of reading acquisition are many and varied. They include case studies, correlational studies, experimental studies, narratives, quasi-experimental studies, surveys, epidemiological studies and many others. The results of many of these studies have been synthesized in several important research syntheses (Adams, 1990; Ehri et al., 2001; National Reading Panel, 2000; Pressley, 1998; Rayner et al., 2002; Reading Coherence Initiative, 1999; Share & Stanovich, 1995; Snow, Burns, & Griffin, 1998; Snowling, 2000; Spear-Swerling & Sternberg, 2001; Stanovich, 2000). These studies were used in a process of establishing converging evidence, a principle that governs the drawing of the conclusion that a particular educational practice is research-based.

The principle of converging evidence is applied in situations requiring a judgment about where the "preponderance of evidence" points. Most areas of science contain competing theories. The extent to which a particular study can be seen as uniquely supporting one particular theory depends on whether other competing explanations have been ruled out. A particular experimental result is never equally relevant to all competing theories. An experiment may be a very strong test of one or two alternative theories but a weak test of others. Thus, research is considered highly convergent when a series of experiments consistently supports a given theory while collectively eliminating the most important competing explanations. Although no single experiment can rule out all alternative explanations, taken collectively, a series of partially diagnostic experiments can lead to a strong conclusion if the data converge.

Contrast this idea of converging evidence with the mistaken view that a problem in science can be solved with a single, crucial experiment, or that a single critical insight can advance theory and overturn all previous knowledge. This view of scientific progress fits nicely with the operation of the news media, in which history is tracked by presenting separate, disconnected "events" in bite-sized units. This is a gross misunderstanding of scientific progress and, if taken too seriously, leads to misconceptions about how conclusions are reached about research-based practices.

One experiment rarely decides an issue, supporting one theory and ruling out all others. Issues are most often decided when the community of scientists gradually begins to agree that the preponderance of evidence supports one alternative theory rather than another. Scientists do not evaluate data from a single experiment that has finally been designed in the perfect way. They most often evaluate data from dozens of experiments, each containing some flaws but providing part of the answer.

Although there are many ways in which an experiment can go wrong (or become confounded ), a scientist with experience working on a particular problem usually has a good idea of what most of the critical factors are, and there are usually only a few. The idea of converging evidence tells us to examine the pattern of flaws running through the research literature because the nature of this pattern can either support or undermine the conclusions that we might draw.

For example, suppose that the findings from a number of different experiments were largely consistent in supporting a particular conclusion. Given the imperfect nature of experiments, we would evaluate the extent and nature of the flaws in these studies. If all the experiments were flawed in a similar way, this circumstance would undermine confidence in the conclusions drawn from them because the consistency of the outcome may simply have resulted from a particular, consistent flaw. On the other hand, if all the experiments were flawed in different ways, our confidence in the conclusions increases because it is less likely that the consistency in the results was due to a contaminating factor that confounded all the experiments. As Anderson and Anderson (1996) note, "When a conceptual hypothesis survives many potential falsifications based on different sets of assumptions, we have a robust effect." (p. 742).

Suppose that five different theoretical summaries (call them A, B, C, D, and E) of a given set of phenomena exist at one time and are investigated in a series of experiments. Suppose that one set of experiments represents a strong test of theories A, B, and C, and that the data largely refute theories A and B and support C. Imagine also that another set of experiments is a particularly strong test of theories C, D, and E, and that the data largely refute theories D and E and support C. In such a situation, we would have strong converging evidence for theory C. Not only do we have data supportive of theory C, but we have data that contradict its major competitors. Note that no one experiment tests all the theories, but taken together, the entire set of experiments allows a strong inference.

In contrast, if the two sets of experiments each represent strong tests of B, C, and E, and the data strongly support C and refute B and E, the overall support for theory C would be less strong than in our previous example. The reason is that, although data supporting theory C have been generated, there is no strong evidence ruling out two viable alternative theories (A and D). Thus research is highly convergent when a series of experiments consistently supports a given theory while collectively eliminating the most important competing explanations. Although no single experiment can rule out all alternative explanations, taken collectively, a series of partially diagnostic experiments can lead to a strong conclusion if the data converge in the manner of our first example.

Increasingly, the combining of evidence from disparate studies to form a conclusion is being done more formally by the use of the statistical technique termed meta-analysis (Cooper & Hedges, 1994; Hedges & Olkin, 1985; Hunter & Schmidt, 1990; Rosenthal, 1995; Schmidt, 1992; Swanson, 1999) which has been used extensively to establish whether various medical practices are research based. In a medical context, meta-analysis:

involves adding together the data from many clinical trials to create a single pool of data big enough to eliminate much of the statistical uncertainty that plagues individual trials...The great virtue of meta-analysis is that clear findings can emerge from a group of studies whose findings are scattered all over the map. (Plotkin,1996, p. 70)

The use of meta-analysis for determining the research validation of educational practices is just the same as in medicine. The effects obtained when one practice is compared against another are expressed in a common statistical metric that allows comparison of effects across studies. The findings are then statistically amalgamated in some standard ways (Cooper & Hedges, 1994; Hedges & Olkin, 1985; Swanson, 1999) and a conclusion about differential efficacy is reached if the amalgamation process passes certain statistical criteria. In some cases, of course, no conclusion can be drawn with confidence, and the result of the meta-analysis is inconclusive.

More and more commentators on the educational research literature are calling for a greater emphasis on meta-analysis as a way of dampening the contentious disputes about conflicting studies that plague education and other behavioral sciences (Kavale & Forness, 1995; Rosnow & Rosenthal, 1989; Schmidt, 1996; Stanovich, 2001; Swanson, 1999). The method is useful for ending disputes that seem to be nothing more than a "he-said, she-said" debate. An emphasis on meta-analysis has often revealed that we actually have more stable and useful findings than is apparent from a perusal of the conflicts in our journals.

The National Reading Panel (2000) found just this in their meta-analysis of the evidence surrounding several issues in reading education. For example, they concluded that the results of a meta-analysis of the results of 66 comparisons from 38 different studies indicated "solid support for the conclusion that systematic phonics instruction makes a bigger contribution to children's growth in reading than alternative programs providing unsystematic or no phonics instruction" (p. 2-84). In another section of their report, the National Reading Panel reported that a meta-analysis of 52 studies of phonemic awareness training indicated that "teaching children to manipulate the sounds in language helps them learn to read. Across the various conditions of teaching, testing, and participant characteristics, the effect sizes were all significantly greater than chance and ranged from large to small, with the majority in the moderate range. Effects of phonemic awareness training on reading lasted well beyond the end of training" (p. 2-5).

A statement by a task force of the American Psychological Association (Wilkinson, 1999) on statistical methods in psychology journals provides an apt summary for this section. The task force stated that investigators should not "interpret a single study's results as having importance independent of the effects reported elsewhere in the relevant literature" (p. 602). Science progresses by convergence upon conclusions. The outcomes of one study can only be interpreted in the context of the present state of the convergence on the particular issue in question.

The logic of the experimental method

Scientific thinking is based on the ideas of comparison, control, and manipulation . In a true experimental study, these characteristics of scientific investigation must be arranged to work in concert.

Comparison alone is not enough to justify a causal inference. In methodology texts, correlational investigations (which involve comparison only) are distinguished from true experimental investigations that warrant much stronger causal inferences because they involve comparison, control, and manipulation. The mere existence of a relationship between two variables does not guarantee that changes in one are causing changes in the other. Correlation does not imply causation.

There are two potential problems with drawing causal inferences from correlational evidence. The first is called the third-variable problem. It occurs when the correlation between the two variables does not indicate a direct causal path between them but arises because both variables are related to a third variable that has not even been measured.

The second reason is called the directionality problem. It creates potential interpretive difficulties because even if two variables have a direct causal relationship, the direction of that relationship is not indicated by the mere presence of the correlation. In short, a correlation between variables A and B could arise because changes in A are causing changes in B or because changes in B are causing changes in A. The mere presence of the correlation does not allow us to decide between these two possibilities.

The heart of the experimental method lies in manipulation and control. In contrast to a correlational study, where the investigator simply observes whether the natural fluctuation in two variables displays a relationship, the investigator in a true experiment manipulates the variable thought to be the cause (the independent variable) and looks for an effect on the variable thought to be the effect (the dependent variable ) while holding all other variables constant by control and randomization. This method removes the third-variable problem because, in the natural world, many different things are related. The experimental method may be viewed as a way of prying apart these naturally occurring relationships. It does so because it isolates one particular variable (the hypothesized cause) by manipulating it and holding everything else constant (control).

When manipulation is combined with a procedure known as random assignment (in which the subjects themselves do not determine which experimental condition they will be in but, instead, are randomly assigned to one of the experimental groups), scientists can rule out alternative explanations of data patterns. By using manipulation, experimental control, and random assignment, investigators construct stronger comparisons so that the outcome eliminates alternative theories and explanations.

The need for both correlational methods and true experiments

As strong as they are methodologically, studies employing true experimental logic are not the only type that can be used to draw conclusions. Correlational studies have value. The results from many different types of investigation, including correlational studies, can be amalgamated to derive a general conclusion. The basis for conclusion rests on the convergence observed from the variety of methods used. This is most certainly true in classroom and curriculum research. It is necessary to amalgamate the results from not only experimental investigations, but correlational studies, nonequivalent control group studies, time series designs, and various other quasi-experimental designs and multivariate correlational designs, All have their strengths and weaknesses. For example, it is often (but not always) the case that experimental investigations are high in internal validity, but limited in external validity, whereas correlational studies are often high in external validity, but low in internal validity.

Internal validity concerns whether we can infer a causal effect for a particular variable. The more a study employs the logic of a true experiment (i.e., includes manipulation, control, and randomization), the more we can make a strong causal inference. External validity concerns the generalizability of the conclusion to the population and setting of interest. Internal and external validity are often traded off across different methodologies. Experimental laboratory investigations are high in internal validity but may not fully address concerns about external validity. Field classroom investigations, on the other hand, are often quite high in external validity but because of the logistical difficulties involved in carrying them out, they are often quite low in internal validity. That is why we need to look for a convergence of results, not just consistency from one method. Convergence increases our confidence in the external and internal validity of our conclusions.

Again, this underscores why correlational studies can contribute to knowledge. First, some variables simply cannot be manipulated for ethical reasons (for instance, human malnutrition or physical disabilities). Other variables, such as birth order, sex, and age, are inherently correlational because they cannot be manipulated, and therefore the scientific knowledge concerning them must be based on correlational evidence. Finally, logistical difficulties in classroom and curriculum research often make it impossible to achieve the logic of the true experiment. However, this circumstance is not unique to educational or psychological research. Astronomers obviously cannot manipulate all the variables affecting the objects they study, yet they are able to arrive at conclusions.

Complex correlational techniques are essential in the absence of experimental research because complex correlational statistics such as multiple regression, path analysis, and structural equation modeling that allow for the partial control of third variables when those variables can be measured. These statistics allow us to recalculate the correlation between two variables after the influence of other variables is removed. If a potential third variable can be measured, complex correlational statistics can help us determine whether that third variable is determining the relationship. These correlational statistics and designs help to rule out certain causal hypotheses, even if they cannot demonstrate the true causal relation definitively.

Stages of scientific investigation: The Role of Case Studies and Qualitative Investigations

The educational literature includes many qualitative investigations that focus less on issues of causal explanation and variable control and more on thick description , in the manner of the anthropologist (Geertz, 1973, 1979). The context of a person's behavior is described as much as possible from the standpoint of the participant. Many different fields (e.g., anthropology, psychology, education) contain case studies where the focus is detailed description and contextualization of the situation of a single participant (or very few participants).

The usefulness of case studies and qualitative investigations is strongly determined by how far scientific investigation has advanced in a particular area. The insights gained from case studies or qualitative investigations may be quite useful in the early stages of an investigation of a certain problem. They can help us determine which variables deserve more intense study by drawing attention to heretofore unrecognized aspects of a person's behavior and by suggesting how understanding of behavior might be sharpened by incorporating the participant's perspective.

However, when we move from the early stages of scientific investigation, where case studies may be very useful, to the more mature stages of theory testing--where adjudicating between causal explanations is the main task--the situation changes drastically. Case studies and qualitative description are not useful at the later stages of scientific investigation because they cannot be used to confirm or disconfirm a particular causal theory. They lack the comparative information necessary to rule out alternative explanations.

Where qualitative investigations are useful relates strongly to a distinction in philosophy of science between the context of discovery and the context of justification . Qualitative research, case studies, and clinical observations support a context of discovery where, as Levin and O'Donnell (2000) note in an educational context, such research must be regarded as "preliminary/exploratory, observational, hypothesis generating" (p. 26). They rightly point to the essential importance of qualitative investigations because "in the early stages of inquiry into a research topic, one has to look before one can leap into designing interventions, making predictions, or testing hypotheses" (p. 26). The orientation provided by qualitative investigations is critical in such cases. Even more important, the results of quantitative investigations--which must sometimes abstract away some of the contextual features of a situation--are often contextualized by the thick situational description provided by qualitative work.

However, in the context of justification, variables must be measured precisely, large groups must be tested to make sure the conclusion generalizes and, most importantly, many variables must be controlled because alternative causal explanations must be ruled out. Gersten (2001) summarizes the value of qualitative research accurately when he says that "despite the rich insights they often provide, descriptive studies cannot be used as evidence for an intervention's efficacy...descriptive research can only suggest innovative strategies to teach students and lay the groundwork for development of such strategies" (p. 47). Qualitative research does, however, help to identify fruitful directions for future experimental studies.

Nevertheless, here is why the sole reliance on qualitative techniques to determine the effectiveness of curricula and instructional strategies has become problematic. As a researcher, you desire to do one of two things.

Objective A

The researcher wishes to make some type of statement about a relationship, however minimal. That is, you at least want to use terms like greater than, or less than, or equal to. You want to say that such and such an educational program or practice is better than another. "Better than" and "worse than" are, of course, quantitative statements--and, in the context of issues about what leads to or fosters greater educational achievement, they are causal statements as well . As quantitative causal statements, the support for such claims obviously must be found in the experimental logic that has been outlined above. To justify such statements, you must adhere to the canons of quantitative research logic.

Objective B

The researcher seeks to adhere to an exclusively qualitative path that abjures statements about relationships and never uses comparative terms of magnitude. The investigator desires to simply engage in thick description of a domain that may well prompt hypotheses when later work moves on to the more quantitative methods that are necessary to justify a causal inference.

Investigators pursuing Objective B are doing essential work. They provide quantitative information with suggestions for richer hypotheses to study. In education, however, investigators sometimes claim to be pursuing Objective B but slide over into Objective A without realizing they have made a crucial switch. They want to make comparative, or quantitative, statements, but have not carried out the proper types of investigation to justify them. They want to say that a certain educational program is better than another (that is, it causes better school outcomes). They want to give educational strictures that are assumed to hold for a population of students, not just to the single or few individuals who were the objects of the qualitative study. They want to condemn an educational practice (and, by inference, deem an alternative quantitatively and causally better). But instead of taking the necessary course of pursuing Objective A, they carry out their investigation in the manner of Objective B.

Let's recall why the use of single case or qualitative description as evidence in support of a particular causal explanation is inappropriate. The idea of alternative explanations is critical to an understanding of theory testing. The goal of experimental design is to structure events so that support of one particular explanation simultaneously disconfirms other explanations. Scientific progress can occur only if the data that are collected rule out some explanations. Science sets up conditions for the natural selection of ideas. Some survive empirical testing and others do not.

This is the honing process by which ideas are sifted so that those that contain the most truth are found. But there must be selection in this process: data collected as support for a particular theory must not leave many other alternative explanations as equally viable candidates. For this reason, scientists construct control or comparison groups in their experimentation. These groups are formed so that, when their results are compared with those from an experimental group, some alternative explanations are ruled out.

Case studies and qualitative description lack the comparative information necessary to prove that a particular theory or educational practice is superior, because they fail to test an alternative; they rule nothing out. Take the seminal work of Jean Piaget for example. His case studies were critical in pointing developmental psychology in new and important directions, but many of his theoretical conclusions and causal explanations did not hold up in controlled experiments (Bjorklund, 1995; Goswami, 1998; Siegler, 1991).

In summary, as educational psychologist Richard Mayer (2000) notes, "the domain of science includes both some quantitative and qualitative methodologies" (p. 39), and the key is to use each where it is most effective (see Kamil, 1995). Likewise, in their recent book on research-based best practices in comprehension instruction, Block and Pressley (2002) argue that future progress in understanding how comprehension works will depend on a healthy interaction between qualitative and quantitative approaches. They point out that getting an initial idea of the comprehension processes involved in hypertext and Web-based environments will involve detailed descriptive studies using think-alouds and assessments of qualitative decision making. Qualitative studies of real reading environments will set the stage for more controlled investigations of causal hypotheses.

The progression to more powerful methods

A final useful concept is the progression to more powerful research methods ("more powerful" in this context meaning more diagnostic of a causal explanation). Research on a particular problem often proceeds from weaker methods (ones less likely to yield a causal explanation) to ones that allow stronger causal inferences. For example, interest in a particular hypothesis may originally emerge from a particular case study of unusual interest. This is the proper role for case studies: to suggest hypotheses for further study with more powerful techniques and to motivate scientists to apply more rigorous methods to a research problem. Thus, following the case studies, researchers often undertake correlational investigations to verify whether the link between variables is real rather than the result of the peculiarities of a few case studies. If the correlational studies support the relationship between relevant variables, then researchers will attempt experiments in which variables are manipulated in order to isolate a causal relationship between the variables.

Summary of principles that support research-based inferences about best practice

Our sketch of the principles that support research-based inferences about best practice in education has revealed that:

  • Science progresses by investigating solvable, or testable, empirical problems.
  • To be testable, a theory must yield predictions that could possible be shown to be wrong.
  • The concepts in the theories in science evolve as evidence accumulates. Scientific knowledge is not infallible knowledge, but knowledge that has at least passed some minimal tests. The theories behind research-based practice can be proven wrong, and therefore they contain a mechanism for growth and advancement.
  • Theories are tested by systematic empiricism. The data obtained from empirical research are in the public domain in the sense that they are presented in a manner that allows replication and criticism by other scientists.
  • Data and theories in science are considered in the public domain only after publication in peer-reviewed scientific journals.
  • Empiricism is systematic because it strives for the logic of control and manipulation that characterizes a true experiment.
  • Correlational techniques are helpful when the logic of an experiment cannot be approximated, but because these techniques only help rule out hypotheses, they are considered weaker than true experimental methods.
  • Researchers use many different methods to arrive at their conclusions, and the strengths and weaknesses of these methods vary. Most often, conclusions are drawn only after a slow accumulation of data from many studies.

Scientific thinking in educational practice: Reason-based practice in the absence of direct evidence

Some areas in educational research, to date, lack a research-based consensus, for a number of reasons. Perhaps the problem or issue has not been researched extensively. Perhaps research into the issue is in the early stages of investigation, where descriptive studies are suggesting interesting avenues, but no controlled research justifying a causal inference has been completed. Perhaps many correlational studies and experiments have been conducted on the issue, but the research evidence has not yet converged in a consistent direction.

Even if teachers know the principles of scientific evaluation described earlier, the research literature sometimes fails to give them clear direction. They will have to fall back on their own reasoning processes as informed by their own teaching experiences. In those cases, teachers still have many ways of reasoning scientifically.

Tracing the link from scientific research to scientific thinking in practice

Scientific thinking in can be done in several ways. Earlier we discussed different types of professional publications that teachers can read to improve their practice. The most important defining feature of these outlets is whether they are peer reviewed. Another defining feature is whether the publication contains primary research rather than presenting opinion pieces or essays on educational issues. If a journal presents primary research, we can evaluate the research using the formal scientific principles outlined above.

If the journal is presenting opinion pieces about what constitutes best practice, we need to trace the link between those opinions and archival peer-reviewed research. We would look to see whether the authors have based their opinions on peer-reviewed research by reading the reference list. Do the authors provide a significant amount of original research citations (is their opinion based on more than one study)? Do the authors cite work other than their own (have the results been replicated)? Are the cited journals peer-reviewed? For example, in the case of best practice for reading instruction, if we came across an article in an opinion-oriented journal such as Intervention in School and Clinic, we might look to see if the authors have cited work that has appeared in such peer-reviewed journals as Journal of Educational Psychology , Elementary School Journal , Journal of Literacy Research , Scientific Studies of Reading , or the Journal of Learning Disabilities .

These same evaluative criteria can be applied to presenters at professional development workshops or papers given at conferences. Are they conversant with primary research in the area on which they are presenting? Can they provide evidence for their methods and does that evidence represent a scientific consensus? Do they understand what is required to justify causal statements? Are they open to the possibility that their claims could be proven false? What evidence would cause them to shift their thinking?

An important principle of scientific evaluation--the connectivity principle (Stanovich, 2001)--can be generalized to scientific thinking in the classroom. Suppose a teacher comes upon a new teaching method, curriculum component, or process. The method is advertised as totally new, which provides an explanation for the lack of direct empirical evidence for the method. A lack of direct empirical evidence should be grounds for suspicion, but should not immediately rule it out. The principle of connectivity means that the teacher now has another question to ask: "OK, there is no direct evidence for this method, but how is the theory behind it (the causal model of the effects it has) connected to the research consensus in the literature surrounding this curriculum area?" Even in the absence of direct empirical evidence on a particular method or technique, there could be a theoretical link to the consensus in the existing literature that would support the method.

For further tips on translating research into classroom practice, see Warby, Greene, Higgins, & Lovitt (1999). They present a format for selecting, reading, and evaluating research articles, and then importing the knowledge gained into the classroom.

Let's take an imaginary example from the domain of treatments for children with extreme reading difficulties. Imagine two treatments have been introduced to a teacher. No direct empirical tests of efficacy have been carried out using either treatment. The first, Treatment A, is a training program to facilitate the awareness of the segmental nature of language at the phonological level. The second, Treatment B, involves giving children training in vestibular sensitivity by having them walk on balance beams while blindfolded. Treatment A and B are equal in one respect--neither has had a direct empirical test of its efficacy, which reflects badly on both. Nevertheless, one of the treatments has the edge when it comes to the principle of connectivity. Treatment A makes contact with a broad consensus in the research literature that children with extraordinary reading difficulties are hampered because of insufficiently developed awareness of the segmental structure of language. Treatment B is not connected to any corresponding research literature consensus. Reason dictates that Treatment A is a better choice, even though neither has been directly tested.

Direct connections with research-based evidence and use of the connectivity principle when direct empirical evidence is absent give us necessary cross-checks on some of the pitfalls that arise when we rely solely on personal experience. Drawing upon personal experience is necessary and desirable in a veteran teacher, but it is not sufficient for making critical judgments about the effectiveness of an instructional strategy or curriculum. The insufficiency of personal experience becomes clear if we consider that the educational judgments--even of veteran teachers--often are in conflict. That is why we have to adjudicate conflicting knowledge claims using the scientific method.

Let us consider two further examples that demonstrate why we need controlled experimentation to verify even the most seemingly definitive personal observations. In the 1990s, considerable media and professional attention were directed at a method for aiding the communicative capacity of autistic individuals. This method is called facilitated communication. Autistic individuals who had previously been nonverbal were reported to have typed highly literate messages on a keyboard when their hands and arms were supported over the typewriter by a so-called facilitator. These startlingly verbal performances by autistic children who had previously shown very limited linguistic behavior raised incredible hopes among many parents of autistic children.

Unfortunately, claims for the efficacy of facilitated communication were disseminated by many media outlets before any controlled studies had been conducted. Since then, many studies have appeared in journals in speech science, linguistics, and psychology and each study has unequivocally demonstrated the same thing: the autistic child's performance is dependent upon tactile cueing from the facilitator. In the experiments, it was shown that when both child and facilitator were looking at the same drawing, the child typed the correct name of the drawing. When the viewing was occluded so that the child and the facilitator were shown different drawings, the child typed the name of the facilitator's drawing, not the one that the child herself was looking at (Beck & Pirovano, 1996; Burgess, Kirsch, Shane, Niederauer, Graham, & Bacon, 1998; Hudson, Melita, & Arnold, 1993; Jacobson, Mulick, & Schwartz, 1995; Wheeler, Jacobson, Paglieri, & Schwartz, 1993). The experimental studies directly contradicted the extensive case studies of the experiences of the facilitators of the children. These individuals invariably deny that they have inadvertently cued the children. Their personal experience, honest and heartfelt though it is, suggests the wrong model for explaining this outcome. The case study evidence told us something about the social connections between the children and their facilitators. But that is something different than what we got from the controlled experimental studies, which provided direct tests of the claim that the technique unlocks hidden linguistic skills in these children. Even if the claim had turned out to be true, the verification of the proof of its truth would not have come from the case studies or personal experiences, but from the necessary controlled studies.

Another example of the need for controlled experimentation to test the insights gleaned from personal experience is provided by the concept of learning styles--the idea that various modality preferences (or variants of this theme in terms of analytic/holistic processing or "learning styles") will interact with instructional methods, allowing teachers to individualize learning. The idea seems to "feel right" to many of us. It does seem to have some face validity, but it has never been demonstrated to work in practice. Its modern incarnation (see Gersten, 2001, Spear-Swerling & Sternberg, 2001) takes a particularly harmful form, one where students identified as auditory learners are matched with phonics instruction and visual and/or kinesthetic learners matched with holistic instruction. The newest form is particularly troublesome because the major syntheses of reading research demonstrate that many children can benefit from phonics-based instruction, not just "auditory" learners (National Reading Panel, 2000; Rayner et al., 2002; Stanovich, 2000). Excluding students identified as "visual/kinesthetic" learners from effective phonics instruction is a bad instructional practice--bad because it is not only not research based, it is actually contradicted by research.

A thorough review of the literature by Arter and Jenkins (1979) found no consistent evidence for the idea that modality strengths and weaknesses could be identified in a reliable and valid way that warranted differential instructional prescriptions. A review of the research evidence by Tarver and Dawson (1978) found likewise that the idea of modality preferences did not hold up to empirical scrutiny. They concluded, "This review found no evidence supporting an interaction between modality preference and method of teaching reading" (p. 17). Kampwirth and Bates (1980) confirmed the conclusions of the earlier reviews, although they stated their conclusions a little more baldly: "Given the rather general acceptance of this idea, and its common-sense appeal, one would presume that there exists a body of evidence to support it. UnfortunatelyÉno such firm evidence exists" (p. 598).

More recently, the idea of modality preferences (also referred to as learning styles, holistic versus analytic processing styles, and right versus left hemispheric processing) has again surfaced in the reading community. The focus of the recent implementations refers more to teaching to strengths, as opposed to remediating weaknesses (the latter being more the focus of the earlier efforts in the learning disabilities field). The research of the 1980s was summarized in an article by Steven Stahl (1988). His conclusions are largely negative because his review of the literature indicates that the methods that have been used in actual implementations of the learning styles idea have not been validated. Stahl concludes: "As intuitively appealing as this notion of matching instruction with learning style may be, past research has turned up little evidence supporting the claim that different teaching methods are more or less effective for children with different reading styles" (p. 317).

Obviously, such research reviews cannot prove that there is no possible implementation of the idea of learning styles that could work. However, the burden of proof in science rests on the investigator who is making a new claim about the nature of the world. It is not incumbent upon critics of a particular claim to show that it "couldn't be true." The question teachers might ask is, "Have the advocates for this new technique provided sufficient proof that it works?" Their burden of responsibility is to provide proof that their favored methods work. Teachers should not allow curricular advocates to avoid this responsibility by introducing confusion about where the burden of proof lies. For example, it is totally inappropriate and illogical to ask "Has anyone proved that it can't work?" One does not "prove a negative" in science. Instead, hypotheses are stated, and then must be tested by those asserting the hypotheses.

Reason-based practice in the classroom

Effective teachers engage in scientific thinking in their classrooms in a variety of ways: when they assess and evaluate student performance, develop Individual Education Plans (IEPs) for their students with disabilities, reflect on their practice, or engage in action research. For example, consider the assessment and evaluation activities in which teachers engage. The scientific mechanisms of systematic empiricism--iterative testing of hypotheses that are revised after the collection of data--can be seen when teachers plan for instruction: they evaluate their students' previous knowledge, develop hypotheses about the best methods for attaining lesson objectives, develop a teaching plan based on those hypotheses, observe the results, and base further instruction on the evidence collected.

This assessment cycle looks even more like the scientific method when teachers (as part of a multidisciplinary team) are developing and implementing an IEP for a student with a disability. The team must assess and evaluate the student's learning strengths and difficulties, develop hypotheses about the learning problems, select curriculum goals and objectives, base instruction on the hypotheses and the goals selected, teach, and evaluate the outcomes of that teaching. If the teaching is successful (goals and objectives are attained), the cycle continues with new goals. If the teaching has been unsuccessful (goals and objectives have not been achieved), the cycle begins again with new hypotheses. We can also see the principle of converging evidence here. No one piece of evidence might be decisive, but collectively the evidence might strongly point in one direction.

Scientific thinking in practice occurs when teachers engage in action research. Action research is research into one's own practice that has, as its main aim, the improvement of that practice. Stokes (1997) discusses how many advances in science came about as a result of "use-inspired research" which draws upon observations in applied settings. According to McNiff, Lomax, and Whitehead (1996), action research shares several characteristics with other types of research: "it leads to knowledge, it provides evidence to support this knowledge, it makes explicit the process of enquiry through which knowledge emerges, and it links new knowledge with existing knowledge" (p. 14). Notice the links to several important concepts: systematic empiricism, publicly verifiable knowledge, converging evidence, and the connectivity principle.

Teachers and Research Commonality in a "what works" epistemology

Many educational researchers have drawn attention to the epistemological commonalities between researchers and teachers (Gersten, Vaughn, Deshler, & Schiller, 1997; Stanovich, 1993/1994). A "what works" epistemology is a critical source of underlying unity in the world views of educators and researchers (Gersten & Dimino, 2001; Gersten, Chard, & Baker, 2000). Empiricism, broadly construed (as opposed to the caricature of white coats, numbers, and test tubes that is often used to discredit scientists) is about watching the world, manipulating it when possible, observing outcomes, and trying to associate outcomes with features observed and with manipulations. This is what the best teachers do. And this is true despite the grain of truth in the statement that "teaching is an art." As Berliner (1987) notes: "No one I know denies the artistic component to teaching. I now think, however, that such artistry should be research-based. I view medicine as an art, but I recognize that without its close ties to science it would be without success, status, or power in our society. Teaching, like medicine, is an art that also can be greatly enhanced by developing a close relationship to science (p. 4)."

In his review of the work of the Committee on the Prevention of Reading Difficulties for the National Research Council of the National Academy of Sciences (Snow, Burns, & Griffin, 1998), Pearson (1999) warned educators that resisting evaluation by hiding behind the "art of teaching" defense will eventually threaten teacher autonomy. Teachers need creativity, but they also need to demonstrate that they know what evidence is, and that they recognize that they practice in a profession based in behavioral science. While making it absolutely clear that he opposes legislative mandates, Pearson (1999) cautions:

We have a professional responsibility to forge best practice out of the raw materials provided by our most current and most valid readings of research...If professional groups wish to retain the privileges of teacher prerogative and choice that we value so dearly, then the price we must pay is constant attention to new knowledge as a vehicle for fine-tuning our individual and collective views of best practice. This is the path that other professions, such as medicine, have taken in order to maintain their professional prerogative, and we must take it, too. My fear is that if the professional groups in education fail to assume this responsibility squarely and openly, then we will find ourselves victims of the most onerous of legislative mandates (p. 245).

Those hostile to a research-based approach to educational practice like to imply that the insights of teachers and those of researchers conflict. Nothing could be farther from the truth. Take reading, for example. Teachers often do observe exactly what the research shows--that most of their children who are struggling with reading have trouble decoding words. In an address to the Reading Hall of Fame at the 1996 meeting of the International Reading Association, Isabel Beck (1996) illustrated this point by reviewing her own intellectual history (see Beck, 1998, for an archival version). She relates her surprise upon coming as an experienced teacher to the Learning Research and Development Center at the University of Pittsburgh and finding "that there were some people there (psychologists) who had not taught anyone to read, yet they were able to describe phenomena that I had observed in the course of teaching reading" (Beck, 1996, p. 5). In fact, what Beck was observing was the triangulation of two empirical approaches to the same issue--two perspectives on the same underlying reality. And she also came to appreciate how these two perspectives fit together: "What I knew were a number of whats--what some kids, and indeed adults, do in the early course of learning to read. And what the psychologists knew were some whys--why some novice readers might do what they do" (pp. 5-6).

Beck speculates on why the disputes about early reading instruction have dragged on so long without resolution and posits that it is due to the power of a particular kind of evidence--evidence from personal observation. The determination of whole language advocates is no doubt sustained because "people keep noticing the fact that some children or perhaps many children--in any event a subset of children--especially those who grow up in print-rich environments, don't seem to need much more of a boost in learning to read than to have their questions answered and to point things out to them in the course of dealing with books and various other authentic literacy acts" (Beck, 1996, p. 8). But Beck points out that it is equally true that proponents of the importance of decoding skills are also fueled by personal observation: "People keep noticing the fact that some children or perhaps many children--in any event a subset of children--don't seem to figure out the alphabetic principle, let alone some of the intricacies involved without having the system directly and systematically presented" (p. 8). But clearly we have lost sight of the basic fact that the two observations are not mutually exclusive--one doesn't negate the other. This is just the type of situation for which the scientific method was invented: a situation requiring a consensual view, triangulated across differing observations by different observers.

Teachers, like scientists, are ruthless pragmatists (Gersten & Dimino, 2001; Gersten, Chard, & Baker, 2000). They believe that some explanations and methods are better than others. They think there is a real world out there--a world in flux, obviously--but still one that is trackable by triangulating observations and observers. They believe that there are valid, if fallible, ways of finding out which educational practices are best. Teachers believe in a world that is predictable and controllable by manipulations that they use in their professional practice, just as scientists do. Researchers and educators are kindred spirits in their approach to knowledge, an important fact that can be used to forge a coalition to bring hard-won research knowledge to light in the classroom.

  • Adams, M. J. (1990). Beginning to read: Thinking and learning about print . Cambridge, MA: MIT Press.
  • Adler, J. E. (1998, January). Open minds and the argument from ignorance. Skeptical Inquirer , 22 (1), 41-44.
  • Anderson, C. A., & Anderson, K. B. (1996). Violent crime rate studies in philosophical context: A destructive testing approach to heat and Southern culture of violence effects. Journal of Personality and Social Psychology , 70 , 740-756.
  • Anderson, R. C., Hiebert, E. H., Scott, J., & Wilkinson, I. (1985). Becoming a nation of readers . Washington, D. C.: National Institute of Education.
  • Arter, A. and Jenkins, J. (1979). Differential diagnosis-prescriptive teaching: A critical appraisal, Review of Educational Research , 49 , 517-555.
  • Beck, A. R., & Pirovano, C. M. (1996). Facilitated communications' performance on a task of receptive language with children and youth with autism. Journal of Autism and Developmental Disorders , 26 , 497-512.
  • Beck, I. L. (1996, April). Discovering reading research: Why I didn't go to law school . Paper presented at the Reading Hall of Fame, International Reading Association, New Orleans.
  • Beck, I. (1998). Understanding beginning reading: A journey through teaching and research. In J. Osborn & F. Lehr (Eds.), Literacy for all: Issues in teaching and learning (pp. 11-31). New York: Guilford Press.
  • Berliner, D. C. (1987). Knowledge is power: A talk to teachers about a revolution in the teaching profession. In D. C. Berliner & B. V. Rosenshine (Eds.), Talks to teachers (pp. 3-33). New York: Random House.
  • Bjorklund, D. F. (1995). Children's thinking: Developmental function and individual differences (Second Edition) . Pacific Grove, CA: Brooks/Cole.
  • Block, C. C., & Pressley, M. (Eds.). (2002). Comprehension instruction: Research-based best practices . New York: Guilford Press.
  • Bronowski, J. (1956). Science and human values . New York: Harper & Row.
  • Bronowski, J. (1973). The ascent of man . Boston: Little, Brown.
  • Bronowski, J. (1977). A sense of the future . Cambridge: MIT Press.
  • Burgess, C. A., Kirsch, I., Shane, H., Niederauer, K., Graham, S., & Bacon, A. (1998). Facilitated communication as an ideomotor response. Psychological Science , 9 , 71-74.
  • Chard, D. J., & Osborn, J. (1999). Phonics and word recognition in early reading programs: Guidelines for accessibility. Learning Disabilities Research & Practice , 14 , 107-117.
  • Cooper, H. & Hedges, L. V. (Eds.), (1994). The handbook of research synthesis . New York: Russell Sage Foundation.
  • Cunningham, P. M., & Allington, R. L. (1994). Classrooms that work: They all can read and write . New York: HarperCollins.
  • Dawkins, R. (1998). Unweaving the rainbow . Boston: Houghton Mifflin.
  • Dennett, D. C. (1995). Darwin's dangerous idea: Evolution and the meanings of life . New York: Simon & Schuster.
  • Dennett, D. C. (1999/2000, Winter). Why getting it right matters. Free Inquiry , 20 (1), 40-43.
  • Ehri, L. C., Nunes, S., Stahl, S., & Willows, D. (2001). Systematic phonics instruction helps students learn to read: Evidence from the National Reading Panel's Meta-Analysis. Review of Educational Research , 71 , 393-447.
  • Foster, E. A., Jobling, M. A., Taylor, P. G., Donnelly, P., Deknijff, P., Renemieremet, J., Zerjal, T., & Tyler-Smith, C. (1998). Jefferson fathered slave's last child. Nature , 396 , 27-28.
  • Fraenkel, J. R., & Wallen, N. R. (1996). How to design and evaluate research in education (Third Edition). New York: McGraw-Hill.
  • Geertz, C. (1973). The interpretation of cultures . New York: Basic Books.
  • Geertz, C. (1979). From the native's point of view: On the nature of anthropological understanding. In P. Rabinow & W. Sullivan (Eds.), Interpretive social science (pp. 225-242). Berkeley: University of California Press.
  • Gersten, R. (2001). Sorting out the roles of research in the improvement of practice. Learning Disabilities: Research & Practice , 16 (1), 45-50.
  • Gersten, R., Chard, D., & Baker, S. (2000). Factors enhancing sustained use of research-based instructional practices. Journal of Learning Disabilities , 33 (5), 445-457.
  • Gersten, R., & Dimino, J. (2001). The realities of translating research into classroom practice. Learning Disabilities: Research & Practice , 16 (2), 120-130.
  • Gersten, R., Vaughn, S., Deshler, D., & Schiller, E. (1997).What we know about using research findings: Implications for improving special education practice. Journal of Learning Disabilities , 30 (5), 466-476.
  • Goswami, U. (1998). Cognition in children . Hove, England: Psychology Press.
  • Gross, P. R., Levitt, N., & Lewis, M. (1997). The flight from science and reason . New York: New York Academy of Science.
  • Hedges, L. V., & Olkin, I. (1985). Statistical Methods for Meta-Analysis . New York: Academic Press.
  • Holton, G., & Roller, D. (1958). Foundations of modern physical science . Reading, MA: Addison-Wesley.
  • Hudson, A., Melita, B., & Arnold, N. (1993). A case study assessing the validity of facilitated communication. Journal of Autism and Developmental Disorders , 23 , 165-173.
  • Hunter, J. E., & Schmidt, F. L. (1990). Methods of meta-analysis: Correcting error and bias in research findings . Newbury Park, CA: Sage.
  • Jacobson, J. W., Mulick, J. A., & Schwartz, A. A. (1995). A history of facilitated communication: Science, pseudoscience, and antiscience. American Psychologist , 50 , 750-765.
  • Kamil, M. L. (1995). Some alternatives to paradigm wars in literacy research. Journal of Reading Behavior , 27 , 243-261.
  • Kampwirth, R., and Bates, E. (1980). Modality preference and teaching method: A review of the research, Academic Therapy , 15 , 597-605.
  • Kavale, K. A., & Forness, S. R. (1995). The nature of learning disabilities: Critical elements of diagnosis and classification . Mahweh, NJ: Lawrence Erlbaum Associates.
  • Levin, J. R., & O'Donnell, A. M. (2000). What to do about educational research's credibility gaps? Issues in Education: Contributions from Educational Psychology , 5 , 1-87.
  • Liberman, A. M. (1999). The reading researcher and the reading teacher need the right theory of speech. Scientific Studies of Reading , 3 , 95-111.
  • Magee, B. (1985). Philosophy and the real world: An introduction to Karl Popper . LaSalle, IL: Open Court.
  • Mayer, R. E. (2000). What is the place of science in educational research? Educational Researcher , 29 (6), 38-39.
  • McNiff, J.,Lomax, P., & Whitehead, J. (1996). You and your action research project . London: Routledge.
  • Medawar, P. B. (1982). Pluto's republic . Oxford: Oxford University Press.
  • Medawar, P. B. (1984). The limits of science . New York: Harper & Row.
  • Medawar, P. B. (1990). The threat and the glory . New York: Harper Collins.
  • Moats, L. (1999). Teaching reading is rocket science . Washington, DC: American Federation of Teachers.
  • National Reading Panel: Reports of the Subgroups. (2000). Teaching children to read: An evidence-based assessment of the scientific research literature on reading and its implications for reading instruction . Washington, DC.
  • Nickerson, R. S. (1998). Confirmation bias: A ubiquitous phenomenon in many guises. Review of General Psychology , 2 , 175-220.
  • Pearson, P. D. (1993). Teaching and learning to read: A research perspective. Language Arts , 70 , 502-511.
  • Pearson, P. D. (1999). A historically based review of preventing reading difficulties in young children. Reading Research Quarterly , 34 , 231-246.
  • Plotkin, D. (1996, June). Good news and bad news about breast cancer. Atlantic Monthly , 53-82.
  • Popper, K. R. (1972). Objective knowledge . Oxford: Oxford University Press.
  • Pressley, M. (1998). Reading instruction that works: The case for balanced teaching . New York: Guilford Press.
  • Pressley, M., Rankin, J., & Yokol, L. (1996). A survey of the instructional practices of outstanding primary-level literacy teachers. Elementary School Journal , 96 , 363-384.
  • Rayner, K. (1998). Eye movements in reading and information processing: 20 Years of research. Psychological Bulletin , 124 , 372-422.
  • Rayner, K., Foorman, B. R., Perfetti, C. A., Pesetsky, D., & Seidenberg, M. S. (2002, March). How should reading be taught? Scientific American , 286 (3), 84-91.
  • Reading Coherence Initiative. (1999). Understanding reading: What research says about how children learn to read . Austin, TX: Southwest Educational Development Laboratory.
  • Rosenthal, R. (1995). Writing meta-analytic reviews. Psychological Bulletin , 118 , 183-192.
  • Rosnow, R. L., & Rosenthal, R. (1989). Statistical procedures and the justification of knowledge in psychological science. American Psychologist , 44 , 1276-1284.
  • Shankweiler, D. (1999). Words to meaning. Scientific Studies of Reading , 3 , 113-127.
  • Share, D. L., & Stanovich, K. E. (1995). Cognitive processes in early reading development: Accommodating individual differences into a model of acquisition. Issues in Education: Contributions from Educational Psychology , 1 , 1-57.
  • Shavelson, R. J., & Towne, L. (Eds.) (2002). Scientific research in education . Washington, DC: National Academy Press.
  • Siegler, R. S. (1991). Children's thinking (Second Edition) . Englewood Cliffs, NJ: Prentice Hall.
  • Snow, C. E., Burns, M. S., & Griffin, P. (Eds.). (1998). Preventing reading difficulties in young children . Washington, DC: National Academy Press.
  • Snowling, M. (2000). Dyslexia (Second Edition) . Oxford: Blackwell.
  • Spear-Swerling, L., & Sternberg, R. J. (2001). What science offers teachers of reading. Learning Disabilities: Research & Practice , 16 (1), 51-57.
  • Stahl, S. (December, 1988). Is there evidence to support matching reading styles and initial reading methods? Phi Delta Kappan , 317-327.
  • Stanovich, K. E. (1993/1994). Romance and reality. The Reading Teacher , 47 (4), 280-291.
  • Stanovich, K. E. (2000). Progress in understanding reading: Scientific foundations and new frontiers . New York: Guilford Press.
  • Stanovich, K. E. (2001). How to think straight about psychology (Sixth Edition). Boston: Allyn & Bacon.
  • Stokes, D. E. (1997). Pasteur's quadrant: Basic science and technological innovation . Washington, DC: Brookings Institution Press.
  • Swanson, H. L. (1999). Interventions for students with learning disabilities: A meta-analysis of treatment outcomes . New York: Guilford Press.
  • Tarver, S. G., & Dawson, E. (1978). Modality preference and the teaching of reading: A review, Journal of Learning Disabilities , 11, 17-29.
  • Vaughn, S., & Dammann, J. E. (2001). Science and sanity in special education. Behavioral Disorders , 27, 21-29.
  • Warby, D. B., Greene, M. T., Higgins, K., & Lovitt, T. C. (1999). Suggestions for translating research into classroom practices. Intervention in School and Clinic , 34 (4), 205-211.
  • Wheeler, D. L., Jacobson, J. W., Paglieri, R. A., & Schwartz, A. A. (1993). An experimental assessment of facilitated communication. Mental Retardation , 31 , 49-60.
  • Wilkinson, L. (1999). Statistical methods in psychology journals: Guidelines and explanations. American Psychologist , 54 , 595-604.
  • Wilson, E. O. (1998). Consilience: The unity of knowledge . New York: Knopf.

For additional copies of this document:

Contact the National Institute for Literacy at ED Pubs PO Box 1398, Jessup, Maryland 20794-1398

Phone 1-800-228-8813 Fax 301-430-1244 [email protected]

NICHD logo

Date Published: 2003 Date Posted: March 2010

Department of Education logo

To read this content please select one of the options below:

Please note you do not have access to teaching notes, chapter 1 quantitative research in education: impact on evidence-based instruction.

Current Issues and Trends in Special Education: Research, Technology, and Teacher Preparation

ISBN : 978-1-84950-954-1 , eISBN : 978-1-84950-955-8

Publication date: 23 April 2010

Quantitative research is based on epistemic beliefs that can be traced back to David Hume. Hume and others who followed in his wake suggested that we can never directly observe cause and effect. Rather we perceive what is called “constant conjunction” or the regularities of relationships among events. Through observing these regularities, we can develop generalizable laws that, once established, describe predictable patterns that can be replicated with reliability. This form of reasoning involves studying groups of individuals and is often called nomothetic and is contrasted with idiographic research that focuses on the uniqueness of the individual. It is clear that large-scale experiments with random assignment to treatment are based on nomothetic models, as are quasi-experimental studies where intact groups of people (e.g., students in a particular classroom) are assigned to treatments.

Brigham, F.J. (2010), "Chapter 1 Quantitative research in education: Impact on evidence-based instruction", Obiakor, F.E. , Bakken, J.P. and Rotatori, n.F. (Ed.) Current Issues and Trends in Special Education: Research, Technology, and Teacher Preparation ( Advances in Special Education, Vol. 20 ), Emerald Group Publishing Limited, Leeds, pp. 3-17. https://doi.org/10.1108/S0270-4013(2010)0000020004

Emerald Group Publishing Limited

Copyright © 2010, Emerald Group Publishing Limited

We’re listening — tell us what you think

Something didn’t work….

Report bugs here

All feedback is valuable

Please share your general feedback

Join us on our journey

Platform update page.

Visit emeraldpublishing.com/platformupdate to discover the latest news and updates

Questions & More Information

Answers to the most commonly asked questions here

The Combination of Qualitative and Quantitative Research Methods in Mathematics Education: A “Mixed Methods” Study on the Development of the Professional Knowledge of Teachers

  • First Online: 01 January 2014

Cite this chapter

how does quantitative research help teachers

  • Udo Kelle 6 &
  • Nils Buchholtz 7  

Part of the book series: Advances in Mathematics Education ((AME))

8842 Accesses

9 Citations

3 Altmetric

Research about education in mathematics is influenced by the ongoing dispute about qualitative and quantitative research methods. Especially in the domain of professional knowledge of teachers one can find a clear distinction between qualitative, interpretive studies on the one hand and large-scale quantitative assessment studies on the other hand. Thereby the question of how professional knowledge of teachers can be measured and whether the applied constructs are developed on a solid theoretical base is heavily debated. Most studies in this area limit themselves to the use of either qualitative or quantitative methods and data. In this chapter we discuss the limitations of such mono-method studies and we show how a combination of research methods within a “mixed methods design” can overcome these problems. Thereby we lay special emphasis on different possibilities a mixed methods approach offers for a mutual validation of both qualitative and quantitative findings. For this purpose, we draw on data and results coming from an empirical study about a teacher training program in mathematics, where quantitative data measuring the development of professional knowledge of student teachers were related to qualitative in-depth interviews about the training program.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
  • Available as EPUB and PDF
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
  • Durable hardcover edition

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Beck and Maier distinguish slightly differently between the “normative” and the “interpretive paradigm” going back on Wilson ( 1970 ).

It should be clear from the preceding discussion that this is not so much a problem of quantitative research per se —it may occur if one strictly follows a hypothetico-deductive approach (which is for many reasons advisable if quantitative methods are applied) and if researchers lack empirically contentful hypotheses, workable theories and/or specific knowledge about the domain under study. The latter is often not so much the fault of uninformed researchers but a consequence of the fact that social action is often structured by culture-bound rules and “local knowledge”.

A methodological adjustment of the treatment groups by measures of treatment evaluation (e.g. propensity score matching) has been omitted so far as the use of elaborate statistical methods to determine treatment effects appeared disproportionate due to the small group sizes. Furthermore, the group differences in Abitur grades are not significant and the relationship of school-related pre-cognitions considering the attendance at Advanced or Basic course merely reflects the pre-cognitions of local convenience samples.

It needs to be noted that performance on the level of individual items can vary due to chance and thus should not be over-interpreted.

Anderson, J. R. (1980). Cognitive psychology and its implications . San Francisco: Freeman.

Google Scholar  

Bammé, A. (2004). Science Wars. Von der akademischen zur postakademischen Wissenschaft [Science Wars. From academic to post-academic science]. Frankfurt/New York: Campus Verlag.

Barton, A. H., & Lazarsfeld, P. F. (1969). Some functions of qualitative analysis in social research. In G. C. McCall & J. L. Simmons (Eds.), Issues in participant observation . Reading: Addison-Wesley.

Baumert, J., Bos, W., & Lehmann, R. (Eds.). (2000a). TIMSSS/III. Dritte Internationale Mathematik- und Naturwissenschaftsstudie – Mathematische und naturwissenschaftliche Bildung am Ende der Schullaufbahn. Vol. 1: Mathematische und naturwissenschaftliche Grundbildung am Ende der Pflichtschulzeit [TIMSSS/III. Third international mathematics and science study – Mathematics and science education at the end of schooling. Vol 1: Mathematical and Scientific Literacy at the end of compulsory schooling]. Opladen: Leske + Budrich.

Baumert, J., Bos, W., & Lehmann, R. (Eds.) (2000b). TIMSSS/III. Dritte Internationale Mathematik- und Naturwissenschaftsstudie—Mathematische und physikalische Kompetenzen am Ende der gymnasialen Oberstufe [TIMSSS / III. Third international mathematics and science study—Mathematical and physical competencies at the end of secondary school]. Opladen: Leske + Budrich.

Baumert, J., Klieme, E., Lehrke, M., & Savelsbergh, E. (2000c). Konzeption und Aussagekraft der TIMSS-Leistungstests [Conception and significance of the TIMSS achievement tests]. Die Deutsche Schule , 92 , 103–115 & 196–217.

Baumert, J., Klieme, E., Neubrand, M., Prenzel, M., Schiefele, U., Schneider, W., Tillmann, K.-J. & Weiß, M. (Eds.) (2001). PISA 2000. Basiskompetenzen von Schülerinnen und Schülern im internationalen Vergleich [PISA 2000. An international comparison of students´ basic competencies of students]. Opladen: Leske + Budrich.

Beck, C., & Maier, H. (1993). Das Interview in der mathematikdidaktischen Forschung [Interviewing in research about mathematics education]. Journal für Mathematik-Didaktik, 14 (2), 147–179.

Article   Google Scholar  

Bender, P. (2004). Die etwas andere Sicht auf den mathematischen Teil der internationalen Vergleichs-Untersuchungen PISA sowie TIMSS und IGLU [A somewhat different view of the mathematical part of the international comparative studies PISA and TIMSS and IGLU]. Mitteilungen der DMV, 12 (2), 101–108.

Bender, P. (2005). PISA, Kompetenzstufen und Mathematik-Didaktik [PISA, levels of competence and mathematics education]. Journal für Mathematik-Didaktik, 26 (3/4), 274–281.

Bender, P. (2006). Was sagen uns PISA & Co, wenn wir uns auf sie einlassen? [What may PISA & Co tell us, if we get involved with them?]. In T. Jahnke & W. Meyerhöfer (Eds.), PISA & Co—Kritik eines Programms . Hildesheim: Franzbecker.

Beutelspacher, A., Danckwerts, R., Nickel, G., Spies, S. & Wickel, G. (2011). Mathematik Neu Denken. Impulse für die Gymnasiallehrerausbildung an Universitäten [Thinking Mathematics in a New Way. Ideas for secondary teacher education at universities]. Wiesbaden: Vieweg +Teubner.

Blömeke, S., Kaiser, G., & Lehmann, R. (Eds.). (2010a). TEDS-M 2008. Professionelle Kompetenz und Lerngelegenheiten angehender Primarstufenlehrkräfte im internationalen Vergleich [An international comparison of professional competence and learning opportunities of prospective primary school teachers]. Münster: Waxmann.

Blömeke, S., Kaiser, G., & Lehmann, R. (Eds.). (2010b). TEDS-M 2008. Professionelle Kompetenz und Lerngelegenheiten angehender Mathematiklehrkräfte für die Sekundarstufe I im internationalen Vergleich [An international comparison of professional competence and learning opportunities of prospective mathematics teachers for secondary school]. Münster: Waxmann.

Blumer, H. (1928). Method in social psychology. Unpublished doctoral dissertation, University of Chicago.

Brannen, J. (Ed.). (1992). Mixing methods: Qualitative and quantitative research . Aldershot: Avebury.

Brewer, J., & Hunter, A. (1989). Multimethod research: A synthesis of styles . Newbury Park: Sage.

Brezinka, W. (1968). Von der Pädagogik zur Erziehungswissenschaft. Vorschläge zur Abgrenzung [From pedagogy to education science. Proposals for a demarcation]. Zeitschrift für Pädagogik, 14 (5), 435–475.

Briedis, K., Egorova, T., Heublein, U., Lörz, M., Middendorff, E., Quast, H., & Spangenberg, H. (2008). Studienaufnahme, Studium und Berufsverbleib von Mathematikern. Einige Grunddaten zum Jahr der Mathematik [Educational and occupational careers of mathematicians. Some basic data for the year of mathematics]. Resource Document. Forum Hochschule 9, Hochschul-Informations-System GmbH. http://www.his.de/pdf/pub_fh/fh-200809.pdf . Accessed 24 Apr 2013.

Bromme, R. (1992). Der Lehrer als Experte: Zur Psychologie des professionellen Wissens [The teacher as an expert: The psychology of professional knowledge]. Bern: Huber.

Bromme, R. (1997). Kompetenzen, Funktionen und unterrichtliches Handeln des Lehrers [The teacher’s competencies, functions and instructional actions of]. In F. E. Weinert (Ed.), Enzyklopädie der Psychologie: Psychologie des Unterrichts und der Schule (Vol. 3, pp. 177–212). Göttingen: Hogrefe.

Bruner, J. S., Goodnow, J. J., & Austin, G. A. (1962). A study of thinking . New York: Science Editions.

Bryman, A. (1988). Quantity and quality in social research . London: Routledge.

Book   Google Scholar  

Bryman, A. (2005). Why do researchers integrate quantitative and qualitative research? Paper presented at the Conference “Mixed Method Research Designs: Advanced Issues and Debates” at the University of Basel, 13 Aug 2005.

Bryman, A. (2008). Why do researchers integrate/combine/mesh/blend/mix/merge/fuse/quantitaive and qualitative research? In M. Bergman (Ed.), Advances in mixed methods research. Theories and applications (pp. 87–100). Los Angeles: Sage.

Buchholtz, N., & Blömeke, S. (2012). Mathematik unterrichten lernen. Zur Wirksamkeit hochschuldidaktischer Innovationen in der Mathematik-Lehrerausbildung [Learning to teach mathematics. The effectiveness of innovations in university teaching in mathematics teacher training]. In D. Bosse, L. Criblez, & T. Hascher (Eds.), Reform der Lehrerbildung in Deutschland, Österreich und der Schweiz. Teil 1: Analyse, Perspektiven und Forschung (pp. 255–276). Immenhausen bei Kassel: Prolog-Verlag.

Buchholtz, N., & Kaiser, G. (2013). Improving mathematics teacher education in Germany: Test results from a longitudinal evaluation of innovative programs. International Journal for Science and Mathematics Education, 11 (4), 949–977.

Buchholtz, N., Scheiner, T., Döhrmann, M., Suhl, U., Kaiser, G., & Blömeke, S. (2012). TEDS-shortM. Teacher Education and Development Study—Short Test on Mathematics Content Knowledge (MCK) and Mathematics Pedagogical Content Knowledge (MPCK). Kurzfassung der mathematischen und mathematikdidaktischen Testinstrumente aus TEDS-M, TEDS-LT und TEDS-Telekom . Hamburg: Universität Hamburg.

Campbell, D. T., & Fiske, D. W. (1959). Convergent and discriminant validation by the multitrait-multimethod matrix. Psychological Bulletin, 2 , 81–105.

Carpenter, T. P., & Fennema, E. (1992). Cognitively guided instruction: Building on the knowledge of students and teachers. International Journal of Educational Research, 17 (5), 457–470.

Carpenter, T. P., Fennema, E., Peterson, P. L., & Carey, D. (1988). Teachers’ pedagogical content knowledge of students’ problem solving. Journal of Research in Mathematics Education, 19 (5), 385–401.

Carpenter, T. P., Fennema, E., Peterson, P. L., Chiang, C. P., & Loef, M. (1989). Using knowledge of children’s mathematics thinking in classroom teaching: An experimental study. American Educational Research Journal, 26 (4), 499–532.

Cresswell, J. W. (1994). Research design. Qualitative and quantitative approaches . Thousand Oaks: Sage.

Denzin, N. K. (1978). The research act. A theoretical introduction to sociological methods . New York: McGraw Hill.

Deutsches PISA Konsortium (Eds.). (2000). Schülerleistungen im internationalen Vergleich. Eine neue Rahmenkonzeption für die Erfassung von Wissen und Fähigkeiten [An international comparison of students’ performance A new conceptual framework for the assessment of knowledge and skills] . Resource document. Max-Planck-Institut für Bildungsforschung. http://www.mpib-berlin.mpg.de/Pisa/Rahmenkonzeptiondt.pdf . Accessed 24 Apr 2013.

Eilerts, K. (2009). Kompetenzorientierung in der Mathematik-Lehrerausbildung. Empirische Untersuchung zu ihrer Implementierung [Competence orientation in mathematics teacher education. An empirical study about its implementation]. Münster: LIT Verlag.

Erzberger, C. (1998). Zahlen und Wörter. Die Verbindung quantitativer und qualitativer Daten und Methoden im Forschungsprozess [Numbers and words. The combination of quantitative and qualitative data and methods in the research process]. Weinheim: Deutscher Studien Verlag.

Erzberger, C., & Kelle, U. (2003). Making inferences in mixed methods: The rules of integration. In A. Tashakkori & C. Teddlie (Eds.), Handbook of mixed methods in social and behavioral sciences (pp. 457–488). Thousand Oaks: Sage.

Erzberger, C., & Prein, G. (1997). Triangulation: Validity and empirically based hypothesis construction. Quality & Quantity, 2 , 141–154.

Fennema, E., Carpenter, T. P., Franke, M. L., Levi, L., Jacobs, V. R., & Empson, S. B. (1996). A longitudinal study of learning to use children’s thinking in mathematics instruction. Journal for Research in Mathematics Education, 27 , 403–434.

Fielding, N. G., & Fielding, J. L. (1986). Linking data. Qualitative research methods (Vol. 4). London: Sage.

Flick, U. (1991). Triangulation [Triangulation]. In U. Flick, E. Kardoff, H. Keupp, L. Rosenstiel, & S. Wolff (Eds.), Handbuch qualitative Sozialforschung (pp. 432–434). München: Psychologie Verlags Union.

Flick, U. (1992). Triangulation revisited: Strategy of validation or alternative? Journal of the Theory of Social Behaviour, 2 , 175–197.

Flick, U. (1998). An introduction to qualitative research . Thousand Oaks/London/New Delhi: Sage.

Freudenthal, H. (1983). Didactical phenomenology of mathematical structures . Dordrecht: Reidel.

Gage, N. (1989). The paradigm wars and their aftermath: A “historical” sketch of research and teaching since 1989. Educational Researcher, 18 , 4–10.

Gellert, U. (2006). Mathematik “in der Welt” und mathematische “Grundbildung”. Zur Konsistenz des mathematikdidaktischen Rahmens von PISA [Mathematics “in the world” and mathematical “basic education.” The consistency of the mathematics educational framework of PISA]. In T. Jahnke & W. Meyerhöfer (Eds.), PISA & Co—Kritik eines Programms . Franzbecker: Hildesheim.

Glaser, B., & Strauss, A. (1967). The discovery of grounded theory. Strategies for qualitative research . New York: Aldine.

Gobo, G. (2004). Sampling, representativeness and generalizability. In C. Seale, G. Gobo, J. F. Gubrium, & D. Silverman (Eds.), Qualitative research practice (pp. 435–456). London: Sage.

Grigutsch, S., Raatz, U., & Törner, G. (1998). Einstellung gegenüber Mathematik bei Mathematiklehrern [Beliefs of mathematics teachers concerning mathematics]. Journal für Mathematik-Didaktik, 19 (1), 3–45.

Guba, E. (1981). Criteria for assessing the trustworthiness of naturalistic inquiries. Educational Technology Research and Development, 29 (2), 75–91.

Hagemeister, V. (1999). Was wurde bei TIMSS erhoben? Über die empirische Basis einer aufregenden Studie [What was surveyed with TIMSS? About the empirical basis of an exciting study]. Die Deutsche Schule, 91 , 160–177.

Hammersley, M. (1989). The dilemma of qualitative method. Herbert Blumer and the Chicago tradition . London: Routledge.

Hammersley, M. (1990). What’s wrong with ethnography? The myth of theoretical description. Sociology, 24 (4), 597–615.

Hartig, J., & Kühnbach, O. (2006). Schätzung von Veränderung mit Plausible Values in mehrdimensionalen Rasch-Modellen [Estimation of change with plausible values in multidimensional Rasch models]. In A. Ittel & H. Merkens (Eds.), Veränderungsmessung und Längsschnittstudien in der der Erziehungswissenschaft (pp. 27–44). Wiesbaden: Verlag für Sozialwissenschaften.

Chapter   Google Scholar  

Hopmann, S., Brinek, G., & Retzl, M. (Eds.). (2007). PISA zufolge PISA – PISA According to PISA. Hält PISA, was es verspricht? – [Does PISA keep it’s promises?]. Wien: LIT-Verlag.

Hunter, A., & Brewer, J. (2003). Multimethod research in sociology. In A. Tashakkori & C. Teddlie (Eds.), Handbook of mixed methods in social and behavioral sciences (pp. 577–594). Thousand Oaks: Sage.

Jablonka, E. (2006). Mathematical Literacy: Die Verflüchtigung eines ambitionierten Testkonstrukts in bedeutungslose PISA-Punkte [Mathematical Literacy: Transforming ambitious test construct into meaningless PISA points]. In T. Jahnke & W. Meyerhöfer (Eds.), PISA & Co—Kritik eines Programms . Hildesheim: Franzbecker.

Jahnke, T. (2009). Kritik empirischer Unvernunft—zur sogenannten Empirischen Bildungsforschung in der Mathematikdidaktik [Critique of empirical irrationality—The so-called empirical education research in mathematics education]. In M. Neubrand (Ed.), Beiträge zum Mathematikunterricht 2009 (pp. 671–674). Münster: WTM-Verlag.

Jahnke, T. (2010). Vom mählichen Verschwinden des Fachs aus der Mathematikdidaktik [The gradual disappearance of the subject from mathematics education]. Mitteilungen der GDM, 89 , 21–24.

Jahnke, T., & Meyerhöfer, W. (Eds.). (2006). PISA & Co—Kritik eines Programms [PISA & Co—Critique of a program]. Hildesheim: Franzbecker.

Jahoda, M., Lazarsfeld, P. F., & Zeisel, H. (1982). Die Arbeitslosen von Marienthal [The unemployed of Marienthal] . Frankfurt am Main: Suhrkamp (Original work published 1933).

Johnson, B., & Turner, L. A. (2003). Data collection strategies in mixed methods research. In A. Tashakkori & C. Teddlie (Eds.), Handbook of mixed methods in social & behavioral research (pp. 297–320). Thousand Oaks: Sage.

Kaiser, G. (2000). Internationale Vergleichsuntersuchungen im Mathematikunterricht—Eine Auseinandersetzung mit ihren Möglichkeiten und Grenzen [International comparative studies in mathematics education—A discussion of their possibilities and limitations]. Journal für Mathematik-Didaktik, 21 (3/4), 171–192.

Kaiser, G., & Buchholtz, N. (2014). Overcoming the gap between university and school mathematics. The impact of an innovative programme in mathematics teacher education at the Justus-Liebig-University in Giessen. In S. Rezat, M. Hattermann, & A. Peter-Koop (Eds.), Transformation – A fundamental idea of mathematics education (S. 85–105). Heidelberg: Springer.

Kelle, U. (2001). Sociological explanations between micro and macro and method triangulation. Resource document. Forum Qualitative Social Research , 2 (1), Art. 5. http://www.qualitative-research.net/index.php/fqs/article/view/966/2109 . Accessed 17 April 2013.

Kelle, U. (2006). Combining qualitative and quantitative methods in research practice: Purposes and advantages. Qualitative Research in Psychology, 3 (4), 293–311.

Kelle, U., & Erzberger, C. (2004). Qualitative and quantitative methods—Not in opposition. In U. Flick, E. von Kardorff, & I. Steinke (Eds.), A companion to qualitative research . London: Sage.

Kelle, U., & Lüdemann, C. (1998). Bridge assumptions in rational choice theory: Methodological problems and possible solutions. In H.-P. Blossfeld & G. Prein (Eds.), Rational choice theory and large-scale data analysis (pp. 112–126). Boulder: Westview Press.

Kießwetter, K. (2002). Unzulänglich vermessen und vermessen unzulänglich: PISA & Co [Inadequately measured and overconfidently inadequate: PISA & Co]. Mitteilungen der DMV, 10 (4), 49–58.

Kirsch, A. (1977). Aspekte des Vereinfachens im Mathematikunterricht [Aspects of simplification in mathematics education]. Didaktik der Mathematik, 2 , 87–101.

Klein, F. (1932). Elementary mathematics from an advanced standpoint. Arithmetic, algebra, analysis (translated from the third German edition by Hedrick, E. R., & Noble, C. A.). New York: Macmillan.

Klieme, E., & Bos, W. (2000). Mathematikleistungen und mathematischer Unterricht in Deutschland und Japan. Triangulation qualitativer und quantitativer Analysen am Beispiel der TIMS-Studie [Performance in mathematics and mathematical education in Germany and Japan. Using TIMSS as an example for triangulating qualitative and quantitative findings ]. Zeitschrift für Erziehungswissenschaft, 3 (3), 359–380.

Köller, O., Baumert, J., & Schnabel, K. (1999). Wege zur Hochschulreife. Offenheit des Systems und Sicherung vergleichbarer Standards [Pathways to university entrance qualification. Openness of the system and protection of comparable standards]. Zeitschrift für Erziehungswissenschaft, 2 (3), 385–422.

König, J., & Blömeke, S. (2010). Pädagogisches Unterrichtswissen (PUW). Dokumentation der Kurzfassung des TEDS-M Testinstruments zur Kompetenzmessung in der ersten Phase der Lehrerausbildung [Pedagogical teaching knowledge (PUW). Documentation of the short version of the TEDS-M test instrument for measuring competence in the first phase of teacher education.]. Berlin: Humboldt-Universität.

Krippendorff, K. (2004). Content analysis: An introduction to its methodology (2nd ed.). Thousand Oaks: Sage.

Kunter, M., Baumert, J., Blum, W., Klusmann, U., Krauss, S., & Neubrand, M. (Eds.). (2011). Professionelle Kompetenz von Lehrkräften. Ergebnisse des Forschungsprogramms COACTIV [Professional competence of teachers. Results of the research program COACTIV]. Münster: Waxmann.

Kuntze, S. (2011). Sichtweisen von Studierenden zum Lehren und Lernen im Mathematikunterricht—“Rich pictures” und Multiple-Choice: Gegenüberstellung zweier Erhebungsformate [Perspectives of students on teaching and learning in mathematics education—“rich pictures” and multiple choice: comparison of two survey formats]. In A. Lindmeier & S. Ufer (Eds.), Beiträge zum Mathematikunterricht 2010 (pp. 521–524). Münster: WTM-Verlag.

Lamnek, S. (1995). Qualitative Sozialforschung. Band 1, Methodologie [Qualitative social research. Volume 1, Methodology]. Weinheim: Psychologie Verlags Union.

Lincoln, Y. S., & Guba, E. G. (1985). Naturalistic inquiry . Beverly Hills, CA: Sage.

Lundberg, G. A. (1929/1942). Social research: A study in methods of gathering data. Green: Longmans.

Mayerhöfer, W. (2004a). Zum Kompetenzstufenmodell von PISA [The model of PISA competence levels]. Journal für Mathematik-Didaktik, 25 (3/4), 294–305.

Mayerhöfer, W. (2004b). Zum Problem des Ratens bei PISA [The problem of guessing in PISA]. Journal für Mathematik-Didaktik, 25 (1), 62–69.

Mayerhöfer, W. (2006). PISA & Co als kulturindustrielle Phänomene [PISA & Co as phenomena of culture industry]. In T. Jahnke & W. Meyerhöfer (Eds.), PISA & Co—Kritik eines Programms . Hildesheim: Franzbecker.

Mayring, P. (2000). Qualitative content analysis [28 paragraphs]. Resource document. Forum: Qualitative Social Research, 1 (2), Art. 20, http://nbnresolving.de/urn:nbn:de:0114-fqs0002204 . Accessed 17 Apr 2013.

Mayring, P. (2008). Qualitative Inhaltsanalyse—Grundlagen und Techniken [Qualitative content analysis—Basics and techniques] (10th ed.). Weinheim/Basel: Beltz Verlag.

Mayring, P. (2010). Die Methodenfrage in der fachdidaktischen Forschung – qualitativ, quantitativ, mixed? [The question of method in didactic research—Qualitative, quantitative, mixed?]. Plenary lecture at the 44th annual meeting of the German Society of Didactics of Mathematics (GDM) in Munich, 08–12. March 2010. Abstract in A. Lindmeier & S. Ufer (Eds.), Beiträge zum Mathematikunterricht 2010 . Münster: WTM-Verlag.

Morse, J. M. (2003). Principles of mixed methods and multimethod research design. In A. Tashakkori & C. Teddlie (Eds.), Handbook of mixed methods in social and behavioral research . Thousand Oaks: Sage.

Nastasi, B. K., Hitchcock, J. H., & Brown, L. M. (2010). An inclusive framework for conceptualizing mixed methods typologies. In A. Tashakkori & C. Teddlie (Eds.), Mixed methods in social and behavioral research (2nd ed.). Thousand Oaks: Sage.

Neubrand, M. (Ed.). (2004). Mathematische Kompetenzen von Schülerinnen und Schülern in Deutschland—Vertiefende Analysen im Rahmen von PISA 2000 [Mathematical competencies of students in Germany— In-depth investigations in the context of PISA 2000]. Wiesbaden: VS Verlag für Sozialwissenschaften.

Patton, M. Q. (2002). Qualitative research & evaluation methods (3rd ed.). Thousand Oaks: Sage.

Piaget, J., & Szeminska, A. (1965). Die Entwicklung des Zahlenbegriffs beim Kinde [The development of the number concept in the child]. Stuttgart: Klett.

Roethlisberger, F. J., & Dickson, W. J. (1939). Management and the worker . Cambridge: Harvard University Press.

Rossman, G. B., & Wilson, B. L. (1985). Numbers and words. Combining quantitative and qualitative methods in a single-scale evaluation study. Evaluation Review, 9 (5), 627–643.

Rossman, G. B., & Wilson, B. L. (1994). Numbers and words revisited: Being “shamelessly eclectic”. Quality & Quantity, 28 (3), 315–327.

Rost, J. (2004). Lehrbuch Testtheorie—Testkonstruktion. [Textbook test theory—test construction] (4th ed.). Bern: Verlag Hans Huber.

Schulz, A. (2010). Ergebnisorientierung als Chance für den Mathematikunterricht? Innovationsprozesse qualitativ und quantitativ erfassen [Result orientation as an opportunity for teaching mathematics? Capturing innovation processes qualitatively and quantitatively]. Munich: Herbert Utz Verlag.

Schwarz, B. (2013). Strukturen von Lehrerprofessionswissen. Ergebnisse einer empirischen Untersuchung an Mathematiklehramtsstudierenden [Structures of teachers’ professional knowledge. Results of an empirical study of mathematics teachers students.]. Wiesbaden: Vieweg + Teubner.

Seale, C. (1999). The quality of qualitative research . London: Sage.

Shulman, L. S. (1986). Those who understand: Knowledge growth in teaching. Educational Researcher, 15 (2), 4–14.

Silverman, D. (2000). Doing qualitative research. A practical handbook . London: Sage.

Steinbring, H. (1998). Mathematikdidaktik: Die Erforschung theoretischen Wissens in sozialen Kontexten des Lehrens und Lernens [Mathematics education: Research on theoretical knowledge in social contexts of teaching and learning]. Zentralblatt für Didaktik der Mathematik, 30 (5), 161–167.

Tashakkori, A., & Teddlie, C. (1998). Mixed methods. Combining qualitative and quantitative approaches (Applied social research methods series, Vol. 46). London: Sage.

Tashakkori, A., & Teddlie, C. (2003). Major issues and controversies in the use of mixed methods in the social and behavioral sciences. In A. Tashakkori & C. Teddlie (Eds.), Handbook of mixed methods in social and behavioral sciences (pp. 3–50). Thousand Oaks: Sage.

Tashakkori, A., & Teddlie, C. (2010). Sage handbook of mixed methods in social and behavioral research (2nd ed.). Los Angeles: Sage.

Warm, T. A. (1989). Weighted likelihood estimation of ability in item response models. Psychometrika, 54 , 427–450.

Webb, E. J., Campbell, D. T., Schwartz, R. D., & Sechrest, L. (1966). Unobstrusive measures: Nonreactive research in the social sciences . Chicago: Rand McNally.

Weinert, F. E. (1999). Konzepte der Kompetenz. Gutachten zum OECD-Projekt “Definition and Selection of Competencies: Theoretical and Conceptual Foundations (DeSeCo)” [Concepts of competence. Report on the OECD project “Definition and Selection of Competencies: Theoretical and Conceptual Foundations (DeSeCo)”.]. Neuchatel: Bundesamt für Statistik.

Wellenreuther, M. (1997). Hypothesenprüfung, Theorieentwicklung und Erkenntnisfortschritt in der Mathematikdidaktik. Ein Plädoyer für Methodenpluralismus [Hypothesis testing, theory development and advances in knowledge in mathematics education. A plea for methodological pluralism]. Journal für Mathematik-Didaktik, 18 (2/3), 186–216.

Wilson, T. P. (1970). Conceptions of interaction and forms of sociological explanation. American Sociological Review, 35 (4), 697–710.

Wittmann, E. C. (1995). Mathematics education as a “design science”. Educational Studies in Mathematics, 29 (4), 355–374.

Witzel, A. (2000). The problem-centered interview [26 paragraphs]. Resource document, Forum: Qualitative Social Research , 1 (1), Art. 22. http://www.qualitative-research.net/index.php/fqs/article/view/1132/2522 . Accessed 17 Apr 2013.

Wuttke, J. (2006). Fehler, Verzerrungen, Unsicherheiten in der PISA-Auswertung [Errors, distortions, uncertainties in PISA assessment]. In T. Jahnke & W. Meyerhöfer (Eds.), PISA & Co—Kritik eines Programms . Hildesheim: Franzbecker.

Wuttke, J. (2009). PISA: Nachträge zu einer nicht geführten Debatte [PISA: supplements to a non-debated debate]. Mitteilungen der GDM, 87 , 22–34.

Znaniecki, F. (1934). The method of sociology . New York: Rinehart.

Download references

Author information

Authors and affiliations.

Helmut-Schmidt-Universität Hamburg, Holstenhoffweg 85, 22043, Hamburg, Germany

Arbeitsbereich Didaktik der Mathematik, Universität Hamburg, Fak. 04, FB ErzWiss 05, Binderstraße 34, 20146, Hamburg, Germany

Nils Buchholtz

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Udo Kelle .

Editor information

Editors and affiliations.

Faculty 3 of Mathematics and Computer Science, University of Bremen, Bremen, Germany

Angelika Bikner-Ahsbahs

Christine Knipping

Mathematics Department, Illinois State University, Normal, Illinois, USA

Norma Presmeg

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer Science+Business Media Dordrecht

About this chapter

Kelle, U., Buchholtz, N. (2015). The Combination of Qualitative and Quantitative Research Methods in Mathematics Education: A “Mixed Methods” Study on the Development of the Professional Knowledge of Teachers. In: Bikner-Ahsbahs, A., Knipping, C., Presmeg, N. (eds) Approaches to Qualitative Research in Mathematics Education. Advances in Mathematics Education. Springer, Dordrecht. https://doi.org/10.1007/978-94-017-9181-6_12

Download citation

DOI : https://doi.org/10.1007/978-94-017-9181-6_12

Published : 05 September 2014

Publisher Name : Springer, Dordrecht

Print ISBN : 978-94-017-9180-9

Online ISBN : 978-94-017-9181-6

eBook Packages : Humanities, Social Sciences and Law Education (R0)

Share this chapter

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Publish with us

Policies and ethics

  • Find a journal
  • Track your research

COMMENTS

  1. Quantitative research in education : Background information

    Educational research has a strong tradition of employing state-of-the-art statistical and psychometric (psychological measurement) techniques. Commonly referred to as quantitative methods, these techniques cover a range of statistical tests and tools. The Sage encyclopedia of educational research, measurement, and evaluation by Bruce B. Frey (Ed.)

  2. How teachers can use research effectively in their classroom

    This article discusses four key considerations for using research well in the classroom, along with initial resources and practical guides to support teachers to engage with research. 1. Research comes from a variety of sources. The educators in our survey told us about the challenges they face in accessing research.

  3. Using Research to Improve Teaching

    Teachers and researchers should work collaboratively to improve student learning. Though researchers in higher education typically conduct formal research and publish their work in journal articles, it's important for teachers to also see themselves as researchers. They engage in qualitative analysis while circulating the room to examine and ...

  4. What are the benefits of educational research for teachers?

    As research becomes embedded in your practice you can gain a range of benefits. Research can: help you find solutions to particular problems arising in your classroom or school. underpin professional learning of knowledge, skills and understanding. connect you with sources of information and networks of professional support.

  5. PDF Using Data to Improve Educational Outcomes

    Teachers/Counselors • Model an openness and willingness to use data to enhance teaching and learning. • Use quantitative and qualitative data sources to improve instruction and better understand student thinking and learning, including test results, portfolios, homework, student conferences, journals, classroom observations, and portfolios.

  6. Critical Quantitative Literacy: An Educational Foundation for Critical

    Quantitative research in the social sciences is undergoing a change. After years of scholarship on the oppressive history of quantitative methods, quantitative scholars are grappling with the ways that our preferred methodology reinforces social injustices (Zuberi, 2001).Among others, the emerging fields of CritQuant (critical quantitative studies) and QuantCrit (quantitative critical race ...

  7. Using Research and Reason in Education: How Teachers Can Use ...

    Qualitative research does, however, help to identify fruitful directions for future experimental studies. ... As quantitative causal statements, the support for such claims obviously must be found in the experimental logic that has been outlined above. ... Teachers and Research Commonality in a "what works" epistemology.

  8. PDF Introduction to quantitative research

    Mixed-methods research is a flexible approach, where the research design is determined by what we want to find out rather than by any predetermined epistemological position. In mixed-methods research, qualitative or quantitative components can predominate, or both can have equal status. 1.4. Units and variables.

  9. Seven Suggestions for Teaching Quantitative Methods

    Indeed, students need quantitative methods to understand quantitative academic research, to conduct this research themselves, and to increase their employability in the labor market. Yet, students often experience quantitative methods as difficult to learn and, in turn, instructors find the topic difficult to teach.

  10. The role of research in teacher education: Reviewing the evidence

    Research can be used to inform the design and structure of teacher education programmes. Teachers and teacher educators can be equipped to engage with and be discerning consumers of research. Teachers and teacher educators may be equipped to conduct their own research, individually and collectively, to investigate the impact of particular ...

  11. Chapter 1 Quantitative research in education: Impact on evidence-based

    Quantitative research is based on epistemic beliefs that can be traced back to David Hume. Hume and others who followed in his wake suggested that we can never directly observe cause and effect. ... Research, Technology, and Teacher Preparation (Advances in Special Education, Vol. 20), Emerald Group Publishing Limited, Leeds, pp. 3-17. https ...

  12. Teaching and learning quantitative research methods in the social

    My book introduces different activities and teaching ideas which can be used to support those students undertaking quantitative social research projects. Each chapter contains an activity for students to complete independently as well as a teaching idea. The datasets used in the examples in the book are all open-access meaning that students are able to replicate the techniques shown.

  13. (PDF) Conducting Quantitative Research in Education

    This book provides a clear and straightforward guide for all those seeking to conduct quantitative research in the field of education, using primary research data samples. While positioned as less ...

  14. (PDF) Quantitative Research in Education

    The. quantitative research methods in education emphasise basic group designs. for research and evaluation, analytic metho ds for exploring re lationships. between categorical and continuous ...

  15. (PDF) Developing teachers' research capacity: the essential role of

    I argue in this article that it is essential for educationists, including teachers and teacher. educators, to take back control of the education policy agenda by producing authorita-. tive ...

  16. On the outcomes of teacher wellbeing: a systematic review of research

    Vo and Allen ( 2022) presented a systematic review of the school-based interventions targeting teacher wellbeing. More recently, Beames et al. ( 2022) published a systematic review and meta-analysis of the intervention programs targeting mental health, professional burnout, and/or teacher wellbeing in schoolteachers.

  17. What is teacher research, and how can you benefit from it?

    Barriers to teacher research. Of course, there are potential barriers to teacher research, including limited resources, unsupportive leadership and economic matters, among other things. But for those who choose to engage, there are many potential benefits of teacher research. Simon outlines various scholars' observations that teacher research:

  18. PDF Effective Teacher Leadership: a Quantitative Study of The Relationship

    EFFECTIVE TEACHER LEADERSHIP: A QUANTITATIVE STUDY OF THE RELATIONSHIP BETWEEN SCHOOL STRUCTURES AND EFFECTIVE TEACHER LEADERS Cathy Galland Dr. Cynthia MacGregor, Dissertation Supervisor ABSTRACT The purpose of this quantitative study was to investigate the relationship between

  19. (PDF) A Quantitative Study of Teachers' Perception of Usefulness of

    The study was aimed to measure teachers' perception of the usefulness of action research skills. The study was quantitative in approach and experimental in nature. The population of the study ...

  20. How Quantitative Research Can Help Senior High School Students

    Senior high school students can benefit from using quantitative research to develop their skills in data analysis, critical thinking, and problem-solving. Quantitative research involves the collection and analysis of data in order to conclude it. This type of research can be used to study various topics, including senior high school students ...

  21. The Combination of Qualitative and Quantitative Research ...

    Research about education in mathematics is influenced by the ongoing dispute about qualitative and quantitative research methods. Especially in the domain of professional knowledge of teachers one can find a clear distinction between qualitative, interpretive studies on the one hand and large-scale quantitative assessment studies on the other hand.

  22. A Quantitative Study of Co-Teaching as an Instructional Model to Serve

    Co-teaching can be accomplished, according to Friend and Cook. (2013), by using one of the six specific approaches from one teach/one observe, station. teach, parallel teach, alternative teach, teaming, and one teach/one assist to help improve. the learning outcomes for students with disabilities.

  23. A Practical Guide to Writing Quantitative and Qualitative Research

    INTRODUCTION. Scientific research is usually initiated by posing evidenced-based research questions which are then explicitly restated as hypotheses.1,2 The hypotheses provide directions to guide the study, solutions, explanations, and expected results.3,4 Both research questions and hypotheses are essentially formulated based on conventional theories and real-world processes, which allow the ...