Abstract
Introduction
Since the beginning of the 21st century, many political and state institutions have drawn on the results of fundamental research, particularly in the humanities and social sciences, to guide decision-making processes or support for public action (Quéré, 2017). Inspired by the paradigms of the biomedical sciences, evidence-based education (EBE) reflects this trend by providing practitioners and actors in the world of education with programs whose effectiveness has been proven in methodologically rigorous scientific studies. In French-speaking Belgium, the OECD’s findings on the basis of the data from Programme for International Student Assessment (PISA) are incontestable: In addition to the fact that this part of Belgium ranks among the low-performing industrialized countries (Organisation for Economic Co-operation and Development [OECD], 2016), its education system tends to increase the differences among students according to their socioeconomic status (Demeuse, 2005; Demeuse et al., 2005; Lafontaine et al., 2019). Under pressure from these results of international surveys, there has been a sense that an in-depth reform of the education system is needed. Historically, however, French-speaking Belgium has drawn little if any inspiration from evidence-based paradigms in its reforms of the education system. The best example is probably the practice of grade retention (i.e., the practice of having students repeat a year) which, despite the vast array of evidence pointing against it since the 1970s (Allen et al., 2009; Baye et al., 2019; Galand et al., 2019; Holmes, 1990; Holmes & Matthews, 1984; Jackson, 1975; Jimerson, 2001; Lorence, 2006), has actually become more common over the last 20 years, affecting 46% of the students between grades 1 and 10 according to the PISA 2015 results (Crahay & Marcoux, 2019; Lafontaine et al., 2019).
In 2015, the double challenge of efficiency and equity led the Minister of Education to undertake a reform which, unprecedentedly for Belgium, was to be driven by the involvement of all educational stakeholders in the system (trade unions, private and public providers of schooling, parent associations, etc.), including the educational science research units of all universities and colleges within the education system.
This article aims to describe the gradual steps taken by the education system in this part of Belgium toward an EBE policy. With this in mind, we outline the history of the evidence-based approach and of the education system in French-speaking Belgium. Next, we analyze structural reforms, putting them into perspective with the best evidence available in educational science. Finally, we describe the reform process in progress in light of the three levers for implementing an evidence-based reform in an educational system, according to Slavin (2017, 2019): (1) making a broad range of proven programs available; (2) promoting educator-friendly research reviews; and (3) providing resources to help schools to implement proven programs.
Evidence-based education
Historically, the first mention of the term “evidence-based” in the scientific literature was in 1991 (Smith & Rennie, 2014), when Gordon Guyatt outlined two opposing approaches to solving a problem situation in the biomedical field: on the one hand, reliance on the medical expertise of established personnel, which Guyatt described as the approach of the past, and on the other hand, the use of research from the scientific literature—the approach of the future (Guyatt, 1991). The following year, Guyatt et al. (1992) published their famous paper “Evidence-Based Medicine. A New Approach to Teaching the Practice of Medicine.” This paper, which some people, such as Zimerman (2013), would describe as a political manifesto, called on the scientific community to make a profound change in its paradigm of clinical practice, with the aim of relying on the best evidence available from empirical studies rather than on the intuition and personal experience of professionals (Guyatt et al., 1992). This new approach, evidence-based medicine, may be defined as “the conscientious, explicit, and judicious use of current best evidence in making decisions about the care of individual patients” (Sackett et al., 1996, p. 71). It involves combining professionals’ individual clinical expertise and skills, the best evidence from the scientific literature, and patients’ preferences and values (Sackett et al., 1996).
Bringing to the fore the best evidence from rigorous scientific studies in clinical decision-making thus has few precedents in the history of biomedical practice. This best evidence, although it often comes from randomized controlled trials (RCTs) and systematic reviews of the literature (Li Wan Po, 1998), may also come from studies with alternative research designs which are better suited to providing answers to the clinical question being asked by the practitioner (Sackett et al., 1996). While experimental studies are appropriate for measuring the efficacy of given treatments, cross-sectional studies may be suitable for evaluating the accuracy of a diagnostic test, and longitudinal studies on certain categories of patients can be useful for optimizing medical personnel’s prognoses (Petticrew & Roberts, 2006; Sackett, 1989; Sackett et al., 1985, 1996). Furthermore, advances in the fundamental sciences such as genetics or chemistry must also be taken into account in professionals’ everyday business (Sackett et al., 1996).
The use of the best evidence from empirical research by practitioners in their clinical decision-making has raised the problem of how to provide and make available to practitioners recommendations from scientific studies. Three issues may be pointed out to explain this difficulty: (1) the quantity of studies, which has increased exponentially over the last few decades, (2) the time practitioners may devote to read it, and (3) the very purpose of these studies. (1) The scientific community has frequently observed the difficulty experienced by practitioners wishing to keep up with the scientific literature. In 1981, Sackett reckoned the literature’s growth rate to be nearly 6%–7% per year in the field of biomedical research. Bastian et al. (2010) estimated that 14 RCTs were being published every day in the field of biomedical research in the late 1970s. Their estimate for the number of experimental studies currently published every day is nearly 75. These figures suggest that the workload of professionals wishing to keep up with the scientific literature is unmanageable, especially when considering (2) their available time for this task. Sackett (1995) estimated, on the basis of a survey of medical specialists in several British hospitals, that practitioners have less than an hour a week to spend reading scientific articles. The last issue (3) is the final aim of the scientific papers. The survey conducted by Williamson et al. (1989) on a two-level random sample of 625 doctors pointed out that, according to the professionals, the scientific literature is often poorly written and poorly organized, so that transposing it to a concrete clinical application is a complex business. As Haynes (1990) noted, making connections between peer-reviewed scientific journal articles and clinical practice is increasingly difficult, because scientific articles are produced as tools for communication between researchers rather than communication with practitioners. The purpose they serve means that such articles rarely reach definite conclusions, instead presenting innovations and hypotheses in need of more extensive study. These issues led Archie Cochrane in 1979 to promote the creation of “critical summaries” of all RCTs in the medical field to support professional practice.
EBE brings this originally biomedical paradigm into the field of educational science. Although the use of experimental methods to study the human and social sciences, including education, is far from new (Cook & Campbell, 1979; Cronbach, 1957; de Landsheere, 1986), educational policies were for a very long time governed by rules, mainly state-imposed and political and financial in nature, or even by fashion, and not by scientific research (Slavin, 2016). It was not until the 1980s and the birth of the “test-based accountability” movement (Slavin, 2016) that reflections on the effectiveness of education began to influence educational policy in the U.S. This movement, which saw the introduction of rewards and sanctions for teachers and schools based on their results in standardized national tests of school performance (Hamilton et al., 2002), did not, however, always have the desired effects. Von der Embse et al. (2016) have shown, using a structural equation model to analyze data from 6,428 teachers collected on the basis of the Teacher Stress Inventory, that these policies significantly increased the general level of stress felt by teachers. This was especially so since these policies were not combined with the dissemination of effective educational programs to support teachers in achieving their targets. The literature review conducted by Herman (1999) confirmed the need to provide more evidence-based programs, since it showed that only 20.8% of school-wide programs could be considered to be “based on strong evidence of effectiveness.” This problem led to the foundation in the late 1990s of the “evidence-based practices in education” movement, brought to the fore in the U.S. by the No Child Left Behind Act (2001); this movement encouraged the use of pedagogical methods that were supported by scientific research and that would meet the increased demand for more effective teaching that had been created by accountability policies (Moran, 2004). At the same time, a similar movement emerged in the UK (Norman, 2016), in response to criticisms of the functioning and results of educational research (Davies, 1999), which rarely used scientifically rigorous (i.e., experimental) designs and produced few education programs that were interesting, usable, and reproducible for practitioners (Baye & Bluge, 2016).
This desire for reproducibility (from both a pragmatic and a financial perspective) has been simultaneously the strength and the stumbling-block of the evidence-based paradigm. Its detractors have accused it of seeking the standardization of teaching, as opposed to genuine appropriation by practitioners of empirical knowledge derived from research (Dupriez, 2015; Lessard, 2007; Saussez & Lessard, 2009; Vandenbroeck, 2018; Vandenbroeck et al., 2012). Some of these researchers (Saussez & Lessard, 2009) have preferred the term “evidence-informed education” to “evidence-based education.” They have argued that, while it is true that evidence can be derived from experimental research, the contributions of other research methods in educational science should not be neglected, in order to avoid any epistemological impoverishment (Hammersley, 2001). Furthermore, as in the fields of social sciences (Petticrew, 2015; Petticrew & Roberts, 2006) and biomedical research (Grant & Booth, 2009; Lorenc et al., 2012), educational science cannot find answers to all the educational questions posed by practitioners using only RCTs or systematic reviews. For example, the effect of structural changes in educational systems can only be measured with large-scale correlational studies such as PISA. Finally, although some researchers claim that education is a more complex process (being culturally and contextually specific) than health care (Hammersley, 1997), Davies (1999) showed that evidence-based medicine, like EBE, has been equal to the challenge of studying complex realities.
The foundations had been laid of a movement which, in the years that followed, led to an increase similar to that observed in medicine (Bastian et al., 2010; Chalmers et al., 2002) in projects in the English-speaking world reviewing experimental and quasi-experimental studies of the effectiveness of educational programs. The most successful of these are
The education system in French-speaking Belgium
In order to outline the changes that should occur to the education system in French-speaking Belgium after the major structural changes planned for the period between 2020 and 2030, it is first necessary to describe its current structure. The structure and current functioning of education in Belgium are defined by the Belgian constitutional reform of 1988, known as the Third State Reform (Parlement de la Fédération Wallonie-Bruxelles, 2011). Reflecting political and territorial changes leading to the Federation of the Belgian linguistic entities (Dutch-speaking, French-speaking, and German-speaking), it grants to the three linguistic Communities almost all competencies in terms of education (de Bouttemont, 2004). Consequently, the French-speaking Community of Belgium has a large degree of autonomy in the organization of compulsory education in its territory, although some federal laws remain applicable throughout Belgium, such as those stipulating that compulsory education should be free of charge, and that education should be compulsory from 6 to 18 years (Beckers, 2006; Demeuse, 2005).
A few current characteristics of the education system in French-speaking Belgium are worth describing, since they are at the center of the reform to come. After presenting these characteristics, we will analyze the plans concerning them in the context of the reform toward a more evidence-based approach:
An approach of
A five-cycle structure (Beckers, 2006) characterized by a
A very full school calendar. Students in French-speaking Belgium spend close to
An
Unfortunately, the education system’s structure and its inherent characteristics do not seem to enable students in French-speaking Belgium to achieve satisfactory results in international surveys, with PISA 2015/2018 showing results below the average for OECD countries in science and reading and close to the average in mathematics (Lafontaine et al., 2019; Quittre et al., 2017). Worse, social determinism is particularly strong and seems to be becoming more pronounced (Monseur & Baye, 2015; Quittre et al., 2017, Lafontaine et al., 2019). Some authors go so far as to conclude: “A characteristic of the French-speaking Belgian school system seems to be emerging: students in vulnerable categories are at higher risk than in most other systems of obtaining relatively weak results” (Demeuse et al., 2005, p. 271).
The Pact for Educational Excellence
In 2015, at the initiative of the Minister of Education, all the educational stakeholders of French-speaking Belgium, from researchers to policymakers, and including the providers of schooling, trade unions, parents’ representatives, and others, came together primarily with a view to reforming the education system in French-speaking Belgium so that it could offer an effective response in the areas of performance and equity where it had previously failed. The various stakeholders were invited to form thematic working groups responsible for synthesizing all the available data on the current state of the education system in relation to certain predefined themes. The findings of these different working groups were condensed by the
A scientific steering committee was created for the reform, in which universities were represented. The representatives of the University of Liège argued for more programs and recognition of the findings of rigorous research, that is, the introduction of evidence in the context of the reform. Our research center was then invited to produce a report introducing the unknown concept of EBE to the scientific and political unit in charge of the ongoing reform (the
Unexpectedly, Baye and Bluge’s (2016) report had a considerable impact, with some of its findings being included in
As we hope that this explicit reference to EBE has impacted the structural changes planned in the reform, we will start our analysis by putting the structural reforms into perspective with the best evidence available in educational science. We will then consider how the process of gathering, selecting, and validating educational tools and programs within the
Are structural changes supported by educational research?
The
Comparison of the situation before and the planned situation after the
The comparison of seven planned structural changes with the best scientific evidence (Column 3) makes it clear that some of the proposals are consistent with scientific evidence of various kinds. Evidence from correlational studies indicates that expanding the compulsory common core from the age of 5 to 15 is consistent with the aim of reducing social inequalities (Demeuse & Baye, 2008; Monsieur & Lafontaine, 2012). Correlational studies also indicate that improving teachers’ skills improves students’ performance (Rivkin et al., 2005; Rockoff, 2004)…assuming reasonably that extending initial teacher training by one year will produce more skilled practitioners.
The other four planned structural changes deserve more comment, either because the amount of rigorous research available on a topic could have justified a much more ambitious reform—such as the grade retention policy—or because, conversely, the lack of rigorous research should lead to much more caution before a given change is implemented widely.
This is the case with the skills-based approach. It remains central to the definition of all standards for compulsory education in French-speaking Belgium. However, specific systematic reviews such as that by Morcke et al. (2013), based on eight studies of the performance of medical students, or the literature review by Lassnigg (2015), based on 33 studies of the effectiveness of competence-based education, have shown the lack of rigorous research into its educational effectiveness. Yet this fashionable approach seems to have maintained its appeal, although the swinging back of the pendulum is also observable with the reintroduction of knowledge in educational standards.
In the same way, the “orientation approach model” (Canzittu, 2018; Canzittu & Demeuse, 2017) lies behind the intention of restoring the image of the academic and vocational streams in schools. Although it is based on educational intentions and principles that may be effective (Dupont et al., 2002; Franquet, 2010), this model lacks the support of a rigorous scientific study with an experimental research design.
A last example is the one of the school calendar change. While its intentions are laudable and it can be said to have psychobiological foundations (FAPEO, 2008; Mouraux, 1992; Testu & Fontaine, 2001), the reform of the school calendar is not based on any evidence of its educational effectiveness in terms of academic achievement (Finnie et al., 2019).
By contrast, given the widespread use of grade retention in French-speaking Belgium, the objective of reducing it by 50% is regrettably inadequate. Although there is some awareness of our education system’s anomalous position as a world leader in terms of grade repetition, it will still be possible for teachers to use this practice in the common core grades (Fédération Wallonie-Bruxelles, 2017). The
How does the process of making programs available contribute to the evidence-based nature of the reform?
The
At the outset of this work, we were expecting to come up with various precise criteria for identifying tools/programs based on studies with experimental or quasi-experimental designs, so that only the most rigorous could be kept. We were amazed to find that out of the 425 tools/programs listed by the working groups in 2017, only 61 had been previously tested in studies conducted by educational researchers. Moreover, of these 61 tools/programs, just 13, or 3% of the total, had been tested according to an experimental design.
The question now arose of how to respond to this state of affairs. Traditionally, evidence-based models primarily advocate empirical research of the experimental type. However, as we saw in the historical review of evidence-based medicine, this paradigm relies above all on using the best evidence currently available on a given issue (Sackett et al., 1996). The low rate of experimental studies conducted in contexts comparable to that of French-speaking Belgium and with pedagogical and didactic tools/programs consistent with its educational programs ruled out a hierarchical (Bouffard, 2012; Laurent et al., 2009) or even a pyramidal (Murad et al., 2016; Pageau, 2016) approach to evidence. Such an approach would have led us to exclude almost 97% of the tools/programs proposed for French-speaking educational professionals in Belgium! In any case, a hierarchical approach is a commonly criticized feature of EBE, and that criticism had also been voiced by some of the researchers in the working groups. Moreover, although such a high exclusion rate would not have been problematic scientifically, politically, it was a real issue. It was therefore essential to accept methodological diversity, and standards that were different from those defined by organizations such as
Inclusive typology of research designs adapted to the work of the
This typology has two distinct purposes. The first goal was to make it possible to produce an inventory of the tools/programs selected by the working groups that would take account of their considerable methodological diversity and to classify each program according to its level of evidence. Second, this typology allowed us to define the minimum standards for each of the methodological approaches chosen, including qualitative approaches. It also provided a framework for the research work carried out by the working groups which, as well as compiling an inventory of current tools/programs, also had the task of studying their implementation in the classroom. Some of the proposed tools/programs had never been the subject of a classroom implementation study. For example, our typology defines a number of methodological quality standards for qualitative case studies reporting on a tool’s implementation in different classroom contexts. Through this typology, we asked that any tested tools/programs should be implemented in at least two classrooms from environments with differing characteristics (school’s socioeconomic status, teaching network, extent of teachers’ experience, etc.).
Three central ideas guided the creation of this typology. First, different types of research complemented each other by answering questions of a different nature (Petticrew & Roberts, 2006). Second, the output of the working groups was valuable in that, at the very least, it identified 425 educational and didactic tools/programs evaluated by a panel of experts consisting of researchers, didactics specialists, and professors from the various universities and university colleges of French-speaking Belgium; it thus enabled practitioners to select tools/programs for use in the classroom on the basis of a preliminary form of validation in the selection of tools/programs instead of relying entirely on their intuition (Thomas, 2004). Third, nonexperimental studies can at least reveal that it is worth carrying out more in-depth and rigorous studies on the actual implementation and effect of tools/programs in a real classroom setting, along similar lines to the “promising programs/projects” in English-speaking countries (Education Endowment Foundation, 2019).
Our first argument in favor of this typology, concerning the complementarity between different types of research by virtue of their differing focuses, was inspired by the work of Petticrew and Roberts (2006), who created a model linking different research methods with the prototypical research questions they sought to answer. As this work was originally intended for the social sciences, we adapted it to the more specific field of educational science and to the context of the work of the
Complementarity among methods, focuses, and research questions.
According to Slavin, “for evidence-based reform to prevail, three conditions must exist.” The first is to make available “a broad range of proven programs in key areas of education, at every grade level” (Slavin, 2019, p. 2). In the
Given this conclusion on the first lever, it is obvious that we do not have enough experimental studies to carry out educator-friendly reviews, the second criterion of Slavin (2017, 2019). We wanted to initiate a change in this direction by selecting two priority areas for improvement: reading and dropout. For each of these key areas, we have selected a program from abroad. These programs are being assessed with a quasi-experimental design. Furthermore, we have selected systematic reviews from the
We are more enthusiastic about the use of Slavin’s third lever in our reform, that is, providing resources to help schools to implement proven programs. In line with the report by Baye and Bluge (2016), widespread promotion of experimentation in schools, associated with a specific and recurrent budget for a selection of projects submitted to rigorous evaluation. This budget is justified by (1) the poor existing resources for policy evaluation and experimentation in French-speaking Belgium, (2) the high potential impact in the long term of experimentation and a systematic evaluation of public policies, (3) the need to assess a large number of initiatives planned in the
Conclusion
One of the many ambitions of the reform of the
While we welcome such a change in the education system as experimental researchers, it remains the case that this
Moreover, we can benefit from recent scientific literature in English. Of course, extensive translation work is necessary to make articles published in the field of educational science accessible to educational actors. This work is under progress.
At this point, we may conclude that the
Supplemental material
Supplemental Material, 20203101_Encu_Appendix_1 - Evidence-Based Education: The (Not So Simple) Case of French-Speaking Belgium
Supplemental Material, 20203101_Encu_Appendix_1 for Evidence-Based Education: The (Not So Simple) Case of French-Speaking Belgium by Dylan Dachet and Ariane Baye in ECNU Review of Education
Footnotes
Declaration of conflicting interests
Funding
Contributorship
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
