Open Access is an initiative that aims to make scientific research freely available to all. To date our community has made over 100 million downloads. It’s based on principles of collaboration, unobstructed discovery, and, most importantly, scientific progression. As PhD students, we found it difficult to access the research we needed, so we decided to create a new Open Access publisher that levels the playing field for scientists across the world. How? By making research easy to access, and puts the academic needs of the researchers before the business interests of publishers.
We are a community of more than 103,000 authors and editors from 3,291 institutions spanning 160 countries, including Nobel Prize winners and some of the world’s most-cited researchers. Publishing on IntechOpen allows authors to earn citations and find new collaborators, meaning more people see your work not only from your own field of study, but from other related fields too.
Brief introduction to this section that descibes Open Access especially from an IntechOpen perspective
Want to get in touch? Contact our London head office or media team here
Our team is growing all the time, so we’re always on the lookout for smart people who want to help us reshape the world of scientific publishing.
Home > Books > Lifelong Learning - Education for the Future World
Special Offer!
Applies to all print copy prices | Offer ends 15 December 2024 Enter the code 20PRINT to unlock the offer
Order this book in hard copy and enjoy a 20% discount! Use the promo code 20PRINT during the ordering process.
And that's not all! You'll also receive a complimentary bag with an inspiring quote.
Code copied to clipboard!
Valid Until: 15 December 2024
Note: All orders include FREE SHIPPING
(DAP Incoterms – Import duties and taxes are not included).
If you require assistance, contact us at [email protected].
Bridging the Gap: Traditional vs. Modern Education (A Value-Based Approach for Multiculturalism)
Submitted: 13 November 2023 Reviewed: 06 December 2023 Published: 18 January 2024
DOI: 10.5772/intechopen.114068
Cite this chapter
There are two ways to cite this chapter:
From the Edited Volume
Lifelong Learning - Education for the Future World
Edited by Filippo Gomez Paloma
To purchase hard copies of this book, please contact the representative in India: CBS Publishers & Distributors Pvt. Ltd. www.cbspd.com | [email protected]
Chapter metrics overview
282 Chapter Downloads
Impact of this chapter
Total Chapter Downloads on intechopen.com
Total Chapter Views on intechopen.com
The dynamic landscape of education has witnessed a profound shift from traditional to modern pedagogical paradigms over the years. The discussion of results delves into the intriguing debate between traditional and modern educational systems (TES and MES), examining them through the lens of a value-based approach. This exploration is crucial in understanding how these two approaches shape the educational experiences of learners, faculty, and impact society at large. Drawing from the literature review and insights from a survey involving 179 students and 28 faculty staff, the work advocates a balanced integration of traditional and modern educational approaches. It underscores the pressing need for a value-based model that harmonizes age-old wisdom with contemporary innovations. The survey reveals student aspirations for a holistic, value-driven education, while the faculty acknowledges challenges and opportunities inherent in bridging this educational gap. In conclusion, the data reinforce the value-based approach, emphasizing its importance in curricula and pedagogy to promote ethical values, critical thinking, and empathy. Furthermore, the findings shed light on practical implementation challenges and offer valuable guidance to educators and policymakers. In an era of transformative education, bridging theory and practice will resonate with both students and faculty who recognize the societal benefits of a balanced synthesis between tradition and modernity.
- traditional education
- modern education
- value-based approach
- multiculturalism
- pedagogical paradigm
- ethical values
- critical thinking
Author Information
Oksana chaika *.
- National University of Life and Environmental Sciences of Ukraine, Kyiv, Ukraine
*Address all correspondence to: [email protected], [email protected]
1. Introduction
The contemporary state of education is subject to be in constant flux, paving new ways and vision for the pedagogical paradigms with technological advancements, evolving societal needs, and being parallelly shaped by historical traditions driven by established educational practices. In recent decades, the dichotomy between traditional and modern educational systems (TES and MES) has emerged as a central point of debate [ 1 , 2 , 3 ]. Traditional education, deeply rooted in heritage, often emphasizes rote learning and discipline [ 4 , 5 ], especially for science majors such as math, physics, etc. that lay primary focus on calculation and comprehension, that is, according to Wang “some subjects that focus more on thinking application and deep calculation should be taught in traditional ways” ([ 4 ], p. 272) and “although traditional lecture-style teaching is more boring than high-tech classrooms, it can be more effective at improving test scores” ([ 4 ], p. 271); it is also stated that “rote learning is a well-practiced approach at all educational level, where education and assessments emphasize on memorization of content rather than comprehension and application of content in real life events” ([ 5 ], p. 114). On the other hand, modern education emphasizes innovation, learner-centric methodologies, and the integration of technology, among which cultural literacy, originality, and creativity play a crucial role, for example, “the content [of cultural literacy] in the pedagogical culture is seen pertinent to the modern FL teacher” and at the same time, it includes “professionalism, non-standard way of thinking/thinking out of the box, originality, high level literacy, expressiveness, logic, lexical wealth and communication, creativity, and culture awareness and readiness for diversity in perception” ([ 2 ], p. 103). This evolving educational landscape necessitates a critical examination to address the challenges and opportunities it presents.
Thus, the research problem at hand is the need to navigate this evolving terrain of education effectively. Understanding the dynamics, strengths, and limitations of TES and MES is crucial. Moreover, it is essential to explore how a value-based approach can bridge the gap and create a more holistic educational experience [ 6 , 7 , 8 ] for students and teachers. In this regard, Dogan believes, “the empowerment of the employees who believe in the organization’s values and respect to employees enable the employees to work at full capacity and to become happy” ([ 6 ], p. 84]); moreover, “making radical changes in the behaviors of individuals based on the values is composed of three stages, including seeing, feeling and change” and for instance, “the stage of ‘seeing’ increases the employees’ awareness towards the values” [ 6 ]. The significance of this research lies in its potential to inform educators, policymakers, and stakeholders on the path towards a more balanced and meaningful educational model.
What are the key characteristics and pedagogical paradigms of TES as defined by faculty and students?
How does MES differ from TES, and what innovations does it bring to the educational landscape?
What is the concept of a value-based approach in education, and how can it be implemented effectively, for example, for foreign language teaching and acquisition (FLT and FLA)?
What is the potential impact of a value-based educational model on learners and society?
To address these research questions, a comprehensive approach, combining literature reviews and the survey, was employed. This mixed-method approach [ 7 ] allowed for a deep exploration of the topic, incorporating both theoretical data and comments by the students and faculty from perspectives of their empirical value as real-world experiences from the educational community.
The research findings break into several sections. The introduction is followed by the review of the literature related to TES and MES, value-based education, in the light of the research agenda. Following that, the methodology section explains the survey design, data collection, and analysis. Then, the focus is laid on a discussion of the value-based approach, including its principles, theoretical framework, and practical implications. Finally, the research explores the challenges and opportunities in implementing a value-based approach in education, offering the educational model to be considered by educators, policymakers, and stakeholders. It concludes by summarizing key findings, emphasizing the importance of a balanced approach to education, and suggesting potential future research directions.
2. Theoretical readings
Education, deeply intertwined with culture and tradition, has a rich historical legacy. TES, often characterized by rote learning, discipline, and the transmission of societal values, can be traced back to ancient civilizations [ 8 , 9 ]. For example, the Confucian educational system in ancient China placed a strong emphasis on the teachings of classic texts, which became a cornerstone of traditional education [ 10 ]. The emergence of MES marked a shift towards learner-centric approaches, innovation, and the integration of technology [ 1 , 11 , 12 ]. In the eighteenth and nineteenth centuries, educational reform movements in Europe, such as those initiated by Rousseau and Pestalozzi, laid the groundwork for contemporary educational principles with Rousseau’s three components in education: “This education comes to us from nature itself, or from other men, or from circumstances” as cited in Bazaluk ([ 13 ], p. 17) and Pestalozzi’s “new understanding of the meaning of human life and a cultural ideal”, as well as the philosopher’s basic principle of education: “education should be built according to the natural course of mental development in a child” [ 13 ]. The implementation of compulsory education and the advent of public schooling in the United States and Europe, including Ukraine, further exemplified this shift [ 14 , 15 , 16 ]. Value-based education, rooted in the philosophical underpinnings of moral and ethical development, seeks to instill values such as empathy, critical thinking, and global awareness in learners, based on fostering poly- and multiculturalism in educational classrooms [ 12 , 16 , 17 ]. This approach draws from various theories, including Kohlberg’s theory of moral development, which emphasizes the importance of nurturing ethical reasoning in education [ 18 , 19 ]. Moreover, McKenzie and Blenkinsop’s “ethic of care” highlights the significance of empathy and relationships in the educational process [ 19 ].
3. Research methodology
The methodology for this study involved the creation of a structured survey instrument to gather data on the perceptions and experiences of individuals within the traditional and modern educational systems. The survey was meticulously designed to explore key aspects of both systems and their potential convergence in a value-based approach [ 17 ]. Development included the formulation of clear and concise questions that addressed the objectives of the research and followed either the Likert scale or open-ended questions that allowed for argumentation and substantiation of opinions. Statistically, the research sample consisted of 179 students and 28 faculty staff members from the National University of Life and Environmental Sciences of Ukraine. A purposive sampling technique was employed to ensure a representative mix of participants with varying educational backgrounds, experiences, and perspectives. Informed consent was obtained from all participants.
Data collection primarily relied on an online survey platform, allowing for the efficient distribution of the questionnaire to the selected participants. The survey comprised both closed-ended and open-ended questions, providing a balance between quantifiable data and in-depth qualitative responses. Participants were given a reasonable time frame to complete the survey to ensure thoughtful and accurate responses. The data collected from the survey was subjected to both quantitative and qualitative analysis. Qualitative data from open-ended questions were subjected to thematic analysis to identify common themes and patterns. The integration of both quantitative and qualitative analyses allowed for a comprehensive understanding of the research questions.
The study adhered to ethical guidelines and protocols. Informed consent was secured from all participants, assuring them of anonymity and confidentiality. Data was stored securely, and all personal identifiers were removed during data analysis to ensure participants’ privacy and confidentiality. The research was conducted with the utmost integrity and respect for the rights and well-being of the participants.
4. Results and discussion
The research findings point to the dichotomy between TES and MES, reinforcing the approaches to education. Thus, the study by Alsubaie investigated the effects of traditional and modern teaching methods on student achievement and found that student-centered modern approaches tend to yield better results [ 20 ]. Contrarily, another study by Chavan and Chavan [ 21 ] emphasized the value of traditional education, on the one hand, in preserving cultural heritage and moral values and certain concerns, on the other, for example, “Within the traditional knowledge system, there was a natural obligation, empathy, and overall mentoring towards students by teachers and reverence and submissiveness was exhibited by students towards teachers. Today, teachers, students, and knowledge are all treated as ‘objects’ whose value depends on the quantitative returns ‘it’ can provide” ([ 21 ], p. 278).
TES focuses on teacher-centered instruction while MES embraces learner-centered and interactive approaches: P18, “I reckon most of the seminars being conducted traditionally are rather good, especially when all the students are asked the questions and no one is bored”, P54, “MES offers personalized learning, technology integration, and a focus on critical thinking and problem-solving”;
TES uses textbooks and printed materials as primary resources as contrasted to MES which utilizes a wide range of digital and multimedia resources: P53, “TES provides a time-tested approach to education. This can offer a sense of stability and familiarity for students. However, MES incorporates modern technology and teaching methods, promoting more interactive learning and adaptability to individual student needs”, P75, “The usage of most of the online pros, such as laptops, presentations, there are more opportunities for the future; Nevertheless, I can’t deny the effectiveness of TES”;
in TES, teachers are seen as the primary source of knowledge and authority, however, in MES, teachers would rather encourage student autonomy and self-directed learning: Participant 14 (P14), “The main advantages [of TES] are the opportunities for students to master the theoretical part of knowledge. This will allow them to use it in the future with possible practical skills”; P36, “TES has limitations, such as potential limitations in adapting to individual learning styles and the rapid pace of technological advancements. MES, on the other hand, offers benefits such as personalized learning, technology integration, and global connectivity”;
TES is mainly characterized by its hierarchical and discipline-oriented structure whereas MES is perceived as flexible and adaptable, and according to the students’ answers in their key focus—often tailored to individual needs of the learner: P27, “TES can provide stronger learning of core knowledge through in-depth interaction with teachers and a focus on core academic subjects. It also often includes a rigorous curriculum and structured learning process, which can promote a disciplined approach to learning”, P122, “TES often prioritizes the basics, such as reading and writing. In addition, the teacher plays a central role, providing direct instruction and guidance, which can be beneficial for students who learn well in a structured learning environment. Also, traditional systems emphasize discipline, punctuality, obedience, and compliance.”, P81, “MES often provide more accessibility through technology, learning experiences, and up-to-date content. They can better cater to individual student needs and adapt to a rapidly changing world, making education more engaging”, and
TES promotes uniformity and conformity in learning while MES encourages creativity and independent thinking of the learner: P162, “TES - Stability and reliability, cultural heritage, specialization, number 1 system”, P171, “Modern education often allows for more flexible scheduling, catering to non-traditional students, working adults, or those with diverse time commitments”, P72, “MES focuses on fostering twenty-first century skills such as critical thinking, creativity, communication, collaboration, and digital literacy”, etc.
Demographic information: participants’ age range.
These and other comments were made by the participants with the following educational background: 29.3% were participants with a college degree (high school) and incomplete bachelor’s that referred mainly to students in their first through fourth years at university; 51.2% were master students; 9.8% were faculty staff with a Ph.D. and/or more advanced degree in science, and ultimately 2.4%—“other”, that stands for a visiting researcher.
Moreover, 78% of participants experienced education in both the systems—TES and MES whereas 22% underlined their good acquaintance and journeys with either TES or MES ( Figure 2 ).
Education experience.
Finally, when the questions regarded the emotional part of the participants’ experience with TES and MES and the information about their education systems awareness and educational preferences, that is, which educational systems they feel most familiar with or have spent the most time in, under the half (41.5%) replied they felt comfortable and familiar with both the systems and the other two groups had a difference of 14.6% as to their understanding and preferences for TES (36.6%) over MES (22%) ( Figure 3 ).
Familiarity with TES and MES.
P44, “TES typically follows a well-defined curriculum, providing a structured and organized approach to learning.”
P56, “Traditional education often leads to recognized degrees, diplomas, or certificates, which are widely accepted by employers and institutions. This provides students with formal qualifications that can enhance their career prospects.”
P32, “In traditional educational settings, students have access to experienced educators who can provide guidance, support, and mentorship.”
P101, “Traditional educational institutions typically provide access to libraries, laboratories, and other resources, which can be essential for hands-on learning and research.”
P49, “MES is better in the practical part of learning of any subject. You receive more applied information and learn new practical skills, which are more similar to real-life situations.” and P94, “An individualized approach to each student and a variety of techniques are the strengths of the modern educational model. This allows the student to feel comfortable and apply knowledge in practice.”
P8, “Modern education leverages technology to enhance learning. This includes e-learning platforms, multimedia resources, and online collaboration tools, which can make education more engaging and accessible.”
P78, “Modern education often allows for more flexible scheduling, catering to non-traditional students, working adults, or those with diverse time commitments.”
P163, “Educational technology allows for the collection and analysis of data to gain insights into student progress and areas of improvement, enabling more targeted interventions.”
These and many other comments, feedback, and remarks allowed for grouping the received data under the keywords and phrases that resulted in Table 1 .
Traditional educational system vs. modern educational system.
Another remarkable aspect of the survey makes it conclude that it is not always that the younger generation, as compared to the teaching/lecturing faculty of more age seniority and less digital proficiency, would favor MES, for instance, let us follow P86, “For me, a person who doesn’t like to explore new things, the traditional system is the best, as it has everything I’m used to.”
Next, like the above, where behavioral patterns and/or values play their crucial role, the below comments shed light on what may be significantly valuable while designing and implementing curricula. The values of critical thinking, future vision, and empathy matter to a greater degree, that is, P49 names critical thinking as a value because “MES places a strong emphasis on developing critical thinking skills, students are encouraged to analyze, evaluate, and synthesize information independently” and adds that “This skill is invaluable in various aspects of life, including problem-solving and decision-making”; P90 focuses on the importance of preparation for the future as “MES equips students with skills relevant to the rapidly changing job market” and “Emphasis on subjects like computer science, coding, and digital literacy prepares students for careers in technology and other evolving fields.”
That underscores the importance of scanning the learning mood and educational preferences in classrooms and outside; in an ever-evolving world, the need for innovation is not limited to technology alone; it extends to the way we approach education and pedagogical systems.
5. Value-based approach in education
Taking the discussion further, the exploration leads to the integration of a value-based educational model and bridging the gap between TES and MES, offering real-world solutions that can enhance quality, progress, and enjoyment.
A value-based educational model is a holistic approach to learning, emphasizing the development of not only academic knowledge but also strong ethical principles and values. In this model, core values such as integrity, empathy, and responsibility are integrated into the educational process, ensuring that students not only excel academically but also grow as responsible, ethical individuals [ 22 ].
Drawing from educational theories such as social learning, constructivism, and moral development, the integration of values into education is a well-structured framework [ 23 , 24 ]. These theories provide a foundation for teaching critical thinking, ethical decision-making, and empathy, allowing students to become well-rounded individuals who can navigate an ever-changing world while adhering to timeless ethical principles [ 13 , 16 , 22 ].
Traditional education often relies on rote learning, while modern education places an emphasis on critical thinking and creativity. The value-based approach bridges this gap by combining the strengths of both, nurturing individuals who can adapt to a changing world while upholding ethical principles [ 1 , 3 , 17 ].
According to the survey, 85.3% of participants believe that values are very important/important in education as contrasted to only 4.9% who do not find them very important almost similar to 2.4% who state they are not important at all as only technical knowledge, expertise, and practical skills matter in the job market today ( Figure 4 ); and 7.3% neither agree nor disagree, where the comments specify that the value component is fully personal and plays little role in education; however, it will play its role in family and work life in future.
Values in education.
To bridge the gap between TES and MES via a value-based approach, participants suggest as follows: P3, “We should try to change the basic course of study and let students decide for themselves what to prioritize. Some people prefer practical knowledge, some prefer theory. We should respect people’s preferences in exploring the approaches that suit them.”; P70, “Implement a blended learning approach that combines elements of TES and MES, this can involve a mix of traditional classroom instruction and technology-enhanced learning experiences.”; P32, “Provide comprehensive and ongoing professional development for teachers to adopt innovative teaching methods and effectively integrate technology into their instructional practices”, which can ensure that teachers are equipped with the skills and knowledge to harness the benefits of both TES and MES in their classrooms; P162, “Design educational frameworks that allow for individualized learning paths” to involve incorporating personalized learning plans, adaptive learning technologies, and student self-assessment; P117, “Place a greater emphasis on developing essential skills such as critical thinking, creativity, communication, collaboration, and digital literacy because these skills are crucial for success in the modern world and can be integrated into both TES and MES approaches.”, etc.
Implementing a value-based educational model has practical implications that positively impact both students and the education system. Students exposed to this approach are more likely to become empathetic, socially responsible individuals who can address complex ethical challenges, leading to improved classroom behavior and a positive school culture [ 3 , 4 , 16 ].
6. Impact on learners and society
The study has brought to the forefront the contrasting characteristics of TES, rooted in tradition and discipline, and MES, driven by innovation and learner-centric methodologies. Participants have expressed a resounding consensus on the need for a more holistic educational model, one that combines the strengths of both systems while emphasizing values such as empathy, tolerance, and cultural diversity.
Speaking about the impact on learners and society, it is required to spotlight (a) the imperative of a balanced approach, (b) the need to transform the educational landscape, and (c) a call to action for educators and policymakers, in particular.
Thus, the implications of the research underscore the critical significance of a balanced educational approach. A value-based educational framework that seamlessly integrates the merits of TES and MES can give rise to a more comprehensive and harmonious learning environment. It fosters not only academic excellence but also the ethical and emotional intelligence essential for thriving in an increasingly interconnected world.
Next, the findings extend beyond the boundaries of this study. They have far-reaching implications for the educational landscape, offering the promise of transformation. The integration of a value-based approach carries the potential to shape a generation of learners who are not only academically proficient but also socially responsible, empathetic, and culturally aware. Finally, the present research calls for consideration, further reflection, and most importantly, action. It calls upon educators, policymakers, and stakeholders to reevaluate the core values that underpin education and the pedagogical methodologies employed. A value-based approach represents an opportunity to redefine the educational landscape, nurturing empathetic, critical, and globally aware individuals. With the onward look to the future, education is not confined to the acquisition of knowledge alone; it is about the transformation of individuals and the enrichment of society. It is believed that with an embraced balanced approach and prioritized value-based education, learners and society at large may reap the benefits of the best of both traditional and modern educational systems.
7. Conclusions and recommendations
In the realm of education, the coexistence of TES and MES presents a fascinating landscape for researchers. Along with the challenges, there are opportunities and lessons learned from the coexistence of these systems, that lead to shape future research directions.
To be realistic, it is necessary to address the obstacles in implementing a value-based approach, especially the two identified with the processed and analyzed research data—resistance to change and assessment methods.
Following the findings, one significant challenge in integrating a value-based approach into both TES and MES is the resistance to change. Educators, parents, and institutions may be reluctant to shift from traditional content-focused teaching to a more values-oriented curriculum. Another challenge arises with the assessment methods. Traditional systems often rely on standardized testing for assessment, making it challenging to evaluate students’ personal values and ethics. It also reads in Halstead and Taylor [ 25 ] that developing effective and standardized methods for assessing values is an ongoing challenge.
However, by laying more focus on incorporating technology in classroom and advancing community engagement, TES and MES can find more opportunities for promoting value-based education. The integration of technology, such as e-learning platforms and educational apps, offers a promising avenue to promote value-based education. Today, more and more researchers are exploring how technology can be used to deliver values-centered content [ 1 , 8 , 9 ]. Collaborating with community organizations and stakeholders provides an opportunity to reinforce values taught in school. There have been many works published where research focuses on strategies to engage local communities in shaping students’ values [ 17 , 23 ].
To design and implement strategies (a) for integrating values into educational curricula, (b) guidance for educators, policymakers, and stakeholders as educators should effectively integrate values into their teaching methods, policymakers should establish these clear guidelines for values-based curricula, and stakeholders should offer support through resources and advocacy.
To consider future research directions for long-term impact, cross-cultural studies, and teacher training.
Future research should focus on assessing the long-term impact of values-based education on students’ personal and professional lives, examining its influence on career success, societal well-being, and global citizenship. Comparative studies across different cultural and socio-economic backgrounds are needed to determine the adaptability and effectiveness of value-based education in diverse settings. In the end, more research should explore effective teacher training programs that prepare educators to deliver values-centered content and foster ethical development in students.
In conclusion, the coexistence of TES and MES presents both challenges and opportunities in the integration of value-based education. Lessons from the survey and literature review highlight the positive impact of values-based education on students, while recommendations emphasize the importance of standardized curricula and support from educators, policymakers, and stakeholders.
A. Survey questionnaire
Section 1: demographic information
1.1. What is your age ?
18–24
25–34
35–44
45–54
55–64
1.2. What is your gender?
Prefer not to say
1.3. What is your educational background?
High school
Bachelor’s degree
Master’s degree
Ph.D. or other advanced degree
Other (please specify)
Section 2: experience and preferences
2.1. Have you experienced education in both traditional and modern educational systems (TES and MES)?
2.2. Which educational system do you feel most familiar with or have spent the most time in?
Traditional educational system (TES)
Modern educational system (MES)
Both equally
2.3. What, in your opinion, are the key strengths of TES in comparison to MES? Why?
2.4. What, in your opinion, are the key strengths of MES in comparison to TES? Why?
Section 3: values in education
3.1. How important do you believe it is to integrate values such as empathy, tolerance, and cultural diversity into the education system?
Very important
Not very important
Not important at all
3.2. Do you think traditional educational systems place enough emphasis on teaching these values?
3.3. Do you think modern educational systems place enough emphasis on teaching these values?
3.4. What other values in your opinion are important for education today? Why?
Section 4: bridging the gap
4.1. In your opinion, what strategies or changes could be made to bridge the gap between TES and MES and create a more balanced educational model?
Section 5: overall impressions
5.1. Overall, do you believe it is possible to create a value-based educational system that successfully combines elements of both TES and MES?
Section 6: additional comments
6.1. Is there anything else you would like to share or any additional comments related to traditional and modern educational systems?
- 1. Dewsbury B, Brame CJ. Inclusive teaching. CBE – Life Sciences Education. 2019; 18 (2):1-5. DOI: 10.1187/cbe.19-01-0021
- 2. Chaika O. Implementing technology of poly-/multicultural education for future foreign language teachers: Integrated approach. International Journal of Philology. 2022; 13 (4):101-109. DOI: 10.31548/philolog13(4_2).2022.011
- 3. Alsubaie MA. The influence of traditional and modern teaching methods on students’ achievements. Journal of Education and Practice. 2015; 6 (19):29-40
- 4. Wang Y. A comparative study on the effectiveness of traditional and modern teaching methods. In: 2022 5th International Conference on Humanities Education and Social Sciences (ICHESS 2022). Zhengzhou: Atlantis Press; 2022. pp. 270-277
- 5. Ahmed A, Ahmed N. Comparative analysis of rote learning on high and low achievers in graduate and undergraduate programs. Journal of Education and Educational Development. 2017; 4 (1):111-129. Available from: https://files.eric.ed.gov/fulltext/EJ1161522.pdf
- 6. Dogan S. Model of values-based management process in schools: A mixed design study. International Journal of Higher Education. 2016; 5 (1):83-96
- 7. Bergman M. Advances in Mixed Methods Research: Theories and Applications. Thousand Oaks, CA: SAGE; 2008
- 8. Takahiro N. Confucian modernity in Japan religion and the state. Journal of Japanese Philosophy. 2020; 6 (1):45-62
- 9. O’Dwyer S. Deflating the ‘Confucian Heritage Culture’ thesis in intercultural and academic English education. Language, Culture and Curriculum. 2017; 30 (2):198-211
- 10. Bahtilla M, Xu H. The influence of Confucius’s educational thoughts on China’s educational system. Open Access Library Journal. 2021; 8 (5):1-17
- 11. Abe H. Borrowing from Japan: China’s first modern educational system. In: China’s Education and the Industrialized World. London: Routledge; 2017. pp. 57-300
- 12. Chaika O, Lytvyniuk O, Trofymenko A, Zaluzhna O, Shepel Y. Fundamentals of poly-/multicultural education for a modern foreign language instructor in the system of their professional and personal growth. AD ALTA: Journal of Interdisciplinary Research. 2022; 11 (02 (Special Issue XXIV)):150-155
- 13. Bazaluk O. Plato’s traditions in modern educational theories. Analele Universității din Craiova. Seria Filosofie. 2019; 1 (43):5-20
- 14. Carpentier V. The history of higher education in modern Europe. In: Tamura EH, editor. The Oxford Handbook of the History of Education. Oxford: Oxford University Press; 2019
- 15. Chicosky CL. Restructuring the modern education system in the United States: A look at the value of compulsory education laws. Brigham Young University Education and Law Journal. 2015; 2015 :7-20. Available from: https://digitalcommons.law.byu.edu/elj/vol2015/iss1/2
- 16. Chaika O, Absaliamova Y, Marchuk V, Dovhaniuk E, Kolesnykov A. Poly- and multiculturalism underpinning personality formation of future foreign language teachers. Amazonia Investiga. 2022; 11 (52):221-231. Available from: https://www.webofscience.com/wos/woscc/full-record/WOS:000806131700024
- 17. Biesta GJ. Why ‘what works’ still won’t work: From evidence-based education to value-based education. Studies in Philosophy and Education. 2010; 29 :491-503
- 18. Mathes EW. An evolutionary perspective on Kohlberg’s theory of moral development. Current Psychology. 2021; 40 (8):3908-3921
- 19. McKenzie M, Blenkinsop S. An ethic of care and educational practice. Journal of Adventure Education & Outdoor Learning. 2006; 6 (2):91-105
- 20. Alsubaie MA. Examples of current issues in the multicultural classroom. Journal of Education and Practice. 2015; 6 (10):86-89
- 21. Chavan JR, Chavan SV. Human evolution–A key focus of transformational teacher. In: Knowledge Globalization Conference; Pune, India. Massachusetts, USA: Knowledge Globalization Institute LLC; 2012. pp. 275-285
- 22. Gilfoil DM, Focht JW. Value-based delivery of education: MOOCs as messengers. American Journal of Business Education. 2015; 8 (3):223-238
- 23. Rushton JP. Moral cognition, behaviorism, and social learning theory. Ethics. 1982; 92 (3):459-467
- 24. Cummings R, Harlow S. The constructivist roots of moral education. The Educational Forum. 2000; 64 (4):300-307
- 25. Halstead M, Taylor MJ. Values in Education and Education in Values. London, Washington, D.C.: The Falmer Press; 1996
© 2023 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Continue reading from the same book
Lifelong learning.
Published: 21 August 2024
By Filippo Gomez Paloma
38 downloads
By Patricia Fidalgo and Joan Thormann
326 downloads
By Hanneke Theelen, Milou De Smet, Dave Van Breukelen...
206 downloads
IntechOpen Author/Editor? To get your discount, log in .
Discounts available on purchase of multiple copies. View rates
Local taxes (VAT) are calculated in later steps, if applicable.
Support: [email protected]
An official website of the United States government
Official websites use .gov A .gov website belongs to an official government organization in the United States.
Secure .gov websites use HTTPS A lock ( Lock Locked padlock icon ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.
- Publications
- Account settings
- Advanced Search
- Journal List
Montessori education's impact on academic and nonacademic outcomes: A systematic review
Justus j randolph, anaya bryson, lakshmi menon, david k henderson, austin kureethara manuel, stephen michaels, debra leigh walls rosenstein, warren mcpherson, rebecca o'grady, angeline s lillard.
- Author information
- Article notes
- Copyright and License information
Correspondence Justus J. Randolph, Georgia Baptist College of Nursing, Mercer University, 3001 Mercer University Dr., Atlanta, GA 30341, USA. Email: [email protected]
Corresponding author.
Collection date 2023 Sep.
This is an open access article under the terms of the http://creativecommons.org/licenses/by/4.0/ License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited.
Montessori education is the oldest and most widely implemented alternative education in the world, yet its effectiveness has not been clearly established.
The primary objective of this review was to examine the effectiveness of Montessori education in improving academic and nonacademic outcomes compared to traditional education. The secondary objectives were to determine the degree to which grade level, Montessori setting (public Montessori vs. private Montessori), random assignment, treatment duration, and length of follow‐up measurements moderate the magnitude of Montessori effects.
Search Methods
We searched for relevant studies in 19 academic databases, in a variety of sources known to publish gray literature, in Montessori‐related journals, and in the references of studies retrieved through these searches. Our search included studies published during or before February 2020. The initial search was performed in March 2014 with a follow‐up search in February 2020.
Selection Criteria
We included articles that compared Montessori education to traditional education, contributed at least one effect size to an academic or nonacademic outcome, provided sufficient data to compute an effect size and its variance, and showed sufficient evidence of baseline equivalency–through random assignment or statistical adjustment–of Montessori and traditional education groups.
Data Collection and Analysis
To synthesize the data, we used a cluster‐robust variance estimation procedure, which takes into account statistical dependencies in the data. Otherwise, we used standard methodological procedures as specified in the Campbell Collaboration reporting and conduct standards.
Main Results
Initial searches yielded 2012 articles, of which 173 were considered in detail to determine whether they met inclusion/exclusion criteria. Of these, 141 were excluded and 32 were included. These 32 studies yielded 204 effect sizes (113 academic and 91 nonacademic) across 132,249 data points. In the 32 studies that met minimum standards for inclusion, including evidence of baseline equivalence, there was evidence that Montessori education outperformed traditional education on a wide variety of academic and nonacademic outcomes. For academic outcomes, Hedges' g effect sizes, where positive values favor Montessori, ranged from 0.26 for general academic ability (with high quality evidence) to 0.06 for social studies. The quality of evidence for language ( g = 0.17) and mathematics ( g = 0.22) was also high. The effect size for a composite of all academic outcomes was 0.24. Science was the only academic outcome that was deemed to have low quality of evidence according to the GRADE approach. Effect sizes for nonacademic outcomes ranged from 0.41 for students' inner experience of school to 0.23 for social skills. Both of these outcomes were deemed as having low quality of evidence. Executive function ( g = 0.36) and creativity ( g = 0.26) had moderate quality of evidence. The effect size for a composite of all nonacademic outcomes was 0.33. Moderator analyses of the composite academic and nonacademic outcomes showed that Montessori education resulted in larger effect sizes for randomized studies compared to nonrandomized studies, for preschool and elementary settings compared to middle school or high school settings, and for private Montessori compared to public Montessori. Moderator analyses for treatment duration and duration from intervention to follow‐up data collection were inconclusive. There was some evidence for a lack of small sample‐size studies in favor of traditional education, which could be an indicator of publication bias. However, a sensitivity analysis indicated that the findings in favor of Montessori education were nonetheless robust.
Authors' Conclusions
Montessori education has a meaningful and positive impact on child outcomes, both academic and nonacademic, relative to outcomes seen when using traditional educational methods.
1. PLAIN LANGUAGE SUMMARY
1.1. montessori education significantly impacts academic and nonacademic outcomes.
Relative to traditional education, Montessori education has modest but meaningful positive effects on children's academic and nonacademic (executive function, creativity and social‐emotional) outcomes. This is indicated by a meta‐analysis of 32 studies in which it was possible to compare traditional business‐as‐usual education to Montessori education.
1.1.1. What is this review about?
How best to educate children is an issue of enduring concern, and Montessori is the most common alternative to the conventional education system. Montessori includes a full system of lessons and hands‐on materials for children from birth to 18 years, presented individually, and embedded in a philosophical framework regarding children's development and its optimal conditions.
The term Montessori is not trademarked, and, therefore, its implementation can vary. We studied the range of variations included in the literature, which likely reflects the range of implementations encountered in the world. We also compared Montessori with a range of control conditions described in the literature as traditional (sometimes referred to as conventional, or business‐as‐usual), reflecting the implementation of traditional education in the real world.
What are the main findings of the review?
Using only studies with evidence of baseline equivalence, this review found that Montessori education had a significant positive impact on academic and nonacademic outcomes. Studies with random assignment, elementary school age level, and private Montessori schools had larger effects.
1.1.2. What studies are included?
From a search yielding over 2,000 studies, the review evaluated 32 of the most rigorous Montessori studies, with publication dates ranging from 1967 to 2020.
Study participants were spread across age levels: preschool, elementary school and middle and high school.
The studies took place in eight countries: the USA (18 studies), Turkey (four studies), Switzerland (three studies) and one each in England, France, Malaysia, Oman, Iran, The Philippines, and Thailand.
1.1.3. How effective is Montessori education?
On academic outcomes, Montessori students performed about 1/4 of a standard deviation better than students in traditional education. The magnitude of these effects could be considered small when compared to findings obtained in tightly‐controlled laboratory studies, but they could be considered to be medium‐large to large when compared to studies in real‐world school contexts involving standardized tests.
Most (28) of the included studies were conducted in schools implementing Montessori as a full program; the remaining Four studies were short‐term add‐ons to otherwise traditional school curricula.
The effect sizes for academic outcomes are similar to those obtained in other studies that compared “No Excuses” charter schools to business‐as‐usual urban schools.
The magnitude of Montessori education's nonacademic effects was slightly stronger than its effects on academic outcomes. Montessori students performed about 1/3 of a standard deviation higher than students in traditional education on nonacademic outcomes, including self‐regulation (executive function), well‐being at school, social skills, and creativity.
The magnitude of Montessori education's effects was greater for randomized than non‐randomized study designs, greater for preschool and elementary school than for middle and high school, and greater for private Montessori compared to public Montessori settings.
1.1.4. What do the findings of this review mean?
Across a wide range of implementations (likely reflecting the range of Montessori implementations in the real world) and in studies of moderate to high quality, Montessori education has a nontrivial impact on children's academic and nonacademic outcomes.
1.1.5. How up to date is this review?
The review authors searched for studies published through February 2020.
2. BACKGROUND
Montessori education is the most widespread alternative education, yet its effectiveness is uncertain. Here we describe the problem addressed by the review, then discuss the Montessori intervention and why it might be effective. Finally, we discuss why it is important to do this review.
2.1. Description of the condition
Dissatisfaction with education has been longstanding, with half or fewer American parents satisfied with K‐12 education in the United States over the last 30 years. The United States as a whole performs poorly on international tests like the PISA, and national tests like the NAEP show little progress over the years and a severe drop in performance with the COVID‐19 pandemic. Achievement is particularly concerning among lower‐income children and children of color (Duncan, 2014 ). A prominent education scholar recently stated, “Preparing all students to meet higher academic standards will require instruction that is different and much better than the instruction that most students receive today” (Duncan, 2014 , p. 141). There is also concern about nonacademic outcomes of children, including how to increase their executive function and social‐emotional skills (Ahmed, 2021 ; Jones, 2015 ). Given that education, as it is usually implemented, has not yielded sufficiently positive outcomes to lead to widespread satisfaction, a question arises as to how alternative forms of education fare in terms of delivered outcomes. One such alternative is Montessori education.
Although Montessori education is the oldest continuously implemented, as well as the most widely implemented alternative education in the world (Lillard, 2019 ), currently used in over 550 public and 3000 private schools in the United States alone, evidence of its outcomes has not been rigorously compiled. Schooling techniques should be based on evidence of what works; this is why the Department of Education established the What Works Clearinghouse. And yet that Clearinghouse's current (yet outdated) entry for Montessori Method states that as of December 2005, it is unable to draw evidence‐based conclusions about the effectiveness of Montessori. Recent reviews also note that its effectiveness has not been clearly established (Ackerman, 2019 ; Marshall, 2017 ). This is a problem, given that parents and school districts want to and should make schooling decisions based on evidence.
The main objective of this review is to determine if Montessori education impacts academic and/or nonacademic outcomes of children, and, thus, whether it should be further explored as a possible type of school reform to address the shortcomings of traditional education. Secondary objectives are to determine if its impacts vary at different ages, in public versus private school settings, with the duration of a child's participation in Montessori, and with the length of time since a Montessori intervention (either fade‐out effects or, by contrast, sleeper effects where the impact strengthens with time). The knowledge gained from this analysis could be important to education policy at national, state, district, and school levels, as well as for parents making individual decisions about their children's education.
2.2. Description of the intervention
Maria Montessori and her collaborators developed a system of education based on observations first of atypically‐developing children, then of lower‐income children, and finally a variety of children in widely disparate cultures (from India to Europe to America) (Montessori, 2017 ). Montessori is currently available in over 150 countries.
Through observation, Montessori developed a distinct philosophy of education that was rooted in her medical training (Trabalzini, 2011 ). She viewed children as biological organisms driven toward their ultimate adult state by internal forces (Montessori, 2012 ; Montessori, 2017 ). When nothing perturbs this development (akin to poor nutrition or other environmental disturbances), she believed children would make optimal choices to propel their own development forward. Thus, in Montessori programs children choose which materials to use at any given time; teacher guidance is given only as needed (i.e., when left on their own, a child's choice is not constructive). Because they are rendered unnecessary (since children have a natural inclination to learn and develop), extrinsic motivators are not employed in Montessori programs. The teacher is able to work with children individually because the Montessori materials themselves do the teaching—they are self‐correcting. And, aligning with children's social tendencies, children are allowed to work together as much as they wish. The only requirement is that children are constructive and that they work through all the materials in a classroom environment during the years when they are in the classroom.
The Montessori system has three elements: the environment, the teacher or guide, and the child (Lillard, 2019a ; Lillard, 2019b ). The system is adapted for different developmental stages (0‐3, 3‐6, 6‐12, and 12‐18) and cultures (Montessori, 2012 ). Yet, because children are biologically the same everywhere and have been for many thousands of years (essentially speaking), the system is highly consistent, such that today's Montessori classroom in Kyoto looks very much the same as one in the highlands of Bhutan, the slums of Mexico City, following nomadic tribes in Kenya, or in 1910 in Rome. Ideally, Montessori classrooms have ample natural light and access to nature (plants and animals in the room, and/or easy access to an outdoor space). Within carefully prepared classroom environments, children encounter an array of brightly colored hands‐on materials, one of each type, arranged neatly on accessible shelves into classroom areas (Math, Language, Music, Art, Sensorial Activities, and so on). Each material is available to every child once they have been taught to use it. Material sets increase in difficulty, serving the youngest to the oldest children in each classroom.
The teacher's role in Montessori is to connect children to the environment by showing them (individually or in small groups) how to use the materials, at a time when each child is judged to be ripe to learn them (Elkind, 2003 ; Jones, 2005 ; Montessori, 1964 ; Montessori, 2012 ; Murray, 2010 ). The teacher spends a great deal of time simply observing the children, judging their ripeness, and figuring out when and how to stoke a child's interest. Teachers undergo a long preparation for their roles, learning about the Montessori philosophy and theory, the subject matter of the classroom, and how to present each material in what is deemed to be a clear and captivating way, as well as developing sensitivity to how children express readiness to learn. Montessori teachers keep records of each child's progress through the sequences of materials, overseeing their learning.
Although often thought of as a private school model for preschool, Montessori actually goes through high school and is also implemented in the public sector. Most of the over 500 public Montessori schools in the United States are Title 1 schools serving children of color (Debs, 2019 ). Initiatives like Educateurs sans Frontières are increasingly bringing Montessori to the global majority.
2.3. How the intervention might work
Montessori education includes philosophical and structural elements; the structural elements were judged by Maria Montessori to be the best way to implement her philosophy. The intervention might work through either or both of these avenues.
Nine philosophical elements of Montessori were described (along with supporting research) in Montessori: The Science Behind the Genius (Lillard, 2017 ). We briefly summarize these here, as important means through which the intervention might work. For an alternative model of how the intervention might work, see the logic model developed by Culclasure and colleagues (Culclasure, 2019 ).
Montessori education involves hands‐on learning; cognition and movement are therefore deeply aligned (see also Laski, 2015 ). Children use materials that convey important concepts and skills that might transfer to improved academic outcomes.
Children in Montessori get to pursue what they are interested in learning at the moment, rather than something a teacher (or state legislature) has chosen for an entire class to learn at once, at a particular moment in time. Interest enhances learning, and being able to do what one is interested in doing also could lead to better nonacademic outcomes, such as positive feelings about or wellbeing in school (Ryan, 2000 ).
Children in Montessori programs choose what they will learn about; they determine how they will spend their time. Research has shown that when children are in environments with more self‐determination, their academic performance improves (Cordova, 1996 ; De Charms, 1976 ). In addition, so does their perceived self‐worth, mastery orientation (Ryan, 1986 ), and creativity (Amabile, 1984 ).
Montessori also places a high priority on concentrated attention and developing executive function (see Diamond, 2011 ). Enhanced self‐regulation early in life predicts a wide range of health‐related and wealth‐related outcomes later in life (Moffitt, 2011 ).
Learning stems from intrinsic motivation; there are no extrinsic motivators encouraging children to work in Montessori classrooms. Intrinsic motivation is desirable in itself, and is also associated with lifelong learning (Cordova, 1996 ; De Charms, 1976 ; Ryan, 1986 ). Creativity is also enhanced by a lack of extrinsic rewards (Amabile, 1984 ).
Montessori learning is situated, so a child who is interested in bugs will study actual bugs, not just read about them in texts. In other cases, Montessori's specially‐developed hands‐on materials make learning situated; e.g., a child learning the Pythagorean theorem gets materials that make the theorem self‐evident. The children also embody the theorem in the schoolyard, using ropes to measure triangles, imagining themselves as ancient Egyptians measuring property lines. Learning is enhanced when it is situated in contexts (Cordova, 1996 ; Lillard, 2017 ).
In Montessori classrooms, children are able to work with peers at will; they learn through imitation, through collaboration, and through peer tutoring. Peers can inspire children to assimilate and accommodate academic and social skills exhibited by those peers (Turner, 1992 ). Many studies show that peer learning is associated with better outcomes (Topping, 2005 ).
Montessori teachers are counseled to work with children in specific ways, to cultivate the sensitive responsiveness that leads to secure attachment, and to take an authoritative approach; such approaches predict better child outcomes (Baumrind, 1989 ). Montessori teachers also facilitate student‐driven creative approaches to solving problems (Ultanir, 2012 ).
The Montessori environment is tightly ordered, with everything in its place. Although children have considerable freedom about how to use their time, exactly how a child uses each material is far from random; there are a series of prescribed steps, from which children are permitted to deviate only when the teacher perceives that deviation to be constructive for their learning and development. Order is also associated with better academic and nonacademic outcomes for children (Lillard, 2017 ).
An educational environment that embodies any or all of these philosophical elements might be expected to result in better outcomes since individual research studies involving each element individually have resulted in better outcomes. Well‐implemented Montessori has all nine of these elements, and thus might improve developmental outcomes.
Over her lifetime, as she developed these philosophical elements, Maria Montessori also arrived at a specific pedagogical structure that she believed was optimal for delivering the pedagogy. When the Montessori structure within which these philosophical elements are intended to be embedded is also included in the intervention, it might further enhance outcomes. These stuctural elements are listed below (Lillard 2019a , 2019b ).
Teachers who are well‐trained to carry out the intervention, who have learned to be sensitively responsive and authoritative in their implementation of the philosophy, have learned to deliver well the full set of Montessori lessons for the age group they are teaching and know how to tend to the carefully prepared Montessori environment. Thorough Montessori teacher training takes a year and comes from teacher trainers who have undergone an extensive decade‐long preparation to convey the philosophy and approach to others.
Classrooms that have children of specific 3‐year‐age spans that are thought to embrace developmental stages (thus all the children in the classroom need particular materials and lessons) and are also thought to be particularly conducive to peer learning.
A full set of specially‐designed Montessori materials that enables hands‐on learning and embodied cognition. These materials might also enhance interest, situate cognition, and evoke a sense of order.
A 2.5–3 h work period in the morning and afternoon, during which children exercise free choice and concentrate deeply, might assist in the development of executive function.
A classroom composition in which there are few adults (and only one trained teacher) and many children, to allow for peer learning and self‐determination. Montessori's ideal ratio was about 1:35, with possibly a nonteaching assistant for young children.
These structural elements are part of what is typically considered high‐fidelity Montessori, but not all Montessori interventions include them. By contrast, the philosophical elements are likely to be in any intervention that is designated as Montessori.
Although we have described the ideal Montessori intervention based on descriptions in Montessori's books, implementation varies in the real world (Daoust, 2004 ; Daoust, 2018 /; Daoust, 2019 ). Likewise, implementation would be expected to vary in research studies. The intervention being assessed here is Montessori as it appears in the body of research that purports to study it, which reflects the variation of Montessori in the real world. The actual implementations described in the included studies are covered in the Types of interventions section.
2.4. Why it is important to do this review
It is important to do this review because better evidence is needed for policymakers and parents to know if the Montessori system produces better or worse outcomes than business‐as‐usual approaches; no prior review has provided definitive evidence. We found only one quantitative meta‐analysis of Montessori education and it included only two studies of Montessori (Borman, 2003 ). This study reviewed nearly 30 comprehensive school reform programs and concluded that Montessori education serves as a reform with “promising evidence of effectiveness,” d = 0.27 [95% confidence interval (CI): 0.19, 0.35], p < 0.01. Other school reform programs that were also classified as having promising evidence of effectiveness in the Borman review were America's Choice, Atlas Communities, Paideia, and the Learning Network. Besides including very few Montessori studies, the Borman review only provided outcomes based on national standardized tests of academic achievement.
There have been several narrative reviews related to Montessori education's impact on academic and nonacademic outcomes, and they have focused especially on preschool‐aged children, termed primary in Montessori circles (e.g., Ackerman, 2019 ; Boehnlein, 1988 ; Boehnlein, 1990 ; Boehnlein, 1996 ; Jones, 2005 ; Marshall, 2017 ; Murray, 2010 ). Murray's ( 2010 ) narrative review of early (1960s and 1970s) and contemporary (2000s to present) Montessori research revealed that recent studies that employed improved statistical methods demonstrated results favorable to Montessori education. Jones ( 2005 ) reviewed Montessori education outcomes research in three areas: (1) the effects of the Montessori approach to at‐risk students, (2) the effects of the Montessori method on exceptional learners including learning disabled, developmentally delayed, and gifted/talented, and (3) comparative analyses of traditional schooling versus Montessori in student achievement and social development. Although thorough, Jones failed to take a systematic approach to searching the literature and did not quantitatively synthesize the research data. Boehnlein (Boehnlein, 1988 ; Boehnlein, 1990 ; Boehnlein, 1996 ) reviewed the Montessori education research that may be of interest to public schools, summarizing the results as follows:
Early research provides evidence that the Montessori method and environment are beneficial to low‐ and middle‐SES children.
Current research corroborates the early findings, in particular, the importance of the Montessori preschool experience.
Of specific importance for best results long‐term are the full 3‐year preschool program, trained Montessori teachers, and multi‐age grouping (Boehnlein, 1988 , p. 476).
Although relatively thorough, the Boehnlein reviews were narrative reviews without a systematic search strategy or quantitative synthesis. The same is true of recent narrative reviews by Ackerman ( 2019 ) and Marshall ( 2017 ). In sum, the narrative reviews of Montessori education lack the systematic quality and rigor afforded by a Campbell Collaboration review, and a systematic review of all the existing literature is needed to resolve the question of whether Montessori has an impact on child outcomes.
3. OBJECTIVES
The primary objective of this review was to examine the effectiveness of Montessori education, compared to traditional education, in improving academic and nonacademic outcomes for prekindergarten to high‐school‐aged students. The secondary objectives were to determine which of the following factors moderate the reported effectiveness of Montessori education: grade level, public versus private Montessori settings, type of assignment to experimental and contrast conditions, treatment duration, and length of follow‐up measurements.
4.1. Criteria for considering studies for this review
4.1.1. types of studies.
The Campbell protocol for this review can be found in Randolph ( 2016 ).
In this subsection, we describe the inclusion and exclusion criteria based on study design features.
We included studies that used group experimental (i.e., with random assignment) and/or quasi‐experimental (i.e., without random assignment) research designs.
We included pretest‐posttest with control group designs, posttest‐only with control group designs, and designs with case‐control matching on a measure of the same construct as the outcome construct.
We excluded pretest‐posttest without control group designs. Studies that used single‐participant, correlational, quantitative descriptive, or qualitative designs were excluded. The portions of mixed‐methods studies that met study criteria were included and the other information was excluded.
- o Group membership was determined through a random process, or
- o Equivalence was established at the baseline for the groups in the analytic sample.
Because of the potential for selection bias, we used strict criteria for establishing baseline equivalency in studies not using random assignment. Quasi‐experimental studies had to meet at least one of the following criteria to be considered for inclusion:
The authors used covariate‐adjusted means where at least one covariate was a measure of the same construct as the outcome. For example, we would consider a study with a mathematics outcome to have baseline equivalency if the authors adjusted for mathematics pretest scores. However, we did NOT consider a study as having baseline equivalency if it only adjusted for covariates that are correlated with the outcome. For example, we would not consider a study with a mathematics outcome to have baseline equivalency if it only adjusted for family income, although family income is known to correlate with academic achievement (Duncan, 2014 ). In short, covariates had to measure the same construct as the outcome for the study to be considered to have baseline equivalency based on covariate adjustments.
The authors matched participants based on a covariate that measures the same construct as the outcome construct.
The authors used gain scores to establish baseline equivalency. The pretest and posttest scores had to use the same measure or an equatable, scaled measure.
The authors provided evidence that there was not a statistically significant difference in pretest scores between Montessori education and traditional education groups.
We excluded studies in which the author did not report enough information to compute standardized mean difference effect sizes. We did not include studies that required us to impute means and standard deviations from medians and ranges/interquartile ranges. For studies published since 2000, we attempted to contact authors to get this information if it was not reported in the study and kept documentation of that information in the inclusion/exclusion data set provided in Randolph ( 2021 ).
Studies were included in the current review if they described their intervention condition as Montessori . Not every study specified the elements of Montessori that were implemented and those that described it to varying degrees. The descriptions provided yielded implementations ranging from full (well‐aligned with what is described in Montessori's books; 10 studies) to weak (four that were merely add‐ons to otherwise traditional programs). This range reflects the wide range of implementation of programs called Montessori in the real world: the term is not trademarked. To be objective, the label alone was used to determine study inclusion. In the Types of Interventions section, we report an analysis of the range of implementations used in the included studies; we also note our grade of the implementation quality in the Intervention characteristics column of Table 1 .
Characteristics of included studies.
Note : None of the included studies had disclosed any declarations of interest.
Abbreviations: PCG, posttest‐only with control group design; PPCG, pretest–posttest with control group design; PPRM, pretest–posttest with matched controls.
4.1.2. Types of participants
We included studies in which the participants were in preschool, elementary, middle school/junior high school, and/or high school. When a study had participants in two or more of these groups, we classified the age group for the study as the age group with the greatest number of participants. See the coding book in the supplemental information (Randolph, 2021 ). The “location, setting, status, or definition of the condition and demographic factors” (Campbell Collaboration, 2019b , p. 8) were not considered as inclusion or exclusion factors. We created some exploratory emergently‐coded variables for demographic factors (e.g., gender, family income, etc.) and country of origin, but those factors were not included as moderators in this analysis. See the Data extraction and management section for more details on the demographic information collected.
4.1.3. Types of interventions
The intervention was defined broadly as Montessori education. We operationalized Montessori education as an intervention in which the study authors claimed to have used the Montessori method of education; this occurred in both public and private school settings. In all studies, Montessori education was a separate measurable intervention. Traditional or business‐as‐usual education was the comparison condition for all studies in the analysis. Next, we describe what these interventions were in the included studies; they align with the range of implementations of Montessori and traditional education in the real world, given that neither term is trademarked.
Montessori interventions
Although the included studies designated the intervention condition as “Montessori,” the studies' methods sections provide varying levels of description of the intervention and although most of those descriptions indicated that the intervention had the philosophical elements of Montessori education, some revealed that they deviated in certain ways from the structural ones. Because of this variation, in response to the first set of reviews we categorized (post‐hoc) the Montessori conditions into five categories. All of the Montessori interventions appeared sensitive to the philosophical elements of Montessori, for example touching on free choice and hands‐on materials in the article Introduction if not also in Methods. Where they varied the most was in structural elements.
Although we performed this categorization, we do not advise considering how effect sizes might vary with the implementation levels because effect sizes stem from many sources including the different effectiveness of the Montessori intervention relative to its control condition; the control or traditional conditions also varied, as they do in the real world. A second reason not to consider these levels for analysis is that the variables that led to the categorizations are our best estimates based on what was reported; they lack precision. In sum, we provide the levels and their descriptions here only to give readers a sense of the range of implementations that were considered in the meta‐analysis.
At the highest level was full implementation, meaning that the school was recognized by a respected association like the Association Montessori Internationale (AMI) or the Swiss Montessori Association or at least had teachers who were fully trained by AMI or the American Montessori Society before the intervention took place. AMI recognition entails the structural elements of having AMI‐trained teachers (trained to implement the philosophy to a high degree), a specific 3‐year‐age range in each classroom (e.g., children ages three to six or six to nine), a 2.5–3 h uninterrupted work period during which children are free to choose their own work, large class sizes and few adults, and a full set of Montessori materials. There are also no grades; the materials are self‐correcting. The Montessori condition in 10 of the 32 studies appeared to meet these criteria (Denervaud, 2019 , 2020 ; Elben, 2015 ; Lillard, 2006 ; Lillard, 2012 ; Lillard, 2017 ; Mix, 2017 ; Rathunde, 2005a ; Rathunde, 2005b ; Yussen, 1980 ).
The next highest level, observed in six studies, was medium implementation; for these, the article mentions some accreditation or teacher training, but the accrediting organization was unspecified (“a local Montessori association,” Prendergast, 1969 ), and teacher training was done by a respected organization like AMS but not all teachers had completed the training at the time of data collection (Mallett, 2015 ). In one case the description of Montessori suggested good implementation, but photos included in the article incorrectly called some commercial toys “Montessori materials” (Faryadi, 2017 ), which put the study in a lower category. In another case, implementation was discussed and a rubric was developed to measure it, and the measure indicated a nontrivial level of deviance from the highest level of implementation among the schools studied (Culclasure, 2018 ). In another case, the UK Montessori Schools Association had accredited the Montessori schools, but they deviated by including pretend play; this may not be a major deviation but did suggest not fully implementing Montessori (Kirkham, 2017 ). One study considered to have a Montessori condition at this medium level had AMS‐trained teachers or teachers in training but used only two ages of children per class instead of three (Manner, 1999 ).
At the next level, observed in seven studies, Montessori was claimed to be implemented as a full classroom program, but there was at least one very serious deviation from the full implementation of the structure, and other elements were unspecified or suggested other problems. For five of these, classrooms had only one age level: 4‐year‐olds (Ansari, 2014 ; Galindo, 2014 ; Jones, 1979 ; Miller, 1983 ; Miller, 1984 ); three of these also had teachers with only a 6‐week training course. For one study, it was stated that some classrooms had mixed grades and it specified two grades, suggesting some other classrooms had only single ages (Besançon, 2013 ). Another study at this level called the intervention modified Montessori and stated that some children had group instruction during part of the session (Coyle, 1968 ; description taken from Concannon, 1966 ). These all suggest some serious deviations from high‐fidelity Montessori.
There were five studies in which the level of Montessori implementation could not be determined: Fleege ( 1967 ), Kayili ( 2016a ), Kayili ( 2016b ), Tobin ( 2015 ), Aydoğan ( 2016 ). In these studies, Montessori philosophy was described appropriately, but there was not enough information about the classrooms themselves to determine whether the structural elements were in place; one was left with very little sense of how Montessori was actually implemented. In the fifth study in this category, not only was the Montessori implementation unspecified, but the intervention appeared to occur over a 7‐week period (Aydoğan, 2016 ); it is unclear if children had had Montessori programming before the 7‐week‐long study.
Finally, there were four studies in which the Montessori condition appeared to be an add‐on curriculum implemented for a limited period of time. Alburaidi ( 2019 ) set up a school Science Hall that had six centers offering hands‐on materials, choice, meaningful activities, order, and other Montessori characteristics; intervention students spent their 45‐min science classes covering specific topics in this hall with a teacher taught to guide in a Montessori manner. Doğru 2015 taught children with ADHD to use Montessori sensorial materials (e.g., tactile boards, binomial cubes) over 8 weeks, for 45 min each week. Hoseinpoor ( 2014 ) offered 12 sessions with a variety of sensorial activities and role‐play activities; it appeared that these built on each other and that each session included the activities from prior sessions. From their introduction, it seems the intervention children were free to choose among these activities. The final study in this group (Juanga, 2015 ) implemented Montessori for a single day using assigned workstations; the environment was prepared, the activities were hands‐on, and a Montessori consultant guided the teacher.
Although there is tremendous variety in these implementations, they reflect the variety of implementations of Montessori in the real world; the term is not trademarked.
Comparison condition: Traditional education
Detailed descriptions of the comparison condition were rarely provided, as if the terms conventional or traditional made clear what they were. At the elementary school level (studies with children ages 6–12), some studies specified that the control condition had teacher‐directed, whole‐class learning involving textbooks, teachers checking and grading work, single‐aged classrooms, and class periods of limited duration (in some cases 45 min). Most studies outside of the United States referred to a national program (such as traditional Swiss or French or Thai pedagogy, or the UK National Curriculum). At the preschool level, some non‐western studies referred to (for example) traditional Malaysian or Philippine, or Iranian conventional education, without reference to a set curriculum; further details suggested these were also teacher‐centered, lecture‐style approaches. By contrast, for some (especially older) studies conducted in the United States, traditional education at the preschool level was described as comprised of free play and pretense (Coyle, 1968 ; Miller, 1983 ; Miller, 1984 ; Prendergast, 1969 ). One study specified that its conventional condition implemented HighScope, which involves centers offering various art, play, and learning activities; this method is modeled after the Perry Preschool Project (Ansari, 2014 ). Others provided little to no information about their conventional preschool condition, referring simply to the control program as pre‐K, traditional, business‐as‐usual, or non‐Montessori. In recent years preschool programs in the United States are more likely to be teacher‐centered and involve little play (Bassok, 2016 ).
In sum, traditional and Montessori conditions reflected a range, across studies, that is reflective of the range of implementations of each in the real world. Although we categorized the Montessori implementation into levels post‐hoc, we advise against considering the relation between effect sizes and these estimated levels of implementation, first because our categorization is imprecise, and second because effect sizes reflect the relative difference across Montessori and control conditions in each study, and control conditions varied as well in ways that are impossible to rank precisely.
4.1.4. Types of outcome measures
We included two broad categories of outcomes: academic and nonacademic. Because of the wide range of nonacademic outcomes, we used an emergent approach to arrive at the specific outcomes being measured in the Montessori literature. We considered academic outcomes to be primary measures and nonacademic outcomes to be secondary measures.
Primary outcomes
The variety of academic outcomes was reduced to five categories of outcomes:
General academic ability. This outcome included measures that did not clearly fit one of the other categories (like math or reading), such as the cognitive subscale of the learning Accomplishment Profile‐Diagnostic (Lap‐D) (see Ansari, 2014 ); this subscale included counting as well as matching and did not report scores separately. It also included three subtests of the Bracken scale (see Galindo, 2014 ) including identifying colors and comparing sizes. Tests of problem solving that were not explicitly mathematical were also categorized as General Academic Ability. Finally, different academic abilities concatenated into a general academic score were included in this category.
Mathematics. Grouped for the mathematics analysis were tests of computation and math concepts, tests of number line estimation and of counting, the Woodcock‐Johnson Applied Problems subtest, and tests administered by states to examine math achievement. Most of the math tests were also standardized, normed measures.
Language/literacy. This cast a broad net around all measures pertaining directly to reading, vocabulary, language use, and language perception, including the Stanford Achievement Tests' language measures, the Woodcock‐Johnson Letter‐Word and Word Attack subtests, and tests of reading comprehension, speech sound discrimination, spelling, and vocabulary. Some tests were administered by states to examine English Language Arts proficiency. Virtually all the literacy tests were standardized, normed measures.
Science. Just three studies included measures of science achievement using a pre‐post design or controlling for pretest scores. Two studies created science tests for the study; the third, which measured children over 3 years, used the South Carolina state tests of science achievement.
Social studies. This outcome included measures such as South Carolina state assessment in social studies.
Individually the academic outcomes had too few studies to do a moderator analysis with cluster–robust variance estimation; therefore, we also created an aggregated academic outcomes variable that comprised studies that contributed at least one effect size to one or more of the academic outcomes listed above.
Secondary outcomes
Although the nonacademic outcomes are listed as secondary outcomes here, we do not intend to assign a hierarchal structure to academic and nonacademic outcomes. After examining the variety of nonacademic outcomes, we found that there were four major nonacademic outcomes:
Creativity. This included measures like Alternate Uses, where children must come up with all possible uses for a common object like a paper clip, the Torrance Test of Creativity (which includes Alternate Uses as one of several tasks), and tests where children need to create a drawing or story and a panel of judges rates their creativity.
Executive function. Executive function was measured directly with a wide variety of tasks and also indirectly with parent or teacher report questionnaires. The direct tasks include, for example, the Simon‐Says‐like game “Head Toes Knees Shoulders”, the Flanker task, and reciting a string of digits backward. An example of a teacher‐ or parent‐report measure of executive function is the Behavior Rating Inventory of Executive Function or BRIEF.
Inner experience of school. Some studies addressed how children experience school, asking how much they like school on a survey or using the experience sampling method, whereby pagers randomly beep children and ask them to rate their immediate emotional and cognitive experience; in such studies, the in‐school data were used from children attending different types of schools.
Social skills. Included in this category were both tests assessing children's social knowledge, including tests of basic social cognition (like tests of emotion recognition and the Theory of Mind scale) and tests assessing knowledge about managing peer relations (e.g., asking how one would respond to a social conflict, as with Rubin's Social Problem Solving test). Also included were teacher and parent ratings of children's social behavior (such as the Deveraux Early Childhood Assessment), and live coding of social behavior (e.g., ambiguous rough and tumble play on a playground).
Individually, the nonacademic outcomes had too few studies to do a moderator analysis with cluster–robust variance estimation; therefore, we created an aggregated nonacademic outcomes variable from studies that contributed at least one effect size to one or more of the nonacademic outcomes listed above.
4.2. Search methods for identification of studies
4.2.1. search methods.
The initial electronic search was conducted in March 2014 with no limits on the date of publication. A follow‐up search was conducted in February 2020 to find any studies published since January 2014. (To be cautious, we intentionally created a 3‐month overlap of the first and second rounds of searching.) No geographic limitation was placed on the search. Results were gathered from electronic databases, the open web, online gray literature websites, and directly from authors and experts in the field.
A comprehensive search strategy was designed and implemented by an academic librarian in an attempt to retrieve all experimental and quasi‐experimental studies that adhere to the inclusion and exclusion criteria. We examined both published and unpublished literature. Articles and gray literature were gathered using online databases that cover education, sociology, and psychology, and recommendations from experts in the field of Montessori education. Only results written in English were considered for inclusion, although no language limiters were utilized in the searches. We employed both free‐text and controlled vocabulary terms in the searches. All permutations of search terms were used during the search process. We complemented our search with a thorough examination of reference lists of relevant retrieved studies, both included and excluded, and contacted experts in the field to identify any ongoing or unpublished studies.
Studies were identified using the following electronic databases and online sources:
Academic Search Complete (EBSCO)
AERA Online Paper Repository
American Montessori Society Montessori Research Library
Arts & Humanities Citation Index (Web of Science)
Dissertations & Theses Global (ProQuest)
Education Full‐Text/Education Research Complete (EBSCO)
Education Journals (ProQuest)
ERIC (EBSCO)
Google Scholar
PsycINFO (EBSCO)
Professional Development Collection (EBSCO)
Research Library (ProQuest)
Social Sciences Citation Index (Web of Science)
Social Sciences Journals (formerly Social Science Database) (ProQuest)
SocINDEX with Full Text (EBSCO)
Sociological Collection (EBSCO)
Teacher Reference Center (EBSCO)
During the search process, we utilized phrase searching and truncation methods to find all variations of relevant search terms. Database thesauri, when available, were used to find controlled vocabulary descriptors and related descriptors which were integrated into search iterations.
The search strategy was customized as needed for each database and was changed to include Montessori classrooms of all grade levels. Details on the search strategy for each source are provided in the Appendix. Search strategies were tailored to the unique controlled vocabularies of each database and were used in conjunction with free text search terms, which can be found in Supporting Information: Appendix 1 . The search strategy included keywords and subject headings pertaining to setting (school), intervention (Montessori), and outcome (academic and nonacademic). The number of search results from each source is provided in Table 2 .
Search results by academic database.
There were limitations related to this search. Namely, the search was limited by the omission of studies published since February 2020 and the studies were also limited to English.
4.2.2. Electronic searches
A comprehensive database search included the following online subscription databases:
EBSCO Academic Search Complete
EBSCO Education Full‐Text
EBSCO Professional Development Collection
EBSCO PsycINFO
EBSCO SocINDEX with Full Text
EBSCO Sociological Collection
EBSCO Teacher Reference Center
ProQuest Dissertations & Theses
ProQuest Education Journals
ProQuest Research Library
ProQuest Social Sciences Journals
Web of Science Arts & Humanities Citation Index
Web of Science Social Sciences Citation Index
Gray literature
Professional Montessori association websites, gray literature, digital repositories, and social science research websites were identified through expert referrals and Internet searches, and conference proceedings, reports, working papers, white papers, and preprints were reviewed to find relevant research. Government documents were identified primarily through the Education Resources Information Center (ERIC) and the US Department of Education's Institute of Education Sciences (IES) and various subsidiaries, projects, and digital libraries. ERIC was searched both using the EBSCOhost platform and government‐hosted websites to ensure an exhaustive search. Websites were searched for relevant studies, reports, and citations. Sources for gray literature included:
American Educational Research Association ( http://www.aera.net/ )
American Montessori Society ( https://amshq.org/ )
Association Montessori Internationale ( https://montessori-ami.org/ )
Association Montessori Internationale/USA ( https://amiusa.org/ )
ERIC ( https://eric.ed.gov/ )
Institute of Education Sciences ( http://ies.ed.gov/ )
Montessori Educational Programs International ( https://www.mepiinc.com/ )
National Center for Montessori in the Public Sector ( http://public-montessori.org/ )
OpenGrey [discontinued] ( http://www.opengrey.eu/ )
Pan American Montessori Society [discontinued] (The website no longer exists.)
Social Science Research Network ( http://papers.ssrn.com/sol3/DisplayAbstractSearch.cfm )
Society for Research on Educational Effectiveness ( https://www.sree.org/ )
Internet search
Google, Google Scholar, and Bing were used to search the open web in an attempt to fill in any gaps left after searching the specialized sources. Results were analyzed until a saturation point was reached, (i.e., until further searching led to no new articles for inclusion).
See Table 2 and Supporting Information: Appendix 1 for further information on Internet searches.
4.2.3. Searching other resources
Contacting other researchers.
Authors of prior studies and other experts in the Montessori method were contacted to obtain unpublished research or to get further clarification on published studies. A record of attempts to contact authors can be found in the notes section of the inclusion and exclusion data set provided in the supplemental information to this review (Randolph, 2021 ).
Prior reviews and reference lists
The prior reviews that were searched are mentioned in the Why it is important to do this review section and there is additional information in the supplemental information provided in Randolph ( 2021 ).
During both the search and selection processes, studies determined to be relevant enough for closer inspection of reference lists were searched for citations to additional studies. For example, a study examining Montessori and non‐Montessori outcomes that had insufficient data and was therefore not included would be examined for references to other studies.
4.2.4. Hand search
Hand searches were conducted with available journals and conference proceedings that had not been adequately electronically indexed. The Journal of Montessori Research & Education, Journal of Montessori Research , and AMI Journal were hand searched by examining tables of contents for all available issues with no data limitations, both in 2014 and 2020.
4.3. Data collection and analysis
In general, we followed the data collection methods in the Cochrane Handbook (Higgins, 2021 ), the MECCIR Conduct Standards (Campbell Collaboration, 2019b ), the MECCIR Reporting Standards (Campbell Collaboration, 2019b ), metafor documentation (Viechtbauer, 2010 ), and the cluster‐robust methods described in Tanner‐Smith ( 2014 ), Tanner‐Smith ( 2016 ), and Tipton ( 2015 ). The metafor (Viechtbauer, 2010 ), metaviz (Kossmeier, 2020 ), robumeta (Fisher, 2017 ), and ClubSandwich (Pustejovsky, 2020 ) packages in R (R Core Team, 2019 ) were used for all meta‐analyses. Supporting R packages used for data preparation can be found in the R code provided in the supplemental information (Randolph, 2021 ).
Following the suggestions of the American Statistical Association (Wasserstein, 2019 ), we refrained from null‐hypothesis statistical significance testing when possible. However, for readers interested in interpreting our results in the null‐hypothesis testing paradigm, we provide Benjamini‐Hochberg‐adjusted critical alpha values (Benjamini, 1995 ; Benjamini, 2005 ). These adjusted critical values and confidence intervals are meant to keep the false discovery rate and false coverage rates, respectively, at an overall 0.05 level. We report the adjusted values separately for primary objectives with 11 main‐effects estimates in the supplemental information (Randolph, 2021 ).
4.3.1. Selection of studies
After the information retrieval expert for this review initially identified potential studies for inclusion based on titles and abstracts, at least two reviewers independently reviewed the full text and/or abstracts of the study to make decisions about study inclusion and exclusion. Disagreements were resolved by consensus.
4.3.2. Data extraction and management
The data specified on the coding sheet were extracted independently by at least two reviewers and a consensus decision was reached when there was disagreement.
Our coding sheet was divided into the following categories: publication characteristics, setting, outcomes, participants, research design (including an indicator of methodological quality and risk of bias), statistical analysis of the study, and characteristics of the effect size. The research design, statistical analysis of the study, and characteristics of effect size were based on the Campbell Collaboration Methods Policy Briefs (Campbell Collaboration, 2019a ; Campbell Collaboration, 2019b ).
We did not include the risk of bias variables related to blinding because blinding was not possible in the educational studies we reviewed. We assumed a high risk of blinding bias for all studies; therefore, we do not have the risk of bias codes related to blinding. We expected quasi‐experimental studies, so we intended to measure what methods and confounding variables the study authors controlled for.
The list below indicates the main variables used in the coding sheet to extract data. The revised coding sheet can be found in the supplemental information provided in Randolph ( 2021 ). This list may differ slightly from the one in the protocol because we used an emergent coding approach to create some variables. We created some emergent codes when we were unsure what categories we would find.
Identification of Studies (e.g., study id)
Publication and Participant Characteristics (e.g., the format of publication, type of school, grade level of participants)
Experimental/Statistical Controls (e.g., covariates)
Effect Size Information (e.g., the g effect size and its variance)
Equivalency at Baseline (e.g., evidence on how equivalency at baseline was established: nonstatistically significant t tests on a pretest, gain scores, covariate adjustments.)
Effect Size Calculations (e.g., the standardized raw mean change approach.)
Notes/other
4.3.3. Assessment of risk of bias in included studies
The RoB 2 tool (Sterne, 2019 ) was used to assess risk of bias in experimental studies and the Robbins‐I tool (Sterne, 2016 ) was used to assess risk of bias in nonexperimental studies. Two authors (DKH, AKM) independently applied these tools to each article and reached consensus agreement in the case of disagreement. A third author (ASL) reviewed the accuracy of the risk‐of‐bias ratings. Documentation of risk‐of‐bias ratings can be found in the supplemental information (Randolph, 2021 ).
RobVis (McGuinness, 2020 ) was used to create traffic‐light plots. Risk of bias includes allocation (selection bias), blinding (performance bias and detection bias), incomplete outcome data (attrition bias), selective reporting (reporting bias), and other potential sources of bias explained in Higgins ( 2021 ). A helpful table explaining the various risk bias domains from Higgins ( 2021 ) is reproduced in Supporting Information: Appendix 2 .
4.3.4. Measures of treatment effect
We used a standardized mean difference effect size (Hedges' g ) as the measure of treatment effect. The effect size was calculated in one of three ways:
using the standardized mean difference approach in the metafor package (Viechtbauer, 2010 ) in R (R Core Team, 2019 ),
using the standardized mean change in raw score approach in the metafor package in R,
or using an “other” approach with Wilson's (Wilson, n.d.) Practical Meta‐Analysis Effect Size Calculator.
For studies using the posttest‐only with control group design, we used the escalc function (measure = “SMD”) of the metafor package in R to calculate the effect size and its unbiased estimate of the sampling variance (vtype = “UB”). For pretest–posttest designs with control groups and covariate adjustments, we used the same method described above using covariate‐adjusted means.
For pretest‐posttest designs with control groups that did not use covariate adjustment or for studies that used case‐control matching on a baseline measure of the outcome construct, we assumed the pretest standard deviation was an unbiased estimate of σ and used a raw mean change score approach as suggested in Morris ( 2002 ) (Equation 6). (The lack of pretest–posttest correlations in many studies was the primary reason we adopted Morris's raw mean change score approach.) Namely, we calculated the posttest–pretest mean differences for each group and used the pretest standard deviation of that group as the measure of variance. For designs with case‐control matching, we used the case‐control group standard deviation as the measure of variance. We then used the standardized mean change raw score parameter (measure = “SMCR”) in the escalc function of the metafor package of R to calculate the standardized mean difference effect size.
Finally, for a small number of studies where there was insufficient information to use the metafor package to calculate an effect size and its variance, we used Wilson's (Wilson, n.d.) Practical Meta‐Analysis Effect Size Calculator to calculate standardized mean difference effect sizes and their variances. The supplemental information provided in Randolph ( 2021 ) has notes on the calculation of effect sizes for these “other” effect size calculations and the R code that was used to calculate effect sizes.
4.3.5. Unit of analysis issues
One of the main difficulties in synthesizing the Montessori literature was accounting for the multiple measures of an outcome's effect size within individual studies. To do this we used the cluster‐robust meta‐analytic approach suggested in Tanner‐Smith ( 2014 ), Tanner‐Smith ( 2016 ), and Tipton ( 2015 ). This allowed us to use every effect size while still accounting for the dependency of effect sizes within studies. Specific information on the cluster‐robust approach can be found in the Data synthesis section.
4.3.6. Dealing with missing data
If outcome data were missing, we attempted to contact the study author(s) to get access to the missing data. If the author(s) never responded to our query, we excluded the study from the analysis. See the supplemental information (Randolph, 2021 ) for a record of which studies were excluded because of missing data and for documentation of attempts to contact study authors. In some cases, the authors sent the original data set and we used that data to generate the data missing in the study itself. A record of these instances is recorded in the data set and spreadsheet of included/excluded studies in the supplemental information as well.
4.3.7. Assessment of heterogeneity
We used several methods to address study heterogeneity largely following the methods in Higgins ( 2021 ). For each outcome, we examined a forest plot containing the effect size estimate and its 95% confidence interval for each study and the weighted mean effect size and its 95% confidence interval. For outcomes with too many effect sizes to visualize in a forest plot with R software, we created a graphic display of study heterogeneity (i.e., a GOSH plot) (Olkin, 2012 ).
Although not included here, we also examined Baujat plots, radial plots, residual plots, and various other diagnostic and leverage tables included in the metafor package to identify studies with atypical heterogeneity. Funnel plots were examined for outcomes with 10 or more studies. We investigated kernel density plots of unweighted effect sizes to examine the distributional characteristics of the individual outcomes. In addition to the visual analysis of heterogeneity, we also examined several statistical measures of heterogeneity: the value of Q, its df , and its related p value, and the value of the I 2 statistic. The R code and data set for these analyses can be found in Randolph ( 2021 ).
4.3.8. Assessment of reporting biases
We used two methods to examine the presence and magnitude of publication bias: funnel plots and a trim and fill analysis,
First, as suggested in Higgins ( 2021 ), we assessed the degree of potential publication bias by visually analyzing funnel plots for outcomes with more than 10 studies and we did not carry out null‐hypothesis‐based statistical significance tests of publication bias.
Second, we used a trim and fill method as an additional tool to detect the presence of bias and to quantify the degree and direction of the bias. A clear explanation of the trim and fill method, from Murad ( 2018 , p. 85) follows:
The trim and fill method is based on the funnel plot in which missing studies are imputed by creating a mirror image of opposite corresponding studies. The adjusted effect size accounting for the missing studies can be used as a sensitivity analysis to determine the presence and magnitude of publication bias. It is important to note that this adjusted effect size is based on strong assumptions about the missing studies and should only be used for the purpose of sensitivity analysis (i.e., should not be considered as a more accurate effect size to be used for decision making).
Specifically, we used the trimfill() function, which is based on the work of Duval ( 2000 ), in R's metafor package (Viechtbauer, 2010 ) using L0, R0, and Q0 parameters and using the Sidak‐Johnson method for random‐effect synthesis. A sensitivity analysis of those parameters, which is not reported here, clearly indicated that the bias was left‐sided. Duval ( 2000 ) recommends using either the L0 or RO estimate; for the sake of brevity, we only report the trim and fill results using the L0 parameter. The results did not differ substantively between L0 and R0.
As suggested in Shi ( 2019 ), we carefully considered how outlying studies might covary with publication bias results through a careful investigation of outlying studies as identified through funnel plots, Baujat plots, residual plots, leverage statistics, and forest plots. If a study was suspected of being an outlier, we reread the full text of the article to find insights into potential sources of heterogeneity; the results of the outlier investigation can be found threaded throughout the Results and Discussion sections.
Bias in the selection of the reported result was assessed using the RoB2 tool (Sterne, 2019 ) for studies with random assignment and the ROBBINS‐I tool (Sterne, 2016 ) for studies without randomized assignment.
4.3.9. Data synthesis
For studies that reported multiple effect sizes for a single outcome, we used a cluster‐robust method for synthesizing effect sizes as described in Tanner‐Smith ( 2014 ), Tanner‐Smith ( 2016 ), and Tipton ( 2015 ) using the robumeta package (Fisher, 2017 ) in R. We used Tipton ( 2015 )'s small‐sample correction and assumed the within‐study effect size correlation ( ρ ) to be 0.80. As suggested in Tipton ( 2015 ), we investigated this assumption with a sensitivity analysis using a range of values of ρ. We also estimated that dependencies between within‐study effect sizes were based more on correlations of effect sizes within studies than hierarchical effects, so we used “correlations” as model weights. The social studies outcome comprised effect sizes from only one study, so we used a random‐effects model there instead of a cluster–robust model. The complete R code for this analysis can be found in the supplemental information provided in Randolph ( 2021 ); see in particular the robust_main() function for details of the cluster–robust variance estimation and diagnostic methods.
For any outcomes with effect sizes from just one study (e.g., social studies) we used a random‐effects model (with REML estimator). See the random_main() function in the R code of Randolph ( 2021 ) for more details.
No data transformations were conducted and no missing data were imputed. (We excluded studies that had insufficient information to calculate an effect size and its variance.) We assessed model quality through an examination of residual plots, leverage charts, funnel plots, forest plots, radial plots, and Baujat plots; see the robust_main() and random_main() functions in the R code in Randolph ( 2021 ) for details.
4.3.10. Subgroup analysis and investigation of heterogeneity
For aggregated academic and aggregated nonacademic outcomes, we conducted the following a priori subgroup analyses using cluster‐robust meta‐regression with the robumeta package (Fisher, 2017 ) using the methods described in Tanner‐Smith ( 2014 ). See the corresponding functions in the R code written for this meta‐analysis (Randolph, 2021 ) for more information; the corresponding functions are listed in italics below.
Duration of follow‐up (in years); mod_followup()
Treatment duration (in weeks); mod_treatment()
Intervention setting (i.e., private or public Montessori); mod_setting()
Student's grade level (preschool, elementary, middle school, or high school); mod_level()
Assignment (random vs. nonrandom); random_method()
For the two continuous moderators (duration of follow‐up and treatment duration), we first performed centering and estimated both between‐study and within‐study effect size estimates. We did not conduct a subgroup analysis for each outcome individually because of the sample size requirements for cluster‐robust meta‐regression. For multinomial outcomes, we conducted an omnibus test of statistical significance as suggested in Tanner‐Smith ( 2014 ).
The social studies outcome only had one study with multiple outcomes so a random‐effects model was used. See the R code in the supplement (Randolph, 2021 ) for more details.
As mentioned in the protocol, we extracted information on student demographic characteristics, such as race/ethnicity, at‐risk status, gifted/talented, or measures of socioeconomic status, but we did not intend to examine these characteristics as moderators in this review. We intended to use these demographic data to richly characterize study participants, to examine whether these variables were used as covariates in study analyses, and to facilitate follow‐up reviews that might examine demographic characteristics as moderators.
4.3.11. Sensitivity analysis
We conducted the following sensitivity analysis. See the related R functions in the supplemental information (Randolph, 2021 ) for more information; the function names are given in italics below.
We compared results between cluster‐robust, random‐effects, and fixed‐effects models; es_calc_method()
We examined how ρ (i.e., the correlation of within‐study effect sizes) covaried with effect sizes; robust_main()
We conducted a leave‐one‐out analysis using the leave1out function in metafor the package; robust_main()
4.3.12. Summary of findings and assessment of the certainty of the evidence
We used the GRADE approach described in Higgins ( 2021 ) to assess the certainty of evidence and summarize findings. The GRADE approach results in one of four ordinal ratings of the certainty of evidence in an outcome: high, moderate, low , or very low .
The first step in the GRADE approach is to establish an initial level of certainty. Randomized studies or studies evaluated using the ROBINS‐I tool (Sterne, 2016 ) for examining risk of bias in nonrandomized studies are given an initial certainty of high certainty . Observational studies not using the ROBINS‐I tool are given an initial certainty rating of low certainty .
Next, review authors downgrade or upgrade the quality of evidence based on several GRADE factors. They downgrade the certainty of evidence based on risk of bias , inconsistency , indirectness , imprecision , and publication bias . Review authors can also upgrade evidence based on a large effect size , evidence of a dose response , or if there are plausible confounding factors that may work together to underestimate a treatment effect. (More detailed information on those factors causing upgrades or downgrades can be found later in this section). The final step is to establish a final level of certainty into one of the four categories mentioned above.
In terms of the number of levels to downgrade or upgrade, we used this guidance from Higgins ( 2021 ):
The highest certainty rating is a body of evidence where there are no concerns in any of the GRADE factors…. Review authors often downgrade evidence to moderate, low, or even very low certainty evidence, depending on the presence of the five [GRADE] factors. Usually, certainty ratings will fall by one level for each factor, up to a maximum of three levels for all factors. If there are severe problems for one domain…., evidence may fall by two levels due to that factor alone… Review authors will generally grade evidence from sound nonrandomized studies as low certainty, even if ROBINS‐I is used. If, however, such studies yield large effects and there is obvious bias explaining those effects, review authors may rate the evidence as moderate or‐‐if the effect is large enough‐‐even as high certainty. (p. 391)
The first author independently assessed the certainty of evidence and other authors reviewed those certainty assessments. Any disagreements were resolved through consensus. The fine details of the GRADE approach used in this review are explained below.
All included studies initially were assumed to have an initial level of “high certainty” because they were all randomized trials or were evaluated with the ROBINS‐1. We then downgraded or upgraded the level of certainty based on the various GRADE factors.
First, we considered risk of bias. We assumed all studies to have low risk of bias. We then downgraded the certainty of evidence by one or two levels based on the RoB 2 (Sterne, 2019 ) for studies using random assignment and the ROBINS‐I tool (Sterne, 2016 ) for studies not using random assignment. We used the following criteria for downgrading studies based on their risk of bias:
A rating of high certainty evidence can be achieved only when most evidence come from studies that meet the criteria for low risk of bias. The certainty of evidence might be downgraded by one level when most of the evidence comes from individual studies either with a crucial limitation for one item, or with some limitations for multiple items. (Higgins, 2021 , p. 392)
Furthermore, it was possible to downgrade two levels when there were very serious limitations defined as a “crucial limitation for one or more criteria sufficient to substantially lower their confidence of an effect” (Higgins, 2021 , p. 393). See the Assessment of risk of bias in included studies section for more details on risk of bias.
After downgrading for risk of bias, we then downgraded for other GRADE factors: inconsistency, indirectness, imprecision, or publication bias and upgraded for large effects, dose–response effects, or opposing residual bias and confounding as described in Higgins ( 2021 ).
We downgraded for inconsistency if an outcome met Higgins' ( 2021 , p. 259) definition of considerable heterogeneity: I 2 values equal to or above 75%.
We downgraded for indirectness when there were indirect comparisons or that could cause “a restricted version of the main review question in terms of population, intervention, comparator, or outcome” as described in Higgins ( 2021 , pp. 393–394).
We downgraded for imprecision up to two levels if an outcome did not meet each of the three following criteria:
The number of participants for that outcome was below the optimal information size. We calculated the optimal information size as the sample size needed for a study given the oft‐used Cohen ( 1988 ) convention for a “small” effect size (i.e., a standardized mean difference effect size of 0.20), one predictor, α = 0.05, β = 0.80, and a two‐sided test. In this case, the optimal information size was 387 as suggested in a sample size table from Randolph ( 2019 ).
The 95% confidence intervals for the standardized mean difference for the outcome included 0.00. We downgraded up to two levels if the 95% CI included small positive effects and small negative effects.
The df resulting from a cluster‐robust analysis was less than 4.00 (Tanner‐Smith, 2014 ; Tanner‐Smith, 2016 ; Tipton, 2015 ).
For outcomes with more than 10 studies, we downgraded for publication bias if the funnel plots showed marked asymmetry.
We upgraded for large effects if the standardized mean difference effect size was greater than 0.80 in absolute value, which corresponds with Cohen ( 1988 )'s convention for a “large” effect size in laboratory studies in the behavioral sciences.
We upgraded if there was evidence of a dose–response effect (i.e., if there was evidence from a meta‐regression that treatment duration had a positive correlation with the effect size).
Finally, we upgraded for opposing residual bias and confounding if we found strong evidence that “all plausible biases from randomized or nonrandomized studies may be working to underestimate an apparent intervention” (Higgins, 2021 , p. 397).
When assigning textual descriptions of the magnitude of effect sizes, we used the conventions of Cohen ( 1988 ) such that standardized mean difference effect sizes of 0.20 are small , 0.50 are medium , and 0.80 or greater are large .
5.1. Description of studies
5.1.1. results of the search.
The systematic search including the initial 2014 and the follow‐up 2020 search yielded 2,012 records from all sources. As shown in the flow diagram in Figure 1 , after the removal of duplicates, abstract screening, and full‐text screening, 32 studies met the criteria for inclusion and were included in this meta‐analysis.
Flow diagram of the literature search for studies included in meta‐analysis. *Studies could have been excluded for more than one reason so the sum of exclusion reasons does not equal 141.
The PRISMA flow diagram combines results from the initial and follow‐up searches of online academic databases ( n = 1864). Also displayed are records retrieved from other sources, including gray literature websites, Montessori professional association websites, hand‐searched journals, reference searches, expert referrals, and open web search engines ( n = 148). After duplicate removal, 1,636 results were screened at the title and abstract level, excluding 1463 records based on inclusion criteria. The full texts of the remaining articles ( n = 173) were screened and 141 were excluded, leaving 32 studies included for quantitative synthesis. Bibliographic data for retrieved studies were managed using Zotero bibliographic software. A complete list of the included and excluded studies, along with detailed information on why each study was excluded is provided in the supplemental information in Randolph ( 2021 ).
5.1.2. Included studies
Table 1 summarizes the studies and key characteristics of the 32 included studies. It includes each study's in‐text citation, a brief description of the participants including country (the default is the United States because over half the studies were conducted there), the study design, a brief description of the intervention, the outcomes used in the meta‐analysis, the dates when the study was done if noted in the article, and the funding source. No study declared a conflict of interest.
Table 3 provides information, in the aggregate, on the characteristics of studies that contributed at least one academic or nonacademic effect size. A summary of those study characteristics is provided in the lists below:
Frequencies and percentages of effect sizes by study characteristics.
Studies with academic outcomes
Studies with academic outcomes tended to use the pretest‐posttest with control group design (Shadish 2002 ). (Studies that used pretest‐posttest without control group designs were excluded).
The majority of effect sizes of academic outcomes came from studies that used random assignment as evidence of baseline equivalency. Other common sources of evidence of baseline equivalency were nonstatistically significant differences on a pretest or using the pretest as a statistical covariate.
Most studies with academic outcomes collected their data within one year of completion of the intervention.
In terms of setting, most studies with academic outcomes were conducted in public settings for both the traditional and Montessori conditions.
Standardized measures of academic achievement were the most frequently used type of measure in studies with academic outcomes.
Studies with academic outcomes tended to be conducted in elementary or pre‐K settings.
The vast majority of studies with academic outcomes were conducted in North America.
Most studies with academic outcomes were published in peer‐reviewed publications.
Studies with nonacademic outcomes
Similar to studies with academic outcomes, nonacademic studies most frequently used the pretest‐posttest with control group design.
In contrast to studies with academic outcomes, studies with nonacademic outcomes tended to use nonstatistically significant pretest measures and/or gain scores as evidence of baseline equivalency. Studies with nonacademic outcomes tended not to use random assignment.
Like studies with academic outcomes, studies with nonacademic outcomes tended to collect data within one year of completion of the intervention.
Studies with nonacademic outcomes were conducted in an approximately equal proportion of public and private settings.
As expected, studies with nonacademic outcomes did not use standardized tests of achievement.
Similar to studies with academic outcomes, the most frequently used settings were in elementary and pre‐K.
Most studies with nonacademic outcomes had first authors from North America or Europe.
Most studies with nonacademic outcomes were published in peer‐reviewed forums.
The list below provides a link to each of the 32 included studies. The supplemental information in Randolph 2021 contains a data set where we extracted the information specified in the coding book; specific details on each study can be found there.
Alburaidi 2019
Ansari 2014
Aydoğan 2016
Besançon 2013
Coyle 1968
Culclasure 2018
Denervaud 2019
Denervaud 2020
Doğru 2015
Elben 2015
Faryadi 2017
Fleege 1967
Galindo 2014
Hoseinpoor 2014
Jones 1979
Juanga 2015
Kayili 2016b
Kayili 2016a
Kirkham 2017
Lillard 2006
Lillard 2012
Lillard 2017
Mallett 2015
Manner 1999
Miller 1983
Miller 1984
Prendergast 1969
Rathunde 2005a
Rathunde 2005b
Tobin 2015
Yussen 1980
5.1.3. Excluded studies
Of the 173 studies that underwent full‐text screening, 141 studies were excluded. See the Excluded studies section for references to the 141 excluded studies. The following list summarizes how many studies were excluded based on each exclusion criterion:
a lack of proof of equivalency of Montessori and traditional groups at baseline ( n = 58),
did not use experimental or quasi‐experimental research design ( n = 38),
insufficient information to calculate an effect size ( n = 44),
a lack of a Montessori‐based intervention ( n = 8),
the absence of a traditional, control group ( n = 11),
a lack of an academic or behavioral outcome ( n = 6 ) ,
was a duplicate study ( n = 3),
or was irretrievable ( n = 6).
Note that exclusion criteria were not mutually exclusive, so a study could have been excluded for one or more reasons. If a study met at least one exclusion criterion, the other exclusion criteria may not have been assessed. An online data set in the supplement (Randolph, 2021 ) to this review has a list of each article considered for inclusion, which inclusion criteria were met by each study, and notes on selected studies. Six studies were irretrievable as shown in the sheet labeled as irretrievable in the included/excluded studies data set in the supplemental online information.
5.2. Risk of bias in included studies
Overall, the risk of bias for the six randomized studies (Figure 2 ) was considered to be low. Similarly, the risk of bias for 26 nonrandomized studies was low (Figure 3 ). Although it is typical for nonrandomized studies to have overall risk‐of‐bias ratings of some concerns or high risk of bias, we believe that our nonrandomized studies typically were at low risk of bias because of the strict inclusion criteria we set. For example, nonrandomized studies were excluded if there was not strong evidence for baseline equivalency, which addresses the domains of confounding and selection of participants in the Robbins‐I tool (Sterne, 2016 ).
Risk of bias for studies with random assignment.
Risk of bias for studies with nonrandomized assignment.
5.2.1. Allocation (selection bias)
Selection bias was deemed to be low in the 32 included studies. For randomized studies, three of the six studies were rated as having low risk for the randomization process; the exceptions were Jones ( 1979 ), Miller ( 1983 ), and Miller ( 1984 ), which were rated as having “some concerns.” For the nonrandomized studies, all 26 studies were deemed to have low risk in terms of confounding and selection of participants. Documentation on these decisions can be found in the online supplemental information (Randolph, 2021 ).
5.2.2. Blinding (performance bias and detection bias)
Because of the nature of the intervention, it was not possible to blind participants to whether they were receiving Montessori or traditional education. Therefore, we did not assess the risk of bias in this domain.
5.2.3. Incomplete outcome data (attrition bias)
All 32 included studies were rated as low risk in terms of attrition bias.
5.2.4. Selective reporting (reporting bias)
All 32 included studies were rated as low risk in terms of selective reporting bias.
5.2.5. Other potential sources of bias
Information on other potential sources of bias not listed above can be found in Figures 2 and 3 . In short, we deemed there to be low risk of bias from other potential sources.
5.3. Effects of interventions
5.3.1. academic outcomes, main effects for academic outcomes.
Table 4 and Figure 4 (a histogram of raw effect sizes for academic outcomes) summarize the main effects of Montessori education versus traditional education for academic outcomes. For readers unfamiliar with the interpretation of meta‐analytic main‐effects tables, we describe the interpretation of each column in Table 4 here before discussing the specific results in the following paragraphs. The first column in Table 4 indicates the outcome. The second column indicates the standardized mean difference effect size, Hedges' g , which is a sample‐size corrected version of Cohen's d . Positive effect sizes favor Montessori education over traditional education, effect sizes of zero indicate the equality of Montessori and traditional education, and negative effect sizes favor traditional education over Montessori education. Since all of the effect sizes in Table 4 are positive, one can interpret the effect size as the number of standard deviations that Montessori students on average performed better than traditional students performed. See (Cohen, 1988 ; Kraft, 2020 ) for additional resources for interpreting the magnitude of an effect size. The third column indicates the 95% confidence intervals for the effect size, which can be interpreted as the plausible range of the population effect size, given chance (Higgins, 2021 ). The fourth column, p , is the two‐tailed probability of the effect size's being equal to 0.00 given chance; a Hedges' g effect size of 0.00 (the null value) would indicate that Montessori education and traditional education are equal. Values of p below 0.05 are typically regarded as statistically significant . The fourth column indicates the df (degrees of freedom) when a cluster‐robust model was used. Tanner‐Smith ( 2014 ) cautioned that cluster–robust results should not be interpreted as being reliable when the df is less than 4.00. The fifth column, I 2 , is a measure of study heterogeneity–the difference in effect sizes among studies–expressed as a percentage. Higgins ( 2021 , p. 259) gives the following rules of thumb for interpreting I 2 :
0%–40%: might not be important;
30%–60%: may represent moderate heterogeneity;
50%–90%: may represent substantial heterogeneity;
75%–100%: considerable heterogeneity.
Main effects of Montessori education versus traditional education academic outcomes.
Note : Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Abbreviation: CI, confidence interval.
A cluster‐robust model was used for all outcomes except social studies, in which a random‐effects model was used. The social studies outcome had effect sizes from only one study.
Kernel density plot of unweighted values of Hedges’ g for individual academic outcomes.
Finally, the fifth and sixth columns of Table 4 indicate the number of studies and the number of effect sizes, respectively, that contributed to each outcome.
As Table 4 shows, with all academic outcomes combined across 24 studies and 113 effect sizes, Montessori students on average performed 0.24 standard deviations higher than students in traditional education programs, 95% CI [0.13, 0.36], with high study heterogeneity, I 2 = 76.54%.
See Figure 5 for a graphical display of study heterogeneity (i.e., a GOSH plot) and effect sizes of academic outcomes from a bootstrap analysis of 1,000,000 samples of academic effect sizes. That GOSH plot shows the results of plotting the effect size and heterogeneity of random subsets of studies. The measure of heterogeneity ( I 2 ) is plotted on the vertical axis and the measure of effect size is plotted on the horizontal axis. So, each data point on the GOSH plot represents the heterogeneity and summary effect size of a random sample of studies. The darker the area, the more likely it is that that is the range in which the population effect size resides. (The population effect size is the unknowable effect of Montessori education that generalizes to all Montessori programs, not just the ones investigated in the studies included here). In Figure 5 , the darkest area varies horizontally between approximately 0.10 and 0.32 in effect size, so, this is the plausible range of the effect of Montessori academic outcomes. This dark area is positioned vertically at around an I 2 value of 90%, indicating high study heterogeneity. On the left‐hand side of Figure 5 , there is a small, light tail extending down to 0% heterogeneity and with an effect size of 0.07, meaning that there is a small statistical probability that the population summary effect size is actually 0.07 with very low heterogeneity. See Olkin ( 2012 ) for more information on the interpretation of the GOSH plot.
Graphical display of study heterogeneity for all academic outcomes combined. In the GOSH plot above, the X ‐axis (Overall Estimate) shows the Hedges’ g effect size. Positive effect sizes favor Montessori education over traditional education. The Y‐axis represents a measure of study heterogeneity, I 2 , where higher values represent more study heterogeneity. The black data points are simulated values of the population effect size and heterogeneity.
We believe there is moderate quality of evidence for this finding as discussed in the Summary of Findings table. (Note that all of our ratings of quality of evidence are based on the GRADE system for rating quality of evidence; see Higgins, 2021 ).
In terms of the individual academic outcomes presented in Table 4 , all academic outcomes were in favor of Montessori education. In summary, there was moderate quality of evidence that Montessori education outperformed traditional education in terms of general academic ability (Figure 6 ) (Hedges' g = 0.26, 95% CI [0.06, 0.46]), high quality of evidence for language/literacy (Figure 7 ) (Hedges' g = 0.17, 95% CI [0.03, 0.31]), high quality of evidence for mathematics (Figure 8 ) (Hedges' g = 0.22, 95% CI [0.06, 0.39]), low quality of evidence for science (Figure 9 ) (Hedges' g = 0.15, 95% CI [−0.61, 0.46]), and moderate quality of evidence for social studies (Figure 10 ) (Hedges' g = 0.05, 95% CI [−0.01, 0.12]). Note that the three social studies outcomes only come from Culclasure ( 2018 ). See the Summary of findings Table 1 for more information on the certainty of evidence. Note that judgments of the quality of evidence throughout this review are based on the GRADE system; see the Quality of the evidence section for details.
Forest plot for general academic ability.
Forest plot for language/literacy. Positive effect sizes favor Montessori education over traditional education.
Forest plot for mathematics. Positive effect sizes favor Montessori education over traditional education.
Forest plot for science. Positive outcomes favor Montessori education over traditional education.
Forest plot for social studies. Positive effect sizes favor Montessori education over traditional education.
Moderator analysis for academic outcomes
Random assignment.
A moderator analysis investigates the degree to which the relation between two outcome variables is influenced by a different study variable. A sensitivity analysis investigates the degree to which methodological or analytical decisions are differentially associated with effect sizes (Higgins, 2021 ). In the following section, we report the results of a moderator analysis for all academic outcomes combined for the following attributes: random vs. nonrandom assignment, grade level, Montessori setting (public vs. private), treatment duration, and length of follow up. We were unable to reliably conduct a moderator and sensitivity analysis for each individual outcome because there was an insufficient number of effect sizes to do so.
Moderator analyses are typically conducted via a technique called meta‐regression (Higgins, 2021 ). An example of a meta‐regression results table comparing the effect sizes of studies with nonrandom assignment to random assignment is presented in Table 5 . For readers who lack familiarity with results from cluster‐robust meta‐regression, we provide some guidance in the paragraphs below.
Cluster robust meta‐regression for random assignment for academic effect sizes.
Note : N of studies = 24, N of effect sizes = 113, assumed ρ = 0.8, I 2 = 76.90. Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Reference category/intercept.
The first column of Table 5 shows the various levels or categories of the attribute that are being investigated and the number of effect sizes included for that category. Here we are investigating the attribute of assignment (i.e., random vs. nonrandom assignment). There were 72 effect sizes from studies with nonrandom assignment and 41 effect sizes from studies with random assignment. The row category denoted with a superscript a is the reference category.
The regression coefficient in the second column of Table 5 is the standardized mean difference effect size (i.e., Hedges' g ). In this case, the effect size for the nonrandom assignment category was 0.19, which means that, when considering only effect sizes from studies with nonrandom assignment, Montessori students on average performed 0.19 standard deviations higher than traditional education students. The effect size coefficient (Hedges' g ) for any row category besides the reference category can be interpreted as the mean effect size difference between that row category and the reference category. In this case, the effect size coefficient for random assignment is 0.28, meaning that the average effect size was 0.28 units higher in randomized studies than in nonrandomized studies. One can find the effect size for any row category besides the reference category by adding the coefficient from the reference category to the coefficient from the row category. In this case, the estimated effect size for randomized studies was 0.47, since 0.19 + 0.28 = 0.47.
The third column in Table 5 shows the 95% confidence interval for the effect size coefficient. Put simply, the 95% confidence intervals show the plausible values of the mean effect size in the entire population. See Higgins ( 2021 ) for a more detailed explanation of the interpretation of confidence intervals around a meta‐analytic point estimate.
The fourth column in Table 5 ( df ) shows the degrees of freedom for each coefficient. According to Tipton 2015 , coefficients with a df less than 4 may not be reliable and should be interpreted with caution. Finally, the last column, p , gives the probability that the population parameter for the effect size coefficient in the second column might be zero, given chance. In short, Hedges' g for the reference category, which will always be the first‐row category in the meta‐regression tables presented here, is the effect size for that reference category. The value in the Hedges' g for any category besides the reference category is the mean effect size difference between that category and the reference category.
Finally, in terms of the substantive interpretation of Table 5 , there are three points of interest. Both randomized and nonrandomized studies had positive academic effect sizes in favor of Montessori education versus traditional education. Randomly assigned studies had significantly larger academic effect sizes than nonrandomized studies; however, the 95% CI [−0.07, 0.63] indicated some statistical uncertainty regarding the degree to which one could accurately generalize these results to the entire population of students.
Grade level
Table 6 shows the meta‐regression results for academic outcomes when disaggregated by grade level. Montessori education outperformed traditional education in all grade levels. The greatest academic effects of Montessori education were found at the elementary level (Hedges' g = 0.36). The mean preschool, middle school, and high school academic effect sizes were 0.16, 0.09, and 0.15 smaller, respectively, than the effect sizes for elementary school. However, all of these mean effect size differences had a high degree of imprecision as evidenced by their wide 95% CIs. A Wald test–an omnibus test for detecting statistically significant differences among means—yielded a p value of 0.761, which indicates that the differences in the effect sizes among grade levels would have been likely given chance. The middle school statistical results had a low df indicating that this moderator effect should be interpreted with caution. In summary, Montessori education had greater academic effects than traditional education at all grade levels, with the strongest effect being at the elementary level; however, there is much statistical uncertainty regarding this result.
Cluster robust meta‐regression for grade level for academic effect sizes.
Note : N of studies = 24, N of effect sizes = 113, assumed ρ = 0.8, I 2 = 78.83. Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Wald test for null difference between preschool, middle and high school grade levels, F (3, 2.82) = 0.41, p = 0.761. Positive effect sizes favor Montessori over traditional education.
Public Montessori versus private Montessori
Both private and public Montessori education had better academic outcomes than traditional education, as Table 7 shows. Relative to traditional education, public Montessori education had academic outcomes 0.13 standard deviations less than private Montessori programs (Hedges' g = 0.28, 95% CI [0.03, 0.52]. However, in terms of the academic effectiveness of public Montessori over traditional education and the relative effectiveness of private Montessori over public Montessori, the wide 95% CI indicates that there is much statistical uncertainty regarding the degree to which these results would generalize to the entire population of students.
Cluster robust meta‐regression for public versus private Montessori setting for academic effect sizes.
Note : N of studies = 20, N of effect sizes = 107, assumed ρ = 0.8, I 2 = 71.89. Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Treatment duration and follow‐up measurements
Table 8 is a meta‐regression table showing Montessori education's relative academic effectiveness over traditional education as a function of treatment duration (i.e., the duration of the Montessori intervention in number of weeks). Within‐study effects refer to effects that estimate growth over time on the same set of students within a study. The interpretation of the within‐subjects effect (Hedges' g = 0.007, 95% CI [−0.087, 0.100]) is that for every 1 week in Montessori education, the effect size increased by 0.007 standard deviations. Figure 11 is a scatterplot of treatment duration and between‐study effect size. Between‐study effects estimate growth over time by comparing treatment duration effects between different studies. The interpretation of the between‐subjects effect (Hedges' g = 0.002, 95% CI [−0.005, 0.009]) is that for every 1 week in Montessori education, the effect size increased by 0.002 standard deviations.
Cluster robust meta‐regression for treatment duration (weeks) for academic effect sizes.
Note : N of studies = 18, N of effect sizes = 67, assumed ρ = 0.8, I 2 = 80.72. Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Scatterplot of treatment duration and observed effect size for academic outcomes. Positive effect sizes favor Montessori education over traditional education.
While both within‐study and between‐study estimates indicate a slightly positive linear relationship between effect size and duration in Montessori education, the wide 95% CIs, the df s less than four, the nonlinearity, and nonconstant variance of the relationship between effect size and duration lead us to conclude that there is too much uncertainty and ambiguity to interpret these treatment duration results with any meaningful degree of certainty. In short, the results for this moderator were inconclusive.
Four studies included in this review used follow‐up measurements of academic outcomes after the Montessori intervention period had ended. We present the results of that analysis in Table 9 . The results ostensibly show a slight fading effect on academic outcomes the longer a student had been out of Montessori education. Specifically, academic effect sizes decreased by 0.07 standard deviations per year after a student finished Montessori education (Hedges' g = −0.07, 95% CI [−0.16, 0.01]). Because of the low df (1.35) and the nonlinearity and nonconstant variance in the scatterplot, we suggest that readers interpret these results as largely inconclusive.
Cluster robust meta‐regression for follow‐up years for academic effect sizes (between).
Note : N of studies = 4, N of effect sizes = 37, assumed ρ = 0.8, I 2 = 31.85. Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Sensitivity analysis for academic outcomes
A sensitivity analysis examines the degree to which methodological or analytical decisions covary with outcomes (Higgins, 2021 ). In this review, we conducted several types of sensitivity analyses, the results of which would be too voluminous to detail here in entirety. Therefore, we summarize some of the results of the sensitivity analysis narratively in this paragraph and concentrate on one particularly important sensitivity analysis in the remainder of this section.
In terms of the ρ parameter (i.e., the estimated correlation between dependent effect sizes) used in cluster‐robust effect size estimation, the results were consistent regardless of the value of the parameter for ρ (i.e., the estimated correlation between dependent effect sizes) we chose. We conducted leave‐one‐out analyses for applicable outcomes and also found that the leave‐one‐out results were consistent with what was reported here.
For main effects analyses, we compared random‐effects and cluster‐robust synthesis methods. The overall results were consistent in terms of point estimates of effect size; however, as expected, the variance estimates differed slightly between random‐effects and cluster‐robust models. The random‐effects models tended to have lower variance than the cluster‐robust models, but we used cluster‐robust models nonetheless because we believed that the benefits of taking into account the dependencies between effect sizes outweighed slightly more accurate variance estimates, as discussed in Tanner‐Smith ( 2014 ).
Fixed‐effect models had smaller 95% CIs than cluster‐robust or random‐effects models, as expected, and were less favorable to Montessori education. We attribute this to the fact that fixed‐effect models tend to weight studies with large samples heavily compared to other models and fixed‐effect models do not take into account studies with multiple effect sizes (Higgins, 2021 ). We believe that Culclasure ( 2018 ) was overweighted in fixed‐effect models because it tended to contribute many effect sizes and had sample sizes in the thousands. As we discuss in the Discussion section, Culclasure ( 2018 ), as well as Ansari ( 2014 ), lacked consistently high treatment fidelity in the implementation of Montessori education and, thus, these studies might likely suppress the effect size in favor of traditional education.
One sensitivity analysis of particular importance was related to our method of effect size estimation. We used either a method for standardized mean differences, a standardized mean raw score method, or an other method where we calculated an effect size using Wilson (n.d.)'s Practical Meta‐Analysis Effect Size Calculator. The results of that sensitivity analysis are shown in Table 10 . For academic outcomes, the mean difference in effect sizes between the reference category and other categories was less than 0.04 standard deviations in absolute value and the results of the omnibus Wald test yielded a p value of 0.955, F (2, 2.26) = 0.05. In this analysis the dfs were low and, therefore, these estimates should be interpreted with caution. Nonetheless, with all of the evidence taken together, we believed that it was reasonable to conclude that the method of effect size estimation was not an important moderator of effect size results and, therefore, it is justifiable to aggregate effect sizes across estimation methods as we did in the majority of analyses in this review.
Cluster robust meta‐regression for effect size calculation method for academic effect sizes.
Note : N of studies = 24, N of effect sizes = 113, assumed ρ = 0.8, I 2 = 72.71. Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Finally, for those interested in null hypothesis testing, we calculated the Benjamini–Hochberg (Benjamini, 1995 ; Benjamini, 2005 ) adjusted α to help readers account for the multiplicity of statistics tests (see Wasserstein, 2019 ), at least for main effects. A list of those adjustments can be found in the online supplement (Randolph, 2021 ).
Possible publication bias for academic outcomes
Publication bias is a type of systematic bias that occurs when studies with null or negative effect sizes are withheld from the research record, either because of authors not submitting null or negative results for publication or journal reviewers or editors tending to not recommend studies with null or negative results for publication. To examine publication bias, we first conducted a visual analysis of a variety of funnel plots for all academic outcomes combined (see the four panels in Figure 12 ). In these funnel plots, the vertical axis represents some measure of sample‐size‐related variability (inverse sample size, raw sample size, square root of the sample size, log sample size) and the horizontal axis represents the effect size with positive effect sizes favoring Montessori education. Each data point within each funnel represents the correspondence of the standard error and effect size for each effect size. In general, larger studies will have less variance and, therefore, will be located near the top of the funnel whereas small studies will be located near the bottom.
Four sample‐size based funnel plot variations for a composite of all academic outcomes combined.
Deviations from a funnel shape may be potential indicators of publication bias. See Higgins ( 2021 ) for more information on the visual analysis of funnel plots. One particular deviation to note is when a funnel plot is asymmetrically in favor of the intervention. Asymmetrical funnel plots with a large number of data points outside of the funnel only on the side favoring the intervention may be an indicator of publication bias in favor of the intervention. As a result of too many data points falling outside the funnel, there may be blank areas on the other side of the funnel. This deviation is seen to some degree in many of the main‐effect funnel plots presented in Figure 12 . There is a preponderance of small studies in favor of Montessori education and an absence of small to medium size studies in favor of traditional education. The result is that the center reference line is also slightly to the right of the couple of studies with many effect sizes with very large sample sizes (e.g., Culclasure, 2018 ). As we explain in the Discussion section, there are some characteristics of the large sample studies that might attenuate their effect sizes in favor of traditional education. In addition, some of the variability of the handful of extreme outliers on the right can be explained through an analysis of study characteristics. The supplemental information (Randolph, 2021 ) contains funnel plots for individual academic outcomes.
In addition to a visual examination of funnel plots, we also conducted a trim and fill analysis. Using Duval's ( 2000 ) trimandfill algorithm enabled us to determine the number of effect sizes that may have been missing due to publication bias, impute the missing effect sizes to minimize funnel plot asymmetry, and then estimate a new effect size with the imputed effect sizes included.
We provide an example of that procedure here starting in terms of the language/literacy outcome. Figure 13 is a funnel plot in which the black data points represent the 46 observed effect sizes for language/literacy. The 11 data points in white on the lefthand side of the funnel plot represent the effect sizes imputed by the trim and fill algorithm to bring symmetry to the funnel plot as a whole. The value of Hedges' g for language/literacy was 0.24 (95% CI: 0.06, 0.36) for the 46 black data points. After imputing the 11 white data points and pooling them with the original effect sizes, the value of Hedges' g decreased to 0.13 (95% CI: 0.03, 0.23) for the new set of 57 effect sizes (i.e., the 46 original effect sizes and 11 imputed effect sizes). Duval ( 2000 ) notes that the trim and fill effect size should not be interpreted as a publication‐bias‐adjusted effect size. Instead, they recommend that one should use the difference between the imputed and unimputed effect sizes to gauge the degree of potential bias for a given outcome. For example, the 0.11 standard deviation difference between the imputed and unimputed effect size estimates is a rough estimate of the degree to which publication bias may be present in the Montessori research regarding language and literacy.
Funnel plot with trim and fill for language/literacy. The black data points represent observed effect sizes. The white data points represent effect sizes that were imputed using Duval's ( 2000 ) trim and fill algorithm to bring symmetry to the funnel plot. Without imputation, Hedges’ g was 0.24 compared to 0.13 when effect sizes were imputed.
Table 11 shows the trim and fill results for the other academic outcomes. Since language/literacy contributed 45 effect sizes when all academic outcomes were combined, it is no surprise that there was statistical evidence of asymmetry for that composite outcome. After imputing 20 left‐side effect sizes, the values of Hedges' g for all academic outcomes combined decreased by about 1/10th of a standard deviation‐‐from 0.24 (95% CI: 0.17, 0.29) to 0.14 (95% CI: 0.07, 0.23).
Academic effect sizes with and without left‐side trim and fill imputation.
Note : Outcomes with an NA indicate that there was enough symmetry that the trim and fill algorithm did not estimate any effect sizes to be missing and, therefore, no effect sizes were imputed. Random‐effects effect size estimates were for imputed and nonimputed estimates to facilitate comparison; therefore, these effect size estimates may differ from the cluster‐robust estimates presented elsewhere. Social studies and science were not included in this analysis because they had less than 10 effect sizes. L0 means left–side imputation.
In summary, we found strong evidence for a publication bias effect on the language literacy outcome. We estimate that the publication bias effect for language/literacy is on the order of about 1/10th of a standard deviation in a way that is systematically biased in favor of Montessori education. However, we believe that the degree of systematic bias in favor of Montessori education is not sufficient to negate the overall finding that Montessori education increases language/literacy outcomes when compared to traditional education.
5.3.2. Nonacademic outcomes
Main effects for nonacademic outcomes.
Table 12 and Figure 14 (a histogram of effect sizes for nonacademic outcomes) show the main effects for all nonacademic outcomes combined and individually for the nonacademic outcomes of creativity, executive function, inner experience of school, and social skills. On average, Montessori students performed 0.33 standard deviations higher than traditional students on nonacademic outcomes, 95% CI [0.16, 0.50], with moderate quality of evidence.
Main effects of Montessori education versus traditional education for all nonacademic.
Note : Cluster robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Kernel density plot of unweighted values of Hedges’ g for individual nonacademic outcomes.
See Figure 15 for a graphical display of study heterogeneity and effect sizes of academic outcomes from a bootstrap analysis of 1,000,000 samples of our nonacademic effect sizes. This plot shows that the plausible values of the summary effect size for nonacademic outcomes range approximately from 0.10 to 0.35 with high study heterogeneity.
Graphical display of study heterogeneity for all nonacademic outcomes combined. In the GOSH plot above, the X ‐axis (Overall Estimate) shows the Hedges’ g effect size. Positive effect sizes favor Montessori education over traditional education. The Y ‐axis represents a measure of study heterogeneity, I 2 , where higher values represent more study heterogeneity. The black data points are simulated values of the population effect size and heterogeneity.
In terms of the individual four nonacademic outcomes shown in Table 12 , there was moderate quality evidence that Montessori education outperformed traditional education. The largest effect sizes were found for executive function (Figure 16 ) (Hedges' g = 0.36, 95% CI [0.15, 0.58]) and inner experience of school (Figure 17 ) (Hedges' g = 0.41, 95% CI [0.19, 0.62]). However, the cluster‐robust df for inner experience of school ( df = 3.79) was slightly under 4.00, indicating that the result should be interpreted with caution (Tanner‐Smith, 2016 ). While there were also small to moderate effects in favor of Montessori education for creativity (Figure 18 ) (Hedges' g = 0.26, 95% CI [−0.21, 0.74]), and social skills (Figure 19 ) (Hedges' g = 0.23, 95% CI [−0.02, 0.49]), there was significant variation in the 95% confidence intervals. There was also evidence of high study heterogeneity for nonacademic outcomes as demonstrated by the high values of I 2 in Table 10 , which ranged from 82.92% for social skills to 66.61% for inner experience of school. Further, while creativity and executive function had moderate quality of evidence, inner experience of school and social skills both had low quality of evidence. See the Summary of findings Table 1 for more information on the certainty of evidence. Overall, there is moderate quality of evidence that Montessori education leads to better performance on nonacademic outcomes than traditional education.
Forest plot for executive function. Positive effect sizes favor Montessori education over traditional education.
Forest plot for inner experience of school. Positive effect sizes favor Montessori education over traditional education.
Forest plot for creativity. Positive effect sizes favor Montessori education over traditional education.
Forest plot for social skills and behavior. Positive effect sizes favor Montessori education over traditional education.
Moderator analysis for nonacademic outcomes
Table 13 shows the meta‐regression results for random versus nonrandom assignment for all nonacademic outcomes combined. Similar to the results for nonacademic outcomes, effect sizes from studies with random assignment had an effect size 0.31 standard deviations greater than effect sizes without random assignment; however, there was significant variance in the plausible range of values for this estimate, 95% CI [−0.54, 1.16] and the df < 4.0.
Cluster‐robust meta‐regression for random assignment for nonacademic effect sizes.
Note : N of studies = 18, N of effect sizes = 91, assumed ρ = 0.8, I 2 = 84.12a. Cluster‐robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Similar to the effects for academic outcomes, Montessori education had greater impacts than traditional education at all measured grade levels on nonacademic outcomes as shown in Table 14 . (There were no included studies with nonacademic effect sizes for high school students). The greatest impacts of Montessori education for nonacademic outcomes were seen at the preschool level (Hedges' g = 0.36, i.e., 0.26 + 0.10 = 0.36). However, a Wald test for an omnibus effect of a grade‐level moderator was not statistically significant ( p = 0.863). In summary, Montessori education had the greatest impacts over traditional education on nonacademic outcomes at the preschool level; however, there was much statistical uncertainty about grade level as a moderator of the effectiveness of Montessori education.
Cluster‐robust meta‐regression for grade level for nonacademic effect sizes.
Note : N of studies = 18, N of effect sizes = 91, assumed ρ = 0.8, I 2 = 85.60. Cluster‐robust estimates with df < 4.00 should be interpreted with caution(Tanner‐Smith, Tipton, & Polanin, 2016 ). Wald test for null difference between preschool and middle school grade levels, F (2, 3.39) = 0.10, p = 0.911. Positive effect sizes favor Montessori over traditional education.
As Table 15 shows, both private and public Montessori education showed greater impacts on nonacademic outcomes than traditional education. This result was especially strong for private Montessori education (Hedges' g = 0.43, 95% CI [0.18, 0.67]). Private Montessori outperformed public Montessori on nonacademic outcomes. There is statistical uncertainty, however, about the degree to which these results would accurately generalize to the population of students.
Cluster‐robust meta‐regression for public versus private Montessori setting for nonacademic effect sizes.
Note : N of studies = 13, N of effect sizes = 70, assumed ρ = 0.8, I 2 = 78.75. Cluster‐robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Table 16 and Figure 20 show the results of an examination of nonacademic effect sizes as a function of treatment duration. The within‐study and between‐study effects were contradictory; the within‐study effect showed a very slight treatment‐duration effect in favor of Montessori education (Hedges' g = 0.001, 95% CI [−0.089, 0.091)] and the between‐study effect showed the same treatment‐duration effect but in favor of traditional education, (Hedges' g = −0.004, 95% CI [−0.011, 0.003] for nonacademic outcomes. For the same reasons explained in the section on treatment duration for academic outcomes, we regard these results as inconclusive. There were too few studies to do a moderator analysis of follow‐up studies for nonacademic outcomes.
Cluster‐robust meta‐regression for treatment duration (weeks) for nonacademic effect sizes.
Note : N of studies = 15, N of effect sizes = 67, assumed ρ = 0.8, I 2 = 86.65. Cluster‐robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Positive effect sizes favor Montessori over traditional education.
Scatterplot of treatment duration (weeks) and observed effect size for nonacademic outcomes. Positive effect sizes favor Montessori education over traditional education.
Sensitivity analysis for nonacademic outcomes
The results of the sensitivity analysis for nonacademic outcomes yielded nearly identical results as for academic outcomes: the difference in results between analytical methods was unremarkable. In summary, the results were consistent regardless of the value of the parameter for ρ that was chosen. The leave‐one‐out results differed only marginally from the effect sizes reported when all studies were included. For main effects analyses, when comparing random‐effects and cluster‐robust synthesis methods, the overall results were consistent in terms of point estimates of effect size; however, as expected, the variance estimates differed slightly between random‐effects and cluster‐robust models. Although not reported here, the fixed‐effect models had smaller 95% CIs and were less favorable of Montessori education, likely due to implementation in the large sample studies that will be discussed later.
In terms of the differences between effect sizes rendered using different estimation methods for nonacademic outcomes (see Table 17 ), there was a large effect size difference between the standardized mean raw score change method and the other method (Hedges' g = −0.45, 95% CI [−2.36, 1.45]) and between the standardized mean difference method and the other method (Hedges' g = −0.54, 95% CI [−2.11, 1.01]), as shown in Table 13 . However, the mean difference between the two standardized methods, which consisted of the majority of effect sizes, was only 0.06 standard deviations and the Wald test for omnibus differences yielded a p value of 0.382, F (2, 1.57) = 1.57. This suggests we are justified in synthesizing results regardless of the method we used to calculate a study's effect size.
Cluster‐robust meta‐regression for effect size calculation method for nonacademic effect sizes.
Note : N of studies = 18, N of effect sizes = 91, assumed ρ = 0.8, I 2 = 83.46. Cluster‐robust estimates with df < 4.00 should be interpreted with caution (Tanner‐Smith, Tipton, & Polanin, 2016 ). Wald test for null difference between standardized mean raw score change and standardized mean difference effect size approaches, F (2, 1.9) = 2.12, p = 0.337. Positive effect sizes favor Montessori over traditional education.
Possible publication bias for nonacademic outcomes
Figure 21 is a four‐panel funnel plot of all nonacademic outcomes combined. The same asymmetry of small studies in favor of Montessori education that was seen for academic outcomes is also seen here for nonacademic outcomes. Similarly, we believe that some of the variance can be explained through an analysis of study characteristics of the extreme outlying studies and of the studies with a very large sample size, as explained in the Discussion. We conclude that there is some slight asymmetry in small studies in favor of Montessori education but it is unclear whether publication bias is the cause. Asymmetry alone is not sufficient evidence for publication bias (Higgins, 2021 ). Overall, because of the small degree of symmetry and the analytical methods used, we conclude that the asymmetry does not change the substantive conclusion that Montessori education is favorable for most nonacademic outcomes. Conservatively, we believe that the range of plausible values of the aggregate effect of Montessori education on nonacademic outcomes is represented well in the GOSH plot shown in Figure 15 .
Four sample‐size based funnel plot variations for nonacademic outcomes.
Table 18 shows the results with a trim and fill analysis to help interpret the degree of publication bias. Similar to the trim and fill analysis conducted for academic outcomes, we estimate that publication bias favors Montessori by about 1/10th of a standard deviation.
Nonacademic effect sizes with and without left‐side trim and fill imputation.
Note : Outcomes with an NA indicate that there was enough symmetry that the trim and fill algorithm did not estimate any effect sizes to be missing and, therefore, no effect sizes were imputed. Random‐effects effect size estimates were for imputed and nonimputed estimates to facilitate comparison; there, these effect size estimates may differ from the cluster‐robust estimates presented elsewhere. L0 means left–side imputation.
6. DISCUSSION
Accumulative evidence is required for scientists and policy decision‐makers to draw conclusions about the efficacy of an intervention (Popper, 1962 /2014). When the evidence is mixed, as was concluded by two recent narrative reviews of Montessori outcomes (Ackerman, 2019 ; Marshall, 2017 ), meta‐analyses can resolve the ambiguity by providing a more precise effect size estimate based on a large set of observations drawn from multiple studies and samples (Rosenthal, 2002 ). The Campbell reviews reflect this view, and the current meta‐analysis was conducted to shed light on the efficacy of Montessori education for children's academic and nonacademic outcomes, testing the hypothesis that Montessori is at least as effective as traditional education. The meta‐analysis included only studies with evidence of baseline equivalence, through random assignment to Montessori, and/or through pretest, matching, or statistical adjustment on key variables. A search of the literature using the search term “Montessori” and snowballing from reference lists resulted in a list of over 2000 published and unpublished studies, but as with other school choice meta‐analyses (e.g., Cheng, 2017 ), the vast majority of the 173 studies deemed worthy of closer inspection were excluded for failing to meet the stringent inclusion criteria that are more suggestive of causal effects.
Thirty‐two studies that were conducted before February 2020 when data collection closed did meet the criteria. They examined a broad array of outcomes, from academic ones like mathematics and literacy to social‐emotional outcomes like creativity and social skills. Because single studies contributed several effects, most analyses used cluster–robust regression. The studies included public and private schools, and participants ranging from preschool through high school; subsequent analyses examined these and random/nonrandom assignment as moderators.
In contrast to the aforementioned recent narrative reviews that were neutral on the efficacy of the Montessori approach, this quantitative meta‐analysis showed that Montessori education had largely positive effects on outcomes, with effect sizes that ranged from small to large in the context of school‐based field research with standardized (non‐custom) measures (Kraft, 2020 ). Below we summarize these findings, separating academic from nonacademic outcomes. Note that we believe our findings might have been unduly influenced (in a direction that is unfavorable to Montessori) by a single study with a very large N that also contributed many effects; this study included schools with varying quality of Montessori implementation as assessed by that study's own fidelity of implementation measurement. This is discussed later.
6.1. Summary of main results
Here we summarize the academic and then the nonacademic findings of our analysis. The outcomes reported here refer generally to all students, not to particular subgroups such as lower‐income children. In addition, we note that several of the included studies have small N s, and studies with small N s typically have larger effect sizes to be published. However, our bias analyses revealed relatively little evidence of publication bias, rendering this less of a concern.
6.1.1. Academic outcomes
Both overall and by individual outcome, the meta‐analysis showed that Montessori's average effect on academic outcomes was uniformly positive for Montessori, with Hedges' g s ranging from 0.26 (general academic ability, including composites) to 0.05 (social studies). Where math was reported separately, the effect size was 0.22 (based on 36 effects from 12 studies), and for literacy, it was 0.17 (based on 45 effects from 16 studies). For research on school programs (which 28 of the 32 included studies concerned), these are noteworthy effect sizes. In a discussion of effect sizes for school studies involving standardized test scores (most of the academic tests included here), the math finding would be considered large and the literacy effect size medium‐large (Kraft, 2020 ). One recent meta‐analysis using school lotteries found the effect of charter school attendance to be much smaller than the effects observed here (Cheng, 2017 ); specifically, it found average effects on ELA (English Language Arts) of about one‐tenth of a standard deviation (0.09), and on mathematics of about two‐tenths of a standard deviation (0.19). Results of charter schools using the “no excuses” model specifically were comparable to those achieved by Montessori (0.25 and 0.17). We note that the charter school analyses typically were of older grades, and effect sizes typically decline across school years (Hill, 2008 ). The Montessori effect sizes for science and social studies were considerably smaller, and the lower end of their confidence intervals was negative, but these analyses included only three and one studies (respectively) and just five and three effect sizes, thus we have less confidence in those findings.
Looked at another way, Montessori education's overall impact on academic performance is about a quarter of a standard deviation, equivalent to about 2 months of school in Grade 1, and the whole school year in Grade 6 (Hill, 2008 ). These equivalences were derived by Hill et al. by looking at the average change in children's scores across a school year; children change much more early in schooling, when learning to read and do elementary math, than they do in older grades. The month equivalence is derived from dividing the average school‐year change scores by the number of months children are in school (typically 9).
The academic results are important because they resolve existing ambiguity regarding Montessori's effects. While many studies of Montessori report positive outcomes, the domains attaining significance are not entirely consistent across studies, and there is at least one relatively highly cited published study with one negative academic effect (Lopata, 2005 ), which did not meet the inclusion criteria for this review. In such circumstances, meta‐analytic results are very important (Rosenthal, 2002 ). Montessori is a constructivist educational approach (Elkind, 2003 ) and is sometimes equated with discovery learning approaches because children have considerable independence, being free to choose what to work on each day (Lillard, 2017 ). Discovery learning approaches typically do not have positive academic outcomes when compared with more traditional direct instruction (Klahr, 2004 ; Mayer, 2004 ). Studies of Montessori outcomes, like those included in this meta‐analysis, most often use traditional direct instruction as the comparison, although some use a more specific counterfactual (like HighScope in the Ansari, 2014 study) and others used conventional U.S. preschools from an earlier era in which preschool was mostly free play. The results of this meta‐analysis indicate that on average, across studies, academic outcomes in Montessori are better than those of traditional education, however it was defined in a given study. The implementation range was broad, as described in Types of interventions, reflecting the broad range of programs dubbed “Montessori” in the real world; the same is true for the traditional school implementations. Thus, this study likely reflects the real‐world difference between Montessori and traditional school. The effect sizes are similar to those rendered in a meta‐analysis of oversubscribed “no excuses” charter schools, in which the counterfactual is typically underperforming urban public schools.
6.1.2. Nonacademic outcomes
The average effect size was slightly stronger for nonacademic outcomes (0.33) than academic ones (0.24). Executive function yielded the largest effect, with a Hedges' g of 0.36 based on 34 effects from 11 studies. This is approaching the near transfer effect of targeted executive function training programs for children (0.44), and far exceeds the effect of far transfer training (0.11, Kassai, 2019 ), which Montessori arguably is: Montessori has no explicit training on common measures of executive function like opposites games. However, it may provide outsized experiences in exercising inhibitory control which is an aspect of executive function (see below). Because executive function is a significant predictor of both concurrent (Jacob, 2015 ) and future outcomes including health, wealth, and criminality (Moffitt, 2011 ), this Montessori result is highly significant. There are several potential routes by which Montessori might impact the development of executive function (Lillard, 2017 ). For example, it has many parallels to mindfulness training, such as emphasizing and cultivating concentrated attention, educating the senses to notice fine gradations in stimuli, taking great care in every movement of the body, and sometimes sitting or walking for a period of time in purposeful silence. Mindfulness training appears to influence executive function through an impact on sustained attention (Leyland, 2019 ; Zoogman, 2015 ), and was observed in a meta‐analysis to influence social‐emotional outcomes more generally (Maynard, 2017 ). Another, simpler route to inhibitory control specifically is that children have to wait in Montessori because there is only one of most resources. Because there is one copy of each material, if another child is using a material, other children who want to use that material must wait. In addition, properly implemented Montessori has only one teacher (Lillard, 2019a ; Lillard, 2019b ), who moves around the classroom helping or teaching children one by one; if a child needs the teacher's help, he or she often must wait for individual attention. Further research on individual children's behaviors in Montessori classrooms could shed light on what aspects of the program might influence executive function development.
Montessori also had a nontrivial impact on one's inner experience of school, which translates to well‐being at school (Hedges' g = 0.41); although this result stems from just 10 effects and five studies and has lower evidence quality, its size makes it likely that there is some impact. It makes theoretical sense that Montessori would lead to higher well‐being. For example, self‐determination is associated with higher well‐being (Ryan, 2000 ), and in Montessori environments, relative to conventional school environments, children are given considerable freedom as long as they use that freedom to constructive ends for their own and others' development (Montessori, 2017 ). People who attended Montessori as children have higher adult well‐being (Lillard, 2021 ) and recall liking school better during childhood (LeBoeuf, 2022 ).
Another nonacademic outcome that is related to Montessori was creativity, with a Hedges' g of 0.26; the 95% CI around this effect ranged from −0.21 to 0.74, so the result, to which six studies contributed 24 effect sizes, should be interpreted cautiously. It is conceivable, but not clear from this evidence, that the freedom to consider possibilities (Rinke, 2013 ), combined with a lack of extrinsic rewards and multiple‐choice tests (see Lillard, 2017 ), fosters creativity in Montessori. Another effect that is less clear is social skills, with a Hedges' g of 0.23 and a 95% CI from −0.02 to 0.49, from nine studies and 23 effect sizes. It is possible that the multi‐aged classrooms and sustained relationships with peers due to looping enable advances in social skills; alternately, advanced social skills could be a by‐product of increased executive function. Again, we have less confidence in the social skills and creativity effects.
Montessori education yielded strong and clear effects on math, literacy, general academic ability, and executive function,
Montessori education's effects on aspects of well being, such as the inner experience of school and school liking, also were strong and appeared reliable, and
Montessori education also appeared to affect social studies, science, creativity, and social skills, but these effects are less clear and need further study.
6.1.3. Potential moderators
Study design: random versus nonrandom assignment.
The major obstacle to drawing conclusions about outcomes from freely chosen school programs is selection bias. One can control for family income, ethnicity, and other factors, but the possibility that families who choose Montessori differ in some other way that accounts for the results remains. Ensuring equality on key variables at pretest is helpful since presumably the family factors that cause a child to be higher on some outcome variable would be present at the outset as well. Controlling for the level of that variable at pretest is also helpful, but if the family produces an exponential growth pattern on that variable, it is uncontrolled. Because family effects that might positively affect child outcomes are confounded with Montessori enrollment, one would expect effects from nonrandom experimental designs to be greater than for random assignment studies. Interestingly, here this was not the case. Rather, results were considerably stronger for studies using random assignment for academic effects, whereby random assignment raised the effect size of 0.19 found with studies that used pretests or control variables to ensure baseline equivalence, to almost 1/3 of a standard deviation (i.e., 0.31). For nonacademic effects, the nonrandom assignment effect size of 0.28 more than doubled, adding 0.31 standard deviations. Just six studies had random assignment; three used data from children randomly assigned to a Montessori intervention at age four through 10th grade, with results published in three included papers (earlier reporting on that same study was not included because effect sizes could not be calculated from the information provided therein), and found strong sleeper effects for children previously enrolled in a Montessori Head Start, especially for boys. Two studies involved lotteries at oversubscribed American public schools. The sixth study was of 60 children enrolled in preschools in a city in Iran; it randomly assigned half the children to a targeted Montessori intervention. It contributed just two effects and the report is not sufficiently clear to allow speculation regarding the reason for its strong effects.
Focusing on the other studies, we speculate that their outcome effects stem from study timing in one case (representing three papers), and implementation of Montessori in the other two. Regarding the former, it might be that even weak Montessori implementation has long‐term effects because philosophical elements like free choice and order can exist even when the structural aspects of implementation are weak; these philosophical elements might not always manifest in differences in the near term, but could manifest later. The Miller studies, which contributed many effect sizes to the random effect, did not have particularly strong Montessori implementation (the original Miller, 1975 study reported an implementation score of 6.5/10), but it did provide for free choice and other philosophical elements, and this early experience at four might have led to long‐term outcome differences relative to the counterfactual (which was a traditional free play school). Another study of long‐term academic outcomes (not included here because it lacked appropriate controls) also found better academic performance for Montessori students (with unclear program implementation) even years after they had left the program (Dohrmann, 2007 ). Other studies suggest that without regard to implementation, Montessori predicts better nonacademic outcomes (LeBoeuf, 2022 ; Lillard, 2021 ).
The two Lillard random lottery studies looked at the immediate effects of Montessori, but the three oversubscribed public schools in the two studies were particularly strong examples of Montessori implementation. Although oversubscription is not itself a clear indicator of school quality (Tuttle, 2012 ; Weiland, 2020 ), here the oversubscribed schools were all AMI (Association Montessori Internationale) recognized Montessori schools. AMI schools adhere to the structural elements (i.e., they hire AMI‐trained teachers, which is an intensive 9‐month, standardized training with a highly prepared and vetted trainer; they have the specific 3‐year age spans, scarce adults and high ratios; and the long work periods and full set of Montessori materials) and they also implement the philosophical elements well. By contrast, teachers trained by the other major training organization, the American Montessori Society (AMS), founded to “Americanize” Montessori education (Rambusch, 1992 ), are relatively more supportive of conventional American education practices like tests, due dates, worksheets, and whole‐class activities (Daoust, 2018 , April), which might dilute the immediate effects of the Montessori program as compared to traditional education. Two studies specifically examined the influence of implementation on outcomes and found that the more classic implementation espoused by AMI is associated with better outcomes than supplemented implementations (Lillard, 2012 ; Lillard, 2016 ).
In sum, we speculate that the randomized trials had better outcomes in this meta‐analysis because sleeper effects on outcomes occur in response to philosophical aspects of Montessori that exist even when structural implementation is weak, and because two of the random studies, while looking at immediate outcomes, had especially high fidelity Montessori implementation.
To shed further light on why random assignment had an unexpectedly stronger effect than nonrandom assignment in this meta‐analysis, more research is needed using random assignment while also coding for Montessori implementation. We would hypothesize that among studies using random assignment, those using a more classic Montessori implementation would have stronger effects.
Academic effects were strongest for children in Elementary school, at 0.36. As is typical, effects were smaller across middle and high school; effect sizes for academic achievement across a school year decline fairly steadily from K to 12 (Hill, 2008 ). However, for Montessori, effects at preschool (0.20) were also smaller than effects at Elementary school (0.36). Yet the preschool effects are still notable. For example, one random‐assignment study of Head Start, with children enrolled either at age three or age four, found that 13 of 22 effects on language, literacy, and math were significant, and of those that achieved significance , the average effect size was 0.18 SD (Barnett, 2011 ), similar to our preschool effect which was not limited to those that achieved significance. The greatest gains for most children are typically seen from Kindergarten to first grade (Hill, 2008 ) when children undergo the famous “5 to 7 shift” (Sameroff, 1996 ) with biological changes augmenting environmentally‐induced ones. In Elementary school, the average growth across each school year is 0.44 (Lipsey, 2012 ), thus Montessori education may add as much as 80% of the gains expected in an entire school year to children's achievement. Although effect sizes were calculated for middle and high school and were considerably smaller, very few studies contributed to effect sizes at those levels giving us less confidence in those effects. Nonacademic effects were strongest for young children, which may indicate that at those ages children's executive function, creativity, and so on are most malleable. There are few developmental studies of these abilities, thus we were unable to view these results comparatively.
Public versus private Montessori
As Table 7 showed, both public and private Montessori educational interventions had better results than traditional education on aggregated academic outcomes. Public Montessori schools were shown to have an effect size 0.13 standard deviations lower than private Montessori schools on aggregate academic outcomes. This difference is equivalent to 29% of the average growth expected in a school year for elementary students (Lipsey, 2012 ). Nonetheless, the 95% CIs showed that the range of plausible values for this estimate varied from public Montessori education's performing 0.36 standard deviations lower than private Montessori education, which is equivalent to 82% of the average yearly academic growth expected in school in elementary education, to public Montessori education's performing 0.12 standard deviations above private Montessori education, which is equivalent to 27% of the average academic growth expected in elementary education. Therefore, we conclude that private Montessori education is likely to achieve academic results as good as or somewhat better than public Montessori education.
As Table 15 showed, both public and private Montessori educational interventions had better results than traditional education on aggregated nonacademic outcomes. Private Montessori had an effect size of 0.43 for aggregated nonacademic outcomes; public Montessori had an effect size of 0.26 standard deviations less than private Montessori. The effect size differences between public Montessori and private Montessori were twice as great for nonacademic outcomes, which had a 0.26 effect size difference in favor of private Montessori, as academic outcomes, which had a 0.13 effect size difference in favor of private Montessori. Because there are few longitudinal studies of the nonacademic outcomes reported here, it is not possible to view these effects in terms of expected yearly growth.
In summary, we conclude that there is preliminary evidence that private Montessori education leads to moderately greater performance in both academic and nonacademic outcomes than public Montessori education. We can also conclude that there is preliminary evidence that this difference in improvement for private Montessori and public Montessori is likely to be greater for nonacademic outcomes than academic outcomes. We argue that the evidence is preliminary because of the statistical imprecision of our estimates.
The public–private difference might be due to implementation. Public schools are required to have children take state exams that are designed for traditional school programs; Montessori schools need to adjust their program to prepare for those tests, and this by necessity dilutes Montessori implementation at public schools. Public schools are also required to follow regulations about recess breaks, special classes in art and sports, add‐in curricula, and teacher‐child ratios and class sizes that compromise the integrity of the Montessori program. Private schools also sometimes make such compromises, but they are more free to determine their programs.
6.2. Overall completeness and applicability of evidence
Our searches included a variety of academic databases, sources known to publish gray literature, Montessori‐related journals, and manual searches of references in retrieved studies. This search led to over 1500 unique records and, of those, 173 were included for full‐text review. Of those, 32 studies met the criteria for inclusion. The exhaustiveness of our search procedure and the number of records found lead us to believe that the Montessori‐related research listed in the excluded or included studies sections here represents a nearly complete, or at least, representative set of the quantitative Montessori research literature published before the specified search period, which ended February 2020. The outcomes also represent a broad variety of academic and nonacademic outcomes.
Also, because of the thoroughness of the search method, we conclude that the results presented here are applicable over a wide variety of traditional education and Montessori settings. However, when deciding the degree to which these results might generalize to particular settings or to future studies, consider these notes:
The majority of studies were conducted in elementary or pre‐K settings.
North American studies were predominant, but there was some degree of international representation from Europe, the Middle East, and Asia.
We only included studies that met strict requirements for demonstrating baseline equivalency; we suspect that including studies that met less stringent baseline‐equivalency requirements would have led to somewhat different effect sizes.
In summary, we are confident that the results presented here are drawn from a complete or nearly complete set of studies published during the search period and that the results are applicable over a wide variety of traditional and Montessori settings, to the most common academic and nonacademic outcomes, and across multiple measures of the given construct.
6.3. Quality of the evidence
The results of the risk of bias assessments indicated that the risk of bias in the included studies was low overall. Nonexperimental studies tend to have moderate or high risk of bias (Higgins, 2021 ), and we attribute our nonexperimental studies' having low risk of bias to our stringent inclusion criteria.
When applying the GRADE criteria, the quality of evidence was downgraded because of lack of precision (e.g., as measured by a nonstatistically significant result), high heterogeneity (e.g., a high I 2 value), and/or asymmetry of funnel plots, which can be an indicator of publication bias. If there were publication bias, we anticipate it would have led to marginally smaller effect sizes since there was a set of studies with small/medium sample sizes and larger effect sizes than would have been expected given chance. We suspect the treatment fidelity may have been compromised in large sample size studies, which tend to have more weight in a meta‐analysis than studies with small sample sizes and which tended to contribute more effect sizes than smaller studies. Therefore, we upgraded the quality of evidence for outcomes that included the large‐sample studies (Culclasure, 2018 and/or Ansari, 2014 ) because we believed it was a plausible confounding factor that may have underestimated the treatment effect. We assumed that greater treatment fidelity should result in greater effect sizes in favor of Montessori education over traditional education.
In summary, we conclude that there is moderate to high quality evidence to support our findings for most academic and nonacademic outcomes. See the Summary of Findings table 1 for more‐detailed information on the quality of evidence.
6.4. Potential biases in the review process
One limitation that could have biased these results is that we were unable to determine the quality of Montessori implementation in five of the included studies, and in several others we based our estimates on article descriptions that varied in completeness. We have noted that this does reflect the state of Montessori education in the real world. If a study claimed to be a test of Montessori education versus traditional education, and had evidence of equivalent baseline by accounting for key variables or by a lottery design in which children whose parents applied to oversubscribed schools were admitted or not admitted at random, then its findings were included. Some of the studies were done at AMI (Association Montessori Internationale) recognized schools (Denervaud, 2019 ; Denervaud, 2020 ; Lillard, 2006 ; Lillard, 2012 ; Lillard, 2017 ; Mix, 2017 ; Rathunde, 2005a ; Rathunde, 2005b ; and Yussen, 1980 ) and another at a school accredited by the Swiss Montessori Association, suggesting high fidelity implementation, but for some studies, implementation quality was known to be of lower quality; for example, both the Ansari ( 2014 ) and the three Miller studies (Jones, 1979 ; Miller, 1983 ; Miller, 1984 ) took place at Montessori preschools that had only 4‐year‐olds, missing the key ingredient of a 3‐year cycle and age grouping.
Culclasure ( 2018 ) is important to discuss in this regard, because it weighed heavily in our analysis due to its many effects (26) and very large sample (thousands of children contributing to each effect). Culclasure studied outcome differences for children in 23 public South Carolina Montessori schools. Only schools that passed minimal criteria for being Montessori were included in the study, but even among those included, implementation varied widely. Culclasure had trained Montessori teachers observe in a random selection of 126 classrooms and rate implementation, and they also had teacher surveys. Although half the programs were considered high fidelity by the study's own metrics, the other half were of medium or low fidelity. The teacher survey indicated that 35% of teachers did not think they had all or even most of the materials they needed to teach Montessori. Also, 90% said they used circle time centered around a weekly theme, suggesting they were more teacher‐directed than is optimal, and 43% said they supplemented the Montessori program with other materials (which Lillard ( 2012 ) and Lillard ( 2016 ) suggest leads to less positive outcomes than does using only Montessori materials). Only 30% introduced the Great Lessons in the first half of the school year, and over half of the teachers surveyed said they felt the quality of the Montessori was declining, and that public school test requirements were a major reason. Thus, although the Montessori implementation was proper in many ways (e.g., very few teachers used extrinsic rewards), this exemplary study that contributed many heavily weighted effect sizes to our study had weaknesses in its Montessori implementation. If those weaknesses reduce effect sizes, then this study alone might have biased results in a negative direction for Montessori.
Because implementation varies widely in Montessori schools (Daoust, 2004 ; Daoust, 2018 ; Daoust, 2019 ), and some studies suggest that better outcomes are achieved when implementation is more closely aligned with Montessori principles and practices (Lillard, 2016 ; Lillard, 2019a ; Lillard, 2019b ), the lack of attention to implementation, or sufficient reporting of implementation fidelity, in some of the studies included in this review is a limitation. The heterogeneity of implementation could be a reason for the asymmetry observed in some of our funnel plots. We expect that effect sizes would be greater had Montessori interventions been completed with greater fidelity. A future analysis might examine whether effect sizes increase as the level of implementation increases.
The effect sizes for the social studies outcome only came from one study (Culclasure, 2018 ). Therefore, we urge readers to note this important limitation when considering the result.
We did not extract data on whether control and experimental groups were assigned to the same school or whether the unit of randomization was at the individual, classroom, or school level. We suggest that this information be extracted and considered in future reviews.
Another potential bias is that the comparison samples/school programs are heterogeneous. It was outside of the scope of this review to compare Montessori to any specific alternative program. Thus, although virtually all the control children were receiving traditional or business–as–usual education, traditional schools vary. Although some specified that this meant teacher‐led, whole‐class learning using mostly lectures and textbooks, others did not. Our comparison could be described as average Montessori versus average traditional programs. We are unable to say which specific Montessori implementation is better or worse–only that compared to a range of other choices, on average, it produces positive effects.
Other sources of potential bias concern the categorization of the measures into the outcome categories we chose. Others might categorize some outcomes differently. Categorization was done without consideration of any potential impact on results, but it is possible that different categorization choices would yield significantly different results. Our database and codes are available for interested readers to review our categorization choices and redo the analyses with alterations to examine their impact.
Another potential limitation of the meta‐analysis is that many studies of Montessori education, including the ones studied here, had relatively small samples (two exceptions are the studies by Ansari, 2014 and Culclasure, 2018 ). Effect sizes tend to be larger with small‐ N studies (Kraft, 2020 ), and this was evident in the funnel plots for many of the outcomes we presented here. Therefore, it is likely that if more studies had been done with large sample sizes, the effect sizes would have been smaller than those reported here. We believe that the GOSH plots (Figures 5 and 15 ) robustly display the plausible ranges of population effect size estimates. Furthermore, we have included the code and data set so that researchers can investigate methodological variations that are too cumbersome to report here (e.g., fixed vs. random effects).
There was some evidence (asymmetry in funnel plots) that may point to publication bias in favor of Montessori. The trim and fill estimates in Tables 11 and 18 provide a general estimate of what the population effect sizes might be if data were imputed to make the funnel plot of effect sizes be symmetrical. We encourage Montessori researchers to attempt to publish the results of methodologically sound studies regardless of whether the results are negative, null, or positive and, thereby, help create a more comprehensive research record.
One might be concerned that one of the authors of this meta‐analysis (Lillard) also authored three of the included studies. Lillard joined the team after the initial analyses were done, and was not involved in devising the selection criteria, nor in deciding which studies to include, nor in the analyses. Thus, the results were obtained without any opportunity for author bias. Her role was limited to interpretation and writing, summarizing the fidelity of implementation of the interventions, as well as reviewing the broader literature and contextualizing the results.
6.5. Agreements and disagreements with other studies or reviews
Very few reviews of the efficacy of Montessori education have been published in peer‐reviewed journals. The results of this meta‐analysis of 32 studies are consistent with an earlier meta‐analysis that only included two Montessori studies, both unpublished, and focused exclusively on achievement outcomes; it calculated a d of 0.27 (Borman, 2003 ), similar to our overall academic effect size (Hedges' g = 0.24). Considering qualitative reviews, which were neutral, the results of this analysis reflect more positively on Montessori. Ackerman ( 2019 ) concluded that “Montessori programs have the potential to enhance young children's learning and development” (p. 11) but that there was no consistent advantage. Her review had included studies with less rigorous designs than those included here, and lack of rigor in the existing experimental base was a main conclusion of the Marshall ( 2017 ) review as well. Marshall stated that random lottery experiments were essential and also that studies' transposing elements of Montessori into other systems is a useful way to figure out what in Montessori leads to benefits. 1
7. AUTHORS' CONCLUSIONS
7.1. implications for practice.
Like the Marshall ( 2017 ) review, many discussions of Montessori end with the question of what causes the benefits. Given that we have a logical positivist science, this is a natural response. But we suggest instead adopting a systems perspective, recognizing that Montessori is a complex system, comprised of a set of attitudes and beliefs, practices, and materials (Lillard, 2019 ). Montessori was very concerned with the training of the teacher, and Montessori teacher training (at least in its traditional form) is said to require a spiritual transformation, creating a disposition of flexibility, restraint, and love of humanity (Whitescarver, 2007 ). The teacher takes a Rousseauian attitude that given the right conditions (including inspiration and curiosity, which the teacher helps to inspire), children will be kind and happy, and will behave in ways that are constructive for self and society. The teacher believes in, indeed loves, every child and is sure of the Montessori system's capacity to achieve this. The practices include structural elements such as specific 3‐year age groupings, a high teacher‐child ratio (about 30:1, plus an assistant for the youngest children), individual and small group lessons with a full set of specially constructed Montessori materials, and 2.5–3 h uninterrupted work periods sufficient for deep concentration. The practices also include philosophical elements like keeping the classroom beautiful and orderly, allowing children to choose freely as long as they are constructive, older children going out to explore the world and giving formal reports about their research activities to the class, and nothing beyond those specific Montessori practices (no grades, tests, or worksheets, no special outside teachers, and so on). Furthermore, in the practice of the AMI, the organization Montessori started to carry on her work, the teacher has gone through intensive training and examination, with a teacher‐trainer who themselves had intensive training in an apprenticeship to become a teacher‐trainer. The materials are sets of hundreds of mostly wooden and glass instruments and paper charts designed by Montessori and her collaborators for each age level to convey specific learning; children engage with these materials with their hands or even their full bodies (for further summary, see Lillard, 2019a ; Lillard, 2019b ). Children are also expected to have had Montessori at the prior level as well, at least from age three on, because Montessori confers specific learning that is bulit on later levels.
Which of these many elements is responsible for better academic and nonacademic outcomes? It may be the wrong question, although there are Montessori programs that eliminate one or more features. If enough programs were found that eliminated a specific one might compare those programs to programs that retain all the standard Montessori features.
For example, the Miller (Jones, 1979 ; Miller, 1983 ; Miller, 1984 ) and Ansari ( 2014 ) studies were not fully authentic implementations of Montessori for at least one reason: the classroom had only 4‐year‐olds. Some other studies also had limited age groups. The Mallett ( 2015 ) and Culclasure ( 2019 ) studies were done in public schools where Montessori preK was not necessarily offered. By contrast, some studies state that the school was recognized by AMI (Denervaud, 2019 ; Denervaud, 2020 ; Lillard, 2006 ; Lillard, 2012 ; Lillard, 2017 ; Mix, 2017 ; Rathunde, 2005a ; Rathunde, 2005b ), each of which suggests the full system was operating, or AMS, which has many elements but a different teacher training and often includes added practices (Daoust, 2018 , April). However, most studies of Montessori do not give enough information for readers to discern if the program had all the elements just described. Measures of Montessori implementation are rare and have yet to be standardized. Furthermore, it is unclear whether objective observers without Montessori training could discern whether important aspects were being properly implemented. For example, an untrained observer would not likely recognize if a teacher was presenting a material properly.
It is the case that people have incorporated elements of Montessori, like looping or no grades or more specifically the practical life materials (Bhatia, 2015 ), in conventional classrooms and often seen better results (see Lillard, 2017 for a review), and this would seem to point at which elements of Montessori are responsible for the results. If conventional teachers could improve child outcomes by simply adapting certain elements of Montessori, that would be more practical than adopting the whole system, which requires retraining many thousands of teachers, purchasing vast amounts of new materials, and eliminating the textbook industry. Adopting even some elements might be worthwhile for improving outcomes. But if the systems perspective is correct, then adopting elements is a weak solution. An alternative, if Montessori is considered sufficiently superior to warrant widespread adoption, is to convert one school or district at a time to the full Montessori system, beginning in lower‐income districts where the need for improvement is greatest.
7.2. Implications for research
Further research on Montessori education should attend carefully to implementation; Montessori programs can vary widely, with those recognized by AMI having the strictest implementation, AMS the next most strict, and others sometimes using the name Montessori without implementing the program to any great degree (Daoust, 2004 ; Daoust, 2018 ; Daoust, 2019 ). Although we caution that Montessori is a system such that the whole is likely not the same as the sum of its parts (Lillard, 2019 ), determining whether and how different implementations influence outcomes is very important.
A second important question for further research not addressed here is examining subsamples such as lower‐income children and global majority children. Montessori education, after all, was first designed to serve the needs of lower income students (Montessori, 1964 ). We extracted some data on these variables for the included studies, but it was outside the scope of this review to do a detailed analysis. We encourage researchers to use our data (Randolph, 2021 ) to carry out this line of inquiry.
Future research should follow children in Montessori longitudinally. Montessori is most frequently a preschool program and follows a pyramid structure with the fewest programs available at high school. Most studies of Montessori have been limited to a single data collection point. A few have collected data twice, at pretest and posttest, and very few have followed a sample over several years (Culclasure, 2018 ; Lillard, 2017 ) or tested people several years after Montessori schooling was complete (Jones, 1979 ; Miller, 1983 ; Miller, 1984 ; see also Dohrmann ( 2007 ), not included here due to lack of baseline equivalency). Recent findings from the Tennessee study of prekindergarten, which showed initially positive gains for children randomly assigned to preK reversed by 3rd grade, raise concerns about early schooling in general (Lipsey, 2018 ). Lipsey ( 2018 ) suggested that this could be due to a lack of curricular alignment between preK and later experiences, with Elementary school teachers focusing on students who had not had preK because they had skill deficits; this lack of focus then led to their later problems. This issue would not be present for children continuing from Montessori preK to Montessori Elementary school programs, but that is very much a concern going from Montessori preK to traditional Elementary programs. Alternatively, Lipsey et al. suggest that the misalignment of preschoolers' developmental needs with the academic emphases of the Elementary schools in which public pre‐K programs are often housed might have contributed to the poor 3rd‐grade outcomes. The current analysis shows that Montessori's nonacademic effects at preschool were stronger than its academic effects (0.39, as opposed to 0.20). This might suggest that Montessori preschool would provide for more whole‐child development and thereby mitigate a later downturn, but longitudinal research is needed to ascertain if that is the case.
Another area important for future research involves experimental study design. There is a dearth of randomized trials and high‐quality nonexperimental designs that adequately account for baseline differences between treatment and control groups. The use of these designs will reduce the selection threat (Shadish, 2002 ), which we consider to be the most likely threat in the Montessori research because it is likely that students enrolled in private schools–the setting for most Montessori programs–will differ at baseline on important academic and nonacademic variables. We suggest that if researchers are unable to implement random assignment, they use a baseline variable, such as a pretest, that is a direct measure of the outcome and to account for baseline differences statistically by using the baseline variable as a covariate or by the use of gain scores.
Finally, in line with suggestions of the American Statistical Association (Wasserstein, 2019 ), we encourage more meta‐analyses of the Montessori research and are making our data publicly available (see Randolph, 2021 ) to encourage replications and extensions of our review. Permission is universally granted for the use of the inclusion/exclusion data and the data extracted from the included studies. Some ideas for follow‐ups to this review are provided below.
It may be meaningful for future Montessori meta‐analyses to be detailed examinations, in which potential moderators are explored, for each of the individual major outcomes with a sufficient number of studies to do so.
It may also be meaningful to use other analytical methods to see if results are consistent across analytical methods. We used a cluster–robust variance estimation procedure that accounts for dependencies in the data and makes accurate point estimates; however, that procedure is less powerful than other synthesis methods and, therefore, will yield variance estimates that are less precise (Tanner‐Smith, 2014 ; Tanner‐Smith, 2016 ; Tipton, 2015 ).
Furthermore, we used a strict methodological criterion in terms of baseline equivalency. It may be useful for authors to use a less strict inclusion criterion and, thereby, be able to include a large number of studies that were excluded because of a lack of strong evidence for baseline equivalency. Our inclusion/exclusion spreadsheet lists the studies that were excluded based on baseline equivalency criteria.
We found evidence of asymmetry in funnel plots, which may be an indicator of publication bias. We suggest that follow‐up studies examine the publication bias in further detail. Trim and fill methods like those discussed in Shi 2019 may be useful for imputing left‐side missing data (L 0 ) and subsequently estimated the degree of bias resulting from study asymmetry.
Finally, we suggest that future research create a rating scale of Montessori and traditional implementation quality and use that as a potential moderator. We imagine that quality of treatment implementation could be an important variable in explaining the high amount of heterogeneity.
CONTRIBUTIONS OF AUTHORS
Justus J. Randolph: Overall methodological design, statistical analysis, and programming, data collection, and extraction, supervision, grant writing and management, primary text writer of methods and results sections. Anaya Bryson: Data collection and extraction, primary text writer of introductory sections. Lakshmi Menon: Data collection and extraction. David K. Henderson: Data collection, management, and extraction, effect size calculation and confirmation, risk‐of‐bias coder, quality control of data. Austin Kureethara Manuel: Data collection, management, and extraction, effect size calculation, quality control of data, risk‐of‐bias coding. Stephen Michaels: Design and implementation of search strategy, study and data management, the primary writer of the search results section. Warren McPherson: Theoretical expert and contributor to the theoretical section. debra leigh walls rosenstein: Theoretical expert and contributor to the theoretical section. Rebecca O'Grady: Data collection and extraction. Angeline S. Lillard: Subject matter and theoretical expert, interpretation, supervision and funding acquisition, introduction and the primary writer of discussion and plain‐language summary.
DECLARATIONS OF INTEREST
Justus J. Randolph's spouse is employed at a private Montessori school. Anaya Bryson has no declarations of interest. David K. Henderson has no declarations of interest. Lakshmi Menon has no declarations of interest. Austin Kureethara Manuel has no declarations of interest. Stephen Michaels has no declarations of interest. Warren McPherson and his spouse were cofounders of a private Montessori school. Mr. McPherson was the Director of the school for over 40 years until 2019. He and his spouse currently serve on the school's Board of Directors, without remuneration. He serves as a Montessori teacher‐trainer, consultant, and speaker and is remunerated for those activities. Debra leigh walls rosenstein has no declarations of interest. Angeline S. Lillard is the author of two studies in the meta‐analysis and was not involved in data extraction/coding/critical appraisal of those studies. She has spoken at many Montessori conferences and received remuneration. She has written a book on Montessori and its relation to developmental science and receives royalties. Finally, she attended Montessori from ages 3 to 6 and took a Montessori training course.
PUBLISHED NOTES
Errata. In figures, the term Hedge's g , should be spelled Hedges' g .
Characteristics of excluded studies 2
SUMMARY OF FINDINGS
1 Summary of findings
Note : Positive standardized mean differences favor Montessori education over traditional education. Because some studies collected data longitudinally in both control and experimental groups and/or used multiple measures of an outcome, we report the total number of observations instead of the number of participants.
The standardized mean difference effect size reported here is Hedges' g using a cluster‐robust method (Tanner‐Smith, 2014 ; Tanner‐Smith, 2016 ; Tipton, 2015 ).
SOURCES OF SUPPORT
Wend Collective, USA
The Wend Collective provided partial support of this review through a grant to Angeline Lillard.
Tift College of Education Seed Grant (2015–2016), USA
A Tift College of Education Seed Grant (Mercer University) provided $2500 to hire Graduate Research Assistants to assist with data collection and extraction (2015–2016).
Tift College of Education Seed Grant (2016–2017), USA
DIFFERENCES BETWEEN PROTOCOL AND REVIEW
There were several differences between the protocol and review:
We made some small changes to the introductory sections because another subject matter expert (Lillard) was added as a co‐author after the protocol had been published.
We updated our overall procedures to comply with the latest version of the MECCIR conduct standards (Campbell Collaboration, 2019a ) and MECCIR reporting standards (Campbell Collaboration, 2019b ) and the second edition of the Cochrane Handbook for Systematic Reviews (Higgins, 2021 ).
We created more specific inclusion and exclusion criteria to help define what qualifies as baseline equivalency.
Studies excluded because there was insufficient information to compute the relevant effect size were not narratively described because of the large number of studies excluded for this reason. However, we have provided readers with access to a spreadsheet that shows reasons for exclusion and notes.
We adopted the ROBINS‐I (Sterne, 2016 ) tool to assess risk of bias in nonrandomized studies.
We included Benjamini‐Hochberg (Benjamini, 1995 ) corrected α values in an online supplement for those interested in null hypothesis statistical testing.
Because almost all studies used continuous outcomes, we used a standardized mean difference effect size (specifically Hedges' g ) as the sole effect size metric.
We provided additional information about three different methods of effect size calculation that were used.
Because multiple methods of effect‐size estimation were used, we conducted a meta‐regression sensitivity analysis to see the degree to which those estimation methods covaried with outcomes.
For outcomes with more than 15 studies, we originally intended to deal with unit‐of‐analysis issues using multilevel meta‐analysis methods described in Viechtbauer ( 2010 ) and the methods in Konstantopoulos ( 2011 ) or the “pre‐defined hierarchy of outcomes” approach. However, because of the strict assumptions of multilevel approaches, we adopted cluster‐robust methods (Tanner‐Smith, 2014 ; Tanner‐Smith, 2016 ; Tipton, 2015 ) that have become more accessible since the protocol had been published.
Similarly, we adopted cluster‐robust models instead of the proposed simple, random effects models for outcomes with 15 or fewer outcomes. The exception is the social studies, in which there was just one study with multiple effect sizes so a random‐effects model was used.
We conducted a sensitivity analysis as suggested in Tanner‐Smith ( 2016 ) and Tipton ( 2015 ) to see how different assumptions about the correlation between within‐study effect sizes ( ρ ) might affect cluster‐robust results.
Because leave‐one‐out sensitivity analyses have become easy to implement since the publication of the protocol, we conducted a leave‐one‐out sensitivity analysis.
We used GOSH plots instead of forest plots when there were too many effect sizes to create a legible forest plot in R.
We visually examined Bajaut plots, radial plots, and various types of residual and fit plots to examine study heterogeneity.
An examination of funnel plots revealed an unexpectedly high degree of asymmetry with the bias in favor of Montessori education (i.e., studies were missing on the left/traditional education side of the funnel plots) on some outcomes. To make better decisions about the quality of evidence, we took the position of Duval ( 2000 ) that a trim‐and‐fill analysis would help us quantify the degree of publication bias while steering away from binary null hypothesis statistical significance tests of publication bias, as suggested by Higgins ( 2021 ). Although we did not intend to conduct a sophisticated analysis for publication bias, we think that the ability to quantify the effects of the potential publication bias on effect sizes warrants this deviation from the protocol.
The protocol was somewhat unclear in the moderator/subgroup analyses that would be conducted, so we clarified them here.
In the protocol, we proposed examining academic and behavioral outcomes. After using an emergent approach to identify specific outcomes, we determined that the term nonacademic outcomes is a more accurate descriptor than behavioral outcomes . We found that the variety of nonacademic outcomes measured in the Montessori literature was so broad, and not limited to just behavioral outcomes, that the best descriptor is simply nonacademic .
Supporting information
Supporting information.
ACKNOWLEDGMENTS
Thanks to Allison Snyder and the research assistants she managed, for their help in pulling together the Characteristics of Included Studies table, among many other helpful tasks. Thanks to the many authors who graciously accommodated our requests for additional data.
Randolph, J. J. , Bryson, A. , Menon, L. , Henderson, D. K. , Kureethara Manuel, A. , Michaels, S. , rosenstein, d. l. w. , McPherson, W. , O'Grady, R. , & Lillard, A. S. (2023). Montessori education's impact on academic and nonacademic outcomes: A systematic review. Campbell Systematic Reviews, 19, e1330. 10.1002/cl2.1330
As the current review was going to press, we became aware of the publication of another Montessori meta‐analysis, which, largely, had effect sizes similar to the ones presented here. See Demangeon, A., Claudel‐Valentin, S., Aubry, A., & Tazouti, Y. (2023). A Meta‐Analysis of the Effects of Montessori Education on Five Fields of Development and Learning in Preschool and School‐Age Children. Contemporary Educational Psychology, 102182. doi:10.1016/j.cedpsych.2023.102182
See the supplemental information (Randolph, 2021 ) for informaton on reasons for exclusion.
INCLUDED STUDIES
Alburaidi 2019.
- Alburaidi, A. , & Ambusaidi, A. (2019). The impact of using activities based on the Montessori approach in science in the academic achievement of fourth grade students. International Journal of Instruction, 12(2), 695–708. [ Google Scholar ]
ANSARI 2014
- Ansari, A. , & Winsler, A. (2014). Montessori public school pre‐k programs and the school readiness of low‐income Black and Latino children. Journal of Educational Psychology, 106(4), 1066–1079. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
AYDOĞAN 2016
- Aydoğan, Y. (2016). Examination of the affect of Montessori education on language development of pre‐school children. Academic Research International, 7(5), 112–119. [ Google Scholar ]
BESANÇON 2013
- Besançon, M. , Todd, L. , & Baptiste, B. (2013). Creative giftedness and educational opportunities. Educational and Child Psychology, 30(2), 79–88. [ Google Scholar ]
- Coyle, S. J. V. , & Josephina, C. S. (1968). A longitudinal assessment of preschool children in haptic learning . Final report.
CULCLASURE 2018
- Culclasure, B. , David, F. , Ginny, R. , & Alexis, S . (2018). An evaluation of Montessori education in South Carolina's public schools. The Riley Institute at Furman University. Unpublished manuscript.
DENERVAUD 2019
- Published and unpublished data
- Denervaud, S. , Jean‐François, K. , Patric, H. , & Edouard, G. (2019). Beyond executive functions, creativity skills benefit academic outcomes: Insights from montessori education. PLoS One , 14(11). [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
DENERVAUD 2020
- Denervaud, S. , Jean‐François, K. , Mary Helen, I.‐Y. , & Patric, H. (2020). Effects of traditional versus Montessori schooling on 4‐ to 15‐year old children's performance monitoring. Mind, Brain, and Education , 14(2). [ Google Scholar ]
- Doğru, S. S. Y. (2015). Efficacy of Montessori education in attention gathering skill of children. Educational Research and Reviews, 10(6), 733–738. [ Google Scholar ]
- Elben, J. M . (2015). A comparison of reading attainment in two first grade classes in a state and a Montessori school in Switzerland [Master's thesis, Massey University]. https://mro.massey.ac.nz/handle/10179/7547
FARYADI 2017
- Faryadi, Q. (2017). The application of Montessori method in learning mathematics: An experimental research. Open Access Library Journal , 4(e4140). [ Google Scholar ]
FLEEGE 1967
- Fleege, U. H. (1967). Montessori pre‐school education . (ED017320). ERIC. http://eric.ed.gov/?id=ED017320
GALINDO 2014
- Galindo, M. I. (2014). A public bilingual prekindergarten Montessori program and its impact on Bracken test literacy outcomes: Principals' beliefs on the results. [Doctoral dissertation, University of Houston]. https://uh-ir.tdl.org/uh-ir/handle/10657/881
HOSEINPOOR 2014
- Hoseinpoor, M. , Mozhgan, A. , & Saeed, S. (2014). Study of effect of education through the Montessori method on attention and perceptual organization and social growth of 5‐6 yr preschool children in the city of Isfahan. International Journal of Review in Life Sciences, 4(12), 148–154. [ Google Scholar ]
- Jones, B. J. , & Miller, L. B. (1979). Four preschool programs: Their lasting effects (ED171415). ERIC. https://eric.ed.gov/?id=ED171415
JUANGA 2015
- Juanga, J. M. V. , & Ressureccion, A. C. (2015). Comparison of academic performance and attention span of children between Montessori and traditional pedagogical approaches of preschools . Proceedings of the International Conference on Information Communication Technologies in Education, 336–347. http://www.icicte.org/ICICTE2015Proceedings(Papers)/9.4%20Final%20Juanga%20_Ressureccion.pdf
KAYILI 2016A
- Kayili, G. , & Ramazan, A. (2016). The effect of Montessori method supported by social skills training program on Turkish kindergarten children's skills of understanding feelings and social problem solving. Journal of Education and Training Studies, 4(12), 81–91. [ Google Scholar ]
KAYILI 2016B
- Kayili, G. (2016). The effect of Montessori Method on cognitive tempo of kindergarten children. Early Child Development and Care, 188(3), 327–335. 10.1080/03004430.2016.1217849 [ DOI ] [ Google Scholar ]
KIRKHAM 2017
- Kirkham, J. A. , & Kidd, E. (2017). The effect of Steiner, Montessori, and national curriculum education upon children's pretence and creativity. The Journal of Creative Behavior, 51(1), 20–34. [ Google Scholar ]
LILLARD 2006
- Lillard, A. , & Else‐Quest, N. (2006). Evaluating Montessori education. Science, 313(5795), 1893–1894. [ DOI ] [ PubMed ] [ Google Scholar ]
LILLARD 2012
- Lillard, A. S. (2012). Preschool children's development in classic Montessori, supplemented montessori, and conventional programs. Journal of School Psychology, 50(3), 379–401. [ DOI ] [ PubMed ] [ Google Scholar ]
LILLARD 2017
- Lillard, A. S. , Heise, M. J. , Richey, E. M. , Tong, X. , Hart, A. , & Bray, P. M. (2017). Montessori preschool elevates and equalizes child outcomes: A longitudinal study. Frontiers in Psychology , 8(1783). 10.3389/fpsyg.2017.01783 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
MALLETT 2015
- Mallett, J. D. , & Schroeder, J. L. (2015). Academic achievement outcomes: A comparison of Montessori and non‐Montessori public elementary school students. Journal of Elementary Education, 25(1), 39–53. [ Google Scholar ]
MANNER 1999
- Manner, J. C . (1999). A comparison of academic achievement of Montessori and non‐Montessori students in a public school setting [Doctoral dissertation, Florida International University]. ProQuest Dissertations Publishing.
MILLER 1983
- Miller, L. B. , & Bizzell, R. P. (1983). Long‐Term effects of four preschool programs: Sixth, seventh, and eighth grades. Child Development , 54(3), 727–741. [ PubMed ] [ Google Scholar ]
MILLER 1984
- Miller, L. B. , & Bizzell, R. P. (1984). Long‐term effects of four preschool programs: Ninth‐ and tenth‐grade results. Child Development, 55(4), 1570–1587. [ PubMed ] [ Google Scholar ]
- Mix, K. S. , Smith, L. B. , Stockton, J. D. S. , Cheng, Y.‐L. , & Barterian, J. A. (2017). Grounding the symbols for place value: Evidence from training and long‐term exposure to base‐10 models. Journal of Cognition & Development, 18(1), 129–151. [ Google Scholar ]
PRENDERGAST 1969
- Prendergast, R. (1969). Pre‐Reading skills developed in Montessori and conventional nursery schools. The Elementary School Journal, 70(3), 135–141. [ Google Scholar ]
RATHUNDE 2005a
- Rathunde, K. , & Mihaly, C. (2005). Middle school students' motivation and quality of experience: A comparison of Montessori and traditional school environments. American Journal of Education, 111(3), 341–371. [ Google Scholar ]
RATHUNDE 2005b
- Rathunde, K. , & Csikszentmihalyi, M. (2005). The social context of middle school: Teachers, friends, and activities in Montessori and traditional school environments. The Elementary School Journal, 106(1), 59–80. [ Google Scholar ]
- Tobin, T. , Prairie, B. , Wenyi, Z. , Paul, H. , & Peter, M. (2015). Improving outcomes for refugee children: A case study on the impact of Montessori education along the Thai‐Burma border. International Education Journal: Comparative Perspectives, 14(3), 138–149. [ Google Scholar ]
YUSSEN 1980
- Yussen, S. R. , Samuel, M. , & Knight, J. W. (1980). Performance of Montessori and traditionally schooled nursery children on social cognitive tasks and memory problems. Contemporary Educational Psychology, 5(2), 124–137. [ Google Scholar ]
EXCLUDED STUDIES
- Ahmad, S. , & Amjad, R. (2018). Social and moral development of students: A comparative study on Montessori and non‐Montessori students. Journal of Elementary Education, 28(1), 53–64. [ Google Scholar ]
AHMADPOUR 2015
- Ahmadpour, N. , & Kraskian, M. A. (2015). The impact of Montessori teaching method on IQ levels of 5‐year old children. Procedia—Social and Behavioral Sciences, 205(Suppl. C), 122–127. [ Google Scholar ]
BAERMAN 2002
- Baerman, M. P. (2002). The effect of the Montessori method on math achievement of first grade students [Master's thesis, Mercer University].
- Bagby, J. , Lucy, B.‐B. , Tracey, S. , Natalie, J. , & Mary, W. (2012). The effects of environment on children's executive function: A study of three private schools. Journal of Research in Childhood Education, 26(4), 418–426. [ Google Scholar ]
BAINES 1973
- Baines, M. R. , & Snortum, J. R. (1973). A time‐sampling analysis of Montessori versus traditional classroom interaction. The Journal of Educational Research, 66(7), 313–316. [ Google Scholar ]
- Bank Street College of Education. (1969). A comparative study of the impact of two contrasting educational approaces in Head Start, 1968‐69 (ED041643). ERIC. https://archive.org/details/ERIC_ED041643
- Banta, T. J. (1970). Research on Montessori and the disadvantaged. In Orem R. C. (Ed.), Montessori and the special child (pp. 171–176). Capricorn Books. [ Google Scholar ]
BEREITER 1967
- Bereiter, C. , & Urbana, I. U . (1967). Acceleration of intellectual development in early childhood (ED014332). ERIC. https://eric.ed.gov/?id=ED014332
BERENDS 2018
- Berends, M. , & Joseph, W. R. (2018). School choice in Indianapolis: Effects of charter, magnet, private, and traditional public schools. Education Finance and Policy, 13(2), 227–255. [ Google Scholar ]
BERGER 1969
- Berger, B . (1969). A longitudinal investigation of Montessori and traditional prekindergarten training with inner city children: A comparative assessment of learning outcomes. Three part study (ED034588). U.S. Department of Health, Education, and Welfare. http://eric.ed.gov/?id=ED034588
BESANÇON 2008
- Besançon, M. , & Todd, L. (2008). Differences in the development of creative competencies in children schooled in diverse learning environments. Learning & Individual Differences, 18(4), 381–389. [ Google Scholar ]
BLANCO‐VEGA 2002
- Blanco‐Vega, C . (2002). Comparing children's self‐esteem levels in Montessori versus traditional schools [Honors thesis, Ball State University]. http://cardinalscholar.bsu.edu/handle/handle/189522
BORMAN 2003
- Borman, G. D. , Hewes, G. M. , Overman, L. T. , & Brown, S. (2003). Comprehensive school reform and achievement: A meta‐analysis. Review of Educational Research, 73(2), 125–230. [ Google Scholar ]
- Brand, H. J. , & Kim, W. (1989). Cognitive and social‐emotional development of children in different preschool environments. Psychological Reports, 65(2), 480–482. [ DOI ] [ PubMed ] [ Google Scholar ]
BROPHY 1973
- Brophy, J. E. , & Choquette, J. J. (1973). Divergent production in Montessori children (ED080212). ERIC. https://eric.ed.gov/?id=ED080212
- Brown, K. E. , & Steele, A. S. L. (2015). Racial discipline disproportionality in Montessori and traditional public schools: A comparative study using the relative rate index. Journal of Montessori Research, 1(1), 14–27. [ Google Scholar ]
- Brown, K. E . (2016). Evaluating the effectiveness of Montessori reading and math instruction for third grade African American students in urban elementary schools (ED572111) [Doctoral dissertation, The University of North Carolina at Charlotte]. https://eric.ed.gov/?id=ED572111
- Byun, W. , Blair, S. N. , & Pate, R. R. (2013). Objectively measured sedentary behavior in preschool children: Comparison between Montessori and traditional preschools. The International Journal of Behavioral Nutrition and Physical Activity , 10(2). 10.1186/1479-5868-10-2 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
CASTELLANOS 2002
- Castellanos, A. G. (2002). A comparison of traditional vs. Montessori education in relation to children's self‐esteem, self‐efficacy, and prosocial behavior [Doctoral dissertation, Carlos Albizu University]. http://search.proquest.com/education/docview/305483456/abstract/D1B17839958F4E38PQ/1
CHATTIN‐MCNICHOLS 1981
- Chattin‐McNichols, J. P. (1981). The effects of Montessori school experience. Young Children, 36(5), 49–66. [ Google Scholar ]
CHISNALL 2007
- Chisnall, N. , & Marguerite, M. (2007). Montessori mathematics in early childhood education. Curriculum Matters, 3, 6–29. [ Google Scholar ]
CLAXTON 1982
- Claxton, S. S . (1982). A comparison of student achievement, student self‐concept, and parental attitude toward traditional and Montessori programs in a public school setting [Doctoral dissertation, North Texas State University]. ProQuest Dissertations Publishing.
COOPMANS 1984
- Coopmans, J. H. (1984). A preliminary study concerning the Middenschool experiment. Studies in Educational Evaluation, 10(3), 329–331. [ Google Scholar ]
- Corry, S. K. (2006). A comparison of Montessori students to general education students as they move from middle school into a traditional high school program [Doctoral dissertation, University of Nebraska]. ProQuest Dissertations Publishing.
- Cox, M. V. , & Rowlands, A. (2000). The effect of three different educational approaches on children's drawing ability: Steiner, Montessori and traditional. British Journal of Educational Psychology, 70(4), 485–503. 10.1348/000709900158263 [ DOI ] [ PubMed ] [ Google Scholar ]
DAWSON 1987
- Dawson, M . (1987). Minority student performance: Is the Montessori magnet school effective? (ED309881). ERIC. http://eric.ed.gov/?id=ED309881
DE LUCA 2006
- De Luca, B. M. , & Steven, H. (2006). Comparing academic achievement in charter schools and public schools: The role of money. Journal of Educational Research & Policy Studies, 6(1), 67–90. [ Google Scholar ]
DHIKSHA 2015
- Dhiksha, J. , & Shivakumara, K. (2015). Emotional intelligence among children of Montessori and traditional method of education. Indian Journal of Applied Research, 5(8), 271–274. [ Google Scholar ]
DIAMOND 2010
- Diamond, A. (2010). The evidence base for improving school outcomes by addressing the whole child and by addressing skills and attitudes, not just content. Early Education & Development, 21(5), 780–793. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
DIAMOND 2011
- Diamond, A. , & Kathleen, L. (2011). Interventions shown to aid executive function development in children 4 to 12 years old. Science, 333(6045), 959–964. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
DI LORENZO 1968
- Di Lorenzo, L. T. (1968). Effects of year‐long pre‐kindergarten programs on intelligence and language of educationally disadvantaged children. Journal of Experimental Education, 36(3), 36–42. [ Google Scholar ]
DI LORENZO 1969
- Di Lorenzo, L. T. (1969). Prekindergarten programs for educationally disadvantaged children . Final Report. (ED038460). ERIC. http://eric.ed.gov/?id=ED038460 [ PubMed ] [ Google Scholar ]
DOHRMANN 2003
- Dohrmann, K. (2003). Outcomes for students in a Montessori program: A longitudinal study of the experience in the Milwaukee public schools . American Montessori International. http://public‐montessori.org/sites/default/files/resources/Outcomes_Milwaukee%20b.pdf
DOHRMANN 2007
- Dohrmann, K. R. , Nishida, T. K. , Gartner, A. , Lipsky, D. K. , & Grimm, K. J. (2007). High school outcomes for students in a public Montessori program. Journal of Research in Childhood Education, 22(2), 205–217. [ Google Scholar ]
DONABELLA 2008
- Donabella, M. A. , & Rule, A. C. (2008). Four seventh grade students who qualify for academic intervention services in mathematics learning multi‐digit multiplication with the Montessori checkerboard. Teaching Exceptional Children Plus , 4(3). Article 2. [ Google Scholar ]
DREYER 1969
- Dreyer, A. S. , & David, R. (1969). Cognitive performance in Montessori and nursery school children. The Journal of Educational Research, 62(9), 411–416. [ Google Scholar ]
- Duax, T. (1989). Preliminary report on the educational effectiveness of a Montessori school in the public sector. The NAMTA Journal, 14(2), 56–62. [ Google Scholar ]
EPSTEIN 1996
- Epstein, A. S. , Schweinhart, L. J. , & McAdoo, L. (1996). Models of early childhood education. High/Scope Press. [ Google Scholar ]
- Ervin, B. , Wash, P. D. , & Mecca, M. E. (2010). A 3‐year study of self‐regulation in Montessori and non‐Montessori classrooms. Montessori Life: A Publication of the American Montessori Society, 22(2), 22–31. [ Google Scholar ]
ESPOSITO 2010
- Esposito, C. L. (2010). School type and mathematics achievement: A comparison of magnet and public secondary schools using the educational longitudinal study of 2002 data set [Doctoral dissertation, University of Connecticut]. https://opencommons.uconn.edu/dissertations/AAI3447464
- Fero, J. R . (1997). A comparison of academic achievement of students taught by the Montessori method and by traditional methods of instruction in the elementary grades [Doctoral dissertation, Montana State University]. http://scholarworks.montana.edu/xmlui/handle/1/7903
FINDLAY 1913
- Findlay, J. J. (1913, May 1). The Montessori system: Report of an investigation recently conducted at the Fielden School. The Educational Times , 203–207.
FLEMING 2019
- Fleming, D. J. , Culclasure, B. T. , & Zhang, D. (2019). The Montessori model and creativity. Journal of Montessori Research, 5(2), 1–14. 10.17161/jomr.v5i2.7695 [ DOI ] [ Google Scholar ]
FLOWER 2006
- Flower, N. R. (2006). A comparison of the occurrence of bullying in a traditional versus a Montessori school setting [Doctoral dissertation, Xavier University]. http://rave.ohiolink.edu/etdc/view?acc_num=xavier1385992331
- Flynn, T. M. (1991). Development of social, personal and cognitive skills of preschool children in Montessori and traditional preschool programs. Early Child Development and Care, 72(1), 117–124. [ Google Scholar ]
- Franc, V. , & Siniša, S . (2015). Differences in phonological awareness of five‐year‐olds from Montessori and regular program preschool institutions. The Faculty of Teacher Education University of Zagreb Conference , (pp. 12–20).
FRANCZAK 2016
- Franczak, I . (2016). Comparative analysis of behavioral engagement and transferable skills in conventional and Montessori schools. Proceedings of The National Conference on Undergraduate Research (NCUR) 2016 University of North Carolina, Asheville. April 7‐9, 2016.
GALLIGER 2009
- Galliger, C. C . (2009). To be or not to be…motivated: A comparison of students' goal orientation within direct instruction and constructivist schools [Doctoral dissertation, Bowling Green State University]. http://rave.ohiolink.edu/etdc/view?acc_num=bgsu1237475910
- Glenn, C. (1993). A comparison of lower and upper elementary Montessori students with a public school sample. The NAMTA Journal, 14(2), 63–68. [ Google Scholar ]
- Glenn, C. M. (2003). The longitudinal assessment study (LAS): Eighteen year follow‐up . Final report. (ED478792). Franciscan Montessori Earth School. https://eric.ed.gov/?id=ED478792
- Gross, R. B. , Green, B. L. , & Clapp, D. F . (1970). The Sands School project: Third year results (ED078958). ERIC. https://files.eric.ed.gov/fulltext/ED078958.pdf [ Google Scholar ]
- Grubb, P. W. (2000). A comparison of concept age gains of kindergarten children in traditional and twenty‐first century classrooms (ED443523). ERIC. http://eric.ed.gov/?id=ED443523
GUIDUBALDI 1974
- Guidubaldi, J. , Bursani, C. A. , Kehle, T. J. , & Sexton, S. E. (1974). Effectiveness of preschool programs as a function of childrens' socioeconomic status (ED 090273). ERIC. https://files.eric.ed.gov/fulltext/ED090273.pdf
- Guven, Y. , Cihat, G. , & Beyzanur, D. A. (2020). Comparison of Sudoku solving skills of preschool children enrolled in the Montessori approach and the national education programs. Journal of Education and Training Studies, 8(3), 32–47. [ Google Scholar ]
HANSON 2009
- Hanson, B. E. E. (2009). An exploratory study on the effectiveness of Montessori constructs and traditional teaching methodology as change agents to increase academic achievement of elementary Black students [Doctoral dissertation, Capella University]. ProQuest Dissertations Publishing.
- Haq, Z. I. A. , Alfilfili, H. H. , & Joradan, J. (2015). The efficiency of an educational program based on Montessori curriculum in developing logical thinking in kindergarten children. European Journal of Research and Reflection in Educational Sciences, 3(1), 8–26. [ Google Scholar ]
HARRIS 1995
- Harris, I. , & Aaron, C. (1995). Comparative study of Peace education approaches and their effectiveness. The NAMTA Journal, 20(2), 133–144. [ Google Scholar ]
HARRIS 2007
- Harris, M. A. (2007). Differences in mathematics scores between students who receive traditional Montessori instruction and students who receive music enriched Montessori instruction. Journal for Learning through the Arts , 3(1). 10.21977/D93110059 [ DOI ] [ Google Scholar ]
- Heise, E. , Boehme, E. , & Koerner, S. B. (2010). Montessori education versus traditional education in primary school: A comparison of the development of orthographical and mathematical skills, intelligence, and creativity. Psychologie in Erziehung Und Unterricht, 57(4), 273–289. 10.2378/peu2010.art19d [ DOI ] [ Google Scholar ]
HICKERSON 1983
- Hickerson, M. K . (1983). A comparison of left and right brain hemisphere processing and brain related sex differences in kindergarten children [Doctoral dissertation, East Tennessee State University]. ProQuest Dissertations Publishing.
- Hobbs, A . (2008). Academic achievement: Montessori and non‐Montessori private school settings [Doctoral dissertation, University of Houston]. ProQuest Dissertations Publishing.
HOJNOSKI 2008
- Hojnoski, R. L. , Margulies, A. S. , Barry, A. , Bose‐Deakins, J. , Sumara, K. M. , & Harman, J. L. (2008). Analysis of two early childhood education settings: Classroom variables and peer verbal interaction. Journal of Research in Childhood Education, 23(2), 193–209. [ Google Scholar ]
- İman, E. D. , Şahin, D. , Akin, D. Z. , & Dilara, Y. (2017). The effect of the Montessori education method on pre‐school children's social competence—behaviour and emotion regulation skills. Early Child Development and Care, 189(9), 1494–1508. [ Google Scholar ]
JARVIS 2015
- Jarvis, J . (2015). The association between preschool type and children's concentration abilities [Honors thesis, Bowling Green State University]. https://scholarworks.bgsu.edu/honorsprojects/161
JOHNSTON 2013
- Johnston, L. C . (2013). Examining Montessori middle school through a self‐determination theory lens: A mixed methods study of the lived experiences of adolescents [Doctoral dissertation, University of California, San Diego]. ProQuest Dissertations Publishing.
- Jones, T . (2005). Comparing Montessori education and conventional education on aspects of creativity [Honors thesis, Syracuse University]. https://surface.syr.edu/honors_capstone/683
- Judge, J. M . (1974). A comparison of preschool children in observational tasks from two programs: Montessori and science—A process approach (ED095013). ERIC. https://eric.ed.gov/?id=ED095013
KAMAKIL 2013
- Kamakil, W. C. (2013). Play and creative drawing in preschool: A comparative study of Montessori and public preschools in Kenya. Herbert Utz Verlag. [ Google Scholar ]
KARNES 1969
- Karnes, M. B. , Hodgins, A. A. , Teska, J. A. , & Kirk, S. A. (1969). Investigations of classroom and at‐home interventions: Research and development program on preschool disadvantaged children (ED036663). U.S. Department of Health, Education, and Welfare. https://eric.ed.gov/?id=ED036663
KARNES 1977
- Karnes, M. B. (1977). The University of Illinois study of the differential effects of five preschool programs (ED136937). ERIC. http://eric.ed.gov/?id=ED136937
KARNES 1978
- Karnes, M. B. , Plummer, C. M. , & Lee, R. C. (1978, March 27). Immediate, short‐term and long‐range effects of five preschool programs for disadvantaged children. Annual Meeting of the American Educational Research Association, Toronto, Canada. [ Google Scholar ]
KARNES 1983
- Karnes, M. B. , Shwedel, A. M. , & Williams, M. B. (1983). A comparison of five approaches for educating young children from low‐income homes. In The Consortium for Longitudinal Studies (Ed.), As the twig is bent: Lasting effects of preschool programs (pp. 133–169). Lawrence Erlbaum Associates. [ Google Scholar ]
KAYILI 2011A
- Kayili, G. , & Ramazan, A. (2011). Examination of the effects of the Montessori method on preschool children's readiness to primary education. Educational Sciences: Theory & Practice, 11(4), 2104–2109. [ Google Scholar ]
KENDALL 1993
- Kendall, S. D. (1993). The development of autonomy in children: An examination of the Montessori educational model. The NAMTA Journal, 18(1), 64–83. [ Google Scholar ]
KIMMINS 1915
- Kimmins, C. W. (1915). A Montessori experiment at Sway, Hampshire. Child Study , 69–72.
KIMMINS 1915A
- Kimmins, C. W. (1915). Some recent Montessori experiments in England. Report on the Conference on New Ideals in Education (pp. 55–69).
KOHLBERG 1968
- Kohlberg, L. (1968). Montessori with the culturally disadvantaged: A cognitive‐developmental interpretation and some research findings. In Hess R. D., & Bear R. M. (Eds.), Early education; Current theory, research, and action (pp. 105–118). Aldine Pub. Co. [ Google Scholar ]
KRAFFT 1998
- Krafft, K. C. , & Berk, L. E. (1998). Private speech in two preschools: Significance of open‐ended activities and make‐believe play for verbal self‐regulation. Early Childhood Research Quarterly, 13(4), 637–658. 10.1016/S0885-2006(99)80065-9 [ DOI ] [ Google Scholar ]
- Krogh, S. L . (1979). Moral development in Montessori and traditional preschool children: Does school type make a difference? (ED190217). ERIC.
- Lail, S.‐F. C . (2017). Students diagnosed with attention deficit hyperactivity disorder: A study of on‐task and off‐task behaviors in traditional versus Montessori classrooms [Doctoral dissertation, University of South Carolina]. ProQuest Dissertations Publishing.
- LaRue, W . (2010). Empowering adolescents: A multiple case study of U.S. Montessori high schools [Doctoral dissertation, Walden University]. http://scholarworks.waldenu.edu/dissertations/731
- Laski, E. V. , Jor'dan, J. R. , Daoust, C. , & Murray, A. K. (2015). What makes mathematics manipulatives effective? Lessons from cognitive science and Montessori education. SAGE Open . 5(2). 10.1177/2158244015589588 [ DOI ] [ Google Scholar ]
- Laski, E. V. , Marina, V. , & Joanna, S. (2016). Longitudinal comparison of place‐value and arithmetic knowledge in Montessori versus non‐Montessori students. Journal of Montessori Research, 2(1), 1–15. [ Google Scholar ]
LILLARD 2005
- Lillard, A. S. (2005). Montessori: The science behind the genius. Oxford University Press. [ Google Scholar ]
LILLARD 2013
- Lillard, A. S. (2013). Playful learning and Montessori education. American Journal of Play, 5(2), 157–186. [ Google Scholar ]
LILLARD 2016
- Lillard, A. S. , & Heise, M. J. (2016). An intervention study: Removing supplemented materials from Montessori classrooms associated with better child outcomes. Journal of Montessori Research, 2(1), 16–26. 10.17161/jomr.v2i1.5678 [ DOI ] [ Google Scholar ]
LILLARD 2021
- Lillard, A. S. , Meyer, M. J. , Vasc, D. , & Fukuda, E. (2021). An association between Montessori education in childhood and adult wellbeing. Frontiers in Psychology, 12, 721943. 10.3389/fpsyg.2021.721943 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
LOPATA 2005
- Lopata, C. , Wallace, N. V. , & Finn, K. V. (2005). Comparison of academic achievement between Montessori and traditional education programs. Journal of Research in Childhood Education, 20(1), 5–13. [ Google Scholar ]
MALLETT 2013
- Mallett, J. D . (2013). Academic achievement outcomes: Montessori and non‐Montessori public elementary students [Master's thesis, Texas A&M University] ProQuest Dissertations Publishing.
MALLETT 2014
- Mallett, J. D . (2014). Longitudinal academic achievement outcomes: Modeling the growth trajectories of Montessori elementary public school students [Doctoral dissertation, Texas A&M University]. ProQuest Dissertations Publishing.
MANNER 2007
- Manner, J. C. (2007). Montessori vs. traditional education in the public sector: Seeking appropriate comparisons of academic achievement. Forum on Public Policy , 2007(2). https://forumonpublicpolicy.com/archivespring07/manner.pdf [ Google Scholar ]
MCCLADDIE 2006
- McCladdie, K . (2006). A comparison of the effectiveness of the Montessori Method of reading instruction and the Balanced Literacy Method for inner city African American students [Doctoral dissertation, Saint Joseph's University]. ProQuest Dissertations Publishing.
MCDURHAM 2011
- McDurham, R . (2011). A comparison of academic achievement for seventh and eighth grade students from Montessori and non‐Montessori school programs [Doctoral dissertation, Tarleton State University]. ProQuest Dissertations Publishing.
MCKINNON 1982
- McKinnon, J. , Flieger, S. , & Patterson, M. (1982). A comparative study of the effects of preschool education on middle class children (ED220179). Puce Area Development Project, ERIC.
MIEZITIS 1971
- Miezitis, S. (1971). The Montessori method: Some recent research. Interchange, 2, 41–59. [ Google Scholar ]
MILLER 1970
- Miller, L. B. , & Dyer, J. L. (1970). Experimental variation of Head Start curricula: A comparison of current approaches (ED045196). ERIC. http://eric.ed.gov/?id=ED045196
MILLER 1971
- Miller, L. B. , & Dyer, J. L. (1971). Two kinds of kindergarten after four types of Head Start (ED050824). ERIC. http://eric.ed.gov/?id=ED050824
MILLER 1972
- Miller, L. B. , & Dyer, J. L. (1972). Four preschool programs: Their dimensions and effects (ED069411). ERIC. http://eric.ed.gov/?id=ED069411
MILLER 1975
- Miller, L. B. , & Dyer, J. L. (1975). Four preschool programs: Their dimensions and effects. Monographs of the Society for Research in Child Development, 40(5—6, Serial 162). [ Google Scholar ]
MORFITT 1937
- Morfitt, M. D. K. (1937). Comparison of individual‐concrete methods and class methods in the teaching of arithmetic. British Journal of Educational Psychology, 7, 196–203. [ Google Scholar ]
MORGAN 1978
- Morgan, S. M. (1978). A comparative assessment of some aspects of the development of the concept of number and arithmetical skills in Montessori and traditional preschools [Doctoral dissertation, Syracuse University].
MROCZKOWSKI 2014
- Mroczkowski, J . (2014). Comparing the normalization of children in traditional and Montessori kindergarten [Master's thesis, University of Wisconsin ‐ River Falls]. http://digital.library.wisc.edu/1793/69077
MURPHY 1976
- Murphy, M. J. , & Goldner, R. P. (1976). Effects of teaching orientation on social interaction in nursery school. Journal of Educational Psychology, 68(6), 725–728. [ Google Scholar ]
- Pate, R. R. , O'Neill, J. R. , Byun, W. , McIver, K. L. , Dowda, M. , & Brown, W. H. (2014). Physical activity in preschool children: Comparison between Montessori and traditional preschools. Journal of School Health, 84(11), 716–721. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
- Peng, H.‐H. (2009). A comparison of the achievement test performance of children who attended Montessori schools and those who attended non‐Montessori schools in Taiwan [Doctoral dissertation, Indiana State University]. ProQuest Dissertations Publishing.
- Peng, H.‐H. , & Md‐Yunus, S. 'ah (2014). Do children in Montessori schools perform better in the achievement test? A Taiwanese perspective. IJEC, 46(2), 299–311. [ Google Scholar ]
PHILLIPS‐SILVER 2018
- Phillips‐Silver, J. , & Daza, M. T. (2018). Cognitive control at age 3: Evaluating executive functions in an equitable Montessori preschool. Frontiers in Education, 3, 106. 10.3389/feduc.2018.00106 [ DOI ] [ Google Scholar ]
RATHUNDE 2003
- Rathunde, K. (2003). A comparison of Montessori and traditional middle schools: Motivation, quality of experience, and social context. The NAMTA Journal, 28(3), 12–53. [ Google Scholar ]
- Reed, M. K. (2000). A comparison of the place value understanding of Montessori and non‐Montessori elementary school students [Doctoral dissertation, The Ohio State University]. ProQuest Dissertations Publishing.
- Reich, C . (1974). Follow‐up of the Montessori programme (ED103134). ERIC. http://eric.ed.gov/?id=ED103134
REUTER 1973
- Reuter, J. , & Yunik, G. (1973). Social interaction in nursery schools. Developmental Psychology, 9(3), 319–325. [ Google Scholar ]
ROBERTS 1976
- Roberts, C. L. C . (1976). An investigation of the influence of preschool programs on children's reading progress: A review of model programs with implications for the future . Paper presented at the Annual Meeting of the International Reading Association World Congress on Reading.
RODRIGUEZ 2003
- Rodriguez, L. , Irby, B. J. , Brown, G. , Lara‐Alecio, R. , & Galloway, M. (2003). An analysis of a public school prekindergarten bilingual Montessori program (ED478568). ERIC. https://eric.ed.gov/?id=ED478568
- Rose, S. E. , Jolley, R. P. , & Charman, A. (2012). An investigation of the expressive and representational drawing development in national curriculum, Steiner, and Montessori schools. Psychology of Aesthetics, Creativity, and the Arts, 6(1), 83–95. 10.1037/a0024460 [ DOI ] [ Google Scholar ]
- Ruijs, N. (2017). The effects of Montessori education: Evidence from admission lotteries. Economics of Education Review, 61, 19–34. 10.1016/j.econedurev.2017.09.001 [ DOI ] [ Google Scholar ]
SALAZAR 2013
- Salazar, M. M . (2013). The impact of Montessori teaching on academic achievement of elementary school students in a central Texas school district: A causal‐comparative inquiry [Doctoral dissertation, Texas A&M University ‐ Corpus Christi]. ProQuest Dissertations Publishing.
SCIARRA 1976
- Sciarra, D. J. , & Dorsey, A . (1976). Nine‐year follow‐up study of Montessori education (ED121929). ERIC. https://files.eric.ed.gov/fulltext/ED121929.pdf
SEBASTIAN 2016
- Sebastian, K. , & Matheen, W. (2016). A comparison of Montessori and traditional school children on cognitive development. The International Journal of Indian Psychology, 4(1), 32–40. [ Google Scholar ]
SEEFELDT 1977
- Seefeldt, C . (1977). Montessori and responsive environment models: A longitudinal study of two preschool programs, phase two (ED139520). ERIC. http://eric.ed.gov/?id=ED139520
SEEFELDT 1981
- Seefeldt, C. (1981). Social and emotional adjustment of first grade children with and without Montessori preschool experience. Child Study Journal, 11(4), 231–246. [ Google Scholar ]
SHANKLAND 2009
- Shankland, R. , França, L. R. , Genolini, C. M. , Guelfi, J.‐D. , & Ionescu, S. (2009). Preliminary study on the role of alternative educational pathways in promoting the use of problem‐focused coping strategies. European Journal of Psychology of Education, 24(4), 499–512. [ Google Scholar ]
SHERNOFF 2013
- Shernoff, D. J. (2013). Models of engaging private schools and the case of Montessori schools. In Optimal learning environments to promote student engagement (pp. 219–246). Springer. 10.1007/978-1-4614-7089-2_10 [ DOI ] [ Google Scholar ]
SHIVAKUMARA 2016
- Shivakumara, K. , Dhiksha, J. , & Nagaraj, O. (2016). Efficacy of Montessori and traditional method of education on self‐concept development of children. International Journal of Educational Policy Research and Review, 3(2), 29–35. [ Google Scholar ]
SIMMONS 1983
- Simmons, C. H. , & Sands‐Dudelczyk, K. (1983). Children helping peers: Altruism and preschool environment. Journal of Psychology: Interdisciplinary and Applied,115(2), 203–207. [ Google Scholar ]
SIMONS 1984
- Simons, J. A. , & Simons, F. A. (1984). Montessori and regular preschools: A comparison . (ED247031). ERIC. http://eric.ed.gov/?id=ED247031
SOMORIN 2016
- Somorin, K. , & Nkeiruka, N. (2016). Differential performance between public and Montessori primary school pupils' in vocational studies: The role of Montessorial teaching in ensuring quality education. African Journal of Studies in Education, 11(2), 240–249. [ Google Scholar ]
STALLINGS 1987
- Stallings, J . (1987). Longitudinal findings for early childhood programs: Focus on direct instruction (ED297874). ERIC. http://eric.ed.gov/?id=ED297874
STEPHENS 1973
- Stephens, M. W. (1973). Effects of type of preschool experience and socioeconomic class on academic achievement motivation . Final report (ED116777). Purdue Research Foundation. http://eric.ed.gov/?id=ED116777
- Stern, C. (1968). Evaluating language curricula for preschool children. Monographs of the Society for Research in Child Development, 33(8), 49–61. [ PubMed ] [ Google Scholar ]
STEVENS 1976
- Stevens, J. H. Jr. (1976). Differential outcomes of early childhood education (ED140935). ERIC. http://eric.ed.gov/?id=ED140935
STODOLSKY 1972
- Stodolsky, S. S. , & Karlson, A. L. (1972). Differential outcomes of a Montessori curriculum. The Elementary School Journal, 72(8), 419–433. [ Google Scholar ]
SZOBIOVÁ 2014
- Szobiová, E. (2014). Child creativity in the context of education at standard and alternative schools in Slovakia. Acta Technologica Dubnicae, 4(2), 19–26. [ Google Scholar ]
TAMMINEN 1967
- Tamminen, A. W. , Weatherman, R. F. , & McKain, C. W. (1967). An evaluation of a preschool training program for culturally deprived children (ED019135). ERIC, U.S. Department of Health, Education, and Welfare. http://eric.ed.gov/?id=ED019135
- Toot, T . (2019). A comparison study: The impacts of Montessori and conventional elementary standards‐based language arts curricula on preschool students' phonemic awareness and reading readiness skills [Doctoral dissertation, University of Findlay]. http://rave.ohiolink.edu/etdc/view?acc_num=findlay1565215112222622
WEIKART 1981
- Weikart, D. P. (1981). Effects of different curricula in early childhood intervention. Educational Evaluation and Policy Analysis, 3(6), 25–35. [ Google Scholar ]
WEXLEY 1974
- Wexley, K. , John, G. , & Kehle, T. (1974). An evaluation of Montessori and day care programs for disadvantaged children. The Journal of Educational Research, 68(3), 95–99. [ Google Scholar ]
- White, J. M. , Yussen, S. R. , & Docherty, E. M. (1976). Performance of Montessori and traditionally schooled nursery children on tasks of seriation, classification, and conservation. Contemporary Educational Psychology, 1(4), 356–368. [ Google Scholar ]
- Witte, E. R. (2013). Individual experiences with Montessori and traditional classrooms: A qualitative comparison [Master's thesis, Kalamazoo College]. https://cache.kzoo.edu/handle/10920/29157
- Yen, S.‐C. , & Ispa, J. M. (2000). Children's temperament and behavior in Montessori and constructivist early childhood programs. Early Education & Development, 11(2), 171–186. [ Google Scholar ]
OTHER REFERENCES
Additional references, ackerman 2019.
- Ackerman, D. J. (2019). The Montessori preschool landscape in the United States: History, programmatic inputs, availability, and effects. ETS Research Report Series , 2019(1), 1–20. 10.1002/ets2.12252 [ DOI ] [ Google Scholar ]
- Ahmed, S. F. , Kuhfeld, M. W. , Davis‐Kean, T. W. , Vandell, P. E. , & Lowe, D. (2021). Preschool executive function and adult outcomes: A developmental cascade model. Developmental Psychology, 57(12), 2234–2249. [ DOI ] [ PubMed ] [ Google Scholar ]
AMABILE 1984
- Amabile, T. M. , & Judith, G. (1984). Children's artistic creativity: Effects of choice in task materials. Personality & Social Psychology Bulletin, 10(2), 209–215. [ Google Scholar ]
AMERICAN MONTESSORI SOCIETY 2016
- American Montessori Society . (2016). Private schools . Retrieved May 18, 2021, from https://amshq.org/School‐Resources/Private
BARNETT 2011
- Barnett, W. S. (2011). Effectiveness of early educational intervention. Science, 333(6045), 975–978. [ DOI ] [ PubMed ] [ Google Scholar ]
BASSOK 2016
- Bassok, D. L. , & Rorem, S. (2016). Is kindergarten the new first grade? AERA Open, 4, 1–21. 10.1177/2332858415616358 [ DOI ] [ Google Scholar ]
BAUMRIND 1989
- Baumrind, D. (1989). Rearing competent children. In Damon W. (Ed.), Child development today and tomorrow (pp. 349–378). Jossey‐Bass. [ Google Scholar ]
BENJAMINI 1995
- Benjamini, Y. , & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society. Series B (Methodological), 57(1), 300–289. [ Google Scholar ]
BENJAMINI 2005
- Benjamini, Y. , & Yekutieli, D. (2005). False discovery rate–adjusted multiple confidence intervals for selected parameters. Journal of the American Statistical Association, 100(469), 71–81. [ Google Scholar ]
BHATIA 2015
- Bhatia, P. , Alan, D. , & Ellen, S.‐B. (2015). Educational gymnastics: The effectiveness of Montessori practical life activities in developing fine motor skills in kindergartners. Early Education and Development, 26(4), 594–607. [ Google Scholar ]
BOEHNLEIN 1988
- Boehnlein, M. M. (1988). Montessori research: Analysis in retrospect. The NAMTA Journal, 13(3), 1–119. [ Google Scholar ]
BOEHNLEIN 1990
- Boehnlein, M. M. (1990). Research and evaluation summary of Montessori programs. In Kahn (Ed.) D., Implementing Montessori education in the public sector (pp. 476–483). North American Montessori Teachers' Association. [ Google Scholar ]
BOEHNLEIN 1996
- Boehnlein, M. M . (1996). Montessori research: Recent trends. In The relevance of Montessori today: Meeting human needs‐Principles to practice . Proceeding of the AMI/USA National Conference (pp. 71–83). ERIC. https://eric.ed.gov/?id=ED423953
BROOKS 1999
- Brooks, J. G. , & Brooks Martin, G. (1999). In search of understanding: The case for constructivist classrooms. Association for Supervision and Curriculum Development. [ Google Scholar ]
CAMPBELL COLLABORATION 2019A
- The Methods Group of the Campbell Collaboration . (2019a). Methodological expectations of Campbell Collaboration intervention reviews: Conduct standards (Campbell Policies and Guidelines Series No. 3). 10.4073/cpg.2016.3 [ DOI ]
CAMPBELL COLLABORATION 2019B
- The Methods Group of the Campbell Collaboration . (2019b). Methodological expectations of Campbell Collaboration intervention reviews: Reporting standards (Campbell Policies and Guidelines Series No. 4), p. 23. 10.4073/cpg.2016.4 [ DOI ]
- Cheng, A. , Collin, H. , Brian, K. , & Mills, J. N. (2017). “No excuses” charter schools: A meta‐analysis of the experimental evidence on student achievement. Journal of School Choice, 11(2), 209–238. [ Google Scholar ]
- Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Academic Press. [ Google Scholar ]
CONCANNON 1966
- Concannon, S. J. (1966). An experimental study of the influence of individual vs. group instruction on spatial abilities in preschool children (ERIC Report No. 80‐630). US Department of Health, Education, and Welfare.
CORDOVA 1996
- Cordova, D. I. , & Lepper, M. R. (1996). Intrinsic motivation and the process of learning: Beneficial effects of contextualization, personalization, and choice. Journal of Educational Psychology, 88(4), 715–730. [ Google Scholar ]
CULCLASURE 2019
- Culclasure, B. T. , Daoust, C. J. , Cote, S. M. , & Zoll, S. (2019). Designing a logic model to inform Montessori research. Journal of Montessori Research, 5(1), 35–74. https://journals.ku.edu/jmr/issue/view/1338 [ Google Scholar ]
DAOUST 2004
- Daoust, C. J . (2004). An examination of implementation practices in Montessori early childhood education [Doctoral dissertation, University of California ‐ Berkeley]. ProQuest Dissertations Publishing.
DAOUST 2018, APRIL
- Daoust, C. J. , & Murray, A. (2018, April). A comparison of elementary classroom practices based on Montessori certification [Poster presentation]. American Montessori Association Annual Conference, Denver, CO, United States. https://amshq.org/‐/media/Files/AMSHQ/Research/Conference‐Handouts/2018/A‐Comparison‐of‐Elementary‐Classroom‐Practices‐Based‐on‐Montessori‐Certification‐2018‐Conference.ashx
DAOUST 2019
- Daoust, C. J. , Murray, A. , & Chen, J. L. (2019). A reexamination of implementation practices in Montessori early childhood education . Paper presented at The Montessori Event, Washington, DC.
DE CHARMS 1976
- DeCharms, R. , & Shea, D. J. (1976). Enhancing motivation: Change in the classroom. Irvington Publishers. [ Google Scholar ]
- Debs, M. (2019). Diverse parents, desirable schools: Public Montessori in an era of school choice. Harvard Education. [ Google Scholar ]
- Diamond, A. , & Lee, K. (2011). How can we help children succeed in the 21st century? What the scientific evidence shows aids executive function development in children 4‐12 years of age. Science, 333, 959–964. 10.1126/science.1204529 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
- Dohrmann, K. R. N. , Gartner, T. K. , Lipsky, A. , Grimm, D. K. , & Kevin, J. (2007). High school outcomes for students in a public Montessori program. Journal of Research in Childhood Education, 22, 205–217. [ Google Scholar ]
DRISCOLL 2005
- Driscoll, M. (2005). Psychology of learning for instruction. Pearson. [ Google Scholar ]
DUNCAN 2014
- Duncan, G. J. , & Murnane, R. J. (2014). Restoring opportunity: The crisis of inequality and the challenge for American Education. Harvard Education Press. [ Google Scholar ]
- Duval, S. J. , & Tweedie, R. L. (2000). A nonparametric “trim and fill” method of accounting for publication bias in meta‐analysis. Journal of the American Statistical Association, 95(449), 89–98. [ Google Scholar ]
ELKIND 2003
- Elkind, D. (2003). Montessori and constructivism. Montessori Life, 15(1), 26–29. [ Google Scholar ]
FISHER 2017
- Fisher, Z. (2017). Robumeta : Robust variance metaregression [Computer program]. Version 2.0.
HIGGINS 2021
- Higgins, J. P. T. , Thomas, J. , Chandler, J. , Cumpston, M. , Li, T. , Page, M. J. , & Welch, V. A (Eds.). (2021). Cochrane handbook for systematic reviews of interventions version 6.2. Cochrane. https://training.cochrane.org/handbook [ Google Scholar ]
- Hill, C. J. , Bloom, H. S. , Black, A. R. , & Lipsey, M. W. (2008). Empirical benchmarks for interpreting effect sizes in research. Child Development Perspectives, 2(3), 172–177. [ Google Scholar ]
HYSLOP‐MARGISON 2008
- Hyslop‐Margison, E. J. , & Strobel, J. (2008). Constructivism and education: Misunderstandings and pedagogical implications. Teacher Educator, 43(1), 72–86. [ Google Scholar ]
IBM CORP. 2020
- IBM . SPSS statistics for Windows. (2020). Version 27.0 [Computer program]. Version 27.0. IBM Corp. [ Google Scholar ]
- Jacob, R. , & Julia, P. (2015). The potential for school‐based interventions that target executive function to improve academic achievement: A review. Review of Educational Research, 85(4), 512–552. [ Google Scholar ]
- Jones, A . (2005). Montessori education in America: An analysis of research conducted from 2000–2005 . Retrieved March 17, 2014, from http://www.plan4preschool.org/documents/montessori‐research‐00‐05.pdf
- Jones, D. E. G. , & Marky, C. M. (2015). Early social‐emotional functioning and public health: The relationship between kindergarten social competence and future wellness. American Journal of Public Health, 105(11), 1541–1548. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
KASSAI 2019
- Kassai, R. , Futo, J. , Demetrovics, Z. , & Takacs, Z. K. (2019). A meta‐analysis of the experimental evidence on the near‐and far‐transfer effects among children's executive function skills. Psychological Bulletin, 145(2), 165–188. [ DOI ] [ PubMed ] [ Google Scholar ]
- Klahr, D. , & Nigam, M. (2004). The equivalence of learning paths in early science instruction: Effects of direct instruction and discovery learning. Psychological Science, 15(10), 661–667. [ DOI ] [ PubMed ] [ Google Scholar ]
KONSTANTOPOULOS 2011
- Konstantopoulos, S. (2011). Fixed effects and variance components estimation in three‐level meta‐analysis. Research Synthesis Methods, 2(1), 61–76. [ DOI ] [ PubMed ] [ Google Scholar ]
KOSSMEIER 2020
- Kossmeier, M., Tran, U. S., & Voracek, M. (2020). Metaviz Forest plots funnel plots and visual funnel plot Inference for metaanalysis [Computer program]. Version 0.3.1.
- Kraft, M. A. (2020). Interpreting effect sizes of education interventions. Educational Researcher, 49(4), 241–253. [ Google Scholar ]
- Laski, E. V. J. 'dan , Daoust, J. R. , Carolyn, M. , & Angela, K. (2015). What makes mathematics manipulatives effective? Lessons from cognitive science and Montessori education. SAGE Open, 5, 1–8. 10.1177/2158244015589588 [ DOI ] [ Google Scholar ]
LEBOEUF 2022
- Snyder, A. , LeBoeuf, L. , & Lillard, A. S. (2022). “My name is Sally Brown, and I hate school!”: A retrospective study of school liking among conventional and Montessori school alumni. Psychology in the Schools, 60, 541–565. 10.1002/pits.22777 [ DOI ] [ Google Scholar ]
LEYLAND 2019
- Leyland, A. , Georgina, R. , & Lisa‐Marie, E. (2019). Experimental effects of mindfulness inductions on self‐regulation: Systematic review and meta‐analysis. Emotion, 19(1), 108–122. [ DOI ] [ PubMed ] [ Google Scholar ]
- Lillard, A. S. , & Heise, M. J. (2016). Removing supplementary materials from Montessori classrooms changed child outcomes. Journal of Montessori Research, 2(1), 17–27. [ Google Scholar ]
- Lillard, A. S. (2017). Montessori: The science behind the genius (3rd ed.). Oxford. [ Google Scholar ]
LILLARD 2019
- Lillard, A. S. S. , & Montessori, A. (2019). Self‐Determination, and a case for radical school reform. Educational Psychology Review, 31(4), 939–965. [ Google Scholar ]
LILLARD 2019A
- Lillard, A. S. , & McHugh, V. (2019). Authentic Montessori: The dottoressa's view at the end of her life part I: The environment. Journal of Montessori Research, 5(1), 1–18. [ Google Scholar ]
LILLARD 2019B
- Lillard, A. S. , & McHugh, V. (2019). Authentic Montessori: The dottoressa's view at the end of her life part II: The teacher and the child. Journal of Montessori Research, 5(1), 19–34. [ Google Scholar ]
LIPSEY 2012
- Lipsey, M. W. , Puzio, K. , Yun, C. , Hebert, M. A. , Steinka‐Fry, K. , Cole, M. W. , Roberts, M. , Anthony, K. S. , & Busick, M. D. (2012). Translating the statistical representation of the effects of education interventions into more readily interpretable forms . National Center for Special Education Research.
LIPSEY 2018
- Lipsey, M. W. , Farran, D. C. , & Durkin, K. (2018). Effects of the Tennessee prekindergarten program on children's achievement and behavior through third grade. Early Childhood Research Quarterly, 45, 155–176. [ Google Scholar ]
LOEFFLER 1992
- Loeffler, M. H. (1992). Montessori and constructivism. In Loeffler M. H., Montessori in contemporary American culture (pp. 101–113). Heinemann. [ Google Scholar ]
MARSHALL 2017
- Marshall, C. (2017). Montessori education: A review of the evidence base. Science of Learning , 2(11). 10.1038/s41539-017-0012-7 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
- Mayer, R. E. (2004). Should there be a three‐strikes rule against pure discovery learning? American Psychologist, 59(1), 14–19. [ DOI ] [ PubMed ] [ Google Scholar ]
MAYNARD 2017
- Maynard, B. R. , Solis, M. R. , Miller, V. L. , & Brendel, K. E. (2017). Mindfulness‐based interventions for improving cognition, academic achievement, behavior, and socioemotional functioning of primary and secondary school students. Campbell Systematic Reviews, 13, 1–144. 10.4073/CSR.2017.5 [ DOI ] [ Google Scholar ]
MCGUINNESS 2020
- McGuinness, L. A. , & Higgins, J. P. T. (2020). Risk‐of‐bias VISualization (robvis): An R package and shiny web app for visualizing risk‐of‐bias assessments. Research Synthesis Methods, 12(1), 55–61. 10.1002/jrsm.1411 [ DOI ] [ PubMed ] [ Google Scholar ]
- Meece, J. L. (2002). Child and adolescent development for educators. McGraw‐Hill. [ Google Scholar ]
- Miller, L. B. , & Dyer, J. L. (1975). Four preschool programs: their dimensions and effects. Monographs of the Society for Research in Child Development , 40(5–6, Serial No. 162). [ Google Scholar ]
MOFFITT 2011
- Moffitt, T. E. , Arseneault, L. , Belsky, D. , Dickson, N. , Hancox, R. J. , Harrington, H. L. , Houts, R. , Poulton, R. , Roberts, B. W. , & Ross, S. (2011). A gradient of childhood self‐control predicts health, wealth, and public safety. Proceedings of the National Academy of Sciences of the United States of America, 108(7), 2693–2698. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
MONTESSORI 1964
- Montessori, M. (1964). The Montessori method. Shocken. [ Google Scholar ]
MONTESSORI 2012
- Montessori, M. (2012). The 1946 London lectures. Montessori‐Pierson Publishing. [ Google Scholar ]
MONTESSORI 2017
- Montessori, M. (2017). The discovery of the child (2nd ed.). Pierson. [ Google Scholar ]
MORRIS 2002
- Morris, S. B. , & DeShon, R. P. (2002). Combining effect size estimates in meta‐analysis with repeated measures and independent‐groups designs. Psychological Methods, 7(1), 105–125. [ DOI ] [ PubMed ] [ Google Scholar ]
- Murad, M. H. , Haitao, C. , Lifeng, L. , & Zhen, W. (2018). The effect of publication bias magnitude and direction on the certainty in evidence. BMJ Evidence‐Based Medicine, 23(3), 84. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
MURRAY 2010
- Murray, A. (2010). Overview of research on Montessori education: An evidence‐based curriculum [white paper]. American Montessori Society. https://amshq.org/‐/media/Files/AMSHQ/Research/White‐Papers/Overview‐of‐Research‐on‐ME.ashx
- Olkin, I. , Dahabreh, I. J. , & Trikalinos, T. A. (2012). GOSH—A graphical display of study heterogeneity. Research Synthesis Methods, 3(3), 214–223. [ DOI ] [ PubMed ] [ Google Scholar ]
PIAGET 1969
- Piaget, J. , Bärbel, I. , & Helen, W. (1969). The psychology of the child. Basic Books. [ Google Scholar ]
POPPER 1962/2014
- Popper, K. (1962/2014). Conjectures and refutations: The growth of scientific knowledge . Routledge. [ Google Scholar ]
PUSTEJOVSKY 2020
- Pustejovsky, J . (2020). ClubSandwich [Computer program]. Version 0.5.3.
R CORE TEAM 2019
- R Core Team . (2019). R: A language and environment for statistical computing . https://www.R-project.org
RAMBUSCH 1992
- Rambusch, N. (1992). Montessori in America: A history. In Loeffler M. H., Montessori in contemporary American culture (pp. 7–16). Heinneman. [ Google Scholar ]
RANDOLPH 2016
- Randolph, J. J. , Anaya, B. , Lakshmi, M. , Michaels, S. , Walls, R. D. L. , & Warren, M. P. (2016). PROTOCOL: Montessori education for improving academic and social/behavioral outcomes for elementary students. Campbell Systematic Reviews, 12(1), 1–32. [ Google Scholar ]
RANDOLPH 2019
- Randolph, J. J. (2019). Quantitative analysis. In Burkholder G. J., Cox K. A., Crawford L. M., & Hitchcock J. H. (Eds.), Research design and methods: An applied guide for the scholar‐practitioner. Sage. [ Google Scholar ]
RANDOLPH 2021
- Randolph, J. J. , Bryson, A. , Henderson, D. , Menon, L. , Kureethara Manuel, A. , Michaels, S. , McPherson, W. , rosenstein, dl , O'Grady, R. , & Lillard, A. (2021, July). Supplemental Information for Campbell Systematic Review of the Montessori Research . https://drive.google.com/drive/folders/1j4YaM5qq1LE4nN7XsfVAmPLfdNF6rYfE?usp=sharing [ DOI ] [ PMC free article ] [ PubMed ]
- Rinke, C. R. , Gimbel, S. J. , & Sophie, H. (2013). Opportunities for inquiry science in Montessori classrooms: Learning from a culture of interest, communication, and explanation. Research in Science Education, 43(4), 1517–1533. [ Google Scholar ]
ROSENTHAL 2002
- Rosenthal, R. , & DiMatteo, M. R. (2002). Meta‐analysis: Recent developments in quantitative methods for literature reviews. Annual Review of Psychology, 52, 59–82. [ DOI ] [ PubMed ] [ Google Scholar ]
- Ryan, R. M. , & Grolnick, W. S. (1986). Origins and pawns in the classroom: Self‐report and projective assessments of individual differences in children's perceptions. Journal of Personality & Social Psychology, 50(3), 550–558. [ Google Scholar ]
- Ryan, R. M. , & Deci, E. L. (2000). Self‐determination theory and the facilitation of intrinsic motivation, social development, and well‐being. American Psychologist, 55(1), 68–78. [ DOI ] [ PubMed ] [ Google Scholar ]
SAMEROFF 1996
- Sameroff, A. J. , & Haith, M. M. (Eds.). (1996). The five to seven year shift: The age of reason and responsibility . University of Chicago Press.
SCHUNK 2012
- Schunk, D. H. (2012). Learning theories: An educational perspective. Pearson. [ Google Scholar ]
SHADISH 2002
- Shadish, W. R. , Cook, T. D. , & Campbell, D. T. (2002). Experimental and quasi‐experimental designs for generalized causal inference. Houghton, Mifflin and Company. [ Google Scholar ]
- Shi, L. , & Lifeng, L. (2019). The trim‐and‐fill method for publication bias: Practical guidelines and recommendations based on a large database of meta‐analyses. Medicine, 98(23):e15987. [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
STERNE 2016
- Sterne, J. A. , Hernán, M. A. , Reeves, B. C. , Savović, J. , Berkman, N. D. , Viswanathan, M. , Henry, D. , Altman, D. G. , Ansari, M. T. , Boutron, I. , Carpenter, J. R. , Chan, A.‐W. , Churchill, R. , Deeks, J. J. , Hróbjartsson, A. , Kirkham, J. , Jüni, P. , Loke, Y. K. , Pigott, T. D. , … Higgins, J. P. (2016). ROBINS‐I: A tool for assessing risk of bias in non‐randomised studies of interventions. BMJ, 355(i4919). 10.1136/bmj.i4919 [ DOI ] [ PMC free article ] [ PubMed ] [ Google Scholar ]
STERNE 2019
- Sterne, J. A. C. , Savović, J. , Page, M. J. , Elbers, R. G. , Blencowe, N. S. , Boutron, I. , Cates, C. J. , Cheng, H.-Y. , Corbett, M. S. , Eldridge, S. M. , Hernán, M. A. , Hopewell, S. , Hróbjartsson, A. , Junqueira, D. R. , Jüni, P. , Kirkham, J. J. , Lasserson, T. , Li, T. , McAleenan, A. , … Higgins, J. P. T. (2019). RoB 2: A revised tool for assessing risk of bias in randomised trials. BMJ, 366, l4898. [ DOI ] [ PubMed ] [ Google Scholar ]
TANNER‐SMITH 2014
- Tanner-Smith, E. E. , & Tipton, E. (2014). Robust variance estimation with dependent effect sizes: Practical considerations including a software tutorial in stata and SPSS. Research Synthesis Methods, 5(1), 13–30. [ DOI ] [ PubMed ] [ Google Scholar ]
TANNER‐SMITH 2016
- Tanner-Smith, E. E. , Tipton, E. , & Polanin, J. R. (2016). Handling complex meta-analytic data structures using robust variance estimates: A tutorial in R. Journal of Developmental and Life Course Criminology, 2(1), 85–112. [ Google Scholar ]
TIPTON 2015
- Tipton, E. (2015). Small sample adjustments for robust variance estimation with meta-regression. Psychological Methods, 20(3), 375–393. [ DOI ] [ PubMed ] [ Google Scholar ]
TOPPING 2005
- Topping, K. J. (2005). Trends in peer learning. Educational Psychology, 25, 631–645. 10.1080/01443410500345172 [ DOI ] [ Google Scholar ]
TRABALZINI 2011
- Trabalzini, P. (2011). Maria Montessori through the seasons of the “method”. North American Montessori Teachers' Association. [ Google Scholar ]
TURNER 1992
- Turner, J. (1992). Montessori's writings versus Montessori practices. In Loeffler M. H. (Ed.), Montessori in contemporary American culture (pp. 17–44). Heinemann. [ Google Scholar ]
TUTTLE 2012
- Tuttle, C. C. , Philip, G. , & Melissa, C. (2012). Using lotteries to evaluate schools of choice: Evidence from a national study of charter schools. Economics of Education Review, 31(2), 237–253. [ Google Scholar ]
ULTANIR 2012
- Ultanir, E. (2012). An epistemological glance at the constructivist approach: Constructivist learning in Dewey, Piaget, and Montessori. International Journal of Instruction, 5(2), 195–212. [ Google Scholar ]
VIECHTBAUER 2010
- Viechtbauer, W. (2010). Conducting meta‐analyses in R with the metafor package. Journal of Statistical Software, 36(3), 1–48. [ Google Scholar ]
WASSERSTEIN 2019
- Wasserstein, R. L. , Schirm, A. L. , & Lazar, N. A. (2019). Moving to a world beyond “p < 0.05”. The American Statistician, 73(Suppl. 1), 1–19. [ Google Scholar ]
WEILAND 2020
- Weiland, C. , Rebecca, U. , Anna, S. , Sara, S. , Shana, R. , & Eleanor, M. (2020). The effects of enrolling in oversubscribed prekindergarten programs through third grade. Child Development, 91(5), 1401–1422. [ DOI ] [ PubMed ] [ Google Scholar ]
WHAT WORKS CLEARINGHOUSE 2014
- What Works Clearinghouse. (2014). What Works Clearinghouse procedures and standards handbook, version 3.0. U.S. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assistance (NCEE). https://ies.ed.gov/ncee/wwc/Docs/referenceresources/wwc_procedures_v3_0_standards_handbook.pdf
WHITESCARVER 2007
- Whitescarver, K. , & Jacqueline, C. (2007). Lessons from the periphery: The role of dispositions in Montessori teacher training. Journal of Educational Controversy, 2(2), 1–12. [ Google Scholar ]
WILSON N.D.
- Wilson, D. B. (n.d.). Practical meta‐analysis effect size calculator . https://www.campbellcollaboration.org/escalc/html/EffectSizeCalculator-SMD-main.php
ZOOGMAN 2015
- Zoogman, S. , Goldberg, S. B. , Hoyt, W. T. , & Miller, L. (2015). Mindfulness interventions with youth: A meta‐analysis. Mindfulness, 6(2), 290–302. [ Google Scholar ]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
- View on publisher site
- PDF (10.6 MB)
- Collections
Similar articles
Cited by other articles, links to ncbi databases.
- Download .nbib .nbib
- Format: AMA APA MLA NLM
IMAGES
VIDEO
COMMENTS
Abstract. The rapid development of the times has also brought more advanced. educational methods and concepts. Many experts advocate replacing traditional. educational methods with modern e ...
American Research Journal of English and Literature Original Article. ISSN 2378-9026 Volume 1, Issue 2, April-2015. www.arjonline.org 1. Education: Traditional Vs. Modern Perspective. Sofe Ahmed 1 ...
The dynamic landscape of education has witnessed a profound shift from traditional to modern pedagogical paradigms over the years. The discussion of results delves into the intriguing debate between traditional and modern educational systems (TES and MES), examining them through the lens of a value-based approach. This exploration is crucial in understanding how these two approaches shape the ...
The research adopted such methods as descriptive, comparative, historical, typological, and statistical for data processing of International Migrant Stock (2020) and National Center for Education ...
Jones reviewed Montessori education outcomes research in three areas: (1) the effects of the Montessori approach to at‐risk students, (2) the effects of the Montessori method on exceptional learners including learning disabled, developmentally delayed, and gifted/talented, and (3) comparative analyses of traditional schooling versus ...
The need for additional, well-designed research on the subject is clear. The present study advances the knowledge base regarding achievement comparisons between Montessori and traditional programs by providing both an examination of former research over the history of the Method, as well as a design which includes controls for some of
1. Topic The literature must relate directly to traditional and inquiry-based learning. 2. Period Studies published between 2002 and 2017, the date for the last meta-analysis related to the subjects under investigation. 3. Research Base Literature should include only empirical studies of both quantitative and qualitative. 4.
learn more in online vs. traditional classes (Allen & Seaman, 2013). Whether or not the perceptions of higher education officials reflect reality is subject to a review of the quantitative work comparing grades achieved in online vs. traditional classes. Previous research that rigorously compares student
Limitations of recent developments in education and schooling - Looking back. Critiques of the industrial model of education and schooling are not new, considering, for example, Robinson's (Citation 2015) work, and others have critiqued how curricula are constructed.Pinar (Citation 2011, Citation 2023), for example, advocates to move from a traditional static, to a more flexible and ...
While modern education harnesses the power of technological advancements to deliver personalized and interactive learning experiences tailored to the diverse needs of students, traditional education often falls short in adapting to individual learning styles, relying heavily on conformity and rote learning.