5. What are the strengths and weaknesses of students in Türkiye?

Analyses of student performance in Chapter 2 revealed substantial improvements over time in mathematics and science in International Association for the Evaluation of Educational Achievement (IEA) Trends in International Mathematics and Science Study (TIMSS) Grades 41 and 8, and in reading, mathematics and science in the OECD Programme for International Student Assessment (PISA). This improvement was driven by a significant decline in the share of low performers in all grades, subjects and in both assessments. Moreover, this improvement brought Türkiye’s performance closer to the OECD average in all subjects of assessment. However, those findings only show the average level of performance and the trends over time. This chapter takes a deeper look, examining student performance in sub-domains of the main subjects of assessment and individual test items.

This chapter relies on an in-depth analysis of performance in different domains and sub-domains of mathematics, science and reading and on success rates on individual test items for those domains and sub-domains. The analysis uses data from TIMSS 2015 and TIMSS 2019 (Grades 4 and 8) (Box 5.2) and from the PISA 2018 international surveys. Note that “domain” and “subject” of assessment are used interchangeably. The domains (and sub-domains) of assessment across TIMSS 2015 Grades 4 and 8 are set out in Table 5.1 and those for TIMSS 2019 in Table 5.2.

In PISA, the analysis in this chapter focuses primarily on reading, since this was the major domain in 2018 (see Chapter 1). The PISA 2018 reading framework is based on the processes of locating information, understanding and evaluating and reflecting across single or multiple-source texts (Table 5.3). For further information about each of these processes, please see PISA 2018 Results (Volume 1): What Students Know and Can Do (OECD, 2019[7]).

This chapter relies on two types of analyses: analysis of student performance on sub-domains of assessment and analyses of student performance on individual test items, which are then mapped to different sub-domains and item characteristics.

Both the TIMSS and PISA assessments assess student performance in a number of sub-domains of the three main subjects: reading, mathematics and science (PISA) and mathematics and science (TIMSS) (see Chapter 1). While student performance across sub-domains tends to be strongly correlated, variation from a country’s average performance in one subject indicates a relative strength or weakness in the country’s performance (e.g. the analysis in TIMSS and PISA might identify a relative strength in geometry because, in comparison with the national average in mathematics, performance in this sub-domain is higher than expected).

Analysing the characteristics of the individual test items reveals if performance is associated with certain item characteristics, such as an item’s content and cognitive sub-domain, its format (multiple-choice or open-ended) and its difficulty. The success rate of each item for students in Türkiye is compared to the average success rate for all participating OECD countries. Conspicuous items, indicating strengths or weaknesses, are those on which students in Türkiye performed particularly well or less well in comparison with the OECD average (Box 5.3).

The analyses are carried out for mathematics and science in Grades 4 and 8 in TIMSS 2015 and 2019 and for mathematics, science and reading in PISA 2018. Strengths or weaknesses identified through those two methods might shed light on different emphasises in teaching or curricula.

Comparing student performance in the different sub-domains in TIMSS 2015 Grade 4 to the overall average for Türkiye:

  • In mathematics, geometry and reasoning are identified as areas of relative weakness.

  • No sub-domain of mathematics was identified as an area of strength (Table 5.4).

  • In science, physical science is an area of strength and life science is a relative weakness (see Table 5.5).

In 2019, students in Türkiye performed 40 points in mathematics and 43 in science higher compared with the country’s performance in 2015. While students performed higher overall on average, the analysis here presents areas of relative weakness or strength in comparison to Türkiye’s overall performance. A country with high average scores may still have areas of relative strength or weakness because students in that country do not perform uniformly across the assessed content.

Comparing student performance in the different sub-domains in TIMSS 2019 Grade 4 to the overall average for Türkiye:

  • In mathematics, data and reasoning are identified as areas of relative weakness (Table 5.6). Compared to TIMSS 2015 Grade 4 , the data sub-domain has become an area of relative weakness while performance in geometry is no longer a weakness and now exceeds the mean performance in mathematics. As in 2015, reasoning in mathematics remains an area of relative weakness.

  • In science, in contrast with 2015, students performed consistently across all assessed areas with no domain emerging as a strength or a weakness (Table 5.7).

In TIMSS 2015, Grade 4 students answered 178 test items in mathematics and 200 in science. The conspicuous items (Box 5.3) in mathematics included 29 test items in which students did relatively well in comparison with the OECD average and 26 in which they performed relatively lower. In science, students did relatively well in 29 test items and relatively lower in 28.

Below, the attributes of those items are discussed with the aim of identifying patterns that characterise the performance of students in Türkiye. It is worth noting that sub-domain analysis identifies whether students are doing better in a particular sub-domain relative to the national average in that domain. In contrast, item analysis compares success rates on individual test items to average success rates of a group of countries – in this case, OECD countries. Hence, the former is based on national comparison within a country of reference while the latter is based on an international comparison between the country of reference and other countries.

In mathematics, the 29 items (out of 178) in which students did particularly well comprise (Figure 5.1):

  • Nineteen items out of 96 covering the content sub-domain of number (representing 20% of all items covering the number sub-domain).

  • Nineteen items out of 74 (representing 26% of items) covering the knowing sub-domain.

In mathematics, the 26 items (out of 178) in which students did particularly less well comprise (Figure 5.1):

  • Nine items out of 96 covering the content sub-domain of geometry (representing 16% of geometry items).

  • Seven items out of 32 (representing 22% of items) covering the reasoning sub-domain.

The test items for the content and cognitive domains “number” and “knowing” in which students did well were relatively easy (i.e. had a level of difficulty lower than the average for their domain). In contrast, the test items in which students did well in geometry had a level of difficulty close to the average for all items in geometry, and those for the cognitive domain of applying were more difficult than the average. The two cognitive functions of applying and reasoning where students were less strong require a higher level of cognitive skills than the knowing sub-domain. Knowing entails familiarity with mathematical concepts while reasoning and applying require logical and systematic thinking and the application of mathematics in a variety of contexts (Vincent-Lancrin et al., 2019[11]).

Students in Türkiye also performed better in multiple-choice questions than open-ended questions. Open-ended questions have been used in international assessments, such as PISA, to assess higher-order skills such as applying knowledge and reasoning. For those questions, students often have to explain their answers and provide solid arguments to support those answers. At the time of the 2019 OECD report OECD Reviews of Evaluation and Assessment in Education: Student Assessment in Türkiye, teaching and testing in Türkiye made frequent use of multiple-choice tests so students may have had less exposure to open-ended questions than their peers in other countries (Kitchen et al., 2019[12]).

In general, the findings may reflect possible variations in curriculum coverage in terms of content domains but also in terms of coverage of knowledge and skills. For instance, students in Türkiye have a relative strength in knowledge about mathematics but weakness when it comes to skills such as mathematical reasoning.

In science, students in Türkiye did well in (Figure 5.2):

  • Physical science (13 out of 77 items, representing 17% of all items covering this domain).

  • Reasoning (6 out of 44 items, representing 14% of all items in this category).

In terms of difficulty, students did well in items that are more difficult than the average of all science items. This is in line with the finding that students in Türkiye perform particularly well in science, which is a national strength, and suggests that teaching and learning support students in mastering even difficult items and content.

  • The items representing weakness are dominated by those covering life science (18 out of 80, representing 23%) (Figure 5.2).

  • About 16% of items (11 out of 70) covering the applying sub-domain were found to represent an area of weakness.

As in mathematics, students did particularly well in multiple-choice science items and less well in open-ended questions.

In the computer-based assessment of TIMSS 2019, Grade 4 students answered 262 test items in mathematics and 240 in science. The conspicuous items in mathematics included 34 test items in which students in Türkiye did relatively well in comparison with the OECD average and 37 in which they performed relatively lower. In science, students did relatively well in comparison with the OECD average in 35 test items and relatively less well in 38.

In mathematics, the 34 items (out of 262) in which students did particularly well comprise (Figure 5.3):

  • Twenty-one items out of 118 covering the content sub-domain of number (representing 18% of all items covering the number sub-domain).

  • Fifteen items out of 110 (representing 14% of items) covering the applying sub-domain.

In mathematics, the 37 items (out of 262) in which students did particularly less well comprise (Figure 5.3):

  • Seven items out of 65 covering the content sub-domain of data (representing 11% of data items).

  • Twenty-four out of 103 (representing 23% of items) covering the knowing sub-domain.

  • Seven items out of 49 (representing 14% of items) covering the reasoning sub-domain.

In mathematics, the number sub-domain was a strength in both TIMSS 2015 and 2019, which might suggest that curriculum coverage, as well as teaching, is particularly strong in this sub-domain across both Grade 4 – at the end of primary education – and Grade 5 – at the start of lower secondary education. The applying sub-domain – which requires students to be able to apply mathematical skills across a variety of contexts – also became a strength in 2019. The change in student performance in the application of mathematical skills between 2015 and 2019 might reflect recent curricula reform, which places greater emphasis on applying knowledge and skills to solve problems (Box 5.4). In terms of difficulty, the items where students performed well in mathematics were more difficult than average, suggesting that students are well prepared to tackle some complex content.

In the cognitive sub-domains, both the 2015 and 2019 analyses identified reasoning as an area of weakness. Reasoning in mathematics requires high-level cognitive skills as individuals must go beyond solving routine problems to integrating knowledge and skills to solve multistep problems in unfamiliar, complex situations (IEA, 2017[3]). The items where students performed less well in data, knowing and reasoning had a higher level of difficulty than the questions on average while students tended to manage easier questions in these sub-domains well.

In science, students in Türkiye did well in (Figure 5.4):

  • Physical science (15 out of 82 items, representing 18% of all items covering this domain).

  • The items representing a weakness are dominated by those covering life science (25 out of 103, representing 24%) (Figure 5.4).

  • About 13% of items covering reasoning (6 out of 45) were found to represent an area of weakness.

Across 2015 and 2019 in science, students in Türkiye performed consistently well in physical science, while the life science sub-domain emerges as a weakness. As in mathematics, the items where students performed well were more difficult than average, again suggesting that students have the knowledge and skills to do well when faced with difficult questions. In contrast, the life science items where students did less well had a lower difficulty than average. As students do not do well even in comparatively easier life science items, their performance might reflect the balance of content in Türkiye’s curriculum with less time devoted to life science. In contrast, the strength of student performance in physical science might suggest that it is a domain that receives greater emphasis. Students’ high performance in physical science is also facilitated by their strength in number skills and knowledge since the two content domains are highly related (Rossdy et al., 2019[15]).

Also, in line with the results in mathematics, there was an improvement in the applying sub-domain in science which was a weakness in 2015. Applying in science requires high-level skills with individuals using their knowledge to compare, contrast and classify groups of objects or materials, relating knowledge of a science concept to a specific context, generating explanations and solving practical problems (IEA, 2017[3]).

Similarly, in line with the performance across different mathematics sub-domains, reasoning in science also emerges as a weakness in 2019. Reasoning in science refers to using evidence and science understanding to analyse, synthesise and generalise, often in unfamiliar situations and complex contexts (IEA, 2017[3]). Students also did not perform well even in easier items, suggesting that they might not be exposed to many activities or content that develops reasoning skills. The identification of reasoning as a weakness in both science and mathematics in 2019 also suggests that the pattern of performance might reflect emphasis in the curriculum and pedagogical approaches.

In line with student performance in Grade 4, physics was a relative strength for students in Türkiye while earth science (which shares some similar content areas with the life science sub-domain in Grade 4, which was a weakness) was a relative weakness (Table 5.9). In mathematics, however, in contrast to student performance in Grade 4, reasoning in mathematics was a relative strength in Grade 8 (Table 5.8). All other differences were not found to be statistically significant.

In line with student performance in Grade 8 in 2015, reasoning in mathematics continues to be an area of relative strength (Table 5.10). In both TIMSS 2015 and 2019, reasoning was a weakness in Grade 4 but became a strength in Grade 8, perhaps suggesting that the curriculum in Türkiye focuses on developing this cognitive domain in lower secondary education.

In line with the pattern that emerged in science in Grade 4 in 2019, performance across all content domains and cognitive domains in Grade 8 was consistent, with no domain being statistically different from the mean (Table 5.11).

In TIMSS 2015, Grade 8 students answered 215 test items in mathematics and 260 in science. The conspicuous items in mathematics included 30 test items in which students did particularly well in comparison with the OECD average and 32 in which they did poorly. In science, students did particularly well in 31 test items and poorly in 32.

Students performed relatively equally, doing well (and less well) across a number of sub-domains. However, items in which students did particularly well included (Figure 5.5):

  • Nine out of 43 items covering geometry (representing 21% of all items in this sub-domain).

  • Seven out of 41 items covering data and chance (representing 17% of all items in this sub-domain).

  • Nine out of 63 items covering algebra (representing 14% of all items in this sub-domain).

  • Eight out of 46 items covering reasoning (representing 17% of all items in this sub-domain).

  • Eighteen out of 68 items covering the number sub-domain (representing 26% of all items in this sub-domain) (Figure 5.5).

  • Fourteen out of 74 items covering the knowing sub-domain (representing 19% of all items in this sub-domain).

As with mathematics and science in Grade 4, students did better in multiple-choice questions and less well in open-ended questions. Moreover, average test item difficulty did not vary much between items representing strengths and those representing weaknesses. The two exceptions were knowing and multiple-choice questions, where the items representing a strength were more difficult than those representing a weakness. Variations in difficulty between items representing strengths and weaknesses could reflect exposure to certain topics in the curriculum, certain teaching practices emphasising particular cognitive skills, or certain assessment tools used more frequently in classrooms in Türkiye, such as multiple-choice questions.

In comparison with Grade 4, there was a reversal in the areas of strength and weakness in Grade 8. Geometry and reasoning, which were areas of weakness in Grade 4, became areas of strength in Grade 8, and the reverse is true for number and knowing. Data from PISA and TIMSS show that students in Türkiye tend to perform less well in mathematics compared with the other domains of reading and science. While the reasons for this cannot be identified through the data alone, this might reflect a less consistent approach across the curriculum and teaching in mathematics in the country. The performance in mathematics contrasts notably with science, where there are clear consistencies in the patterns of strength and weakness across grades and domains in Türkiye.

  • Twelve conspicuous physics items out of 62 (representing 19% of all physics items) fell into the group of items representing strengths (Figure 5.6)

  • Eleven out of 56 items covering earth science (representing 20% of all items in this sub-domain) fell into the group of items representing weaknesses (Figure 5.6).

  • Thirteen out of 101 items covering the cognitive domain of applying (representing 13% of those items) also fell into this group.

The test items covering physics in which students did particularly well were more difficult than those representing weaknesses. This seems counterintuitive and could be for a number of reasons. One might be because students make less effort in items that they perceive to be easier, another might be differences in emphasis in the physics curriculum or approaches to teaching and learning in physics in Türkiye. For the rest of the content and cognitive domains the average difficulty did not vary hugely between items representing strengths and those representing weaknesses. In comparison with Grade 4, physics remains an area of strength while applying scientific knowledge remains an area of weakness.

Items in which students did particularly well included (Figure 5.7):

  • Ten out of 74 items covering algebra (representing 14% of all items in this sub-domain).

  • Twenty out of 93 items covering knowing (representing 22% of all items in this sub-domain).

  • Six out of 53 items covering reasoning (representing 11% of all items in this sub-domain).

Items in which students did less well included (Figure 5.7):

  • Thirteen out of 60 items covering geometry (representing 22% of all items in this sub-domain).

  • Eight out of 59 items covering data and probability (representing 14% of all items in this sub-domain).

  • Twenty-two out of 134 items covering applying (representing 16% of all items in this sub-domain).

From both the analysis of 2015 and 2019, algebra emerges as a strong area of content in Türkiye in Grade 8. The consistency of algebra as a strength might reflect the focus of the curriculum in Türkiye. However, aside from this, there is no clear pattern in the strengths and weaknesses in mathematics over 2015 and 2019 or in comparison with Grade 4. This might reflect the impact of curriculum changes or more broadly a less consistent approach to mathematics teaching across schooling.

In Grade 8 mathematics, students performed better in multiple-choice questions. This is consistent with the results from 2015 and might be related to the frequent use of multiple-choice questions in classroom tests at school in Türkiye (Kitchen et al., 2019[12]). Regarding difficulty, in mathematics, students did well in items that were more difficult on average compared to other items, suggesting that they master these areas, such as number in Grade 4 and algebra in Grade 8. At the same time, the items in which students performed less well were more difficult on average, indicating that students are still able to respond well to questions covering the basis of geometry, data and probability and applying.

Items in which students did particularly well included (Figure 5.8):

  • Sixteen out of 63 items covering chemistry (representing 25% of all items in this sub-domain).

Items in which students did less well included (Figure 5.8):

  • Twenty-two out of 110 items covering biology (representing 20% of all items in this sub-domain).

  • Eight out of 61 items covering earth science (representing 13% of all items in this sub-domain).

  • Twenty-six out of 124 items covering knowing (representing 21% of all items in this sub-domain).

In Grade 8 in 2019, students performed particularly well in chemistry. Students in Türkiye are strong in both algebra and chemistry, which reflects the shared skills that underpin these subjects – chemistry uses algebra to express relationships between quantities and substances (Cunningham and Whelan, 2014[16]). In line with Grade 4 in 2019, students improved in the domain of applying and this is might be related to the new curriculum implemented in 2018 (Box 5.4). In science, earth science was an area of weakness in both 2015 and 2019, which might point to this sub-domain having less coverage in Türkiye’s curriculum.

In science, students performed well in items that were relatively easy compared to the other items in the same domains showing that even in the weaker ones – biology, earth science and the cognitive domains of knowing items – students are able to respond to easier questions covering the basics. However, students struggled with more difficult items, which might suggest that the curriculum or pedagogy does not cover or prepare students for topics that are more difficult. Only in chemistry, students performed well in difficult items. This reflects the high performance of students in Türkiye in chemistry compared to other OECD countries.

At the age of 15, almost all students in Türkiye were in upper secondary education. Students who sat the PISA 2018 test responded to a series of test items in mathematics, reading and science selected from a pool of 243 reading items, 70 mathematics items and 115 science items. Since reading was the major domain of assessment in PISA 2018, more test items were used to assess this domain than for mathematics and science, so the analysis in this chapter focuses on reading. Moreover, sub-scales were constructed only for reading and not for mathematics and science.

In each of the domains – reading, mathematics and science – PISA items differ in terms of content and cognitive process (Table 5.3). Items are also differentiated by proficiency (with levels ascending with difficulty) and question format. Chapter 1 provides a full description of the assessment framework. The reading test items were characterised by a number of attributes such as: cognitive process (understand, locate information, evaluate and reflect), text source (single or multiple texts), item format (i.e. simple multiple-choice, complex multiple-choice, open-ended), text type (e.g. argumentative, narrative, exposition) and text format (i.e. continuous, non-continuous and mixed). The design of the reading test is described in detail in the PISA 2018 assessment and analytical framework (OECD, 2019[17]).

Comparing student performance in the different reading sub-domains to the overall average for Türkiye, understanding, evaluating and reflecting, in addition to proficiency with single texts, were found to be areas in which students did better than the national average for reading as a whole (Table 5.12). It is possible that those results are driven by curriculum coverage or by student exposure to teaching practices that emphasise certain cognitive skills.

  • In terms of cognitive processes, only one process stood out: evaluate and reflect (Figure 5.9). Out of 63 test items, 15 items (24%) were identified as representing a strength against only four items representing a weakness (6%). In evaluating and reflecting, students should reflect on the content and form of the text and critically assess the quality and validity of the information therein.

The PISA items can also be differentiated in terms of the types of sources that students were required to engage with. Particular strengths among students in Türkiye included:

  • Out of the 45 test items covering narration texts, 11 items (24% of all items covering this text type) were identified as conspicuous items in which Turkish students did exceptionally well. The information in narration texts refers to the properties of objects in time. Narration texts typically require students to answer questions relating to when, in what sequence and why characters in stories behave as they do.

  • Six out of 38 items covering argumentative text (representing 16% of those items) were found to represent an area of strength. Argumentation texts present the relationship among concepts or propositions and often answer “why” questions.

  • Turkish students did particularly well in test items covering continuous texts, with 26 out of 158 items identified as representing an area of strength (representing 16% of all items covering this type). Continuous texts are formed by sentences organised into paragraphs. Examples of continuous texts include newspaper reports, essays, novels, short stories, reviews and letters. In contrast, no test items covering non-continuous texts fell into the category representing strength. Examples of non-continuous texts include lists, tables, graphs, diagrams, advertisements, schedules, catalogues, indices and forms.

No major differences in average difficulty were observed between items in the category representing a strength and those representing a weakness with the exception of the cognitive process of evaluating and reflecting. For this sub-domain, the items representing an area of strength were easier than those representing a weakness. This could reflect variations in students’ effort when answering test questions, as they may put more effort into answering difficult questions rather than easy ones. It could also reflect exposure to certain teaching practices or particular topics in the curriculum that are associated with the cognitive process of evaluating and reflecting.

  • Twelve out of 49 items (24%) covering the cognitive process of locating information fell into the category of conspicuous items representing a weakness (Figure 5.9). Locating information consists of accessing and retrieving information within a piece of text and searching for and selecting relevant text.

Other weaknesses reflected the types of sources students were required to engage with including description, multiple-source and non-continuous texts.

  • Six out of 30 items (20%) covering description texts were identified as conspicuously weak items. Description texts are texts where the information refers to the properties of objects in space. Such texts typically provide an answer to “what” questions.

  • Students also did worse in test items covering multiple-source texts with 38% of all items covering this type (8 out of 21 items) being identified as representing an area of weakness.

  • Students did particularly worse in test items covering non-continuous texts such as lists, tables, graphs, diagrams, advertisements, schedules, catalogues and forms. Six out of 30 (20%) test items fell into this category.

These results could reflect certain exposure at school to particular types of text or instructional strategies. Students are probably more exposed to continuous texts such as newspapers, reports, essays and novels than non-continuous texts such as tables, graphs, diagrams, indices and forms.

Finally, students found it more difficult to engage with complex, multiple-choice questions, with eight out of 32 test items (25%) being identified as representing an area of weakness. No major differences in average difficulty were observed between items in the category representing a strength and those representing a weakness.

Since mathematics and science were not the major domains of assessment in PISA 2018, fewer items were included in the assessment and fewer were identified as conspicuous. Therefore:

  • In science, students did well in the content domain of physical systems (6 items representing 16% of all items in this group) and did worse in items covering living systems (8 items representing 17% of all items in this group).

  • Students also did particularly well in test items covering the cognitive process of “evaluate and design scientific enquiry” with 5 out of 30 items (17%) identified as representing an area of strength.

  • Eleven out of 49 (24%) test items covering the cognitive process of explaining phenomena scientifically were identified as conspicuous items representing a weakness.

Students did well in open-ended science questions and less well in complex multiple-choice ones. In physics and in the cognitive domain of explaining phenomena scientifically, the items in which students did particularly well were slightly more difficult than those in which they did poorly. In mathematics, only 14 conspicuous items (8 representing strength and 6 representing weakness) were identified out of 70 test items without a clear pattern emerging in terms of item characteristics.

By analysing the results from TIMSS Grade 4 and 8 in 2015 and 2019, as well as PISA 2018, physics emerges as an area of strength across all levels of schooling (except in Grade 8 in 2019). This may be related to the curriculum’s focus as well as strong pedagogical practices that support students to do particularly well in this domain. Students also performed very well in chemistry in Grade 8 in 2019. Earth science and life science are areas of weakness for students in Türkiye in all grades. These two domains are closely related and the students’ performance may reflect relatively less exposure to this kind of content. From TIMSS 2015 to TIMSS 2019, applying scientific knowledge improved, possibly reflecting the curriculum reform that took place in 2017/18.

In mathematics, in contrast with science, there is less consistency across areas of strength and weakness over the different levels of school and different assessments. However, sub-domains number in Grade 4 and algebra in Grade 8 were identified as areas of strength. Students’ strength in these domains may facilitate their competency in the complex science domains of physics and chemistry since these mathematical skills underpin the skills required in these two science domains (Rossdy et al., 2019[15]). Overall, strengths and weaknesses that emerge from this chapter’s analysis of the international assessments might be associated with curricula coverage or teaching and learning practices in Türkiye and might warrant further analysis nationally.

References

[16] Cunningham, A. and R. Whelan (2014), “Maths for chemists”, University of Birmingham, University of Leeds, https://www.birmingham.ac.uk/Documents/college-eps/college/stem/Student-Summer-Education-Internships/Maths-for-Chemists-Booklet.pdf.

[2] Fishbein, B. et al. (2018), “The TIMSS 2019 Item Equivalence Study: Examining mode effects for computer-based assessment and implications for measuring trends”, Large-scale Assessments in Education, Vol. 6/11, https://doi.org/10.1186/s40536-018-0064-z.

[14] Government of Turkey (2019), Strategic Plan 2019-2023 (MEB 2019-2023 Stratejik Plan)ı, Government of Turkey, https://www.meb.gov.tr/stratejik_plan/ (accessed on 27 June 2022).

[1] IEA (2020), Methods and Procedures: TIMSS 2019 Technical Report, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, https://timssandpirls.bc.edu/timss2019/methods/ (accessed on 6 December 2021).

[10] IEA (2020), TIMSS 2019, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, https://timssandpirls.bc.edu/timss2019/international-database/ (accessed on 20 December 2021).

[19] IEA (2020), TIMSS 2019 International Results in Mathematics and Science, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, https://timss2019.org/reports/home-contexts/# (accessed on 6 December 2021).

[3] IEA (2017), TIMSS 2019 Assessment Frameworks, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, https://timssandpirls.bc.edu/timss2019/frameworks/framework-chapters/mathematics-framework/.

[9] IEA (2016), TIMSS 2015, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, https://timss2015.org/timss-2015/about-timss-2015/ (accessed on 21 May 2021).

[20] IEA (2016), TIMSS 2015 International Reports, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, http://timssandpirls.bc.edu/timss2015/international-results/ (accessed on 21 May 2021).

[6] IEA (2013), TIMSS 2015 Assessment Frameworks, TIMSS & PIRLS International Study Center, Lynch School of Education, Boston College, and International Association for the Evaluation of Educational Achievement, https://timssandpirls.bc.edu/timss2015/frameworks.html (accessed on 28 July 2021).

[12] Kitchen, H. et al. (2019), OECD Reviews of Evaluation and Assessment in Education: Student Assessment in Turkey, OECD Reviews of Evaluation and Assessment in Education, OECD Publishing, Paris, https://doi.org/10.1787/5edc0abe-en.

[5] Martin, M. et al. (2016), TIMSS 2015 International Results in Science, TIMSS & PIRLS International Study Center, Boston.

[4] Mullis, I. et al. (2016), Student Achievement, International Study Center, Lynch School of Education, Boston College, http://timssandpirls.bc.edu/timss2015/international-results/timss-2015/mathematics/student-achievement/ (accessed on 5 March 2018).

[18] OECD (2021), “PISA: Programme for International Student Assessment”, OECD Education Statistics (database), https://doi.org/10.1787/data-00365-en (accessed on 21 May 2021).

[17] OECD (2019), PISA 2018 Assessment and Analytical Framework, PISA, OECD Publishing, Paris, https://doi.org/10.1787/b25efab8-en.

[7] OECD (2019), PISA 2018 Results (Volume I): What Students Know and Can Do, PISA, OECD Publishing, Paris, https://doi.org/10.1787/5f07c754-en.

[8] OECD (2013), Lessons from PISA 2012 for the United States, Strong Performers and Successful Reformers in Education, OECD Publishing, Paris, https://doi.org/10.1787/9789264207585-en.

[15] Rossdy, M. et al. (2019), “The role of physics and mathematics in influencing science students’ performance”, in Proceedings of the Second International Conference on the Future of ASEAN (ICoFA) 2017 - Volume 1, https://link.springer.com/chapter/10.1007/978-981-10-8730-1_40.

[11] Vincent-Lancrin, S. et al. (2019), Fostering Students’ Creativity and Critical Thinking: What it Means in School, Educational Research and Innovation, OECD Publishing, Paris, https://doi.org/10.1787/62212c37-en.

[13] World Bank (2021), “Turkey’s improvements in the quality of learning: TIMSS 2019 results”, World Bank, Washington, DC, https://www.worldbank.org/en/country/turkey/brief/turkeys-improvements-in-the-quality-of-learning.

Note

← 1. In this report, the terminology of “TIMSS, Grade 4” is used throughout since this is the official name of the assessment. However, the data refer to Grade 5 students in lower secondary education in Türkiye.

Metadata, Legal and Rights

This document, as well as any data and map included herein, are without prejudice to the status of or sovereignty over any territory, to the delimitation of international frontiers and boundaries and to the name of any territory, city or area. Extracts from publications may be subject to additional disclaimers, which are set out in the complete version of the publication, available at the link provided.

© OECD 2022

The use of this work, whether digital or print, is governed by the Terms and Conditions to be found at https://www.oecd.org/termsandconditions.